From 3feb87b0054d23d8452b4fa557e9400781c3440d Mon Sep 17 00:00:00 2001 From: Andrew Purtell Date: Wed, 16 Aug 2017 10:52:18 -0700 Subject: [PATCH] HBASE-18431 Mitigate compatibility concerns between branch-1.3 and branch-1.4 - Refactor TableName into its own proto module and fix up users - Move SnapshotDescription from Snapshot.proto back to HBase.proto - Restore FastLongHistogram and TestFastLongHistogram; deprecate FastLongHistogram - Move DeleteQueryTracker back to o.a.h.h.regionserver --- .../org/apache/hadoop/hbase/client/Admin.java | 3 +- .../hadoop/hbase/client/HBaseAdmin.java | 5 +- .../hadoop/hbase/protobuf/ProtobufUtil.java | 9 +- .../ClientSnapshotDescriptionUtils.java | 2 +- .../snapshot/CorruptedSnapshotException.java | 2 +- .../snapshot/HBaseSnapshotException.java | 2 +- .../snapshot/RestoreSnapshotException.java | 2 +- .../snapshot/SnapshotCreationException.java | 2 +- .../SnapshotDoesNotExistException.java | 2 +- .../snapshot/SnapshotExistsException.java | 2 +- .../hbase/client/TestSnapshotFromAdmin.java | 2 +- .../hadoop/hbase/util/FastLongHistogram.java | 319 +++ .../hbase/util/TestFastLongHistogram.java | 132 + hbase-protocol/pom.xml | 1 + .../generated/AccessControlProtos.java | 92 +- .../protobuf/generated/ClientProtos.java | 2 +- .../hbase/protobuf/generated/HBaseProtos.java | 2315 ++++++++++++----- .../generated/MasterProcedureProtos.java | 586 ++--- .../protobuf/generated/MasterProtos.java | 1689 ++++++------ .../generated/SecureBulkLoadProtos.java | 94 +- .../protobuf/generated/SnapshotProtos.java | 1551 +---------- .../hbase/protobuf/generated/TableProtos.java | 607 +++++ .../hbase/protobuf/generated/WALProtos.java | 145 +- .../protobuf/generated/ZooKeeperProtos.java | 171 +- .../src/main/protobuf/AccessControl.proto | 2 +- hbase-protocol/src/main/protobuf/HBase.proto | 28 +- hbase-protocol/src/main/protobuf/Master.proto | 1 + .../src/main/protobuf/MasterProcedure.proto | 1 + .../src/main/protobuf/SecureBulkLoad.proto | 1 + .../src/main/protobuf/Snapshot.proto | 19 - hbase-protocol/src/main/protobuf/Table.proto | 33 + hbase-protocol/src/main/protobuf/WAL.proto | 1 + .../src/main/protobuf/ZooKeeper.proto | 1 + .../hbase/tmpl/master/MasterStatusTmpl.jamon | 2 +- .../BaseMasterAndRegionObserver.java | 2 +- .../hbase/coprocessor/BaseMasterObserver.java | 2 +- .../hbase/coprocessor/BaseRegionObserver.java | 2 +- .../hbase/coprocessor/MasterObserver.java | 2 +- .../hbase/coprocessor/RegionObserver.java | 2 +- .../TableSnapshotInputFormatImpl.java | 2 +- .../hbase/master/MasterCoprocessorHost.java | 2 +- .../hbase/master/MasterRpcServices.java | 4 +- .../hadoop/hbase/master/SnapshotSentinel.java | 2 +- .../master/snapshot/CloneSnapshotHandler.java | 2 +- .../DisabledTableSnapshotHandler.java | 2 +- .../snapshot/EnabledTableSnapshotHandler.java | 2 +- .../snapshot/MasterSnapshotVerifier.java | 2 +- .../snapshot/RestoreSnapshotHandler.java | 2 +- .../master/snapshot/SnapshotManager.java | 4 +- .../master/snapshot/TakeSnapshotHandler.java | 2 +- .../{querymatcher => }/DeleteTracker.java | 2 +- .../hadoop/hbase/regionserver/HRegion.java | 2 +- .../regionserver/RegionCoprocessorHost.java | 2 +- .../CompactionScanQueryMatcher.java | 1 + ...DropDeletesCompactionScanQueryMatcher.java | 1 + .../querymatcher/LegacyScanQueryMatcher.java | 3 +- .../MajorCompactionScanQueryMatcher.java | 1 + .../MinorCompactionScanQueryMatcher.java | 1 + .../NormalUserScanQueryMatcher.java | 1 + .../querymatcher/ScanDeleteTracker.java | 1 + .../querymatcher/ScanQueryMatcher.java | 3 +- .../StripeCompactionScanQueryMatcher.java | 1 + .../snapshot/FlushSnapshotSubprocedure.java | 2 +- .../snapshot/RegionServerSnapshotManager.java | 2 +- .../security/access/AccessController.java | 2 +- .../visibility/VisibilityController.java | 2 +- .../hadoop/hbase/snapshot/CreateSnapshot.java | 2 +- .../hadoop/hbase/snapshot/ExportSnapshot.java | 2 +- .../hbase/snapshot/RestoreSnapshotHelper.java | 2 +- .../snapshot/SnapshotDescriptionUtils.java | 2 +- .../hadoop/hbase/snapshot/SnapshotInfo.java | 2 +- .../hbase/snapshot/SnapshotManifest.java | 2 +- .../hbase/snapshot/SnapshotManifestV1.java | 2 +- .../hbase/snapshot/SnapshotManifestV2.java | 2 +- .../hbase/snapshot/SnapshotReferenceUtil.java | 2 +- .../hbase-webapps/master/snapshot.jsp | 2 +- .../hbase-webapps/master/snapshotsStats.jsp | 2 +- .../hbase/client/TestSnapshotFromClient.java | 2 +- .../hbase/coprocessor/TestMasterObserver.java | 2 +- .../cleaner/TestSnapshotFromMaster.java | 2 +- .../hbase/regionserver/TestHRegionInfo.java | 4 +- .../querymatcher/TestScanDeleteTracker.java | 2 +- .../security/access/TestAccessController.java | 2 +- .../access/TestWithDisabledAuthorization.java | 2 +- .../hbase/snapshot/SnapshotTestingUtils.java | 2 +- .../hbase/snapshot/TestExportSnapshot.java | 2 +- .../snapshot/TestFlushSnapshotFromClient.java | 2 +- .../snapshot/TestRegionSnapshotTask.java | 10 +- .../TestRestoreFlushSnapshotFromClient.java | 2 +- .../snapshot/TestRestoreSnapshotHelper.java | 2 +- .../snapshot/TestSnapshotClientRetries.java | 2 +- .../TestSnapshotDescriptionUtils.java | 2 +- .../hbase/snapshot/TestSnapshotManifest.java | 2 +- 93 files changed, 4247 insertions(+), 3706 deletions(-) create mode 100644 hbase-common/src/main/java/org/apache/hadoop/hbase/util/FastLongHistogram.java create mode 100644 hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestFastLongHistogram.java create mode 100644 hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java create mode 100644 hbase-protocol/src/main/protobuf/Table.proto rename hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/{querymatcher => }/DeleteTracker.java (98%) diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java index 5b3744a33ec..fa9594a9dbd 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java @@ -43,9 +43,8 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.client.security.SecurityCapability; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.quotas.QuotaFilter; import org.apache.hadoop.hbase.quotas.QuotaRetriever; import org.apache.hadoop.hbase.quotas.QuotaSettings; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index 2cbeb9ae5af..84871ed3bea 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -95,6 +95,7 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest; @@ -156,7 +157,7 @@ import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.StopMasterRequest import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.TruncateTableRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest; import org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.TableProtos; import org.apache.hadoop.hbase.quotas.QuotaFilter; import org.apache.hadoop.hbase.quotas.QuotaRetriever; import org.apache.hadoop.hbase.quotas.QuotaSettings; @@ -3174,7 +3175,7 @@ public class HBaseAdmin implements Admin { public TableName[] call(int callTimeout) throws Exception { HBaseRpcController controller = rpcControllerFactory.newController(); controller.setCallTimeout(callTimeout); - List tableNames = + List tableNames = master.listTableNamesByNamespace(controller, ListTableNamesByNamespaceRequest. newBuilder().setNamespaceName(name).build()) .getTableNameList(); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index b704f4adf4f..fa44cdd49a3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -145,6 +145,7 @@ import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterService; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerReportRequest; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest; +import org.apache.hadoop.hbase.protobuf.generated.TableProtos; import org.apache.hadoop.hbase.protobuf.generated.WALProtos; import org.apache.hadoop.hbase.protobuf.generated.WALProtos.BulkLoadDescriptor; import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor; @@ -2975,18 +2976,18 @@ public final class ProtobufUtil { ", type=" + proto.getMutateType().toString(); } - public static TableName toTableName(HBaseProtos.TableName tableNamePB) { + public static TableName toTableName(TableProtos.TableName tableNamePB) { return TableName.valueOf(tableNamePB.getNamespace().asReadOnlyByteBuffer(), tableNamePB.getQualifier().asReadOnlyByteBuffer()); } - public static HBaseProtos.TableName toProtoTableName(TableName tableName) { - return HBaseProtos.TableName.newBuilder() + public static TableProtos.TableName toProtoTableName(TableName tableName) { + return TableProtos.TableName.newBuilder() .setNamespace(ByteStringer.wrap(tableName.getNamespace())) .setQualifier(ByteStringer.wrap(tableName.getQualifier())).build(); } - public static TableName[] getTableNameArray(List tableNamesList) { + public static TableName[] getTableNameArray(List tableNamesList) { if (tableNamesList == null) { return new TableName[0]; } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/ClientSnapshotDescriptionUtils.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/ClientSnapshotDescriptionUtils.java index 686dbb81cbd..bba166b8358 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/ClientSnapshotDescriptionUtils.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/ClientSnapshotDescriptionUtils.java @@ -21,7 +21,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.util.Bytes; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.java index a1f59115ae9..d29c89cb58e 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java index a88d8203837..8889cfcb628 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/HBaseSnapshotException.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** * General exception base class for when a snapshot fails diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotException.java index 8ed42e07e03..25cd583b03a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotException.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** * Thrown when a snapshot could not be restored due to a server-side error when restoring it. diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotCreationException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotCreationException.java index d5501010a3d..324d41f3c6a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotCreationException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotCreationException.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** * Thrown when a snapshot could not be created due to a server-side error when diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDoesNotExistException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDoesNotExistException.java index 6006136cae8..6ba45bddc92 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDoesNotExistException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDoesNotExistException.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotExistsException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotExistsException.java index 0344aff56ca..ff3cdcb87c8 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotExistsException.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotExistsException.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hbase.snapshot; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** * Thrown when a snapshot exists but should not diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java index afa589249b0..6385c27dfa4 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java @@ -32,11 +32,11 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.ipc.HBaseRpcController; import org.apache.hadoop.hbase.ipc.RpcControllerFactory; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SnapshotResponse; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.Test; import org.junit.experimental.categories.Category; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/FastLongHistogram.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/FastLongHistogram.java new file mode 100644 index 00000000000..2473bb49a29 --- /dev/null +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/FastLongHistogram.java @@ -0,0 +1,319 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.util; + +import java.util.concurrent.atomic.AtomicBoolean; +import java.util.concurrent.atomic.AtomicLong; + +import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.classification.InterfaceStability; + +/** + * FastLongHistogram is a thread-safe class that estimate distribution of data and computes the + * quantiles. + */ +@Deprecated +@InterfaceAudience.Public +@InterfaceStability.Evolving +public class FastLongHistogram { + + /** + * Default number of bins. + */ + public static final int DEFAULT_NBINS = 255; + + public static final double[] DEFAULT_QUANTILES = + new double[]{0.25, 0.5, 0.75, 0.90, 0.95, 0.98, 0.99, 0.999}; + + /** + * Bins is a class containing a list of buckets(or bins) for estimation histogram of some data. + */ + private static class Bins { + private final Counter[] counts; + // inclusive + private final long binsMin; + // exclusive + private final long binsMax; + private final long bins10XMax; + private final AtomicLong min = new AtomicLong(Long.MAX_VALUE); + private final AtomicLong max = new AtomicLong(0L); + + private final Counter count = new Counter(0); + private final Counter total = new Counter(0); + + // set to true when any of data has been inserted to the Bins. It is set after the counts are + // updated. + private final AtomicBoolean hasData = new AtomicBoolean(false); + + /** + * The constructor for creating a Bins without any prior data. + */ + public Bins(int numBins) { + counts = createCounters(numBins + 3); + this.binsMin = 1L; + + // These two numbers are total guesses + // and should be treated as highly suspect. + this.binsMax = 1000; + this.bins10XMax = binsMax * 10; + } + + /** + * The constructor for creating a Bins with last Bins. + */ + public Bins(Bins last, int numOfBins, double minQ, double maxQ) { + long[] values = last.getQuantiles(new double[] { minQ, maxQ }); + long wd = values[1] - values[0] + 1; + // expand minQ and maxQ in two ends back assuming uniform distribution + this.binsMin = Math.max(0L, (long) (values[0] - wd * minQ)); + long binsMax = (long) (values[1] + wd * (1 - maxQ)) + 1; + // make sure each of bins is at least of width 1 + this.binsMax = Math.max(binsMax, this.binsMin + numOfBins); + this.bins10XMax = Math.max((long) (values[1] + (binsMax - 1) * 9), this.binsMax + 1); + + this.counts = createCounters(numOfBins + 3); + } + + private Counter[] createCounters(int num) { + Counter[] counters = new Counter[num]; + for (int i = 0; i < num; i++) { + counters[i] = new Counter(); + } + return counters; + } + + private int getIndex(long value) { + if (value < this.binsMin) { + return 0; + } else if (value > this.bins10XMax) { + return this.counts.length - 1; + } else if (value >= this.binsMax) { + return this.counts.length - 2; + } + // compute the position + return 1 + (int) ((value - this.binsMin) * (this.counts.length - 3) / + (this.binsMax - this.binsMin)); + + } + + /** + * Adds a value to the histogram. + */ + public void add(long value, long count) { + if (value < 0) { + // The whole computation is completely thrown off if there are negative numbers + // + // Normally we would throw an IllegalArgumentException however this is the metrics + // system and it should be completely safe at all times. + // So silently throw it away. + return; + } + AtomicUtils.updateMin(min, value); + AtomicUtils.updateMax(max, value); + + this.count.add(count); + this.total.add(value * count); + + int pos = getIndex(value); + this.counts[pos].add(count); + + // hasData needs to be updated as last + this.hasData.set(true); + } + + /** + * Computes the quantiles give the ratios. + */ + public long[] getQuantiles(double[] quantiles) { + if (!this.hasData.get()) { + // No data yet. + return new long[quantiles.length]; + } + + // Make a snapshot of lowerCounter, higherCounter and bins.counts to counts. + // This is not synchronized, but since the counter are accumulating, the result is a good + // estimation of a snapshot. + long[] counts = new long[this.counts.length]; + long total = 0L; + for (int i = 0; i < this.counts.length; i++) { + counts[i] = this.counts[i].get(); + total += counts[i]; + } + + int rIndex = 0; + double qCount = total * quantiles[0]; + long cum = 0L; + + long[] res = new long[quantiles.length]; + countsLoop: for (int i = 0; i < counts.length; i++) { + // mn and mx define a value range + long mn, mx; + if (i == 0) { + mn = this.min.get(); + mx = this.binsMin; + } else if (i == counts.length - 1) { + mn = this.bins10XMax; + mx = this.max.get(); + } else if (i == counts.length - 2) { + mn = this.binsMax; + mx = this.bins10XMax; + } else { + mn = this.binsMin + (i - 1) * (this.binsMax - this.binsMin) / (this.counts.length - 3); + mx = this.binsMin + i * (this.binsMax - this.binsMin) / (this.counts.length - 3); + } + + if (mx < this.min.get()) { + continue; + } + if (mn > this.max.get()) { + break; + } + mn = Math.max(mn, this.min.get()); + mx = Math.min(mx, this.max.get()); + + // lastCum/cum are the corresponding counts to mn/mx + double lastCum = cum; + cum += counts[i]; + + // fill the results for qCount is within current range. + while (qCount <= cum) { + if (cum == lastCum) { + res[rIndex] = mn; + } else { + res[rIndex] = (long) ((qCount - lastCum) * (mx - mn) / (cum - lastCum) + mn); + } + + // move to next quantile + rIndex++; + if (rIndex >= quantiles.length) { + break countsLoop; + } + qCount = total * quantiles[rIndex]; + } + } + // In case quantiles contains values >= 100% + for (; rIndex < quantiles.length; rIndex++) { + res[rIndex] = this.max.get(); + } + + return res; + } + + + long getNumAtOrBelow(long val) { + final int targetIndex = getIndex(val); + long totalToCurrentIndex = 0; + for (int i = 0; i <= targetIndex; i++) { + totalToCurrentIndex += this.counts[i].get(); + } + return totalToCurrentIndex; + } + } + + // The bins counting values. It is replaced with a new one in calling of reset(). + private volatile Bins bins; + + /** + * Constructor. + */ + public FastLongHistogram() { + this(DEFAULT_NBINS); + } + + /** + * Constructor. + * @param numOfBins the number of bins for the histogram. A larger value results in more precise + * results but with lower efficiency, and vice versus. + */ + public FastLongHistogram(int numOfBins) { + this.bins = new Bins(numOfBins); + } + + /** + * Constructor setting the bins assuming a uniform distribution within a range. + * @param numOfBins the number of bins for the histogram. A larger value results in more precise + * results but with lower efficiency, and vice versus. + * @param min lower bound of the region, inclusive. + * @param max higher bound of the region, inclusive. + */ + public FastLongHistogram(int numOfBins, long min, long max) { + this(numOfBins); + Bins bins = new Bins(numOfBins); + bins.add(min, 1); + bins.add(max, 1); + this.bins = new Bins(bins, numOfBins, 0.01, 0.999); + } + + private FastLongHistogram(Bins bins) { + this.bins = bins; + } + + /** + * Adds a value to the histogram. + */ + public void add(long value, long count) { + this.bins.add(value, count); + } + + /** + * Computes the quantiles give the ratios. + */ + public long[] getQuantiles(double[] quantiles) { + return this.bins.getQuantiles(quantiles); + } + + public long[] getQuantiles() { + return this.bins.getQuantiles(DEFAULT_QUANTILES); + } + + public long getMin() { + long min = this.bins.min.get(); + return min == Long.MAX_VALUE ? 0 : min; // in case it is not initialized + } + + public long getMax() { + return this.bins.max.get(); + } + + public long getCount() { + return this.bins.count.get(); + } + + public long getMean() { + Bins bins = this.bins; + long count = bins.count.get(); + long total = bins.total.get(); + if (count == 0) { + return 0; + } + return total / count; + } + + public long getNumAtOrBelow(long value) { + return this.bins.getNumAtOrBelow(value); + } + + /** + * Resets the histogram for new counting. + */ + public FastLongHistogram reset() { + Bins oldBins = this.bins; + this.bins = new Bins(this.bins, this.bins.counts.length - 3, 0.01, 0.99); + return new FastLongHistogram(oldBins); + } +} diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestFastLongHistogram.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestFastLongHistogram.java new file mode 100644 index 00000000000..7304e2db23e --- /dev/null +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestFastLongHistogram.java @@ -0,0 +1,132 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.util; + +import java.util.Arrays; +import java.util.Random; + + +import org.apache.hadoop.hbase.testclassification.SmallTests; +import org.junit.Assert; +import org.junit.Test; +import org.junit.experimental.categories.Category; + +import static org.junit.Assert.assertEquals; + +/** + * Testcases for FastLongHistogram. + */ +@Category(SmallTests.class) +public class TestFastLongHistogram { + + private static void doTestUniform(FastLongHistogram hist) { + long[] VALUES = { 0, 10, 20, 30, 40, 50 }; + double[] qs = new double[VALUES.length]; + for (int i = 0; i < qs.length; i++) { + qs[i] = (double) VALUES[i] / VALUES[VALUES.length - 1]; + } + + for (int i = 0; i < 10; i++) { + for (long v : VALUES) { + hist.add(v, 1); + } + long[] vals = hist.getQuantiles(qs); + System.out.println(Arrays.toString(vals)); + for (int j = 0; j < qs.length; j++) { + Assert.assertTrue(j + "-th element org: " + VALUES[j] + ", act: " + vals[j], + Math.abs(vals[j] - VALUES[j]) <= 10); + } + hist.reset(); + } + } + + @Test + public void testUniform() { + FastLongHistogram hist = new FastLongHistogram(100, 0, 50); + doTestUniform(hist); + } + + @Test + public void testAdaptionOfChange() { + // assumes the uniform distribution + FastLongHistogram hist = new FastLongHistogram(100, 0, 100); + + Random rand = new Random(); + + for (int n = 0; n < 10; n++) { + for (int i = 0; i < 900; i++) { + hist.add(rand.nextInt(100), 1); + } + + // add 10% outliers, this breaks the assumption, hope bin10xMax works + for (int i = 0; i < 100; i++) { + hist.add(1000 + rand.nextInt(100), 1); + } + + long[] vals = hist.getQuantiles(new double[] { 0.25, 0.75, 0.95 }); + System.out.println(Arrays.toString(vals)); + if (n == 0) { + Assert.assertTrue("Out of possible value", vals[0] >= 0 && vals[0] <= 50); + Assert.assertTrue("Out of possible value", vals[1] >= 50 && vals[1] <= 100); + Assert.assertTrue("Out of possible value", vals[2] >= 900 && vals[2] <= 1100); + } + + hist.reset(); + } + } + + + @Test + public void testGetNumAtOrBelow() { + long[] VALUES = { 1, 10, 20, 30, 40, 50 }; + + FastLongHistogram h = new FastLongHistogram(); + for (long v : VALUES) { + for (int i = 0; i < 100; i++) { + h.add(v, 1); + } + } + + h.add(Integer.MAX_VALUE, 1); + + h.reset(); + + for (long v : VALUES) { + for (int i = 0; i < 100; i++) { + h.add(v, 1); + } + } + // Add something way out there to make sure it doesn't throw off the counts. + h.add(Integer.MAX_VALUE, 1); + + assertEquals(100, h.getNumAtOrBelow(1)); + assertEquals(200, h.getNumAtOrBelow(11)); + assertEquals(601, h.getNumAtOrBelow(Long.MAX_VALUE)); + } + + + @Test + public void testSameValues() { + FastLongHistogram hist = new FastLongHistogram(100); + + hist.add(50, 100); + + hist.reset(); + doTestUniform(hist); + } +} diff --git a/hbase-protocol/pom.xml b/hbase-protocol/pom.xml index 7291d15e64e..8588a8c28bf 100644 --- a/hbase-protocol/pom.xml +++ b/hbase-protocol/pom.xml @@ -197,6 +197,7 @@ RowProcessor.proto SecureBulkLoad.proto Snapshot.proto + Table.proto Tracing.proto VisibilityLabels.proto WAL.proto diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java index b72e6e5cae6..161282cc3c6 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AccessControlProtos.java @@ -1285,11 +1285,11 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional bytes family = 2; /** @@ -1377,11 +1377,11 @@ public final class AccessControlProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -1477,7 +1477,7 @@ public final class AccessControlProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -1487,13 +1487,13 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -1552,7 +1552,7 @@ public final class AccessControlProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); family_ = com.google.protobuf.ByteString.EMPTY; qualifier_ = com.google.protobuf.ByteString.EMPTY; action_ = java.util.Collections.emptyList(); @@ -1797,7 +1797,7 @@ public final class AccessControlProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -1926,9 +1926,9 @@ public final class AccessControlProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -1938,7 +1938,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -1948,7 +1948,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -1965,7 +1965,7 @@ public final class AccessControlProtos { * optional .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -1978,12 +1978,12 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -1999,7 +1999,7 @@ public final class AccessControlProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -2010,7 +2010,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -2018,7 +2018,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -2029,11 +2029,11 @@ public final class AccessControlProtos { * optional .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -7462,11 +7462,11 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional bytes namespace_name = 3; /** @@ -7541,11 +7541,11 @@ public final class AccessControlProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -7616,7 +7616,7 @@ public final class AccessControlProtos { // optional .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 2; */ @@ -7626,13 +7626,13 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -7654,7 +7654,7 @@ public final class AccessControlProtos { private void initFields() { type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global; - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); namespaceName_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; @@ -7881,7 +7881,7 @@ public final class AccessControlProtos { type_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Type.Global; bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -8027,9 +8027,9 @@ public final class AccessControlProtos { } // optional .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 2; */ @@ -8039,7 +8039,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -8049,7 +8049,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -8066,7 +8066,7 @@ public final class AccessControlProtos { * optional .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -8079,12 +8079,12 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -8100,7 +8100,7 @@ public final class AccessControlProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -8111,7 +8111,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -8119,7 +8119,7 @@ public final class AccessControlProtos { /** * optional .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -8130,11 +8130,11 @@ public final class AccessControlProtos { * optional .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -10499,7 +10499,7 @@ public final class AccessControlProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\023AccessControl.proto\022\010hbase.pb\032\013HBase.p" + + "\n\023AccessControl.proto\022\010hbase.pb\032\013Table.p" + "roto\"\314\002\n\nPermission\022\'\n\004type\030\001 \002(\0162\031.hbas" + "e.pb.Permission.Type\0225\n\021global_permissio" + "n\030\002 \001(\0132\032.hbase.pb.GlobalPermission\022;\n\024n" + @@ -10648,7 +10648,7 @@ public final class AccessControlProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), }, assigner); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java index 12950f72f54..4bd3f5fd7bd 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java @@ -22646,7 +22646,7 @@ public final class ClientProtos { * */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CursorOrBuilder> getCursorFieldBuilder() { if (cursorBuilder_ == null) { cursorBuilder_ = new com.google.protobuf.SingleFieldBuilder< diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java index 3c4fb616d42..b4c6c04c609 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java @@ -266,562 +266,6 @@ public final class HBaseProtos { // @@protoc_insertion_point(enum_scope:hbase.pb.TimeUnit) } - public interface TableNameOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required bytes namespace = 1; - /** - * required bytes namespace = 1; - */ - boolean hasNamespace(); - /** - * required bytes namespace = 1; - */ - com.google.protobuf.ByteString getNamespace(); - - // required bytes qualifier = 2; - /** - * required bytes qualifier = 2; - */ - boolean hasQualifier(); - /** - * required bytes qualifier = 2; - */ - com.google.protobuf.ByteString getQualifier(); - } - /** - * Protobuf type {@code hbase.pb.TableName} - * - *
-   **
-   * Table Name
-   * 
- */ - public static final class TableName extends - com.google.protobuf.GeneratedMessage - implements TableNameOrBuilder { - // Use TableName.newBuilder() to construct. - private TableName(com.google.protobuf.GeneratedMessage.Builder builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final TableName defaultInstance; - public static TableName getDefaultInstance() { - return defaultInstance; - } - - public TableName getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private TableName( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - namespace_ = input.readBytes(); - break; - } - case 18: { - bitField0_ |= 0x00000002; - qualifier_ = input.readBytes(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder.class); - } - - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public TableName parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new TableName(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private int bitField0_; - // required bytes namespace = 1; - public static final int NAMESPACE_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString namespace_; - /** - * required bytes namespace = 1; - */ - public boolean hasNamespace() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required bytes namespace = 1; - */ - public com.google.protobuf.ByteString getNamespace() { - return namespace_; - } - - // required bytes qualifier = 2; - public static final int QUALIFIER_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString qualifier_; - /** - * required bytes qualifier = 2; - */ - public boolean hasQualifier() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * required bytes qualifier = 2; - */ - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; - } - - private void initFields() { - namespace_ = com.google.protobuf.ByteString.EMPTY; - qualifier_ = com.google.protobuf.ByteString.EMPTY; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasNamespace()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasQualifier()) { - memoizedIsInitialized = 0; - return false; - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, namespace_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, qualifier_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, namespace_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, qualifier_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) obj; - - boolean result = true; - result = result && (hasNamespace() == other.hasNamespace()); - if (hasNamespace()) { - result = result && getNamespace() - .equals(other.getNamespace()); - } - result = result && (hasQualifier() == other.hasQualifier()); - if (hasQualifier()) { - result = result && getQualifier() - .equals(other.getQualifier()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasNamespace()) { - hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; - hash = (53 * hash) + getNamespace().hashCode(); - } - if (hasQualifier()) { - hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; - hash = (53 * hash) + getQualifier().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.TableName} - * - *
-     **
-     * Table Name
-     * 
- */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - namespace_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000001); - qualifier_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_TableName_descriptor; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName build() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.namespace_ = namespace_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.qualifier_ = qualifier_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) return this; - if (other.hasNamespace()) { - setNamespace(other.getNamespace()); - } - if (other.hasQualifier()) { - setQualifier(other.getQualifier()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasNamespace()) { - - return false; - } - if (!hasQualifier()) { - - return false; - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // required bytes namespace = 1; - private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY; - /** - * required bytes namespace = 1; - */ - public boolean hasNamespace() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required bytes namespace = 1; - */ - public com.google.protobuf.ByteString getNamespace() { - return namespace_; - } - /** - * required bytes namespace = 1; - */ - public Builder setNamespace(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - namespace_ = value; - onChanged(); - return this; - } - /** - * required bytes namespace = 1; - */ - public Builder clearNamespace() { - bitField0_ = (bitField0_ & ~0x00000001); - namespace_ = getDefaultInstance().getNamespace(); - onChanged(); - return this; - } - - // required bytes qualifier = 2; - private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; - /** - * required bytes qualifier = 2; - */ - public boolean hasQualifier() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * required bytes qualifier = 2; - */ - public com.google.protobuf.ByteString getQualifier() { - return qualifier_; - } - /** - * required bytes qualifier = 2; - */ - public Builder setQualifier(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - qualifier_ = value; - onChanged(); - return this; - } - /** - * required bytes qualifier = 2; - */ - public Builder clearQualifier() { - bitField0_ = (bitField0_ & ~0x00000002); - qualifier_ = getDefaultInstance().getQualifier(); - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:hbase.pb.TableName) - } - - static { - defaultInstance = new TableName(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:hbase.pb.TableName) - } - public interface TableSchemaOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -833,11 +277,11 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // repeated .hbase.pb.BytesBytesPair attributes = 2; /** @@ -972,11 +416,11 @@ public final class HBaseProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -1059,7 +503,7 @@ public final class HBaseProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -1069,13 +513,13 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -1188,7 +632,7 @@ public final class HBaseProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); attributes_ = java.util.Collections.emptyList(); columnFamilies_ = java.util.Collections.emptyList(); configuration_ = java.util.Collections.emptyList(); @@ -1449,7 +893,7 @@ public final class HBaseProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -1684,9 +1128,9 @@ public final class HBaseProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -1696,7 +1140,7 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -1706,7 +1150,7 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -1723,7 +1167,7 @@ public final class HBaseProtos { * optional .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -1736,12 +1180,12 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -1757,7 +1201,7 @@ public final class HBaseProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -1768,7 +1212,7 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -1776,7 +1220,7 @@ public final class HBaseProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -1787,11 +1231,11 @@ public final class HBaseProtos { * optional .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -3769,11 +3213,11 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional bytes start_key = 3; /** @@ -3887,11 +3331,11 @@ public final class HBaseProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -3982,7 +3426,7 @@ public final class HBaseProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -3992,13 +3436,13 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -4084,7 +3528,7 @@ public final class HBaseProtos { private void initFields() { regionId_ = 0L; - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); startKey_ = com.google.protobuf.ByteString.EMPTY; endKey_ = com.google.protobuf.ByteString.EMPTY; offline_ = false; @@ -4390,7 +3834,7 @@ public final class HBaseProtos { regionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -4575,9 +4019,9 @@ public final class HBaseProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -4587,7 +4031,7 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -4597,7 +4041,7 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4614,7 +4058,7 @@ public final class HBaseProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -4627,12 +4071,12 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -4648,7 +4092,7 @@ public final class HBaseProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -4659,7 +4103,7 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -4667,7 +4111,7 @@ public final class HBaseProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -4678,11 +4122,11 @@ public final class HBaseProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -17658,11 +17102,1494 @@ public final class HBaseProtos { // @@protoc_insertion_point(class_scope:hbase.pb.RegionServerInfo) } - private static com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_TableName_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_hbase_pb_TableName_fieldAccessorTable; + public interface SnapshotDescriptionOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required string name = 1; + /** + * required string name = 1; + */ + boolean hasName(); + /** + * required string name = 1; + */ + java.lang.String getName(); + /** + * required string name = 1; + */ + com.google.protobuf.ByteString + getNameBytes(); + + // optional string table = 2; + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + boolean hasTable(); + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + java.lang.String getTable(); + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + com.google.protobuf.ByteString + getTableBytes(); + + // optional int64 creation_time = 3 [default = 0]; + /** + * optional int64 creation_time = 3 [default = 0]; + */ + boolean hasCreationTime(); + /** + * optional int64 creation_time = 3 [default = 0]; + */ + long getCreationTime(); + + // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + boolean hasType(); + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type getType(); + + // optional int32 version = 5; + /** + * optional int32 version = 5; + */ + boolean hasVersion(); + /** + * optional int32 version = 5; + */ + int getVersion(); + + // optional string owner = 6; + /** + * optional string owner = 6; + */ + boolean hasOwner(); + /** + * optional string owner = 6; + */ + java.lang.String getOwner(); + /** + * optional string owner = 6; + */ + com.google.protobuf.ByteString + getOwnerBytes(); + + // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + boolean hasUsersAndPermissions(); + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions(); + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder(); + } + /** + * Protobuf type {@code hbase.pb.SnapshotDescription} + * + *
+   **
+   * Description of the snapshot to take
+   * 
+ */ + public static final class SnapshotDescription extends + com.google.protobuf.GeneratedMessage + implements SnapshotDescriptionOrBuilder { + // Use SnapshotDescription.newBuilder() to construct. + private SnapshotDescription(com.google.protobuf.GeneratedMessage.Builder builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private SnapshotDescription(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final SnapshotDescription defaultInstance; + public static SnapshotDescription getDefaultInstance() { + return defaultInstance; + } + + public SnapshotDescription getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private SnapshotDescription( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + name_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + table_ = input.readBytes(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + creationTime_ = input.readInt64(); + break; + } + case 32: { + int rawValue = input.readEnum(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type value = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type.valueOf(rawValue); + if (value == null) { + unknownFields.mergeVarintField(4, rawValue); + } else { + bitField0_ |= 0x00000008; + type_ = value; + } + break; + } + case 40: { + bitField0_ |= 0x00000010; + version_ = input.readInt32(); + break; + } + case 50: { + bitField0_ |= 0x00000020; + owner_ = input.readBytes(); + break; + } + case 58: { + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder subBuilder = null; + if (((bitField0_ & 0x00000040) == 0x00000040)) { + subBuilder = usersAndPermissions_.toBuilder(); + } + usersAndPermissions_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.PARSER, extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(usersAndPermissions_); + usersAndPermissions_ = subBuilder.buildPartial(); + } + bitField0_ |= 0x00000040; + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder.class); + } + + public static com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + public SnapshotDescription parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SnapshotDescription(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + /** + * Protobuf enum {@code hbase.pb.SnapshotDescription.Type} + */ + public enum Type + implements com.google.protobuf.ProtocolMessageEnum { + /** + * DISABLED = 0; + */ + DISABLED(0, 0), + /** + * FLUSH = 1; + */ + FLUSH(1, 1), + /** + * SKIPFLUSH = 2; + */ + SKIPFLUSH(2, 2), + ; + + /** + * DISABLED = 0; + */ + public static final int DISABLED_VALUE = 0; + /** + * FLUSH = 1; + */ + public static final int FLUSH_VALUE = 1; + /** + * SKIPFLUSH = 2; + */ + public static final int SKIPFLUSH_VALUE = 2; + + + public final int getNumber() { return value; } + + public static Type valueOf(int value) { + switch (value) { + case 0: return DISABLED; + case 1: return FLUSH; + case 2: return SKIPFLUSH; + default: return null; + } + } + + public static com.google.protobuf.Internal.EnumLiteMap + internalGetValueMap() { + return internalValueMap; + } + private static com.google.protobuf.Internal.EnumLiteMap + internalValueMap = + new com.google.protobuf.Internal.EnumLiteMap() { + public Type findValueByNumber(int number) { + return Type.valueOf(number); + } + }; + + public final com.google.protobuf.Descriptors.EnumValueDescriptor + getValueDescriptor() { + return getDescriptor().getValues().get(index); + } + public final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptorForType() { + return getDescriptor(); + } + public static final com.google.protobuf.Descriptors.EnumDescriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDescriptor().getEnumTypes().get(0); + } + + private static final Type[] VALUES = values(); + + public static Type valueOf( + com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + if (desc.getType() != getDescriptor()) { + throw new java.lang.IllegalArgumentException( + "EnumValueDescriptor is not for this type."); + } + return VALUES[desc.getIndex()]; + } + + private final int index; + private final int value; + + private Type(int index, int value) { + this.index = index; + this.value = value; + } + + // @@protoc_insertion_point(enum_scope:hbase.pb.SnapshotDescription.Type) + } + + private int bitField0_; + // required string name = 1; + public static final int NAME_FIELD_NUMBER = 1; + private java.lang.Object name_; + /** + * required string name = 1; + */ + public boolean hasName() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * required string name = 1; + */ + public java.lang.String getName() { + java.lang.Object ref = name_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + name_ = s; + } + return s; + } + } + /** + * required string name = 1; + */ + public com.google.protobuf.ByteString + getNameBytes() { + java.lang.Object ref = name_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + name_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional string table = 2; + public static final int TABLE_FIELD_NUMBER = 2; + private java.lang.Object table_; + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + public boolean hasTable() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + public java.lang.String getTable() { + java.lang.Object ref = table_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + table_ = s; + } + return s; + } + } + /** + * optional string table = 2; + * + *
+     * not needed for delete, but checked for in taking snapshot
+     * 
+ */ + public com.google.protobuf.ByteString + getTableBytes() { + java.lang.Object ref = table_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + table_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional int64 creation_time = 3 [default = 0]; + public static final int CREATION_TIME_FIELD_NUMBER = 3; + private long creationTime_; + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public boolean hasCreationTime() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public long getCreationTime() { + return creationTime_; + } + + // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + public static final int TYPE_FIELD_NUMBER = 4; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type type_; + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public boolean hasType() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type getType() { + return type_; + } + + // optional int32 version = 5; + public static final int VERSION_FIELD_NUMBER = 5; + private int version_; + /** + * optional int32 version = 5; + */ + public boolean hasVersion() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + /** + * optional int32 version = 5; + */ + public int getVersion() { + return version_; + } + + // optional string owner = 6; + public static final int OWNER_FIELD_NUMBER = 6; + private java.lang.Object owner_; + /** + * optional string owner = 6; + */ + public boolean hasOwner() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + /** + * optional string owner = 6; + */ + public java.lang.String getOwner() { + java.lang.Object ref = owner_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + owner_ = s; + } + return s; + } + } + /** + * optional string owner = 6; + */ + public com.google.protobuf.ByteString + getOwnerBytes() { + java.lang.Object ref = owner_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + owner_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + public static final int USERS_AND_PERMISSIONS_FIELD_NUMBER = 7; + private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_; + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public boolean hasUsersAndPermissions() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { + return usersAndPermissions_; + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { + return usersAndPermissions_; + } + + private void initFields() { + name_ = ""; + table_ = ""; + creationTime_ = 0L; + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type.FLUSH; + version_ = 0; + owner_ = ""; + usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasName()) { + memoizedIsInitialized = 0; + return false; + } + if (hasUsersAndPermissions()) { + if (!getUsersAndPermissions().isInitialized()) { + memoizedIsInitialized = 0; + return false; + } + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, getNameBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, getTableBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeInt64(3, creationTime_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeEnum(4, type_.getNumber()); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeInt32(5, version_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeBytes(6, getOwnerBytes()); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + output.writeMessage(7, usersAndPermissions_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, getNameBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, getTableBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeInt64Size(3, creationTime_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeEnumSize(4, type_.getNumber()); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeInt32Size(5, version_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(6, getOwnerBytes()); + } + if (((bitField0_ & 0x00000040) == 0x00000040)) { + size += com.google.protobuf.CodedOutputStream + .computeMessageSize(7, usersAndPermissions_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription) obj; + + boolean result = true; + result = result && (hasName() == other.hasName()); + if (hasName()) { + result = result && getName() + .equals(other.getName()); + } + result = result && (hasTable() == other.hasTable()); + if (hasTable()) { + result = result && getTable() + .equals(other.getTable()); + } + result = result && (hasCreationTime() == other.hasCreationTime()); + if (hasCreationTime()) { + result = result && (getCreationTime() + == other.getCreationTime()); + } + result = result && (hasType() == other.hasType()); + if (hasType()) { + result = result && + (getType() == other.getType()); + } + result = result && (hasVersion() == other.hasVersion()); + if (hasVersion()) { + result = result && (getVersion() + == other.getVersion()); + } + result = result && (hasOwner() == other.hasOwner()); + if (hasOwner()) { + result = result && getOwner() + .equals(other.getOwner()); + } + result = result && (hasUsersAndPermissions() == other.hasUsersAndPermissions()); + if (hasUsersAndPermissions()) { + result = result && getUsersAndPermissions() + .equals(other.getUsersAndPermissions()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasName()) { + hash = (37 * hash) + NAME_FIELD_NUMBER; + hash = (53 * hash) + getName().hashCode(); + } + if (hasTable()) { + hash = (37 * hash) + TABLE_FIELD_NUMBER; + hash = (53 * hash) + getTable().hashCode(); + } + if (hasCreationTime()) { + hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getCreationTime()); + } + if (hasType()) { + hash = (37 * hash) + TYPE_FIELD_NUMBER; + hash = (53 * hash) + hashEnum(getType()); + } + if (hasVersion()) { + hash = (37 * hash) + VERSION_FIELD_NUMBER; + hash = (53 * hash) + getVersion(); + } + if (hasOwner()) { + hash = (37 * hash) + OWNER_FIELD_NUMBER; + hash = (53 * hash) + getOwner().hashCode(); + } + if (hasUsersAndPermissions()) { + hash = (37 * hash) + USERS_AND_PERMISSIONS_FIELD_NUMBER; + hash = (53 * hash) + getUsersAndPermissions().hashCode(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code hbase.pb.SnapshotDescription} + * + *
+     **
+     * Description of the snapshot to take
+     * 
+ */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder.class); + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + getUsersAndPermissionsFieldBuilder(); + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + name_ = ""; + bitField0_ = (bitField0_ & ~0x00000001); + table_ = ""; + bitField0_ = (bitField0_ & ~0x00000002); + creationTime_ = 0L; + bitField0_ = (bitField0_ & ~0x00000004); + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type.FLUSH; + bitField0_ = (bitField0_ & ~0x00000008); + version_ = 0; + bitField0_ = (bitField0_ & ~0x00000010); + owner_ = ""; + bitField0_ = (bitField0_ & ~0x00000020); + if (usersAndPermissionsBuilder_ == null) { + usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); + } else { + usersAndPermissionsBuilder_.clear(); + } + bitField0_ = (bitField0_ & ~0x00000040); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; + } + + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription build() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.name_ = name_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.table_ = table_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.creationTime_ = creationTime_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.type_ = type_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + to_bitField0_ |= 0x00000010; + } + result.version_ = version_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + to_bitField0_ |= 0x00000020; + } + result.owner_ = owner_; + if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + to_bitField0_ |= 0x00000040; + } + if (usersAndPermissionsBuilder_ == null) { + result.usersAndPermissions_ = usersAndPermissions_; + } else { + result.usersAndPermissions_ = usersAndPermissionsBuilder_.build(); + } + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) return this; + if (other.hasName()) { + bitField0_ |= 0x00000001; + name_ = other.name_; + onChanged(); + } + if (other.hasTable()) { + bitField0_ |= 0x00000002; + table_ = other.table_; + onChanged(); + } + if (other.hasCreationTime()) { + setCreationTime(other.getCreationTime()); + } + if (other.hasType()) { + setType(other.getType()); + } + if (other.hasVersion()) { + setVersion(other.getVersion()); + } + if (other.hasOwner()) { + bitField0_ |= 0x00000020; + owner_ = other.owner_; + onChanged(); + } + if (other.hasUsersAndPermissions()) { + mergeUsersAndPermissions(other.getUsersAndPermissions()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + if (!hasName()) { + + return false; + } + if (hasUsersAndPermissions()) { + if (!getUsersAndPermissions().isInitialized()) { + + return false; + } + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + private int bitField0_; + + // required string name = 1; + private java.lang.Object name_ = ""; + /** + * required string name = 1; + */ + public boolean hasName() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * required string name = 1; + */ + public java.lang.String getName() { + java.lang.Object ref = name_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + name_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * required string name = 1; + */ + public com.google.protobuf.ByteString + getNameBytes() { + java.lang.Object ref = name_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + name_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * required string name = 1; + */ + public Builder setName( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + name_ = value; + onChanged(); + return this; + } + /** + * required string name = 1; + */ + public Builder clearName() { + bitField0_ = (bitField0_ & ~0x00000001); + name_ = getDefaultInstance().getName(); + onChanged(); + return this; + } + /** + * required string name = 1; + */ + public Builder setNameBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + name_ = value; + onChanged(); + return this; + } + + // optional string table = 2; + private java.lang.Object table_ = ""; + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public boolean hasTable() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public java.lang.String getTable() { + java.lang.Object ref = table_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + table_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public com.google.protobuf.ByteString + getTableBytes() { + java.lang.Object ref = table_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + table_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public Builder setTable( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + table_ = value; + onChanged(); + return this; + } + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public Builder clearTable() { + bitField0_ = (bitField0_ & ~0x00000002); + table_ = getDefaultInstance().getTable(); + onChanged(); + return this; + } + /** + * optional string table = 2; + * + *
+       * not needed for delete, but checked for in taking snapshot
+       * 
+ */ + public Builder setTableBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + table_ = value; + onChanged(); + return this; + } + + // optional int64 creation_time = 3 [default = 0]; + private long creationTime_ ; + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public boolean hasCreationTime() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public long getCreationTime() { + return creationTime_; + } + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public Builder setCreationTime(long value) { + bitField0_ |= 0x00000004; + creationTime_ = value; + onChanged(); + return this; + } + /** + * optional int64 creation_time = 3 [default = 0]; + */ + public Builder clearCreationTime() { + bitField0_ = (bitField0_ & ~0x00000004); + creationTime_ = 0L; + onChanged(); + return this; + } + + // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type.FLUSH; + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public boolean hasType() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type getType() { + return type_; + } + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public Builder setType(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000008; + type_ = value; + onChanged(); + return this; + } + /** + * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; + */ + public Builder clearType() { + bitField0_ = (bitField0_ & ~0x00000008); + type_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type.FLUSH; + onChanged(); + return this; + } + + // optional int32 version = 5; + private int version_ ; + /** + * optional int32 version = 5; + */ + public boolean hasVersion() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + /** + * optional int32 version = 5; + */ + public int getVersion() { + return version_; + } + /** + * optional int32 version = 5; + */ + public Builder setVersion(int value) { + bitField0_ |= 0x00000010; + version_ = value; + onChanged(); + return this; + } + /** + * optional int32 version = 5; + */ + public Builder clearVersion() { + bitField0_ = (bitField0_ & ~0x00000010); + version_ = 0; + onChanged(); + return this; + } + + // optional string owner = 6; + private java.lang.Object owner_ = ""; + /** + * optional string owner = 6; + */ + public boolean hasOwner() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + /** + * optional string owner = 6; + */ + public java.lang.String getOwner() { + java.lang.Object ref = owner_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + owner_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * optional string owner = 6; + */ + public com.google.protobuf.ByteString + getOwnerBytes() { + java.lang.Object ref = owner_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + owner_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * optional string owner = 6; + */ + public Builder setOwner( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000020; + owner_ = value; + onChanged(); + return this; + } + /** + * optional string owner = 6; + */ + public Builder clearOwner() { + bitField0_ = (bitField0_ & ~0x00000020); + owner_ = getDefaultInstance().getOwner(); + onChanged(); + return this; + } + /** + * optional string owner = 6; + */ + public Builder setOwnerBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000020; + owner_ = value; + onChanged(); + return this; + } + + // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); + private com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> usersAndPermissionsBuilder_; + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public boolean hasUsersAndPermissions() { + return ((bitField0_ & 0x00000040) == 0x00000040); + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { + if (usersAndPermissionsBuilder_ == null) { + return usersAndPermissions_; + } else { + return usersAndPermissionsBuilder_.getMessage(); + } + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public Builder setUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { + if (usersAndPermissionsBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + usersAndPermissions_ = value; + onChanged(); + } else { + usersAndPermissionsBuilder_.setMessage(value); + } + bitField0_ |= 0x00000040; + return this; + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public Builder setUsersAndPermissions( + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder builderForValue) { + if (usersAndPermissionsBuilder_ == null) { + usersAndPermissions_ = builderForValue.build(); + onChanged(); + } else { + usersAndPermissionsBuilder_.setMessage(builderForValue.build()); + } + bitField0_ |= 0x00000040; + return this; + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public Builder mergeUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { + if (usersAndPermissionsBuilder_ == null) { + if (((bitField0_ & 0x00000040) == 0x00000040) && + usersAndPermissions_ != org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance()) { + usersAndPermissions_ = + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.newBuilder(usersAndPermissions_).mergeFrom(value).buildPartial(); + } else { + usersAndPermissions_ = value; + } + onChanged(); + } else { + usersAndPermissionsBuilder_.mergeFrom(value); + } + bitField0_ |= 0x00000040; + return this; + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public Builder clearUsersAndPermissions() { + if (usersAndPermissionsBuilder_ == null) { + usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); + onChanged(); + } else { + usersAndPermissionsBuilder_.clear(); + } + bitField0_ = (bitField0_ & ~0x00000040); + return this; + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder getUsersAndPermissionsBuilder() { + bitField0_ |= 0x00000040; + onChanged(); + return getUsersAndPermissionsFieldBuilder().getBuilder(); + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { + if (usersAndPermissionsBuilder_ != null) { + return usersAndPermissionsBuilder_.getMessageOrBuilder(); + } else { + return usersAndPermissions_; + } + } + /** + * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; + */ + private com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> + getUsersAndPermissionsFieldBuilder() { + if (usersAndPermissionsBuilder_ == null) { + usersAndPermissionsBuilder_ = new com.google.protobuf.SingleFieldBuilder< + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder>( + usersAndPermissions_, + getParentForChildren(), + isClean()); + usersAndPermissions_ = null; + } + return usersAndPermissionsBuilder_; + } + + // @@protoc_insertion_point(builder_scope:hbase.pb.SnapshotDescription) + } + + static { + defaultInstance = new SnapshotDescription(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotDescription) + } + private static com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_TableSchema_descriptor; private static @@ -17773,6 +18700,11 @@ public final class HBaseProtos { private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_hbase_pb_RegionServerInfo_fieldAccessorTable; + private static com.google.protobuf.Descriptors.Descriptor + internal_static_hbase_pb_SnapshotDescription_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { @@ -17782,204 +18714,211 @@ public final class HBaseProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\013HBase.proto\022\010hbase.pb\032\nCell.proto\"1\n\tT" + - "ableName\022\021\n\tnamespace\030\001 \002(\014\022\021\n\tqualifier" + - "\030\002 \002(\014\"\314\001\n\013TableSchema\022\'\n\ntable_name\030\001 \001" + - "(\0132\023.hbase.pb.TableName\022,\n\nattributes\030\002 " + - "\003(\0132\030.hbase.pb.BytesBytesPair\0225\n\017column_" + - "families\030\003 \003(\0132\034.hbase.pb.ColumnFamilySc" + - "hema\022/\n\rconfiguration\030\004 \003(\0132\030.hbase.pb.N" + - "ameStringPair\"\201\001\n\022ColumnFamilySchema\022\014\n\004" + - "name\030\001 \002(\014\022,\n\nattributes\030\002 \003(\0132\030.hbase.p" + - "b.BytesBytesPair\022/\n\rconfiguration\030\003 \003(\0132", - "\030.hbase.pb.NameStringPair\"\243\001\n\nRegionInfo" + - "\022\021\n\tregion_id\030\001 \002(\004\022\'\n\ntable_name\030\002 \002(\0132" + - "\023.hbase.pb.TableName\022\021\n\tstart_key\030\003 \001(\014\022" + - "\017\n\007end_key\030\004 \001(\014\022\017\n\007offline\030\005 \001(\010\022\r\n\005spl" + - "it\030\006 \001(\010\022\025\n\nreplica_id\030\007 \001(\005:\0010\":\n\014Favor" + - "edNodes\022*\n\014favored_node\030\001 \003(\0132\024.hbase.pb" + - ".ServerName\"\236\001\n\017RegionSpecifier\022;\n\004type\030" + - "\001 \002(\0162-.hbase.pb.RegionSpecifier.RegionS" + - "pecifierType\022\r\n\005value\030\002 \002(\014\"?\n\023RegionSpe" + - "cifierType\022\017\n\013REGION_NAME\020\001\022\027\n\023ENCODED_R", - "EGION_NAME\020\002\"%\n\tTimeRange\022\014\n\004from\030\001 \001(\004\022" + - "\n\n\002to\030\002 \001(\004\"W\n\025ColumnFamilyTimeRange\022\025\n\r" + - "column_family\030\001 \002(\014\022\'\n\ntime_range\030\002 \002(\0132" + - "\023.hbase.pb.TimeRange\"A\n\nServerName\022\021\n\tho" + - "st_name\030\001 \002(\t\022\014\n\004port\030\002 \001(\r\022\022\n\nstart_cod" + - "e\030\003 \001(\004\"\033\n\013Coprocessor\022\014\n\004name\030\001 \002(\t\"-\n\016" + - "NameStringPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 " + - "\002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002(\t\022\r\n\005va" + - "lue\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005first\030\001 \002" + - "(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64Pair\022\014\n\004n", - "ame\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\206\001\n\024ProcedureDe" + - "scription\022\021\n\tsignature\030\001 \002(\t\022\020\n\010instance" + - "\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\022/\n\rconf" + - "iguration\030\004 \003(\0132\030.hbase.pb.NameStringPai" + - "r\"\n\n\010EmptyMsg\"\033\n\007LongMsg\022\020\n\010long_msg\030\001 \002" + - "(\003\"\037\n\tDoubleMsg\022\022\n\ndouble_msg\030\001 \002(\001\"\'\n\rB" + - "igDecimalMsg\022\026\n\016bigdecimal_msg\030\001 \002(\014\"5\n\004" + - "UUID\022\026\n\016least_sig_bits\030\001 \002(\004\022\025\n\rmost_sig" + - "_bits\030\002 \002(\004\"T\n\023NamespaceDescriptor\022\014\n\004na" + - "me\030\001 \002(\014\022/\n\rconfiguration\030\002 \003(\0132\030.hbase.", - "pb.NameStringPair\"\235\001\n\013VersionInfo\022\017\n\007ver" + - "sion\030\001 \002(\t\022\013\n\003url\030\002 \002(\t\022\020\n\010revision\030\003 \002(" + - "\t\022\014\n\004user\030\004 \002(\t\022\014\n\004date\030\005 \002(\t\022\024\n\014src_che" + - "cksum\030\006 \002(\t\022\025\n\rversion_major\030\007 \001(\r\022\025\n\rve" + - "rsion_minor\030\010 \001(\r\"Q\n\020RegionServerInfo\022\020\n" + - "\010infoPort\030\001 \001(\005\022+\n\014version_info\030\002 \001(\0132\025." + - "hbase.pb.VersionInfo*r\n\013CompareType\022\010\n\004L" + - "ESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tN" + - "OT_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GREA" + - "TER\020\005\022\t\n\005NO_OP\020\006*n\n\010TimeUnit\022\017\n\013NANOSECO", - "NDS\020\001\022\020\n\014MICROSECONDS\020\002\022\020\n\014MILLISECONDS\020" + - "\003\022\013\n\007SECONDS\020\004\022\013\n\007MINUTES\020\005\022\t\n\005HOURS\020\006\022\010" + - "\n\004DAYS\020\007B>\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\013HBaseProtosH\001\240\001\001" + "\n\013HBase.proto\022\010hbase.pb\032\nCell.proto\032\013Tab" + + "le.proto\032\023AccessControl.proto\"\314\001\n\013TableS" + + "chema\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Tab" + + "leName\022,\n\nattributes\030\002 \003(\0132\030.hbase.pb.By" + + "tesBytesPair\0225\n\017column_families\030\003 \003(\0132\034." + + "hbase.pb.ColumnFamilySchema\022/\n\rconfigura" + + "tion\030\004 \003(\0132\030.hbase.pb.NameStringPair\"\201\001\n" + + "\022ColumnFamilySchema\022\014\n\004name\030\001 \002(\014\022,\n\natt" + + "ributes\030\002 \003(\0132\030.hbase.pb.BytesBytesPair\022" + + "/\n\rconfiguration\030\003 \003(\0132\030.hbase.pb.NameSt", + "ringPair\"\243\001\n\nRegionInfo\022\021\n\tregion_id\030\001 \002" + + "(\004\022\'\n\ntable_name\030\002 \002(\0132\023.hbase.pb.TableN" + + "ame\022\021\n\tstart_key\030\003 \001(\014\022\017\n\007end_key\030\004 \001(\014\022" + + "\017\n\007offline\030\005 \001(\010\022\r\n\005split\030\006 \001(\010\022\025\n\nrepli" + + "ca_id\030\007 \001(\005:\0010\":\n\014FavoredNodes\022*\n\014favore" + + "d_node\030\001 \003(\0132\024.hbase.pb.ServerName\"\236\001\n\017R" + + "egionSpecifier\022;\n\004type\030\001 \002(\0162-.hbase.pb." + + "RegionSpecifier.RegionSpecifierType\022\r\n\005v" + + "alue\030\002 \002(\014\"?\n\023RegionSpecifierType\022\017\n\013REG" + + "ION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002\"%\n\tT", + "imeRange\022\014\n\004from\030\001 \001(\004\022\n\n\002to\030\002 \001(\004\"W\n\025Co" + + "lumnFamilyTimeRange\022\025\n\rcolumn_family\030\001 \002" + + "(\014\022\'\n\ntime_range\030\002 \002(\0132\023.hbase.pb.TimeRa" + + "nge\"A\n\nServerName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004" + + "port\030\002 \001(\r\022\022\n\nstart_code\030\003 \001(\004\"\033\n\013Coproc" + + "essor\022\014\n\004name\030\001 \002(\t\"-\n\016NameStringPair\022\014\n" + + "\004name\030\001 \002(\t\022\r\n\005value\030\002 \002(\t\",\n\rNameBytesP" + + "air\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002 \001(\014\"/\n\016Byte" + + "sBytesPair\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002 \002(" + + "\014\",\n\rNameInt64Pair\022\014\n\004name\030\001 \001(\t\022\r\n\005valu", + "e\030\002 \001(\003\"\206\001\n\024ProcedureDescription\022\021\n\tsign" + + "ature\030\001 \002(\t\022\020\n\010instance\030\002 \001(\t\022\030\n\rcreatio" + + "n_time\030\003 \001(\003:\0010\022/\n\rconfiguration\030\004 \003(\0132\030" + + ".hbase.pb.NameStringPair\"\n\n\010EmptyMsg\"\033\n\007" + + "LongMsg\022\020\n\010long_msg\030\001 \002(\003\"\037\n\tDoubleMsg\022\022" + + "\n\ndouble_msg\030\001 \002(\001\"\'\n\rBigDecimalMsg\022\026\n\016b" + + "igdecimal_msg\030\001 \002(\014\"5\n\004UUID\022\026\n\016least_sig" + + "_bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002 \002(\004\"T\n\023Na" + + "mespaceDescriptor\022\014\n\004name\030\001 \002(\014\022/\n\rconfi" + + "guration\030\002 \003(\0132\030.hbase.pb.NameStringPair", + "\"\235\001\n\013VersionInfo\022\017\n\007version\030\001 \002(\t\022\013\n\003url" + + "\030\002 \002(\t\022\020\n\010revision\030\003 \002(\t\022\014\n\004user\030\004 \002(\t\022\014" + + "\n\004date\030\005 \002(\t\022\024\n\014src_checksum\030\006 \002(\t\022\025\n\rve" + + "rsion_major\030\007 \001(\r\022\025\n\rversion_minor\030\010 \001(\r" + + "\"Q\n\020RegionServerInfo\022\020\n\010infoPort\030\001 \001(\005\022+" + + "\n\014version_info\030\002 \001(\0132\025.hbase.pb.VersionI" + + "nfo\"\223\002\n\023SnapshotDescription\022\014\n\004name\030\001 \002(" + + "\t\022\r\n\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003 \001(\003:" + + "\0010\0227\n\004type\030\004 \001(\0162\".hbase.pb.SnapshotDesc" + + "ription.Type:\005FLUSH\022\017\n\007version\030\005 \001(\005\022\r\n\005", + "owner\030\006 \001(\t\022<\n\025users_and_permissions\030\007 \001" + + "(\0132\035.hbase.pb.UsersAndPermissions\".\n\004Typ" + + "e\022\014\n\010DISABLED\020\000\022\t\n\005FLUSH\020\001\022\r\n\tSKIPFLUSH\020" + + "\002*r\n\013CompareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQ" + + "UAL\020\001\022\t\n\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREAT" + + "ER_OR_EQUAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006*n\n" + + "\010TimeUnit\022\017\n\013NANOSECONDS\020\001\022\020\n\014MICROSECON" + + "DS\020\002\022\020\n\014MILLISECONDS\020\003\022\013\n\007SECONDS\020\004\022\013\n\007M" + + "INUTES\020\005\022\t\n\005HOURS\020\006\022\010\n\004DAYS\020\007B>\n*org.apa" + + "che.hadoop.hbase.protobuf.generatedB\013HBa", + "seProtosH\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_hbase_pb_TableName_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_hbase_pb_TableName_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_hbase_pb_TableName_descriptor, - new java.lang.String[] { "Namespace", "Qualifier", }); internal_static_hbase_pb_TableSchema_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_TableSchema_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_TableSchema_descriptor, new java.lang.String[] { "TableName", "Attributes", "ColumnFamilies", "Configuration", }); internal_static_hbase_pb_ColumnFamilySchema_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_ColumnFamilySchema_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_ColumnFamilySchema_descriptor, new java.lang.String[] { "Name", "Attributes", "Configuration", }); internal_static_hbase_pb_RegionInfo_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(2); internal_static_hbase_pb_RegionInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_RegionInfo_descriptor, new java.lang.String[] { "RegionId", "TableName", "StartKey", "EndKey", "Offline", "Split", "ReplicaId", }); internal_static_hbase_pb_FavoredNodes_descriptor = - getDescriptor().getMessageTypes().get(4); + getDescriptor().getMessageTypes().get(3); internal_static_hbase_pb_FavoredNodes_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_FavoredNodes_descriptor, new java.lang.String[] { "FavoredNode", }); internal_static_hbase_pb_RegionSpecifier_descriptor = - getDescriptor().getMessageTypes().get(5); + getDescriptor().getMessageTypes().get(4); internal_static_hbase_pb_RegionSpecifier_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_RegionSpecifier_descriptor, new java.lang.String[] { "Type", "Value", }); internal_static_hbase_pb_TimeRange_descriptor = - getDescriptor().getMessageTypes().get(6); + getDescriptor().getMessageTypes().get(5); internal_static_hbase_pb_TimeRange_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_TimeRange_descriptor, new java.lang.String[] { "From", "To", }); internal_static_hbase_pb_ColumnFamilyTimeRange_descriptor = - getDescriptor().getMessageTypes().get(7); + getDescriptor().getMessageTypes().get(6); internal_static_hbase_pb_ColumnFamilyTimeRange_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_ColumnFamilyTimeRange_descriptor, new java.lang.String[] { "ColumnFamily", "TimeRange", }); internal_static_hbase_pb_ServerName_descriptor = - getDescriptor().getMessageTypes().get(8); + getDescriptor().getMessageTypes().get(7); internal_static_hbase_pb_ServerName_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_ServerName_descriptor, new java.lang.String[] { "HostName", "Port", "StartCode", }); internal_static_hbase_pb_Coprocessor_descriptor = - getDescriptor().getMessageTypes().get(9); + getDescriptor().getMessageTypes().get(8); internal_static_hbase_pb_Coprocessor_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_Coprocessor_descriptor, new java.lang.String[] { "Name", }); internal_static_hbase_pb_NameStringPair_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(9); internal_static_hbase_pb_NameStringPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_NameStringPair_descriptor, new java.lang.String[] { "Name", "Value", }); internal_static_hbase_pb_NameBytesPair_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(10); internal_static_hbase_pb_NameBytesPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_NameBytesPair_descriptor, new java.lang.String[] { "Name", "Value", }); internal_static_hbase_pb_BytesBytesPair_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(11); internal_static_hbase_pb_BytesBytesPair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_BytesBytesPair_descriptor, new java.lang.String[] { "First", "Second", }); internal_static_hbase_pb_NameInt64Pair_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(12); internal_static_hbase_pb_NameInt64Pair_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_NameInt64Pair_descriptor, new java.lang.String[] { "Name", "Value", }); internal_static_hbase_pb_ProcedureDescription_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(13); internal_static_hbase_pb_ProcedureDescription_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_ProcedureDescription_descriptor, new java.lang.String[] { "Signature", "Instance", "CreationTime", "Configuration", }); internal_static_hbase_pb_EmptyMsg_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(14); internal_static_hbase_pb_EmptyMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_EmptyMsg_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_LongMsg_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(15); internal_static_hbase_pb_LongMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_LongMsg_descriptor, new java.lang.String[] { "LongMsg", }); internal_static_hbase_pb_DoubleMsg_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(16); internal_static_hbase_pb_DoubleMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_DoubleMsg_descriptor, new java.lang.String[] { "DoubleMsg", }); internal_static_hbase_pb_BigDecimalMsg_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(17); internal_static_hbase_pb_BigDecimalMsg_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_BigDecimalMsg_descriptor, new java.lang.String[] { "BigdecimalMsg", }); internal_static_hbase_pb_UUID_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(18); internal_static_hbase_pb_UUID_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_UUID_descriptor, new java.lang.String[] { "LeastSigBits", "MostSigBits", }); internal_static_hbase_pb_NamespaceDescriptor_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(19); internal_static_hbase_pb_NamespaceDescriptor_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_NamespaceDescriptor_descriptor, new java.lang.String[] { "Name", "Configuration", }); internal_static_hbase_pb_VersionInfo_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(20); internal_static_hbase_pb_VersionInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_VersionInfo_descriptor, new java.lang.String[] { "Version", "Url", "Revision", "User", "Date", "SrcChecksum", "VersionMajor", "VersionMinor", }); internal_static_hbase_pb_RegionServerInfo_descriptor = - getDescriptor().getMessageTypes().get(22); + getDescriptor().getMessageTypes().get(21); internal_static_hbase_pb_RegionServerInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_RegionServerInfo_descriptor, new java.lang.String[] { "InfoPort", "VersionInfo", }); + internal_static_hbase_pb_SnapshotDescription_descriptor = + getDescriptor().getMessageTypes().get(22); + internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_hbase_pb_SnapshotDescription_descriptor, + new java.lang.String[] { "Name", "Table", "CreationTime", "Type", "Version", "Owner", "UsersAndPermissions", }); return null; } }; @@ -17987,6 +18926,8 @@ public final class HBaseProtos { .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.protobuf.generated.CellProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.getDescriptor(), }, assigner); } diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProcedureProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProcedureProtos.java index d40c1f78254..6a8cd7dea69 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProcedureProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProcedureProtos.java @@ -3814,11 +3814,11 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 3; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional .hbase.pb.TableSchema table_schema = 4; /** @@ -3929,11 +3929,11 @@ public final class MasterProcedureProtos { break; } case 26: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000004) == 0x00000004)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -4045,7 +4045,7 @@ public final class MasterProcedureProtos { // optional .hbase.pb.TableName table_name = 3; public static final int TABLE_NAME_FIELD_NUMBER = 3; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 3; */ @@ -4055,13 +4055,13 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 3; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -4126,7 +4126,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); preserveSplits_ = false; - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); regionInfo_ = java.util.Collections.emptyList(); } @@ -4416,7 +4416,7 @@ public final class MasterProcedureProtos { preserveSplits_ = false; bitField0_ = (bitField0_ & ~0x00000002); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -4760,9 +4760,9 @@ public final class MasterProcedureProtos { } // optional .hbase.pb.TableName table_name = 3; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 3; */ @@ -4772,7 +4772,7 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -4782,7 +4782,7 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -4799,7 +4799,7 @@ public final class MasterProcedureProtos { * optional .hbase.pb.TableName table_name = 3; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -4812,12 +4812,12 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000004) == 0x00000004) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -4833,7 +4833,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -4844,7 +4844,7 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000004; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -4852,7 +4852,7 @@ public final class MasterProcedureProtos { /** * optional .hbase.pb.TableName table_name = 3; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -4863,11 +4863,11 @@ public final class MasterProcedureProtos { * optional .hbase.pb.TableName table_name = 3; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -5269,11 +5269,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // repeated .hbase.pb.RegionInfo region_info = 3; /** @@ -5365,11 +5365,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -5452,7 +5452,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -5462,13 +5462,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -5510,7 +5510,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); regionInfo_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; @@ -5756,7 +5756,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -6035,9 +6035,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -6047,7 +6047,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -6057,7 +6057,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -6074,7 +6074,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -6087,12 +6087,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -6108,7 +6108,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -6119,7 +6119,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -6127,7 +6127,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -6138,11 +6138,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -8492,11 +8492,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required .hbase.pb.ColumnFamilySchema columnfamily_schema = 3; /** @@ -8591,11 +8591,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -8693,7 +8693,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -8703,13 +8703,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -8759,7 +8759,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnfamilySchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); unmodifiedTableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); } @@ -9034,7 +9034,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -9314,9 +9314,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -9326,7 +9326,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -9336,7 +9336,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -9353,7 +9353,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -9366,12 +9366,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -9387,7 +9387,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -9398,7 +9398,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -9406,7 +9406,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -9417,11 +9417,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -9700,11 +9700,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required .hbase.pb.ColumnFamilySchema columnfamily_schema = 3; /** @@ -9799,11 +9799,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -9901,7 +9901,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -9911,13 +9911,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -9967,7 +9967,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnfamilySchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); unmodifiedTableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); } @@ -10242,7 +10242,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -10522,9 +10522,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -10534,7 +10534,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -10544,7 +10544,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -10561,7 +10561,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -10574,12 +10574,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -10595,7 +10595,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -10606,7 +10606,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -10614,7 +10614,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -10625,11 +10625,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -10908,11 +10908,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required bytes columnfamily_name = 3; /** @@ -11003,11 +11003,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -11097,7 +11097,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -11107,13 +11107,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -11157,7 +11157,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnfamilyName_ = com.google.protobuf.ByteString.EMPTY; unmodifiedTableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); } @@ -11427,7 +11427,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -11695,9 +11695,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -11707,7 +11707,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -11717,7 +11717,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -11734,7 +11734,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -11747,12 +11747,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -11768,7 +11768,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -11779,7 +11779,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -11787,7 +11787,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -11798,11 +11798,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -12000,11 +12000,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required bool skip_table_state_check = 3; /** @@ -12081,11 +12081,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -12162,7 +12162,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -12172,13 +12172,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -12200,7 +12200,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); skipTableStateCheck_ = false; } private byte memoizedIsInitialized = -1; @@ -12446,7 +12446,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -12691,9 +12691,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -12703,7 +12703,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -12713,7 +12713,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -12730,7 +12730,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -12743,12 +12743,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -12764,7 +12764,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -12775,7 +12775,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -12783,7 +12783,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -12794,11 +12794,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -12876,11 +12876,11 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 2; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required bool skip_table_state_check = 3; /** @@ -12957,11 +12957,11 @@ public final class MasterProcedureProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -13038,7 +13038,7 @@ public final class MasterProcedureProtos { // required .hbase.pb.TableName table_name = 2; public static final int TABLE_NAME_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -13048,13 +13048,13 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -13076,7 +13076,7 @@ public final class MasterProcedureProtos { private void initFields() { userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); skipTableStateCheck_ = false; } private byte memoizedIsInitialized = -1; @@ -13322,7 +13322,7 @@ public final class MasterProcedureProtos { } bitField0_ = (bitField0_ & ~0x00000001); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -13567,9 +13567,9 @@ public final class MasterProcedureProtos { } // required .hbase.pb.TableName table_name = 2; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 2; */ @@ -13579,7 +13579,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -13589,7 +13589,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -13606,7 +13606,7 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -13619,12 +13619,12 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -13640,7 +13640,7 @@ public final class MasterProcedureProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -13651,7 +13651,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -13659,7 +13659,7 @@ public final class MasterProcedureProtos { /** * required .hbase.pb.TableName table_name = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -13670,11 +13670,11 @@ public final class MasterProcedureProtos { * required .hbase.pb.TableName table_name = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -15392,144 +15392,145 @@ public final class MasterProcedureProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\025MasterProcedure.proto\022\010hbase.pb\032\013HBase" + - ".proto\032\tRPC.proto\"\234\001\n\024CreateTableStateDa" + - "ta\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb.UserInf" + - "ormation\022+\n\014table_schema\030\002 \002(\0132\025.hbase.p" + - "b.TableSchema\022)\n\013region_info\030\003 \003(\0132\024.hba" + - "se.pb.RegionInfo\"\332\001\n\024ModifyTableStateDat" + - "a\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb.UserInfo" + - "rmation\0226\n\027unmodified_table_schema\030\002 \001(\013" + - "2\025.hbase.pb.TableSchema\0224\n\025modified_tabl" + - "e_schema\030\003 \002(\0132\025.hbase.pb.TableSchema\022&\n", - "\036delete_column_family_in_modify\030\004 \002(\010\"\340\001" + - "\n\026TruncateTableStateData\022,\n\tuser_info\030\001 " + - "\002(\0132\031.hbase.pb.UserInformation\022\027\n\017preser" + - "ve_splits\030\002 \002(\010\022\'\n\ntable_name\030\003 \001(\0132\023.hb" + - "ase.pb.TableName\022+\n\014table_schema\030\004 \001(\0132\025" + - ".hbase.pb.TableSchema\022)\n\013region_info\030\005 \003" + - "(\0132\024.hbase.pb.RegionInfo\"\230\001\n\024DeleteTable" + - "StateData\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb." + - "UserInformation\022\'\n\ntable_name\030\002 \002(\0132\023.hb" + - "ase.pb.TableName\022)\n\013region_info\030\003 \003(\0132\024.", - "hbase.pb.RegionInfo\"W\n\030CreateNamespaceSt" + - "ateData\022;\n\024namespace_descriptor\030\001 \002(\0132\035." + - "hbase.pb.NamespaceDescriptor\"\237\001\n\030ModifyN" + - "amespaceStateData\022;\n\024namespace_descripto" + - "r\030\001 \002(\0132\035.hbase.pb.NamespaceDescriptor\022F" + - "\n\037unmodified_namespace_descriptor\030\002 \001(\0132" + - "\035.hbase.pb.NamespaceDescriptor\"o\n\030Delete" + - "NamespaceStateData\022\026\n\016namespace_name\030\001 \002" + - "(\t\022;\n\024namespace_descriptor\030\002 \001(\0132\035.hbase" + - ".pb.NamespaceDescriptor\"\344\001\n\030AddColumnFam", - "ilyStateData\022,\n\tuser_info\030\001 \002(\0132\031.hbase." + - "pb.UserInformation\022\'\n\ntable_name\030\002 \002(\0132\023" + - ".hbase.pb.TableName\0229\n\023columnfamily_sche" + - "ma\030\003 \002(\0132\034.hbase.pb.ColumnFamilySchema\0226" + - "\n\027unmodified_table_schema\030\004 \001(\0132\025.hbase." + - "pb.TableSchema\"\347\001\n\033ModifyColumnFamilySta" + - "teData\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb.Use" + - "rInformation\022\'\n\ntable_name\030\002 \002(\0132\023.hbase" + - ".pb.TableName\0229\n\023columnfamily_schema\030\003 \002" + - "(\0132\034.hbase.pb.ColumnFamilySchema\0226\n\027unmo", - "dified_table_schema\030\004 \001(\0132\025.hbase.pb.Tab" + - "leSchema\"\307\001\n\033DeleteColumnFamilyStateData" + - "\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb.UserInfor" + - "mation\022\'\n\ntable_name\030\002 \002(\0132\023.hbase.pb.Ta" + - "bleName\022\031\n\021columnfamily_name\030\003 \002(\014\0226\n\027un" + - "modified_table_schema\030\004 \001(\0132\025.hbase.pb.T" + - "ableSchema\"\215\001\n\024EnableTableStateData\022,\n\tu" + - "ser_info\030\001 \002(\0132\031.hbase.pb.UserInformatio" + - "n\022\'\n\ntable_name\030\002 \002(\0132\023.hbase.pb.TableNa" + - "me\022\036\n\026skip_table_state_check\030\003 \002(\010\"\216\001\n\025D", - "isableTableStateData\022,\n\tuser_info\030\001 \002(\0132" + - "\031.hbase.pb.UserInformation\022\'\n\ntable_name" + - "\030\002 \002(\0132\023.hbase.pb.TableName\022\036\n\026skip_tabl" + - "e_state_check\030\003 \002(\010\"\201\002\n\024ServerCrashState" + - "Data\022)\n\013server_name\030\001 \002(\0132\024.hbase.pb.Ser" + - "verName\022\036\n\026distributed_log_replay\030\002 \001(\010\022" + - "7\n\031regions_on_crashed_server\030\003 \003(\0132\024.hba" + - "se.pb.RegionInfo\022.\n\020regions_assigned\030\004 \003" + - "(\0132\024.hbase.pb.RegionInfo\022\025\n\rcarrying_met" + - "a\030\005 \001(\010\022\036\n\020should_split_wal\030\006 \001(\010:\004true*", - "\330\001\n\020CreateTableState\022\036\n\032CREATE_TABLE_PRE" + - "_OPERATION\020\001\022 \n\034CREATE_TABLE_WRITE_FS_LA" + - "YOUT\020\002\022\034\n\030CREATE_TABLE_ADD_TO_META\020\003\022\037\n\033" + - "CREATE_TABLE_ASSIGN_REGIONS\020\004\022\"\n\036CREATE_" + - "TABLE_UPDATE_DESC_CACHE\020\005\022\037\n\033CREATE_TABL" + - "E_POST_OPERATION\020\006*\207\002\n\020ModifyTableState\022" + - "\030\n\024MODIFY_TABLE_PREPARE\020\001\022\036\n\032MODIFY_TABL" + - "E_PRE_OPERATION\020\002\022(\n$MODIFY_TABLE_UPDATE" + - "_TABLE_DESCRIPTOR\020\003\022&\n\"MODIFY_TABLE_REMO" + - "VE_REPLICA_COLUMN\020\004\022!\n\035MODIFY_TABLE_DELE", - "TE_FS_LAYOUT\020\005\022\037\n\033MODIFY_TABLE_POST_OPER" + - "ATION\020\006\022#\n\037MODIFY_TABLE_REOPEN_ALL_REGIO" + - "NS\020\007*\212\002\n\022TruncateTableState\022 \n\034TRUNCATE_" + - "TABLE_PRE_OPERATION\020\001\022#\n\037TRUNCATE_TABLE_" + - "REMOVE_FROM_META\020\002\022\"\n\036TRUNCATE_TABLE_CLE" + - "AR_FS_LAYOUT\020\003\022#\n\037TRUNCATE_TABLE_CREATE_" + - "FS_LAYOUT\020\004\022\036\n\032TRUNCATE_TABLE_ADD_TO_MET" + - "A\020\005\022!\n\035TRUNCATE_TABLE_ASSIGN_REGIONS\020\006\022!" + - "\n\035TRUNCATE_TABLE_POST_OPERATION\020\007*\337\001\n\020De" + - "leteTableState\022\036\n\032DELETE_TABLE_PRE_OPERA", - "TION\020\001\022!\n\035DELETE_TABLE_REMOVE_FROM_META\020" + - "\002\022 \n\034DELETE_TABLE_CLEAR_FS_LAYOUT\020\003\022\"\n\036D" + - "ELETE_TABLE_UPDATE_DESC_CACHE\020\004\022!\n\035DELET" + - "E_TABLE_UNASSIGN_REGIONS\020\005\022\037\n\033DELETE_TAB" + - "LE_POST_OPERATION\020\006*\320\001\n\024CreateNamespaceS" + - "tate\022\034\n\030CREATE_NAMESPACE_PREPARE\020\001\022%\n!CR" + - "EATE_NAMESPACE_CREATE_DIRECTORY\020\002\022)\n%CRE" + - "ATE_NAMESPACE_INSERT_INTO_NS_TABLE\020\003\022\036\n\032" + - "CREATE_NAMESPACE_UPDATE_ZK\020\004\022(\n$CREATE_N" + - "AMESPACE_SET_NAMESPACE_QUOTA\020\005*z\n\024Modify", - "NamespaceState\022\034\n\030MODIFY_NAMESPACE_PREPA" + - "RE\020\001\022$\n MODIFY_NAMESPACE_UPDATE_NS_TABLE" + - "\020\002\022\036\n\032MODIFY_NAMESPACE_UPDATE_ZK\020\003*\332\001\n\024D" + - "eleteNamespaceState\022\034\n\030DELETE_NAMESPACE_" + - "PREPARE\020\001\022)\n%DELETE_NAMESPACE_DELETE_FRO" + - "M_NS_TABLE\020\002\022#\n\037DELETE_NAMESPACE_REMOVE_" + - "FROM_ZK\020\003\022\'\n#DELETE_NAMESPACE_DELETE_DIR" + - "ECTORIES\020\004\022+\n\'DELETE_NAMESPACE_REMOVE_NA" + - "MESPACE_QUOTA\020\005*\331\001\n\024AddColumnFamilyState" + - "\022\035\n\031ADD_COLUMN_FAMILY_PREPARE\020\001\022#\n\037ADD_C", - "OLUMN_FAMILY_PRE_OPERATION\020\002\022-\n)ADD_COLU" + - "MN_FAMILY_UPDATE_TABLE_DESCRIPTOR\020\003\022$\n A" + - "DD_COLUMN_FAMILY_POST_OPERATION\020\004\022(\n$ADD" + - "_COLUMN_FAMILY_REOPEN_ALL_REGIONS\020\005*\353\001\n\027" + - "ModifyColumnFamilyState\022 \n\034MODIFY_COLUMN" + - "_FAMILY_PREPARE\020\001\022&\n\"MODIFY_COLUMN_FAMIL" + - "Y_PRE_OPERATION\020\002\0220\n,MODIFY_COLUMN_FAMIL" + - "Y_UPDATE_TABLE_DESCRIPTOR\020\003\022\'\n#MODIFY_CO" + - "LUMN_FAMILY_POST_OPERATION\020\004\022+\n\'MODIFY_C" + - "OLUMN_FAMILY_REOPEN_ALL_REGIONS\020\005*\226\002\n\027De", - "leteColumnFamilyState\022 \n\034DELETE_COLUMN_F" + - "AMILY_PREPARE\020\001\022&\n\"DELETE_COLUMN_FAMILY_" + - "PRE_OPERATION\020\002\0220\n,DELETE_COLUMN_FAMILY_" + - "UPDATE_TABLE_DESCRIPTOR\020\003\022)\n%DELETE_COLU" + - "MN_FAMILY_DELETE_FS_LAYOUT\020\004\022\'\n#DELETE_C" + - "OLUMN_FAMILY_POST_OPERATION\020\005\022+\n\'DELETE_" + - "COLUMN_FAMILY_REOPEN_ALL_REGIONS\020\006*\350\001\n\020E" + - "nableTableState\022\030\n\024ENABLE_TABLE_PREPARE\020" + - "\001\022\036\n\032ENABLE_TABLE_PRE_OPERATION\020\002\022)\n%ENA" + - "BLE_TABLE_SET_ENABLING_TABLE_STATE\020\003\022$\n ", - "ENABLE_TABLE_MARK_REGIONS_ONLINE\020\004\022(\n$EN" + - "ABLE_TABLE_SET_ENABLED_TABLE_STATE\020\005\022\037\n\033" + - "ENABLE_TABLE_POST_OPERATION\020\006*\362\001\n\021Disabl" + - "eTableState\022\031\n\025DISABLE_TABLE_PREPARE\020\001\022\037" + - "\n\033DISABLE_TABLE_PRE_OPERATION\020\002\022+\n\'DISAB" + - "LE_TABLE_SET_DISABLING_TABLE_STATE\020\003\022&\n\"" + - "DISABLE_TABLE_MARK_REGIONS_OFFLINE\020\004\022*\n&" + - "DISABLE_TABLE_SET_DISABLED_TABLE_STATE\020\005" + - "\022 \n\034DISABLE_TABLE_POST_OPERATION\020\006*\234\002\n\020S" + - "erverCrashState\022\026\n\022SERVER_CRASH_START\020\001\022", - "\035\n\031SERVER_CRASH_PROCESS_META\020\002\022\034\n\030SERVER" + - "_CRASH_GET_REGIONS\020\003\022\036\n\032SERVER_CRASH_NO_" + - "SPLIT_LOGS\020\004\022\033\n\027SERVER_CRASH_SPLIT_LOGS\020" + - "\005\022#\n\037SERVER_CRASH_PREPARE_LOG_REPLAY\020\006\022\027" + - "\n\023SERVER_CRASH_ASSIGN\020\010\022\037\n\033SERVER_CRASH_" + - "WAIT_ON_ASSIGN\020\t\022\027\n\023SERVER_CRASH_FINISH\020" + - "dBK\n*org.apache.hadoop.hbase.protobuf.ge" + - "neratedB\025MasterProcedureProtosH\001\210\001\001\240\001\001" + "\n\025MasterProcedure.proto\022\010hbase.pb\032\013Table" + + ".proto\032\013HBase.proto\032\tRPC.proto\"\234\001\n\024Creat" + + "eTableStateData\022,\n\tuser_info\030\001 \002(\0132\031.hba" + + "se.pb.UserInformation\022+\n\014table_schema\030\002 " + + "\002(\0132\025.hbase.pb.TableSchema\022)\n\013region_inf" + + "o\030\003 \003(\0132\024.hbase.pb.RegionInfo\"\332\001\n\024Modify" + + "TableStateData\022,\n\tuser_info\030\001 \002(\0132\031.hbas" + + "e.pb.UserInformation\0226\n\027unmodified_table" + + "_schema\030\002 \001(\0132\025.hbase.pb.TableSchema\0224\n\025" + + "modified_table_schema\030\003 \002(\0132\025.hbase.pb.T", + "ableSchema\022&\n\036delete_column_family_in_mo" + + "dify\030\004 \002(\010\"\340\001\n\026TruncateTableStateData\022,\n" + + "\tuser_info\030\001 \002(\0132\031.hbase.pb.UserInformat" + + "ion\022\027\n\017preserve_splits\030\002 \002(\010\022\'\n\ntable_na" + + "me\030\003 \001(\0132\023.hbase.pb.TableName\022+\n\014table_s" + + "chema\030\004 \001(\0132\025.hbase.pb.TableSchema\022)\n\013re" + + "gion_info\030\005 \003(\0132\024.hbase.pb.RegionInfo\"\230\001" + + "\n\024DeleteTableStateData\022,\n\tuser_info\030\001 \002(" + + "\0132\031.hbase.pb.UserInformation\022\'\n\ntable_na" + + "me\030\002 \002(\0132\023.hbase.pb.TableName\022)\n\013region_", + "info\030\003 \003(\0132\024.hbase.pb.RegionInfo\"W\n\030Crea" + + "teNamespaceStateData\022;\n\024namespace_descri" + + "ptor\030\001 \002(\0132\035.hbase.pb.NamespaceDescripto" + + "r\"\237\001\n\030ModifyNamespaceStateData\022;\n\024namesp" + + "ace_descriptor\030\001 \002(\0132\035.hbase.pb.Namespac" + + "eDescriptor\022F\n\037unmodified_namespace_desc" + + "riptor\030\002 \001(\0132\035.hbase.pb.NamespaceDescrip" + + "tor\"o\n\030DeleteNamespaceStateData\022\026\n\016names" + + "pace_name\030\001 \002(\t\022;\n\024namespace_descriptor\030" + + "\002 \001(\0132\035.hbase.pb.NamespaceDescriptor\"\344\001\n", + "\030AddColumnFamilyStateData\022,\n\tuser_info\030\001" + + " \002(\0132\031.hbase.pb.UserInformation\022\'\n\ntable" + + "_name\030\002 \002(\0132\023.hbase.pb.TableName\0229\n\023colu" + + "mnfamily_schema\030\003 \002(\0132\034.hbase.pb.ColumnF" + + "amilySchema\0226\n\027unmodified_table_schema\030\004" + + " \001(\0132\025.hbase.pb.TableSchema\"\347\001\n\033ModifyCo" + + "lumnFamilyStateData\022,\n\tuser_info\030\001 \002(\0132\031" + + ".hbase.pb.UserInformation\022\'\n\ntable_name\030" + + "\002 \002(\0132\023.hbase.pb.TableName\0229\n\023columnfami" + + "ly_schema\030\003 \002(\0132\034.hbase.pb.ColumnFamilyS", + "chema\0226\n\027unmodified_table_schema\030\004 \001(\0132\025" + + ".hbase.pb.TableSchema\"\307\001\n\033DeleteColumnFa" + + "milyStateData\022,\n\tuser_info\030\001 \002(\0132\031.hbase" + + ".pb.UserInformation\022\'\n\ntable_name\030\002 \002(\0132" + + "\023.hbase.pb.TableName\022\031\n\021columnfamily_nam" + + "e\030\003 \002(\014\0226\n\027unmodified_table_schema\030\004 \001(\013" + + "2\025.hbase.pb.TableSchema\"\215\001\n\024EnableTableS" + + "tateData\022,\n\tuser_info\030\001 \002(\0132\031.hbase.pb.U" + + "serInformation\022\'\n\ntable_name\030\002 \002(\0132\023.hba" + + "se.pb.TableName\022\036\n\026skip_table_state_chec", + "k\030\003 \002(\010\"\216\001\n\025DisableTableStateData\022,\n\tuse" + + "r_info\030\001 \002(\0132\031.hbase.pb.UserInformation\022" + + "\'\n\ntable_name\030\002 \002(\0132\023.hbase.pb.TableName" + + "\022\036\n\026skip_table_state_check\030\003 \002(\010\"\201\002\n\024Ser" + + "verCrashStateData\022)\n\013server_name\030\001 \002(\0132\024" + + ".hbase.pb.ServerName\022\036\n\026distributed_log_" + + "replay\030\002 \001(\010\0227\n\031regions_on_crashed_serve" + + "r\030\003 \003(\0132\024.hbase.pb.RegionInfo\022.\n\020regions" + + "_assigned\030\004 \003(\0132\024.hbase.pb.RegionInfo\022\025\n" + + "\rcarrying_meta\030\005 \001(\010\022\036\n\020should_split_wal", + "\030\006 \001(\010:\004true*\330\001\n\020CreateTableState\022\036\n\032CRE" + + "ATE_TABLE_PRE_OPERATION\020\001\022 \n\034CREATE_TABL" + + "E_WRITE_FS_LAYOUT\020\002\022\034\n\030CREATE_TABLE_ADD_" + + "TO_META\020\003\022\037\n\033CREATE_TABLE_ASSIGN_REGIONS" + + "\020\004\022\"\n\036CREATE_TABLE_UPDATE_DESC_CACHE\020\005\022\037" + + "\n\033CREATE_TABLE_POST_OPERATION\020\006*\207\002\n\020Modi" + + "fyTableState\022\030\n\024MODIFY_TABLE_PREPARE\020\001\022\036" + + "\n\032MODIFY_TABLE_PRE_OPERATION\020\002\022(\n$MODIFY" + + "_TABLE_UPDATE_TABLE_DESCRIPTOR\020\003\022&\n\"MODI" + + "FY_TABLE_REMOVE_REPLICA_COLUMN\020\004\022!\n\035MODI", + "FY_TABLE_DELETE_FS_LAYOUT\020\005\022\037\n\033MODIFY_TA" + + "BLE_POST_OPERATION\020\006\022#\n\037MODIFY_TABLE_REO" + + "PEN_ALL_REGIONS\020\007*\212\002\n\022TruncateTableState" + + "\022 \n\034TRUNCATE_TABLE_PRE_OPERATION\020\001\022#\n\037TR" + + "UNCATE_TABLE_REMOVE_FROM_META\020\002\022\"\n\036TRUNC" + + "ATE_TABLE_CLEAR_FS_LAYOUT\020\003\022#\n\037TRUNCATE_" + + "TABLE_CREATE_FS_LAYOUT\020\004\022\036\n\032TRUNCATE_TAB" + + "LE_ADD_TO_META\020\005\022!\n\035TRUNCATE_TABLE_ASSIG" + + "N_REGIONS\020\006\022!\n\035TRUNCATE_TABLE_POST_OPERA" + + "TION\020\007*\337\001\n\020DeleteTableState\022\036\n\032DELETE_TA", + "BLE_PRE_OPERATION\020\001\022!\n\035DELETE_TABLE_REMO" + + "VE_FROM_META\020\002\022 \n\034DELETE_TABLE_CLEAR_FS_" + + "LAYOUT\020\003\022\"\n\036DELETE_TABLE_UPDATE_DESC_CAC" + + "HE\020\004\022!\n\035DELETE_TABLE_UNASSIGN_REGIONS\020\005\022" + + "\037\n\033DELETE_TABLE_POST_OPERATION\020\006*\320\001\n\024Cre" + + "ateNamespaceState\022\034\n\030CREATE_NAMESPACE_PR" + + "EPARE\020\001\022%\n!CREATE_NAMESPACE_CREATE_DIREC" + + "TORY\020\002\022)\n%CREATE_NAMESPACE_INSERT_INTO_N" + + "S_TABLE\020\003\022\036\n\032CREATE_NAMESPACE_UPDATE_ZK\020" + + "\004\022(\n$CREATE_NAMESPACE_SET_NAMESPACE_QUOT", + "A\020\005*z\n\024ModifyNamespaceState\022\034\n\030MODIFY_NA" + + "MESPACE_PREPARE\020\001\022$\n MODIFY_NAMESPACE_UP" + + "DATE_NS_TABLE\020\002\022\036\n\032MODIFY_NAMESPACE_UPDA" + + "TE_ZK\020\003*\332\001\n\024DeleteNamespaceState\022\034\n\030DELE" + + "TE_NAMESPACE_PREPARE\020\001\022)\n%DELETE_NAMESPA" + + "CE_DELETE_FROM_NS_TABLE\020\002\022#\n\037DELETE_NAME" + + "SPACE_REMOVE_FROM_ZK\020\003\022\'\n#DELETE_NAMESPA" + + "CE_DELETE_DIRECTORIES\020\004\022+\n\'DELETE_NAMESP" + + "ACE_REMOVE_NAMESPACE_QUOTA\020\005*\331\001\n\024AddColu" + + "mnFamilyState\022\035\n\031ADD_COLUMN_FAMILY_PREPA", + "RE\020\001\022#\n\037ADD_COLUMN_FAMILY_PRE_OPERATION\020" + + "\002\022-\n)ADD_COLUMN_FAMILY_UPDATE_TABLE_DESC" + + "RIPTOR\020\003\022$\n ADD_COLUMN_FAMILY_POST_OPERA" + + "TION\020\004\022(\n$ADD_COLUMN_FAMILY_REOPEN_ALL_R" + + "EGIONS\020\005*\353\001\n\027ModifyColumnFamilyState\022 \n\034" + + "MODIFY_COLUMN_FAMILY_PREPARE\020\001\022&\n\"MODIFY" + + "_COLUMN_FAMILY_PRE_OPERATION\020\002\0220\n,MODIFY" + + "_COLUMN_FAMILY_UPDATE_TABLE_DESCRIPTOR\020\003" + + "\022\'\n#MODIFY_COLUMN_FAMILY_POST_OPERATION\020" + + "\004\022+\n\'MODIFY_COLUMN_FAMILY_REOPEN_ALL_REG", + "IONS\020\005*\226\002\n\027DeleteColumnFamilyState\022 \n\034DE" + + "LETE_COLUMN_FAMILY_PREPARE\020\001\022&\n\"DELETE_C" + + "OLUMN_FAMILY_PRE_OPERATION\020\002\0220\n,DELETE_C" + + "OLUMN_FAMILY_UPDATE_TABLE_DESCRIPTOR\020\003\022)" + + "\n%DELETE_COLUMN_FAMILY_DELETE_FS_LAYOUT\020" + + "\004\022\'\n#DELETE_COLUMN_FAMILY_POST_OPERATION" + + "\020\005\022+\n\'DELETE_COLUMN_FAMILY_REOPEN_ALL_RE" + + "GIONS\020\006*\350\001\n\020EnableTableState\022\030\n\024ENABLE_T" + + "ABLE_PREPARE\020\001\022\036\n\032ENABLE_TABLE_PRE_OPERA" + + "TION\020\002\022)\n%ENABLE_TABLE_SET_ENABLING_TABL", + "E_STATE\020\003\022$\n ENABLE_TABLE_MARK_REGIONS_O" + + "NLINE\020\004\022(\n$ENABLE_TABLE_SET_ENABLED_TABL" + + "E_STATE\020\005\022\037\n\033ENABLE_TABLE_POST_OPERATION" + + "\020\006*\362\001\n\021DisableTableState\022\031\n\025DISABLE_TABL" + + "E_PREPARE\020\001\022\037\n\033DISABLE_TABLE_PRE_OPERATI" + + "ON\020\002\022+\n\'DISABLE_TABLE_SET_DISABLING_TABL" + + "E_STATE\020\003\022&\n\"DISABLE_TABLE_MARK_REGIONS_" + + "OFFLINE\020\004\022*\n&DISABLE_TABLE_SET_DISABLED_" + + "TABLE_STATE\020\005\022 \n\034DISABLE_TABLE_POST_OPER" + + "ATION\020\006*\234\002\n\020ServerCrashState\022\026\n\022SERVER_C", + "RASH_START\020\001\022\035\n\031SERVER_CRASH_PROCESS_MET" + + "A\020\002\022\034\n\030SERVER_CRASH_GET_REGIONS\020\003\022\036\n\032SER" + + "VER_CRASH_NO_SPLIT_LOGS\020\004\022\033\n\027SERVER_CRAS" + + "H_SPLIT_LOGS\020\005\022#\n\037SERVER_CRASH_PREPARE_L" + + "OG_REPLAY\020\006\022\027\n\023SERVER_CRASH_ASSIGN\020\010\022\037\n\033" + + "SERVER_CRASH_WAIT_ON_ASSIGN\020\t\022\027\n\023SERVER_" + + "CRASH_FINISH\020dBK\n*org.apache.hadoop.hbas" + + "e.protobuf.generatedB\025MasterProcedurePro" + + "tosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -15620,6 +15621,7 @@ public final class MasterProcedureProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.RPCProtos.getDescriptor(), }, assigner); diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java index 2661dc1c499..b3ae957f624 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java @@ -101,11 +101,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required .hbase.pb.ColumnFamilySchema column_families = 2; /** @@ -193,11 +193,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -270,7 +270,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -280,13 +280,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -345,7 +345,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; @@ -599,7 +599,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -738,9 +738,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -750,7 +750,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -760,7 +760,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -777,7 +777,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -790,12 +790,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -811,7 +811,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -822,7 +822,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -830,7 +830,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -841,11 +841,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -1397,11 +1397,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required bytes column_name = 2; /** @@ -1485,11 +1485,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -1554,7 +1554,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -1564,13 +1564,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -1623,7 +1623,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnName_ = com.google.protobuf.ByteString.EMPTY; nonceGroup_ = 0L; nonce_ = 0L; @@ -1872,7 +1872,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -1999,9 +1999,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2011,7 +2011,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -2021,7 +2021,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2038,7 +2038,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -2051,12 +2051,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -2072,7 +2072,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -2083,7 +2083,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -2091,7 +2091,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -2102,11 +2102,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -2577,11 +2577,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required .hbase.pb.ColumnFamilySchema column_families = 2; /** @@ -2669,11 +2669,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -2746,7 +2746,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2756,13 +2756,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -2821,7 +2821,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; @@ -3075,7 +3075,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -3214,9 +3214,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -3226,7 +3226,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -3236,7 +3236,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -3253,7 +3253,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -3266,12 +3266,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -3287,7 +3287,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -3298,7 +3298,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -3306,7 +3306,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -3317,11 +3317,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -10319,11 +10319,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional uint64 nonce_group = 2 [default = 0]; /** @@ -10397,11 +10397,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -10461,7 +10461,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -10471,13 +10471,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -10514,7 +10514,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; } @@ -10742,7 +10742,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -10856,9 +10856,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -10868,7 +10868,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -10878,7 +10878,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -10895,7 +10895,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -10908,12 +10908,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -10929,7 +10929,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -10940,7 +10940,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -10948,7 +10948,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -10959,11 +10959,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -11494,11 +11494,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName tableName = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional bool preserveSplits = 2 [default = false]; /** @@ -11582,11 +11582,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -11651,7 +11651,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName tableName = 1; public static final int TABLENAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName tableName = 1; */ @@ -11661,13 +11661,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -11720,7 +11720,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); preserveSplits_ = false; nonceGroup_ = 0L; nonce_ = 0L; @@ -11965,7 +11965,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -12088,9 +12088,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName tableName = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName tableName = 1; */ @@ -12100,7 +12100,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -12110,7 +12110,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -12127,7 +12127,7 @@ public final class MasterProtos { * required .hbase.pb.TableName tableName = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -12140,12 +12140,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -12161,7 +12161,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -12172,7 +12172,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -12180,7 +12180,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -12191,11 +12191,11 @@ public final class MasterProtos { * required .hbase.pb.TableName tableName = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -12663,11 +12663,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional uint64 nonce_group = 2 [default = 0]; /** @@ -12741,11 +12741,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -12805,7 +12805,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -12815,13 +12815,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -12858,7 +12858,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; } @@ -13086,7 +13086,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -13200,9 +13200,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -13212,7 +13212,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -13222,7 +13222,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -13239,7 +13239,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -13252,12 +13252,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -13273,7 +13273,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -13284,7 +13284,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -13292,7 +13292,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -13303,11 +13303,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -13838,11 +13838,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional uint64 nonce_group = 2 [default = 0]; /** @@ -13916,11 +13916,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -13980,7 +13980,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -13990,13 +13990,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -14033,7 +14033,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; } @@ -14261,7 +14261,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -14375,9 +14375,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -14387,7 +14387,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -14397,7 +14397,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -14414,7 +14414,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -14427,12 +14427,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -14448,7 +14448,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -14459,7 +14459,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -14467,7 +14467,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -14478,11 +14478,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -15013,11 +15013,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required .hbase.pb.TableSchema table_schema = 2; /** @@ -15105,11 +15105,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -15182,7 +15182,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -15192,13 +15192,13 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -15257,7 +15257,7 @@ public final class MasterProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); tableSchema_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance(); nonceGroup_ = 0L; nonce_ = 0L; @@ -15511,7 +15511,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -15650,9 +15650,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -15662,7 +15662,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -15672,7 +15672,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -15689,7 +15689,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -15702,12 +15702,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -15723,7 +15723,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -15734,7 +15734,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -15742,7 +15742,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -15753,11 +15753,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -23390,12 +23390,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - java.util.List + java.util.List getTableNameList(); /** * repeated .hbase.pb.TableName tableName = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index); /** * repeated .hbase.pb.TableName tableName = 1; */ @@ -23403,12 +23403,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - java.util.List + java.util.List getTableNameOrBuilderList(); /** * repeated .hbase.pb.TableName tableName = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder( int index); } /** @@ -23464,10 +23464,10 @@ public final class MasterProtos { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - tableName_ = new java.util.ArrayList(); + tableName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); + tableName_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry)); break; } } @@ -23514,17 +23514,17 @@ public final class MasterProtos { // repeated .hbase.pb.TableName tableName = 1; public static final int TABLENAME_FIELD_NUMBER = 1; - private java.util.List tableName_; + private java.util.List tableName_; /** * repeated .hbase.pb.TableName tableName = 1; */ - public java.util.List getTableNameList() { + public java.util.List getTableNameList() { return tableName_; } /** * repeated .hbase.pb.TableName tableName = 1; */ - public java.util.List + public java.util.List getTableNameOrBuilderList() { return tableName_; } @@ -23537,13 +23537,13 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) { return tableName_.get(index); } /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder( int index) { return tableName_.get(index); } @@ -23854,22 +23854,22 @@ public final class MasterProtos { private int bitField0_; // repeated .hbase.pb.TableName tableName = 1; - private java.util.List tableName_ = + private java.util.List tableName_ = java.util.Collections.emptyList(); private void ensureTableNameIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - tableName_ = new java.util.ArrayList(tableName_); + tableName_ = new java.util.ArrayList(tableName_); bitField0_ |= 0x00000001; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * repeated .hbase.pb.TableName tableName = 1; */ - public java.util.List getTableNameList() { + public java.util.List getTableNameList() { if (tableNameBuilder_ == null) { return java.util.Collections.unmodifiableList(tableName_); } else { @@ -23889,7 +23889,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(int index) { if (tableNameBuilder_ == null) { return tableName_.get(index); } else { @@ -23900,7 +23900,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder setTableName( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -23917,7 +23917,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder setTableName( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.set(index, builderForValue.build()); @@ -23930,7 +23930,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder addTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -23947,7 +23947,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -23964,7 +23964,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.add(builderForValue.build()); @@ -23978,7 +23978,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.add(index, builderForValue.build()); @@ -23992,7 +23992,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName tableName = 1; */ public Builder addAllTableName( - java.lang.Iterable values) { + java.lang.Iterable values) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); super.addAll(values, tableName_); @@ -24031,14 +24031,14 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder( int index) { return getTableNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder( int index) { if (tableNameBuilder_ == null) { return tableName_.get(index); } else { @@ -24048,7 +24048,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public java.util.List + public java.util.List getTableNameOrBuilderList() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilderList(); @@ -24059,31 +24059,31 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder() { return getTableNameFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName tableName = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNameBuilder( int index) { return getTableNameFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName tableName = 1; */ - public java.util.List + public java.util.List getTableNameBuilderList() { return getTableNameFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), @@ -36539,7 +36539,7 @@ public final class MasterProtos { public final boolean isInitialized() { if (!hasCleanerChoreRan()) { - + return false; } return true; @@ -36981,7 +36981,7 @@ public final class MasterProtos { public final boolean isInitialized() { if (!hasOn()) { - + return false; } return true; @@ -38195,7 +38195,7 @@ public final class MasterProtos { public final boolean isInitialized() { if (!hasValue()) { - + return false; } return true; @@ -38275,11 +38275,11 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.SnapshotRequest} @@ -38333,11 +38333,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -38387,7 +38387,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; public static final int SNAPSHOT_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -38397,18 +38397,18 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } private void initFields() { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -38602,7 +38602,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -38698,9 +38698,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -38710,7 +38710,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -38720,7 +38720,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -38737,7 +38737,7 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -38750,12 +38750,12 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -38771,7 +38771,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -38782,7 +38782,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -38790,7 +38790,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -38801,11 +38801,11 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -39612,12 +39612,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - java.util.List + java.util.List getSnapshotsList(); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshots(int index); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @@ -39625,12 +39625,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - java.util.List + java.util.List getSnapshotsOrBuilderList(); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index); } /** @@ -39686,10 +39686,10 @@ public final class MasterProtos { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - snapshots_ = new java.util.ArrayList(); + snapshots_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - snapshots_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry)); + snapshots_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry)); break; } } @@ -39736,17 +39736,17 @@ public final class MasterProtos { // repeated .hbase.pb.SnapshotDescription snapshots = 1; public static final int SNAPSHOTS_FIELD_NUMBER = 1; - private java.util.List snapshots_; + private java.util.List snapshots_; /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public java.util.List getSnapshotsList() { + public java.util.List getSnapshotsList() { return snapshots_; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public java.util.List + public java.util.List getSnapshotsOrBuilderList() { return snapshots_; } @@ -39759,13 +39759,13 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index) { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshots(int index) { return snapshots_.get(index); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index) { return snapshots_.get(index); } @@ -40076,22 +40076,22 @@ public final class MasterProtos { private int bitField0_; // repeated .hbase.pb.SnapshotDescription snapshots = 1; - private java.util.List snapshots_ = + private java.util.List snapshots_ = java.util.Collections.emptyList(); private void ensureSnapshotsIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - snapshots_ = new java.util.ArrayList(snapshots_); + snapshots_ = new java.util.ArrayList(snapshots_); bitField0_ |= 0x00000001; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotsBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotsBuilder_; /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public java.util.List getSnapshotsList() { + public java.util.List getSnapshotsList() { if (snapshotsBuilder_ == null) { return java.util.Collections.unmodifiableList(snapshots_); } else { @@ -40111,7 +40111,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index) { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshots(int index) { if (snapshotsBuilder_ == null) { return snapshots_.get(index); } else { @@ -40122,7 +40122,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder setSnapshots( - int index, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -40139,7 +40139,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder setSnapshots( - int index, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.set(index, builderForValue.build()); @@ -40152,7 +40152,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public Builder addSnapshots(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder addSnapshots(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -40169,7 +40169,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( - int index, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -40186,7 +40186,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.add(builderForValue.build()); @@ -40200,7 +40200,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( - int index, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.add(index, builderForValue.build()); @@ -40214,7 +40214,7 @@ public final class MasterProtos { * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addAllSnapshots( - java.lang.Iterable values) { + java.lang.Iterable values) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); super.addAll(values, snapshots_); @@ -40253,14 +40253,14 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotsBuilder( + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotsBuilder( int index) { return getSnapshotsFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index) { if (snapshotsBuilder_ == null) { return snapshots_.get(index); } else { @@ -40270,7 +40270,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public java.util.List + public java.util.List getSnapshotsOrBuilderList() { if (snapshotsBuilder_ != null) { return snapshotsBuilder_.getMessageOrBuilderList(); @@ -40281,31 +40281,31 @@ public final class MasterProtos { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder addSnapshotsBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder addSnapshotsBuilder() { return getSnapshotsFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder addSnapshotsBuilder( + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder addSnapshotsBuilder( int index) { return getSnapshotsFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ - public java.util.List + public java.util.List getSnapshotsBuilderList() { return getSnapshotsFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotsFieldBuilder() { if (snapshotsBuilder_ == null) { snapshotsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshots_, ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), @@ -40337,11 +40337,11 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.DeleteSnapshotRequest} @@ -40395,11 +40395,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -40449,7 +40449,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; public static final int SNAPSHOT_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -40459,18 +40459,18 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } private void initFields() { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -40664,7 +40664,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -40760,9 +40760,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -40772,7 +40772,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -40782,7 +40782,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -40799,7 +40799,7 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -40812,12 +40812,12 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -40833,7 +40833,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -40844,7 +40844,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -40852,7 +40852,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -40863,11 +40863,11 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -41236,11 +41236,11 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); // optional bool restoreACL = 2 [default = false]; /** @@ -41304,11 +41304,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -41363,7 +41363,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; public static final int SNAPSHOT_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -41373,13 +41373,13 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } @@ -41400,7 +41400,7 @@ public final class MasterProtos { } private void initFields() { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); restoreACL_ = false; } private byte memoizedIsInitialized = -1; @@ -41611,7 +41611,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -41716,9 +41716,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.SnapshotDescription snapshot = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -41728,7 +41728,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -41738,7 +41738,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -41755,7 +41755,7 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -41768,12 +41768,12 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -41789,7 +41789,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -41800,7 +41800,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -41808,7 +41808,7 @@ public final class MasterProtos { /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -41819,11 +41819,11 @@ public final class MasterProtos { * required .hbase.pb.SnapshotDescription snapshot = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -42225,11 +42225,11 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsSnapshotDoneRequest} @@ -42288,11 +42288,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -42342,7 +42342,7 @@ public final class MasterProtos { private int bitField0_; // optional .hbase.pb.SnapshotDescription snapshot = 1; public static final int SNAPSHOT_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -42352,18 +42352,18 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } private void initFields() { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -42560,7 +42560,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -42654,9 +42654,9 @@ public final class MasterProtos { private int bitField0_; // optional .hbase.pb.SnapshotDescription snapshot = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -42666,7 +42666,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -42676,7 +42676,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -42693,7 +42693,7 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -42706,12 +42706,12 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -42727,7 +42727,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -42738,7 +42738,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -42746,7 +42746,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -42757,11 +42757,11 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -42802,11 +42802,11 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsSnapshotDoneResponse} @@ -42865,11 +42865,11 @@ public final class MasterProtos { break; } case 18: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -42935,7 +42935,7 @@ public final class MasterProtos { // optional .hbase.pb.SnapshotDescription snapshot = 2; public static final int SNAPSHOT_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ @@ -42945,19 +42945,19 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } private void initFields() { done_ = false; - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -43167,7 +43167,7 @@ public final class MasterProtos { done_ = false; bitField0_ = (bitField0_ & ~0x00000001); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -43301,9 +43301,9 @@ public final class MasterProtos { } // optional .hbase.pb.SnapshotDescription snapshot = 2; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ @@ -43313,7 +43313,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -43323,7 +43323,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -43340,7 +43340,7 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -43353,12 +43353,12 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -43374,7 +43374,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -43385,7 +43385,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000002; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -43393,7 +43393,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -43404,11 +43404,11 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 2; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -43439,11 +43439,11 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsRestoreSnapshotDoneRequest} @@ -43497,11 +43497,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = snapshot_.toBuilder(); } - snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); + snapshot_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); @@ -43551,7 +43551,7 @@ public final class MasterProtos { private int bitField0_; // optional .hbase.pb.SnapshotDescription snapshot = 1; public static final int SNAPSHOT_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -43561,18 +43561,18 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { return snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_; } private void initFields() { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -43764,7 +43764,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); } else { snapshotBuilder_.clear(); } @@ -43858,9 +43858,9 @@ public final class MasterProtos { private int bitField0_; // optional .hbase.pb.SnapshotDescription snapshot = 1; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @@ -43870,7 +43870,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_; } else { @@ -43880,7 +43880,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder setSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -43897,7 +43897,7 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); @@ -43910,12 +43910,12 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription value) { + public Builder mergeSnapshot(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - snapshot_ != org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { + snapshot_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } @@ -43931,7 +43931,7 @@ public final class MasterProtos { */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { - snapshot_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); + snapshot_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.getDefaultInstance(); onChanged(); } else { snapshotBuilder_.clear(); @@ -43942,7 +43942,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); @@ -43950,7 +43950,7 @@ public final class MasterProtos { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { @@ -43961,11 +43961,11 @@ public final class MasterProtos { * optional .hbase.pb.SnapshotDescription snapshot = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescriptionOrBuilder>( snapshot_, getParentForChildren(), isClean()); @@ -44430,11 +44430,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetSchemaAlterStatusRequest} @@ -44488,11 +44488,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -44542,7 +44542,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -44552,18 +44552,18 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -44757,7 +44757,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -44853,9 +44853,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -44865,7 +44865,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -44875,7 +44875,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -44892,7 +44892,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -44905,12 +44905,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -44926,7 +44926,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -44937,7 +44937,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -44945,7 +44945,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -44956,11 +44956,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -45511,12 +45511,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - java.util.List + java.util.List getTableNamesList(); /** * repeated .hbase.pb.TableName table_names = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index); /** * repeated .hbase.pb.TableName table_names = 1; */ @@ -45524,12 +45524,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - java.util.List + java.util.List getTableNamesOrBuilderList(); /** * repeated .hbase.pb.TableName table_names = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index); // optional string regex = 2; @@ -45625,10 +45625,10 @@ public final class MasterProtos { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - tableNames_ = new java.util.ArrayList(); + tableNames_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - tableNames_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); + tableNames_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry)); break; } case 18: { @@ -45691,17 +45691,17 @@ public final class MasterProtos { private int bitField0_; // repeated .hbase.pb.TableName table_names = 1; public static final int TABLE_NAMES_FIELD_NUMBER = 1; - private java.util.List tableNames_; + private java.util.List tableNames_; /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List getTableNamesList() { + public java.util.List getTableNamesList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesOrBuilderList() { return tableNames_; } @@ -45714,13 +45714,13 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index) { return tableNames_.get(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { return tableNames_.get(index); } @@ -46217,22 +46217,22 @@ public final class MasterProtos { private int bitField0_; // repeated .hbase.pb.TableName table_names = 1; - private java.util.List tableNames_ = + private java.util.List tableNames_ = java.util.Collections.emptyList(); private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - tableNames_ = new java.util.ArrayList(tableNames_); + tableNames_ = new java.util.ArrayList(tableNames_); bitField0_ |= 0x00000001; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNamesBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNamesBuilder_; /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List getTableNamesList() { + public java.util.List getTableNamesList() { if (tableNamesBuilder_ == null) { return java.util.Collections.unmodifiableList(tableNames_); } else { @@ -46252,7 +46252,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { @@ -46263,7 +46263,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -46280,7 +46280,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.set(index, builderForValue.build()); @@ -46293,7 +46293,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public Builder addTableNames(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder addTableNames(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -46310,7 +46310,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -46327,7 +46327,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(builderForValue.build()); @@ -46341,7 +46341,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(index, builderForValue.build()); @@ -46355,7 +46355,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addAllTableNames( - java.lang.Iterable values) { + java.lang.Iterable values) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); super.addAll(values, tableNames_); @@ -46394,14 +46394,14 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNamesBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNamesBuilder( int index) { return getTableNamesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { @@ -46411,7 +46411,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesOrBuilderList() { if (tableNamesBuilder_ != null) { return tableNamesBuilder_.getMessageOrBuilderList(); @@ -46422,31 +46422,31 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNamesBuilder() { return getTableNamesFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNamesBuilder( int index) { return getTableNamesFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesBuilderList() { return getTableNamesFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNamesFieldBuilder() { if (tableNamesBuilder_ == null) { tableNamesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableNames_, ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), @@ -48140,12 +48140,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - java.util.List + java.util.List getTableNamesList(); /** * repeated .hbase.pb.TableName table_names = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index); /** * repeated .hbase.pb.TableName table_names = 1; */ @@ -48153,12 +48153,12 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - java.util.List + java.util.List getTableNamesOrBuilderList(); /** * repeated .hbase.pb.TableName table_names = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index); } /** @@ -48214,10 +48214,10 @@ public final class MasterProtos { } case 10: { if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - tableNames_ = new java.util.ArrayList(); + tableNames_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - tableNames_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); + tableNames_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry)); break; } } @@ -48264,17 +48264,17 @@ public final class MasterProtos { // repeated .hbase.pb.TableName table_names = 1; public static final int TABLE_NAMES_FIELD_NUMBER = 1; - private java.util.List tableNames_; + private java.util.List tableNames_; /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List getTableNamesList() { + public java.util.List getTableNamesList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesOrBuilderList() { return tableNames_; } @@ -48287,13 +48287,13 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index) { return tableNames_.get(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { return tableNames_.get(index); } @@ -48604,22 +48604,22 @@ public final class MasterProtos { private int bitField0_; // repeated .hbase.pb.TableName table_names = 1; - private java.util.List tableNames_ = + private java.util.List tableNames_ = java.util.Collections.emptyList(); private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) == 0x00000001)) { - tableNames_ = new java.util.ArrayList(tableNames_); + tableNames_ = new java.util.ArrayList(tableNames_); bitField0_ |= 0x00000001; } } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNamesBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNamesBuilder_; /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List getTableNamesList() { + public java.util.List getTableNamesList() { if (tableNamesBuilder_ == null) { return java.util.Collections.unmodifiableList(tableNames_); } else { @@ -48639,7 +48639,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableNames(int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { @@ -48650,7 +48650,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -48667,7 +48667,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.set(index, builderForValue.build()); @@ -48680,7 +48680,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public Builder addTableNames(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder addTableNames(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -48697,7 +48697,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -48714,7 +48714,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(builderForValue.build()); @@ -48728,7 +48728,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( - int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + int index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(index, builderForValue.build()); @@ -48742,7 +48742,7 @@ public final class MasterProtos { * repeated .hbase.pb.TableName table_names = 1; */ public Builder addAllTableNames( - java.lang.Iterable values) { + java.lang.Iterable values) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); super.addAll(values, tableNames_); @@ -48781,14 +48781,14 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNamesBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNamesBuilder( int index) { return getTableNamesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { @@ -48798,7 +48798,7 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesOrBuilderList() { if (tableNamesBuilder_ != null) { return tableNamesBuilder_.getMessageOrBuilderList(); @@ -48809,31 +48809,31 @@ public final class MasterProtos { /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNamesBuilder() { return getTableNamesFieldBuilder().addBuilder( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder( + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder addTableNamesBuilder( int index) { return getTableNamesFieldBuilder().addBuilder( - index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); + index, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ - public java.util.List + public java.util.List getTableNamesBuilderList() { return getTableNamesFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNamesFieldBuilder() { if (tableNamesBuilder_ == null) { tableNamesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableNames_, ((bitField0_ & 0x00000001) == 0x00000001), getParentForChildren(), @@ -56372,11 +56372,11 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 4; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional bool remove_all = 5; /** @@ -56479,11 +56479,11 @@ public final class MasterProtos { break; } case 34: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000008) == 0x00000008)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -56685,7 +56685,7 @@ public final class MasterProtos { // optional .hbase.pb.TableName table_name = 4; public static final int TABLE_NAME_FIELD_NUMBER = 4; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 4; */ @@ -56695,13 +56695,13 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 4; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -56763,7 +56763,7 @@ public final class MasterProtos { userName_ = ""; userGroup_ = ""; namespace_ = ""; - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); removeAll_ = false; bypassGlobals_ = false; throttle_ = org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance(); @@ -57067,7 +57067,7 @@ public final class MasterProtos { namespace_ = ""; bitField0_ = (bitField0_ & ~0x00000004); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -57451,9 +57451,9 @@ public final class MasterProtos { } // optional .hbase.pb.TableName table_name = 4; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 4; */ @@ -57463,7 +57463,7 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -57473,7 +57473,7 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -57490,7 +57490,7 @@ public final class MasterProtos { * optional .hbase.pb.TableName table_name = 4; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -57503,12 +57503,12 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -57524,7 +57524,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -57535,7 +57535,7 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000008; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -57543,7 +57543,7 @@ public final class MasterProtos { /** * optional .hbase.pb.TableName table_name = 4; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -57554,11 +57554,11 @@ public final class MasterProtos { * optional .hbase.pb.TableName table_name = 4; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -58110,11 +58110,11 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampRequest} @@ -58168,11 +58168,11 @@ public final class MasterProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -58222,7 +58222,7 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -58232,18 +58232,18 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -58437,7 +58437,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -58533,9 +58533,9 @@ public final class MasterProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -58545,7 +58545,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -58555,7 +58555,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -58572,7 +58572,7 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -58585,12 +58585,12 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -58606,7 +58606,7 @@ public final class MasterProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -58617,7 +58617,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -58625,7 +58625,7 @@ public final class MasterProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -58636,11 +58636,11 @@ public final class MasterProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -66180,342 +66180,342 @@ public final class MasterProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\014Master.proto\022\010hbase.pb\032\013HBase.proto\032\014C" + - "lient.proto\032\023ClusterStatus.proto\032\023ErrorH" + - "andling.proto\032\017Procedure.proto\032\013Quota.pr" + - "oto\032\016Snapshot.proto\"\234\001\n\020AddColumnRequest" + - "\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.TableNam" + - "e\0225\n\017column_families\030\002 \002(\0132\034.hbase.pb.Co" + - "lumnFamilySchema\022\026\n\013nonce_group\030\003 \001(\004:\0010" + - "\022\020\n\005nonce\030\004 \001(\004:\0010\"\023\n\021AddColumnResponse\"" + - "}\n\023DeleteColumnRequest\022\'\n\ntable_name\030\001 \002" + - "(\0132\023.hbase.pb.TableName\022\023\n\013column_name\030\002", - " \002(\014\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 " + - "\001(\004:\0010\"\026\n\024DeleteColumnResponse\"\237\001\n\023Modif" + - "yColumnRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hba" + - "se.pb.TableName\0225\n\017column_families\030\002 \002(\013" + - "2\034.hbase.pb.ColumnFamilySchema\022\026\n\013nonce_" + - "group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"\026\n\024Modi" + - "fyColumnResponse\"n\n\021MoveRegionRequest\022)\n" + - "\006region\030\001 \002(\0132\031.hbase.pb.RegionSpecifier" + - "\022.\n\020dest_server_name\030\002 \001(\0132\024.hbase.pb.Se" + - "rverName\"\024\n\022MoveRegionResponse\"\222\001\n\035Dispa", - "tchMergingRegionsRequest\022+\n\010region_a\030\001 \002" + - "(\0132\031.hbase.pb.RegionSpecifier\022+\n\010region_" + - "b\030\002 \002(\0132\031.hbase.pb.RegionSpecifier\022\027\n\010fo" + - "rcible\030\003 \001(\010:\005false\" \n\036DispatchMergingRe" + - "gionsResponse\"@\n\023AssignRegionRequest\022)\n\006" + - "region\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\"" + - "\026\n\024AssignRegionResponse\"X\n\025UnassignRegio" + + "\n\014Master.proto\022\010hbase.pb\032\013Table.proto\032\013H" + + "Base.proto\032\014Client.proto\032\023ClusterStatus." + + "proto\032\023ErrorHandling.proto\032\017Procedure.pr" + + "oto\032\013Quota.proto\032\016Snapshot.proto\"\234\001\n\020Add" + + "ColumnRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbas" + + "e.pb.TableName\0225\n\017column_families\030\002 \002(\0132" + + "\034.hbase.pb.ColumnFamilySchema\022\026\n\013nonce_g" + + "roup\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"\023\n\021AddCo" + + "lumnResponse\"}\n\023DeleteColumnRequest\022\'\n\nt" + + "able_name\030\001 \002(\0132\023.hbase.pb.TableName\022\023\n\013", + "column_name\030\002 \002(\014\022\026\n\013nonce_group\030\003 \001(\004:\001" + + "0\022\020\n\005nonce\030\004 \001(\004:\0010\"\026\n\024DeleteColumnRespo" + + "nse\"\237\001\n\023ModifyColumnRequest\022\'\n\ntable_nam" + + "e\030\001 \002(\0132\023.hbase.pb.TableName\0225\n\017column_f" + + "amilies\030\002 \002(\0132\034.hbase.pb.ColumnFamilySch" + + "ema\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001" + + "(\004:\0010\"\026\n\024ModifyColumnResponse\"n\n\021MoveReg" + + "ionRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Re" + + "gionSpecifier\022.\n\020dest_server_name\030\002 \001(\0132" + + "\024.hbase.pb.ServerName\"\024\n\022MoveRegionRespo", + "nse\"\222\001\n\035DispatchMergingRegionsRequest\022+\n" + + "\010region_a\030\001 \002(\0132\031.hbase.pb.RegionSpecifi" + + "er\022+\n\010region_b\030\002 \002(\0132\031.hbase.pb.RegionSp" + + "ecifier\022\027\n\010forcible\030\003 \001(\010:\005false\" \n\036Disp" + + "atchMergingRegionsResponse\"@\n\023AssignRegi" + + "onRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Reg" + + "ionSpecifier\"\026\n\024AssignRegionResponse\"X\n\025" + + "UnassignRegionRequest\022)\n\006region\030\001 \002(\0132\031." + + "hbase.pb.RegionSpecifier\022\024\n\005force\030\002 \001(\010:" + + "\005false\"\030\n\026UnassignRegionResponse\"A\n\024Offl", + "ineRegionRequest\022)\n\006region\030\001 \002(\0132\031.hbase" + + ".pb.RegionSpecifier\"\027\n\025OfflineRegionResp" + + "onse\"\177\n\022CreateTableRequest\022+\n\014table_sche" + + "ma\030\001 \002(\0132\025.hbase.pb.TableSchema\022\022\n\nsplit" + + "_keys\030\002 \003(\014\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005n" + + "once\030\004 \001(\004:\0010\"&\n\023CreateTableResponse\022\017\n\007" + + "proc_id\030\001 \001(\004\"g\n\022DeleteTableRequest\022\'\n\nt" + + "able_name\030\001 \002(\0132\023.hbase.pb.TableName\022\026\n\013" + + "nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"&" + + "\n\023DeleteTableResponse\022\017\n\007proc_id\030\001 \001(\004\"\207", + "\001\n\024TruncateTableRequest\022&\n\ttableName\030\001 \002" + + "(\0132\023.hbase.pb.TableName\022\035\n\016preserveSplit" + + "s\030\002 \001(\010:\005false\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020" + + "\n\005nonce\030\004 \001(\004:\0010\"\027\n\025TruncateTableRespons" + + "e\"g\n\022EnableTableRequest\022\'\n\ntable_name\030\001 " + + "\002(\0132\023.hbase.pb.TableName\022\026\n\013nonce_group\030" + + "\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"&\n\023EnableTabl" + + "eResponse\022\017\n\007proc_id\030\001 \001(\004\"h\n\023DisableTab" + + "leRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb" + + ".TableName\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005no", + "nce\030\003 \001(\004:\0010\"\'\n\024DisableTableResponse\022\017\n\007" + + "proc_id\030\001 \001(\004\"\224\001\n\022ModifyTableRequest\022\'\n\n" + + "table_name\030\001 \002(\0132\023.hbase.pb.TableName\022+\n" + + "\014table_schema\030\002 \002(\0132\025.hbase.pb.TableSche" + + "ma\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(" + + "\004:\0010\"\025\n\023ModifyTableResponse\"~\n\026CreateNam" + + "espaceRequest\022:\n\023namespaceDescriptor\030\001 \002" + + "(\0132\035.hbase.pb.NamespaceDescriptor\022\026\n\013non" + + "ce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\031\n\027C" + + "reateNamespaceResponse\"Y\n\026DeleteNamespac", + "eRequest\022\025\n\rnamespaceName\030\001 \002(\t\022\026\n\013nonce" + + "_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\031\n\027Del" + + "eteNamespaceResponse\"~\n\026ModifyNamespaceR" + + "equest\022:\n\023namespaceDescriptor\030\001 \002(\0132\035.hb" + + "ase.pb.NamespaceDescriptor\022\026\n\013nonce_grou" + + "p\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\031\n\027ModifyNa" + + "mespaceResponse\"6\n\035GetNamespaceDescripto" + + "rRequest\022\025\n\rnamespaceName\030\001 \002(\t\"\\\n\036GetNa" + + "mespaceDescriptorResponse\022:\n\023namespaceDe" + + "scriptor\030\001 \002(\0132\035.hbase.pb.NamespaceDescr", + "iptor\"!\n\037ListNamespaceDescriptorsRequest" + + "\"^\n ListNamespaceDescriptorsResponse\022:\n\023" + + "namespaceDescriptor\030\001 \003(\0132\035.hbase.pb.Nam" + + "espaceDescriptor\"?\n&ListTableDescriptors" + + "ByNamespaceRequest\022\025\n\rnamespaceName\030\001 \002(" + + "\t\"U\n\'ListTableDescriptorsByNamespaceResp" + + "onse\022*\n\013tableSchema\030\001 \003(\0132\025.hbase.pb.Tab" + + "leSchema\"9\n ListTableNamesByNamespaceReq" + + "uest\022\025\n\rnamespaceName\030\001 \002(\t\"K\n!ListTable" + + "NamesByNamespaceResponse\022&\n\ttableName\030\001 ", + "\003(\0132\023.hbase.pb.TableName\"\021\n\017ShutdownRequ" + + "est\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReq" + + "uest\"\024\n\022StopMasterResponse\"\034\n\032IsInMainte" + + "nanceModeRequest\"8\n\033IsInMaintenanceModeR" + + "esponse\022\031\n\021inMaintenanceMode\030\001 \002(\010\"\037\n\016Ba" + + "lanceRequest\022\r\n\005force\030\001 \001(\010\"\'\n\017BalanceRe" + + "sponse\022\024\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalan" + + "cerRunningRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchro" + + "nous\030\002 \001(\010\"8\n\032SetBalancerRunningResponse" + + "\022\032\n\022prev_balance_value\030\001 \001(\010\"\032\n\030IsBalanc", + "erEnabledRequest\",\n\031IsBalancerEnabledRes" + + "ponse\022\017\n\007enabled\030\001 \002(\010\"w\n\035SetSplitOrMerg" + + "eEnabledRequest\022\017\n\007enabled\030\001 \002(\010\022\023\n\013sync" + + "hronous\030\002 \001(\010\0220\n\014switch_types\030\003 \003(\0162\032.hb" + + "ase.pb.MasterSwitchType\"4\n\036SetSplitOrMer" + + "geEnabledResponse\022\022\n\nprev_value\030\001 \003(\010\"O\n" + + "\034IsSplitOrMergeEnabledRequest\022/\n\013switch_" + + "type\030\001 \002(\0162\032.hbase.pb.MasterSwitchType\"0" + + "\n\035IsSplitOrMergeEnabledResponse\022\017\n\007enabl" + + "ed\030\001 \002(\010\"\022\n\020NormalizeRequest\"+\n\021Normaliz", + "eResponse\022\026\n\016normalizer_ran\030\001 \002(\010\")\n\033Set" + + "NormalizerRunningRequest\022\n\n\002on\030\001 \002(\010\"=\n\034" + + "SetNormalizerRunningResponse\022\035\n\025prev_nor" + + "malizer_value\030\001 \001(\010\"\034\n\032IsNormalizerEnabl" + + "edRequest\".\n\033IsNormalizerEnabledResponse" + + "\022\017\n\007enabled\030\001 \002(\010\"\027\n\025RunCatalogScanReque" + + "st\"-\n\026RunCatalogScanResponse\022\023\n\013scan_res" + + "ult\030\001 \001(\005\"-\n\033EnableCatalogJanitorRequest" + + "\022\016\n\006enable\030\001 \002(\010\"2\n\034EnableCatalogJanitor" + + "Response\022\022\n\nprev_value\030\001 \001(\010\" \n\036IsCatalo", + "gJanitorEnabledRequest\"0\n\037IsCatalogJanit" + + "orEnabledResponse\022\r\n\005value\030\001 \002(\010\"\030\n\026RunC" + + "leanerChoreRequest\"4\n\027RunCleanerChoreRes" + + "ponse\022\031\n\021cleaner_chore_ran\030\001 \002(\010\"+\n\035SetC" + + "leanerChoreRunningRequest\022\n\n\002on\030\001 \002(\010\"4\n" + + "\036SetCleanerChoreRunningResponse\022\022\n\nprev_" + + "value\030\001 \001(\010\"\036\n\034IsCleanerChoreEnabledRequ" + + "est\".\n\035IsCleanerChoreEnabledResponse\022\r\n\005" + + "value\030\001 \002(\010\"B\n\017SnapshotRequest\022/\n\010snapsh" + + "ot\030\001 \002(\0132\035.hbase.pb.SnapshotDescription\"", + ",\n\020SnapshotResponse\022\030\n\020expected_timeout\030" + + "\001 \002(\003\"\036\n\034GetCompletedSnapshotsRequest\"Q\n" + + "\035GetCompletedSnapshotsResponse\0220\n\tsnapsh" + + "ots\030\001 \003(\0132\035.hbase.pb.SnapshotDescription" + + "\"H\n\025DeleteSnapshotRequest\022/\n\010snapshot\030\001 " + + "\002(\0132\035.hbase.pb.SnapshotDescription\"\030\n\026De" + + "leteSnapshotResponse\"d\n\026RestoreSnapshotR" + + "equest\022/\n\010snapshot\030\001 \002(\0132\035.hbase.pb.Snap" + + "shotDescription\022\031\n\nrestoreACL\030\002 \001(\010:\005fal" + + "se\"\031\n\027RestoreSnapshotResponse\"H\n\025IsSnaps", + "hotDoneRequest\022/\n\010snapshot\030\001 \001(\0132\035.hbase" + + ".pb.SnapshotDescription\"^\n\026IsSnapshotDon" + + "eResponse\022\023\n\004done\030\001 \001(\010:\005false\022/\n\010snapsh" + + "ot\030\002 \001(\0132\035.hbase.pb.SnapshotDescription\"" + + "O\n\034IsRestoreSnapshotDoneRequest\022/\n\010snaps" + + "hot\030\001 \001(\0132\035.hbase.pb.SnapshotDescription" + + "\"4\n\035IsRestoreSnapshotDoneResponse\022\023\n\004don" + + "e\030\001 \001(\010:\005false\"F\n\033GetSchemaAlterStatusRe" + + "quest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Tab" + + "leName\"T\n\034GetSchemaAlterStatusResponse\022\035", + "\n\025yet_to_update_regions\030\001 \001(\r\022\025\n\rtotal_r" + + "egions\030\002 \001(\r\"\213\001\n\032GetTableDescriptorsRequ" + + "est\022(\n\013table_names\030\001 \003(\0132\023.hbase.pb.Tabl" + + "eName\022\r\n\005regex\030\002 \001(\t\022!\n\022include_sys_tabl" + + "es\030\003 \001(\010:\005false\022\021\n\tnamespace\030\004 \001(\t\"J\n\033Ge" + + "tTableDescriptorsResponse\022+\n\014table_schem" + + "a\030\001 \003(\0132\025.hbase.pb.TableSchema\"[\n\024GetTab" + + "leNamesRequest\022\r\n\005regex\030\001 \001(\t\022!\n\022include" + + "_sys_tables\030\002 \001(\010:\005false\022\021\n\tnamespace\030\003 " + + "\001(\t\"A\n\025GetTableNamesResponse\022(\n\013table_na", + "mes\030\001 \003(\0132\023.hbase.pb.TableName\"\031\n\027GetClu" + + "sterStatusRequest\"K\n\030GetClusterStatusRes" + + "ponse\022/\n\016cluster_status\030\001 \002(\0132\027.hbase.pb" + + ".ClusterStatus\"\030\n\026IsMasterRunningRequest" + + "\"4\n\027IsMasterRunningResponse\022\031\n\021is_master" + + "_running\030\001 \002(\010\"I\n\024ExecProcedureRequest\0221" + + "\n\tprocedure\030\001 \002(\0132\036.hbase.pb.ProcedureDe" + + "scription\"F\n\025ExecProcedureResponse\022\030\n\020ex" + + "pected_timeout\030\001 \001(\003\022\023\n\013return_data\030\002 \001(" + + "\014\"K\n\026IsProcedureDoneRequest\0221\n\tprocedure", + "\030\001 \001(\0132\036.hbase.pb.ProcedureDescription\"`" + + "\n\027IsProcedureDoneResponse\022\023\n\004done\030\001 \001(\010:" + + "\005false\0220\n\010snapshot\030\002 \001(\0132\036.hbase.pb.Proc" + + "edureDescription\",\n\031GetProcedureResultRe" + + "quest\022\017\n\007proc_id\030\001 \002(\004\"\371\001\n\032GetProcedureR" + + "esultResponse\0229\n\005state\030\001 \002(\0162*.hbase.pb." + + "GetProcedureResultResponse.State\022\022\n\nstar" + + "t_time\030\002 \001(\004\022\023\n\013last_update\030\003 \001(\004\022\016\n\006res" + + "ult\030\004 \001(\014\0224\n\texception\030\005 \001(\0132!.hbase.pb." + + "ForeignExceptionMessage\"1\n\005State\022\r\n\tNOT_", + "FOUND\020\000\022\013\n\007RUNNING\020\001\022\014\n\010FINISHED\020\002\"M\n\025Ab" + + "ortProcedureRequest\022\017\n\007proc_id\030\001 \002(\004\022#\n\025" + + "mayInterruptIfRunning\030\002 \001(\010:\004true\"6\n\026Abo" + + "rtProcedureResponse\022\034\n\024is_procedure_abor" + + "ted\030\001 \002(\010\"\027\n\025ListProceduresRequest\"@\n\026Li" + + "stProceduresResponse\022&\n\tprocedure\030\001 \003(\0132" + + "\023.hbase.pb.Procedure\"\315\001\n\017SetQuotaRequest" + + "\022\021\n\tuser_name\030\001 \001(\t\022\022\n\nuser_group\030\002 \001(\t\022" + + "\021\n\tnamespace\030\003 \001(\t\022\'\n\ntable_name\030\004 \001(\0132\023" + + ".hbase.pb.TableName\022\022\n\nremove_all\030\005 \001(\010\022", + "\026\n\016bypass_globals\030\006 \001(\010\022+\n\010throttle\030\007 \001(" + + "\0132\031.hbase.pb.ThrottleRequest\"\022\n\020SetQuota" + + "Response\"J\n\037MajorCompactionTimestampRequ" + + "est\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Table" + + "Name\"U\n(MajorCompactionTimestampForRegio" + "nRequest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Regi" + - "onSpecifier\022\024\n\005force\030\002 \001(\010:\005false\"\030\n\026Una" + - "ssignRegionResponse\"A\n\024OfflineRegionRequ", - "est\022)\n\006region\030\001 \002(\0132\031.hbase.pb.RegionSpe" + - "cifier\"\027\n\025OfflineRegionResponse\"\177\n\022Creat" + - "eTableRequest\022+\n\014table_schema\030\001 \002(\0132\025.hb" + - "ase.pb.TableSchema\022\022\n\nsplit_keys\030\002 \003(\014\022\026" + - "\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010" + - "\"&\n\023CreateTableResponse\022\017\n\007proc_id\030\001 \001(\004" + - "\"g\n\022DeleteTableRequest\022\'\n\ntable_name\030\001 \002" + - "(\0132\023.hbase.pb.TableName\022\026\n\013nonce_group\030\002" + - " \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"&\n\023DeleteTable" + - "Response\022\017\n\007proc_id\030\001 \001(\004\"\207\001\n\024TruncateTa", - "bleRequest\022&\n\ttableName\030\001 \002(\0132\023.hbase.pb" + - ".TableName\022\035\n\016preserveSplits\030\002 \001(\010:\005fals" + - "e\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004" + - ":\0010\"\027\n\025TruncateTableResponse\"g\n\022EnableTa" + - "bleRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.p" + - "b.TableName\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005n" + - "once\030\003 \001(\004:\0010\"&\n\023EnableTableResponse\022\017\n\007" + - "proc_id\030\001 \001(\004\"h\n\023DisableTableRequest\022\'\n\n" + - "table_name\030\001 \002(\0132\023.hbase.pb.TableName\022\026\n" + - "\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"", - "\'\n\024DisableTableResponse\022\017\n\007proc_id\030\001 \001(\004" + - "\"\224\001\n\022ModifyTableRequest\022\'\n\ntable_name\030\001 " + - "\002(\0132\023.hbase.pb.TableName\022+\n\014table_schema" + - "\030\002 \002(\0132\025.hbase.pb.TableSchema\022\026\n\013nonce_g" + - "roup\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"\025\n\023Modif" + - "yTableResponse\"~\n\026CreateNamespaceRequest" + - "\022:\n\023namespaceDescriptor\030\001 \002(\0132\035.hbase.pb" + - ".NamespaceDescriptor\022\026\n\013nonce_group\030\002 \001(" + - "\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\031\n\027CreateNamespac" + - "eResponse\"Y\n\026DeleteNamespaceRequest\022\025\n\rn", - "amespaceName\030\001 \002(\t\022\026\n\013nonce_group\030\002 \001(\004:" + - "\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\031\n\027DeleteNamespaceR" + - "esponse\"~\n\026ModifyNamespaceRequest\022:\n\023nam" + - "espaceDescriptor\030\001 \002(\0132\035.hbase.pb.Namesp" + - "aceDescriptor\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n" + - "\005nonce\030\003 \001(\004:\0010\"\031\n\027ModifyNamespaceRespon" + - "se\"6\n\035GetNamespaceDescriptorRequest\022\025\n\rn" + - "amespaceName\030\001 \002(\t\"\\\n\036GetNamespaceDescri" + - "ptorResponse\022:\n\023namespaceDescriptor\030\001 \002(" + - "\0132\035.hbase.pb.NamespaceDescriptor\"!\n\037List", - "NamespaceDescriptorsRequest\"^\n ListNames" + - "paceDescriptorsResponse\022:\n\023namespaceDesc" + - "riptor\030\001 \003(\0132\035.hbase.pb.NamespaceDescrip" + - "tor\"?\n&ListTableDescriptorsByNamespaceRe" + - "quest\022\025\n\rnamespaceName\030\001 \002(\t\"U\n\'ListTabl" + - "eDescriptorsByNamespaceResponse\022*\n\013table" + - "Schema\030\001 \003(\0132\025.hbase.pb.TableSchema\"9\n L" + - "istTableNamesByNamespaceRequest\022\025\n\rnames" + - "paceName\030\001 \002(\t\"K\n!ListTableNamesByNamesp" + - "aceResponse\022&\n\ttableName\030\001 \003(\0132\023.hbase.p", - "b.TableName\"\021\n\017ShutdownRequest\"\022\n\020Shutdo" + - "wnResponse\"\023\n\021StopMasterRequest\"\024\n\022StopM" + - "asterResponse\"\034\n\032IsInMaintenanceModeRequ" + - "est\"8\n\033IsInMaintenanceModeResponse\022\031\n\021in" + - "MaintenanceMode\030\001 \002(\010\"\037\n\016BalanceRequest\022" + - "\r\n\005force\030\001 \001(\010\"\'\n\017BalanceResponse\022\024\n\014bal" + - "ancer_ran\030\001 \002(\010\"<\n\031SetBalancerRunningReq" + - "uest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\"8\n" + - "\032SetBalancerRunningResponse\022\032\n\022prev_bala" + - "nce_value\030\001 \001(\010\"\032\n\030IsBalancerEnabledRequ", - "est\",\n\031IsBalancerEnabledResponse\022\017\n\007enab" + - "led\030\001 \002(\010\"w\n\035SetSplitOrMergeEnabledReque" + - "st\022\017\n\007enabled\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010" + - "\0220\n\014switch_types\030\003 \003(\0162\032.hbase.pb.Master" + - "SwitchType\"4\n\036SetSplitOrMergeEnabledResp" + - "onse\022\022\n\nprev_value\030\001 \003(\010\"O\n\034IsSplitOrMer" + - "geEnabledRequest\022/\n\013switch_type\030\001 \002(\0162\032." + - "hbase.pb.MasterSwitchType\"0\n\035IsSplitOrMe" + - "rgeEnabledResponse\022\017\n\007enabled\030\001 \002(\010\"\022\n\020N" + - "ormalizeRequest\"+\n\021NormalizeResponse\022\026\n\016", - "normalizer_ran\030\001 \002(\010\")\n\033SetNormalizerRun" + - "ningRequest\022\n\n\002on\030\001 \002(\010\"=\n\034SetNormalizer" + - "RunningResponse\022\035\n\025prev_normalizer_value" + - "\030\001 \001(\010\"\034\n\032IsNormalizerEnabledRequest\".\n\033" + - "IsNormalizerEnabledResponse\022\017\n\007enabled\030\001" + - " \002(\010\"\027\n\025RunCatalogScanRequest\"-\n\026RunCata" + - "logScanResponse\022\023\n\013scan_result\030\001 \001(\005\"-\n\033" + - "EnableCatalogJanitorRequest\022\016\n\006enable\030\001 " + - "\002(\010\"2\n\034EnableCatalogJanitorResponse\022\022\n\np" + - "rev_value\030\001 \001(\010\" \n\036IsCatalogJanitorEnabl", - "edRequest\"0\n\037IsCatalogJanitorEnabledResp" + - "onse\022\r\n\005value\030\001 \002(\010\"\030\n\026RunCleanerChoreRe" + - "quest\"4\n\027RunCleanerChoreResponse\022\031\n\021clea" + - "ner_chore_ran\030\001 \002(\010\"+\n\035SetCleanerChoreRu" + - "nningRequest\022\n\n\002on\030\001 \002(\010\"4\n\036SetCleanerCh" + - "oreRunningResponse\022\022\n\nprev_value\030\001 \001(\010\"\036" + - "\n\034IsCleanerChoreEnabledRequest\".\n\035IsClea" + - "nerChoreEnabledResponse\022\r\n\005value\030\001 \002(\010\"B" + - "\n\017SnapshotRequest\022/\n\010snapshot\030\001 \002(\0132\035.hb" + - "ase.pb.SnapshotDescription\",\n\020SnapshotRe", - "sponse\022\030\n\020expected_timeout\030\001 \002(\003\"\036\n\034GetC" + - "ompletedSnapshotsRequest\"Q\n\035GetCompleted" + - "SnapshotsResponse\0220\n\tsnapshots\030\001 \003(\0132\035.h" + - "base.pb.SnapshotDescription\"H\n\025DeleteSna" + - "pshotRequest\022/\n\010snapshot\030\001 \002(\0132\035.hbase.p" + - "b.SnapshotDescription\"\030\n\026DeleteSnapshotR" + - "esponse\"d\n\026RestoreSnapshotRequest\022/\n\010sna" + - "pshot\030\001 \002(\0132\035.hbase.pb.SnapshotDescripti" + - "on\022\031\n\nrestoreACL\030\002 \001(\010:\005false\"\031\n\027Restore" + - "SnapshotResponse\"H\n\025IsSnapshotDoneReques", - "t\022/\n\010snapshot\030\001 \001(\0132\035.hbase.pb.SnapshotD" + - "escription\"^\n\026IsSnapshotDoneResponse\022\023\n\004" + - "done\030\001 \001(\010:\005false\022/\n\010snapshot\030\002 \001(\0132\035.hb" + - "ase.pb.SnapshotDescription\"O\n\034IsRestoreS" + - "napshotDoneRequest\022/\n\010snapshot\030\001 \001(\0132\035.h" + - "base.pb.SnapshotDescription\"4\n\035IsRestore" + - "SnapshotDoneResponse\022\023\n\004done\030\001 \001(\010:\005fals" + - "e\"F\n\033GetSchemaAlterStatusRequest\022\'\n\ntabl" + - "e_name\030\001 \002(\0132\023.hbase.pb.TableName\"T\n\034Get" + - "SchemaAlterStatusResponse\022\035\n\025yet_to_upda", - "te_regions\030\001 \001(\r\022\025\n\rtotal_regions\030\002 \001(\r\"" + - "\213\001\n\032GetTableDescriptorsRequest\022(\n\013table_" + - "names\030\001 \003(\0132\023.hbase.pb.TableName\022\r\n\005rege" + - "x\030\002 \001(\t\022!\n\022include_sys_tables\030\003 \001(\010:\005fal" + - "se\022\021\n\tnamespace\030\004 \001(\t\"J\n\033GetTableDescrip" + - "torsResponse\022+\n\014table_schema\030\001 \003(\0132\025.hba" + - "se.pb.TableSchema\"[\n\024GetTableNamesReques" + - "t\022\r\n\005regex\030\001 \001(\t\022!\n\022include_sys_tables\030\002" + - " \001(\010:\005false\022\021\n\tnamespace\030\003 \001(\t\"A\n\025GetTab" + - "leNamesResponse\022(\n\013table_names\030\001 \003(\0132\023.h", - "base.pb.TableName\"\031\n\027GetClusterStatusReq" + - "uest\"K\n\030GetClusterStatusResponse\022/\n\016clus" + - "ter_status\030\001 \002(\0132\027.hbase.pb.ClusterStatu" + - "s\"\030\n\026IsMasterRunningRequest\"4\n\027IsMasterR" + - "unningResponse\022\031\n\021is_master_running\030\001 \002(" + - "\010\"I\n\024ExecProcedureRequest\0221\n\tprocedure\030\001" + - " \002(\0132\036.hbase.pb.ProcedureDescription\"F\n\025" + - "ExecProcedureResponse\022\030\n\020expected_timeou" + - "t\030\001 \001(\003\022\023\n\013return_data\030\002 \001(\014\"K\n\026IsProced" + - "ureDoneRequest\0221\n\tprocedure\030\001 \001(\0132\036.hbas", - "e.pb.ProcedureDescription\"`\n\027IsProcedure" + - "DoneResponse\022\023\n\004done\030\001 \001(\010:\005false\0220\n\010sna" + - "pshot\030\002 \001(\0132\036.hbase.pb.ProcedureDescript" + - "ion\",\n\031GetProcedureResultRequest\022\017\n\007proc" + - "_id\030\001 \002(\004\"\371\001\n\032GetProcedureResultResponse" + - "\0229\n\005state\030\001 \002(\0162*.hbase.pb.GetProcedureR" + - "esultResponse.State\022\022\n\nstart_time\030\002 \001(\004\022" + - "\023\n\013last_update\030\003 \001(\004\022\016\n\006result\030\004 \001(\014\0224\n\t" + - "exception\030\005 \001(\0132!.hbase.pb.ForeignExcept" + - "ionMessage\"1\n\005State\022\r\n\tNOT_FOUND\020\000\022\013\n\007RU", - "NNING\020\001\022\014\n\010FINISHED\020\002\"M\n\025AbortProcedureR" + - "equest\022\017\n\007proc_id\030\001 \002(\004\022#\n\025mayInterruptI" + - "fRunning\030\002 \001(\010:\004true\"6\n\026AbortProcedureRe" + - "sponse\022\034\n\024is_procedure_aborted\030\001 \002(\010\"\027\n\025" + - "ListProceduresRequest\"@\n\026ListProceduresR" + - "esponse\022&\n\tprocedure\030\001 \003(\0132\023.hbase.pb.Pr" + - "ocedure\"\315\001\n\017SetQuotaRequest\022\021\n\tuser_name" + - "\030\001 \001(\t\022\022\n\nuser_group\030\002 \001(\t\022\021\n\tnamespace\030" + - "\003 \001(\t\022\'\n\ntable_name\030\004 \001(\0132\023.hbase.pb.Tab" + - "leName\022\022\n\nremove_all\030\005 \001(\010\022\026\n\016bypass_glo", - "bals\030\006 \001(\010\022+\n\010throttle\030\007 \001(\0132\031.hbase.pb." + - "ThrottleRequest\"\022\n\020SetQuotaResponse\"J\n\037M" + - "ajorCompactionTimestampRequest\022\'\n\ntable_" + - "name\030\001 \002(\0132\023.hbase.pb.TableName\"U\n(Major" + - "CompactionTimestampForRegionRequest\022)\n\006r" + - "egion\030\001 \002(\0132\031.hbase.pb.RegionSpecifier\"@" + - "\n MajorCompactionTimestampResponse\022\034\n\024co" + - "mpaction_timestamp\030\001 \002(\003\"\035\n\033SecurityCapa" + - "bilitiesRequest\"\354\001\n\034SecurityCapabilities" + - "Response\022G\n\014capabilities\030\001 \003(\01621.hbase.p", - "b.SecurityCapabilitiesResponse.Capabilit" + - "y\"\202\001\n\nCapability\022\031\n\025SIMPLE_AUTHENTICATIO" + - "N\020\000\022\031\n\025SECURE_AUTHENTICATION\020\001\022\021\n\rAUTHOR" + - "IZATION\020\002\022\026\n\022CELL_AUTHORIZATION\020\003\022\023\n\017CEL" + - "L_VISIBILITY\020\004*(\n\020MasterSwitchType\022\t\n\005SP" + - "LIT\020\000\022\t\n\005MERGE\020\0012\232+\n\rMasterService\022e\n\024Ge" + - "tSchemaAlterStatus\022%.hbase.pb.GetSchemaA" + - "lterStatusRequest\032&.hbase.pb.GetSchemaAl" + - "terStatusResponse\022b\n\023GetTableDescriptors" + - "\022$.hbase.pb.GetTableDescriptorsRequest\032%", - ".hbase.pb.GetTableDescriptorsResponse\022P\n" + - "\rGetTableNames\022\036.hbase.pb.GetTableNamesR" + - "equest\032\037.hbase.pb.GetTableNamesResponse\022" + - "Y\n\020GetClusterStatus\022!.hbase.pb.GetCluste" + - "rStatusRequest\032\".hbase.pb.GetClusterStat" + - "usResponse\022V\n\017IsMasterRunning\022 .hbase.pb" + - ".IsMasterRunningRequest\032!.hbase.pb.IsMas" + - "terRunningResponse\022D\n\tAddColumn\022\032.hbase." + - "pb.AddColumnRequest\032\033.hbase.pb.AddColumn" + - "Response\022M\n\014DeleteColumn\022\035.hbase.pb.Dele", - "teColumnRequest\032\036.hbase.pb.DeleteColumnR" + - "esponse\022M\n\014ModifyColumn\022\035.hbase.pb.Modif" + - "yColumnRequest\032\036.hbase.pb.ModifyColumnRe" + - "sponse\022G\n\nMoveRegion\022\033.hbase.pb.MoveRegi" + - "onRequest\032\034.hbase.pb.MoveRegionResponse\022" + - "k\n\026DispatchMergingRegions\022\'.hbase.pb.Dis" + - "patchMergingRegionsRequest\032(.hbase.pb.Di" + - "spatchMergingRegionsResponse\022M\n\014AssignRe" + - "gion\022\035.hbase.pb.AssignRegionRequest\032\036.hb" + - "ase.pb.AssignRegionResponse\022S\n\016UnassignR", - "egion\022\037.hbase.pb.UnassignRegionRequest\032 " + - ".hbase.pb.UnassignRegionResponse\022P\n\rOffl" + - "ineRegion\022\036.hbase.pb.OfflineRegionReques" + - "t\032\037.hbase.pb.OfflineRegionResponse\022J\n\013De" + - "leteTable\022\034.hbase.pb.DeleteTableRequest\032" + - "\035.hbase.pb.DeleteTableResponse\022P\n\rtrunca" + - "teTable\022\036.hbase.pb.TruncateTableRequest\032" + - "\037.hbase.pb.TruncateTableResponse\022J\n\013Enab" + - "leTable\022\034.hbase.pb.EnableTableRequest\032\035." + - "hbase.pb.EnableTableResponse\022M\n\014DisableT", - "able\022\035.hbase.pb.DisableTableRequest\032\036.hb" + - "ase.pb.DisableTableResponse\022J\n\013ModifyTab" + - "le\022\034.hbase.pb.ModifyTableRequest\032\035.hbase" + - ".pb.ModifyTableResponse\022J\n\013CreateTable\022\034" + - ".hbase.pb.CreateTableRequest\032\035.hbase.pb." + - "CreateTableResponse\022A\n\010Shutdown\022\031.hbase." + - "pb.ShutdownRequest\032\032.hbase.pb.ShutdownRe" + - "sponse\022G\n\nStopMaster\022\033.hbase.pb.StopMast" + - "erRequest\032\034.hbase.pb.StopMasterResponse\022" + - "h\n\031IsMasterInMaintenanceMode\022$.hbase.pb.", - "IsInMaintenanceModeRequest\032%.hbase.pb.Is" + - "InMaintenanceModeResponse\022>\n\007Balance\022\030.h" + - "base.pb.BalanceRequest\032\031.hbase.pb.Balanc" + - "eResponse\022_\n\022SetBalancerRunning\022#.hbase." + - "pb.SetBalancerRunningRequest\032$.hbase.pb." + - "SetBalancerRunningResponse\022\\\n\021IsBalancer" + - "Enabled\022\".hbase.pb.IsBalancerEnabledRequ" + - "est\032#.hbase.pb.IsBalancerEnabledResponse" + - "\022k\n\026SetSplitOrMergeEnabled\022\'.hbase.pb.Se" + - "tSplitOrMergeEnabledRequest\032(.hbase.pb.S", - "etSplitOrMergeEnabledResponse\022h\n\025IsSplit" + - "OrMergeEnabled\022&.hbase.pb.IsSplitOrMerge" + - "EnabledRequest\032\'.hbase.pb.IsSplitOrMerge" + - "EnabledResponse\022D\n\tNormalize\022\032.hbase.pb." + - "NormalizeRequest\032\033.hbase.pb.NormalizeRes" + - "ponse\022e\n\024SetNormalizerRunning\022%.hbase.pb" + - ".SetNormalizerRunningRequest\032&.hbase.pb." + - "SetNormalizerRunningResponse\022b\n\023IsNormal" + - "izerEnabled\022$.hbase.pb.IsNormalizerEnabl" + - "edRequest\032%.hbase.pb.IsNormalizerEnabled", - "Response\022S\n\016RunCatalogScan\022\037.hbase.pb.Ru" + - "nCatalogScanRequest\032 .hbase.pb.RunCatalo" + - "gScanResponse\022e\n\024EnableCatalogJanitor\022%." + - "hbase.pb.EnableCatalogJanitorRequest\032&.h" + - "base.pb.EnableCatalogJanitorResponse\022n\n\027" + - "IsCatalogJanitorEnabled\022(.hbase.pb.IsCat" + - "alogJanitorEnabledRequest\032).hbase.pb.IsC" + - "atalogJanitorEnabledResponse\022V\n\017RunClean" + - "erChore\022 .hbase.pb.RunCleanerChoreReques" + - "t\032!.hbase.pb.RunCleanerChoreResponse\022k\n\026", - "SetCleanerChoreRunning\022\'.hbase.pb.SetCle" + - "anerChoreRunningRequest\032(.hbase.pb.SetCl" + - "eanerChoreRunningResponse\022h\n\025IsCleanerCh" + - "oreEnabled\022&.hbase.pb.IsCleanerChoreEnab" + - "ledRequest\032\'.hbase.pb.IsCleanerChoreEnab" + - "ledResponse\022^\n\021ExecMasterService\022#.hbase" + - ".pb.CoprocessorServiceRequest\032$.hbase.pb" + - ".CoprocessorServiceResponse\022A\n\010Snapshot\022" + - "\031.hbase.pb.SnapshotRequest\032\032.hbase.pb.Sn" + - "apshotResponse\022h\n\025GetCompletedSnapshots\022", - "&.hbase.pb.GetCompletedSnapshotsRequest\032" + - "\'.hbase.pb.GetCompletedSnapshotsResponse" + - "\022S\n\016DeleteSnapshot\022\037.hbase.pb.DeleteSnap" + - "shotRequest\032 .hbase.pb.DeleteSnapshotRes" + - "ponse\022S\n\016IsSnapshotDone\022\037.hbase.pb.IsSna" + - "pshotDoneRequest\032 .hbase.pb.IsSnapshotDo" + - "neResponse\022V\n\017RestoreSnapshot\022 .hbase.pb" + - ".RestoreSnapshotRequest\032!.hbase.pb.Resto" + - "reSnapshotResponse\022h\n\025IsRestoreSnapshotD" + - "one\022&.hbase.pb.IsRestoreSnapshotDoneRequ", - "est\032\'.hbase.pb.IsRestoreSnapshotDoneResp" + - "onse\022P\n\rExecProcedure\022\036.hbase.pb.ExecPro" + - "cedureRequest\032\037.hbase.pb.ExecProcedureRe" + - "sponse\022W\n\024ExecProcedureWithRet\022\036.hbase.p" + - "b.ExecProcedureRequest\032\037.hbase.pb.ExecPr" + - "ocedureResponse\022V\n\017IsProcedureDone\022 .hba" + - "se.pb.IsProcedureDoneRequest\032!.hbase.pb." + - "IsProcedureDoneResponse\022V\n\017ModifyNamespa" + - "ce\022 .hbase.pb.ModifyNamespaceRequest\032!.h" + - "base.pb.ModifyNamespaceResponse\022V\n\017Creat", - "eNamespace\022 .hbase.pb.CreateNamespaceReq" + - "uest\032!.hbase.pb.CreateNamespaceResponse\022" + - "V\n\017DeleteNamespace\022 .hbase.pb.DeleteName" + - "spaceRequest\032!.hbase.pb.DeleteNamespaceR" + - "esponse\022k\n\026GetNamespaceDescriptor\022\'.hbas" + - "e.pb.GetNamespaceDescriptorRequest\032(.hba" + - "se.pb.GetNamespaceDescriptorResponse\022q\n\030" + - "ListNamespaceDescriptors\022).hbase.pb.List" + - "NamespaceDescriptorsRequest\032*.hbase.pb.L" + - "istNamespaceDescriptorsResponse\022\206\001\n\037List", - "TableDescriptorsByNamespace\0220.hbase.pb.L" + - "istTableDescriptorsByNamespaceRequest\0321." + - "hbase.pb.ListTableDescriptorsByNamespace" + - "Response\022t\n\031ListTableNamesByNamespace\022*." + - "hbase.pb.ListTableNamesByNamespaceReques" + - "t\032+.hbase.pb.ListTableNamesByNamespaceRe" + - "sponse\022A\n\010SetQuota\022\031.hbase.pb.SetQuotaRe" + - "quest\032\032.hbase.pb.SetQuotaResponse\022x\n\037get" + - "LastMajorCompactionTimestamp\022).hbase.pb." + - "MajorCompactionTimestampRequest\032*.hbase.", - "pb.MajorCompactionTimestampResponse\022\212\001\n(" + - "getLastMajorCompactionTimestampForRegion" + - "\0222.hbase.pb.MajorCompactionTimestampForR" + - "egionRequest\032*.hbase.pb.MajorCompactionT" + - "imestampResponse\022_\n\022getProcedureResult\022#" + - ".hbase.pb.GetProcedureResultRequest\032$.hb" + - "ase.pb.GetProcedureResultResponse\022h\n\027get" + - "SecurityCapabilities\022%.hbase.pb.Security" + - "CapabilitiesRequest\032&.hbase.pb.SecurityC" + - "apabilitiesResponse\022S\n\016AbortProcedure\022\037.", - "hbase.pb.AbortProcedureRequest\032 .hbase.p" + - "b.AbortProcedureResponse\022S\n\016ListProcedur" + - "es\022\037.hbase.pb.ListProceduresRequest\032 .hb" + - "ase.pb.ListProceduresResponseBB\n*org.apa" + - "che.hadoop.hbase.protobuf.generatedB\014Mas" + - "terProtosH\001\210\001\001\240\001\001" + "onSpecifier\"@\n MajorCompactionTimestampR" + + "esponse\022\034\n\024compaction_timestamp\030\001 \002(\003\"\035\n" + + "\033SecurityCapabilitiesRequest\"\354\001\n\034Securit" + + "yCapabilitiesResponse\022G\n\014capabilities\030\001 ", + "\003(\01621.hbase.pb.SecurityCapabilitiesRespo" + + "nse.Capability\"\202\001\n\nCapability\022\031\n\025SIMPLE_" + + "AUTHENTICATION\020\000\022\031\n\025SECURE_AUTHENTICATIO" + + "N\020\001\022\021\n\rAUTHORIZATION\020\002\022\026\n\022CELL_AUTHORIZA" + + "TION\020\003\022\023\n\017CELL_VISIBILITY\020\004*(\n\020MasterSwi" + + "tchType\022\t\n\005SPLIT\020\000\022\t\n\005MERGE\020\0012\232+\n\rMaster" + + "Service\022e\n\024GetSchemaAlterStatus\022%.hbase." + + "pb.GetSchemaAlterStatusRequest\032&.hbase.p" + + "b.GetSchemaAlterStatusResponse\022b\n\023GetTab" + + "leDescriptors\022$.hbase.pb.GetTableDescrip", + "torsRequest\032%.hbase.pb.GetTableDescripto" + + "rsResponse\022P\n\rGetTableNames\022\036.hbase.pb.G" + + "etTableNamesRequest\032\037.hbase.pb.GetTableN" + + "amesResponse\022Y\n\020GetClusterStatus\022!.hbase" + + ".pb.GetClusterStatusRequest\032\".hbase.pb.G" + + "etClusterStatusResponse\022V\n\017IsMasterRunni" + + "ng\022 .hbase.pb.IsMasterRunningRequest\032!.h" + + "base.pb.IsMasterRunningResponse\022D\n\tAddCo" + + "lumn\022\032.hbase.pb.AddColumnRequest\032\033.hbase" + + ".pb.AddColumnResponse\022M\n\014DeleteColumn\022\035.", + "hbase.pb.DeleteColumnRequest\032\036.hbase.pb." + + "DeleteColumnResponse\022M\n\014ModifyColumn\022\035.h" + + "base.pb.ModifyColumnRequest\032\036.hbase.pb.M" + + "odifyColumnResponse\022G\n\nMoveRegion\022\033.hbas" + + "e.pb.MoveRegionRequest\032\034.hbase.pb.MoveRe" + + "gionResponse\022k\n\026DispatchMergingRegions\022\'" + + ".hbase.pb.DispatchMergingRegionsRequest\032" + + "(.hbase.pb.DispatchMergingRegionsRespons" + + "e\022M\n\014AssignRegion\022\035.hbase.pb.AssignRegio" + + "nRequest\032\036.hbase.pb.AssignRegionResponse", + "\022S\n\016UnassignRegion\022\037.hbase.pb.UnassignRe" + + "gionRequest\032 .hbase.pb.UnassignRegionRes" + + "ponse\022P\n\rOfflineRegion\022\036.hbase.pb.Offlin" + + "eRegionRequest\032\037.hbase.pb.OfflineRegionR" + + "esponse\022J\n\013DeleteTable\022\034.hbase.pb.Delete" + + "TableRequest\032\035.hbase.pb.DeleteTableRespo" + + "nse\022P\n\rtruncateTable\022\036.hbase.pb.Truncate" + + "TableRequest\032\037.hbase.pb.TruncateTableRes" + + "ponse\022J\n\013EnableTable\022\034.hbase.pb.EnableTa" + + "bleRequest\032\035.hbase.pb.EnableTableRespons", + "e\022M\n\014DisableTable\022\035.hbase.pb.DisableTabl" + + "eRequest\032\036.hbase.pb.DisableTableResponse" + + "\022J\n\013ModifyTable\022\034.hbase.pb.ModifyTableRe" + + "quest\032\035.hbase.pb.ModifyTableResponse\022J\n\013" + + "CreateTable\022\034.hbase.pb.CreateTableReques" + + "t\032\035.hbase.pb.CreateTableResponse\022A\n\010Shut" + + "down\022\031.hbase.pb.ShutdownRequest\032\032.hbase." + + "pb.ShutdownResponse\022G\n\nStopMaster\022\033.hbas" + + "e.pb.StopMasterRequest\032\034.hbase.pb.StopMa" + + "sterResponse\022h\n\031IsMasterInMaintenanceMod", + "e\022$.hbase.pb.IsInMaintenanceModeRequest\032" + + "%.hbase.pb.IsInMaintenanceModeResponse\022>" + + "\n\007Balance\022\030.hbase.pb.BalanceRequest\032\031.hb" + + "ase.pb.BalanceResponse\022_\n\022SetBalancerRun" + + "ning\022#.hbase.pb.SetBalancerRunningReques" + + "t\032$.hbase.pb.SetBalancerRunningResponse\022" + + "\\\n\021IsBalancerEnabled\022\".hbase.pb.IsBalanc" + + "erEnabledRequest\032#.hbase.pb.IsBalancerEn" + + "abledResponse\022k\n\026SetSplitOrMergeEnabled\022" + + "\'.hbase.pb.SetSplitOrMergeEnabledRequest", + "\032(.hbase.pb.SetSplitOrMergeEnabledRespon" + + "se\022h\n\025IsSplitOrMergeEnabled\022&.hbase.pb.I" + + "sSplitOrMergeEnabledRequest\032\'.hbase.pb.I" + + "sSplitOrMergeEnabledResponse\022D\n\tNormaliz" + + "e\022\032.hbase.pb.NormalizeRequest\032\033.hbase.pb" + + ".NormalizeResponse\022e\n\024SetNormalizerRunni" + + "ng\022%.hbase.pb.SetNormalizerRunningReques" + + "t\032&.hbase.pb.SetNormalizerRunningRespons" + + "e\022b\n\023IsNormalizerEnabled\022$.hbase.pb.IsNo" + + "rmalizerEnabledRequest\032%.hbase.pb.IsNorm", + "alizerEnabledResponse\022S\n\016RunCatalogScan\022" + + "\037.hbase.pb.RunCatalogScanRequest\032 .hbase" + + ".pb.RunCatalogScanResponse\022e\n\024EnableCata" + + "logJanitor\022%.hbase.pb.EnableCatalogJanit" + + "orRequest\032&.hbase.pb.EnableCatalogJanito" + + "rResponse\022n\n\027IsCatalogJanitorEnabled\022(.h" + + "base.pb.IsCatalogJanitorEnabledRequest\032)" + + ".hbase.pb.IsCatalogJanitorEnabledRespons" + + "e\022V\n\017RunCleanerChore\022 .hbase.pb.RunClean" + + "erChoreRequest\032!.hbase.pb.RunCleanerChor", + "eResponse\022k\n\026SetCleanerChoreRunning\022\'.hb" + + "ase.pb.SetCleanerChoreRunningRequest\032(.h" + + "base.pb.SetCleanerChoreRunningResponse\022h" + + "\n\025IsCleanerChoreEnabled\022&.hbase.pb.IsCle" + + "anerChoreEnabledRequest\032\'.hbase.pb.IsCle" + + "anerChoreEnabledResponse\022^\n\021ExecMasterSe" + + "rvice\022#.hbase.pb.CoprocessorServiceReque" + + "st\032$.hbase.pb.CoprocessorServiceResponse" + + "\022A\n\010Snapshot\022\031.hbase.pb.SnapshotRequest\032" + + "\032.hbase.pb.SnapshotResponse\022h\n\025GetComple", + "tedSnapshots\022&.hbase.pb.GetCompletedSnap" + + "shotsRequest\032\'.hbase.pb.GetCompletedSnap" + + "shotsResponse\022S\n\016DeleteSnapshot\022\037.hbase." + + "pb.DeleteSnapshotRequest\032 .hbase.pb.Dele" + + "teSnapshotResponse\022S\n\016IsSnapshotDone\022\037.h" + + "base.pb.IsSnapshotDoneRequest\032 .hbase.pb" + + ".IsSnapshotDoneResponse\022V\n\017RestoreSnapsh" + + "ot\022 .hbase.pb.RestoreSnapshotRequest\032!.h" + + "base.pb.RestoreSnapshotResponse\022h\n\025IsRes" + + "toreSnapshotDone\022&.hbase.pb.IsRestoreSna", + "pshotDoneRequest\032\'.hbase.pb.IsRestoreSna" + + "pshotDoneResponse\022P\n\rExecProcedure\022\036.hba" + + "se.pb.ExecProcedureRequest\032\037.hbase.pb.Ex" + + "ecProcedureResponse\022W\n\024ExecProcedureWith" + + "Ret\022\036.hbase.pb.ExecProcedureRequest\032\037.hb" + + "ase.pb.ExecProcedureResponse\022V\n\017IsProced" + + "ureDone\022 .hbase.pb.IsProcedureDoneReques" + + "t\032!.hbase.pb.IsProcedureDoneResponse\022V\n\017" + + "ModifyNamespace\022 .hbase.pb.ModifyNamespa" + + "ceRequest\032!.hbase.pb.ModifyNamespaceResp", + "onse\022V\n\017CreateNamespace\022 .hbase.pb.Creat" + + "eNamespaceRequest\032!.hbase.pb.CreateNames" + + "paceResponse\022V\n\017DeleteNamespace\022 .hbase." + + "pb.DeleteNamespaceRequest\032!.hbase.pb.Del" + + "eteNamespaceResponse\022k\n\026GetNamespaceDesc" + + "riptor\022\'.hbase.pb.GetNamespaceDescriptor" + + "Request\032(.hbase.pb.GetNamespaceDescripto" + + "rResponse\022q\n\030ListNamespaceDescriptors\022)." + + "hbase.pb.ListNamespaceDescriptorsRequest" + + "\032*.hbase.pb.ListNamespaceDescriptorsResp", + "onse\022\206\001\n\037ListTableDescriptorsByNamespace" + + "\0220.hbase.pb.ListTableDescriptorsByNamesp" + + "aceRequest\0321.hbase.pb.ListTableDescripto" + + "rsByNamespaceResponse\022t\n\031ListTableNamesB" + + "yNamespace\022*.hbase.pb.ListTableNamesByNa" + + "mespaceRequest\032+.hbase.pb.ListTableNames" + + "ByNamespaceResponse\022A\n\010SetQuota\022\031.hbase." + + "pb.SetQuotaRequest\032\032.hbase.pb.SetQuotaRe" + + "sponse\022x\n\037getLastMajorCompactionTimestam" + + "p\022).hbase.pb.MajorCompactionTimestampReq", + "uest\032*.hbase.pb.MajorCompactionTimestamp" + + "Response\022\212\001\n(getLastMajorCompactionTimes" + + "tampForRegion\0222.hbase.pb.MajorCompaction" + + "TimestampForRegionRequest\032*.hbase.pb.Maj" + + "orCompactionTimestampResponse\022_\n\022getProc" + + "edureResult\022#.hbase.pb.GetProcedureResul" + + "tRequest\032$.hbase.pb.GetProcedureResultRe" + + "sponse\022h\n\027getSecurityCapabilities\022%.hbas" + + "e.pb.SecurityCapabilitiesRequest\032&.hbase" + + ".pb.SecurityCapabilitiesResponse\022S\n\016Abor", + "tProcedure\022\037.hbase.pb.AbortProcedureRequ" + + "est\032 .hbase.pb.AbortProcedureResponse\022S\n" + + "\016ListProcedures\022\037.hbase.pb.ListProcedure" + + "sRequest\032 .hbase.pb.ListProceduresRespon" + + "seBB\n*org.apache.hadoop.hbase.protobuf.g" + + "eneratedB\014MasterProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -67218,6 +67218,7 @@ public final class MasterProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(), diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java index 538e03117ee..8521ba883fa 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java @@ -2517,11 +2517,11 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.PrepareBulkLoadRequest} @@ -2575,11 +2575,11 @@ public final class SecureBulkLoadProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -2629,7 +2629,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2639,18 +2639,18 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -2844,7 +2844,7 @@ public final class SecureBulkLoadProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -2940,9 +2940,9 @@ public final class SecureBulkLoadProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2952,7 +2952,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -2962,7 +2962,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2979,7 +2979,7 @@ public final class SecureBulkLoadProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -2992,12 +2992,12 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -3013,7 +3013,7 @@ public final class SecureBulkLoadProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -3024,7 +3024,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -3032,7 +3032,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -3043,11 +3043,11 @@ public final class SecureBulkLoadProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -4857,30 +4857,31 @@ public final class SecureBulkLoadProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013HBase." + - "proto\032\014Client.proto\"\266\001\n\033SecureBulkLoadHF" + - "ilesRequest\022>\n\013family_path\030\001 \003(\0132).hbase" + - ".pb.BulkLoadHFileRequest.FamilyPath\022\026\n\016a" + - "ssign_seq_num\030\002 \001(\010\022+\n\010fs_token\030\003 \002(\0132\031." + - "hbase.pb.DelegationToken\022\022\n\nbulk_token\030\004" + - " \002(\t\".\n\034SecureBulkLoadHFilesResponse\022\016\n\006" + - "loaded\030\001 \002(\010\"V\n\017DelegationToken\022\022\n\nident" + - "ifier\030\001 \001(\014\022\020\n\010password\030\002 \001(\014\022\014\n\004kind\030\003 " + - "\001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026PrepareBulkLoadR", - "equest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Ta" + - "bleName\"-\n\027PrepareBulkLoadResponse\022\022\n\nbu" + - "lk_token\030\001 \002(\t\",\n\026CleanupBulkLoadRequest" + - "\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoadR" + - "esponse2\256\002\n\025SecureBulkLoadService\022V\n\017Pre" + - "pareBulkLoad\022 .hbase.pb.PrepareBulkLoadR" + - "equest\032!.hbase.pb.PrepareBulkLoadRespons" + - "e\022e\n\024SecureBulkLoadHFiles\022%.hbase.pb.Sec" + - "ureBulkLoadHFilesRequest\032&.hbase.pb.Secu" + - "reBulkLoadHFilesResponse\022V\n\017CleanupBulkL", - "oad\022 .hbase.pb.CleanupBulkLoadRequest\032!." + - "hbase.pb.CleanupBulkLoadResponseBJ\n*org." + - "apache.hadoop.hbase.protobuf.generatedB\024" + - "SecureBulkLoadProtosH\001\210\001\001\240\001\001" + "\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013Table." + + "proto\032\013HBase.proto\032\014Client.proto\"\266\001\n\033Sec" + + "ureBulkLoadHFilesRequest\022>\n\013family_path\030" + + "\001 \003(\0132).hbase.pb.BulkLoadHFileRequest.Fa" + + "milyPath\022\026\n\016assign_seq_num\030\002 \001(\010\022+\n\010fs_t" + + "oken\030\003 \002(\0132\031.hbase.pb.DelegationToken\022\022\n" + + "\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadHFile" + + "sResponse\022\016\n\006loaded\030\001 \002(\010\"V\n\017DelegationT" + + "oken\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password\030\002 \001" + + "(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026Pre", + "pareBulkLoadRequest\022\'\n\ntable_name\030\001 \002(\0132" + + "\023.hbase.pb.TableName\"-\n\027PrepareBulkLoadR" + + "esponse\022\022\n\nbulk_token\030\001 \002(\t\",\n\026CleanupBu" + + "lkLoadRequest\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027Cle" + + "anupBulkLoadResponse2\256\002\n\025SecureBulkLoadS" + + "ervice\022V\n\017PrepareBulkLoad\022 .hbase.pb.Pre" + + "pareBulkLoadRequest\032!.hbase.pb.PrepareBu" + + "lkLoadResponse\022e\n\024SecureBulkLoadHFiles\022%" + + ".hbase.pb.SecureBulkLoadHFilesRequest\032&." + + "hbase.pb.SecureBulkLoadHFilesResponse\022V\n", + "\017CleanupBulkLoad\022 .hbase.pb.CleanupBulkL" + + "oadRequest\032!.hbase.pb.CleanupBulkLoadRes" + + "ponseBJ\n*org.apache.hadoop.hbase.protobu" + + "f.generatedB\024SecureBulkLoadProtosH\001\210\001\001\240\001" + + "\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -4935,6 +4936,7 @@ public final class SecureBulkLoadProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(), }, assigner); diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java index 2e11b4a2be7..9805d505082 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java @@ -8,1494 +8,6 @@ public final class SnapshotProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } - public interface SnapshotDescriptionOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required string name = 1; - /** - * required string name = 1; - */ - boolean hasName(); - /** - * required string name = 1; - */ - java.lang.String getName(); - /** - * required string name = 1; - */ - com.google.protobuf.ByteString - getNameBytes(); - - // optional string table = 2; - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - boolean hasTable(); - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - java.lang.String getTable(); - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - com.google.protobuf.ByteString - getTableBytes(); - - // optional int64 creation_time = 3 [default = 0]; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - boolean hasCreationTime(); - /** - * optional int64 creation_time = 3 [default = 0]; - */ - long getCreationTime(); - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - boolean hasType(); - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType(); - - // optional int32 version = 5; - /** - * optional int32 version = 5; - */ - boolean hasVersion(); - /** - * optional int32 version = 5; - */ - int getVersion(); - - // optional string owner = 6; - /** - * optional string owner = 6; - */ - boolean hasOwner(); - /** - * optional string owner = 6; - */ - java.lang.String getOwner(); - /** - * optional string owner = 6; - */ - com.google.protobuf.ByteString - getOwnerBytes(); - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - boolean hasUsersAndPermissions(); - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions(); - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder(); - } - /** - * Protobuf type {@code hbase.pb.SnapshotDescription} - * - *
-   **
-   * Description of the snapshot to take
-   * 
- */ - public static final class SnapshotDescription extends - com.google.protobuf.GeneratedMessage - implements SnapshotDescriptionOrBuilder { - // Use SnapshotDescription.newBuilder() to construct. - private SnapshotDescription(com.google.protobuf.GeneratedMessage.Builder builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private SnapshotDescription(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SnapshotDescription defaultInstance; - public static SnapshotDescription getDefaultInstance() { - return defaultInstance; - } - - public SnapshotDescription getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private SnapshotDescription( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - name_ = input.readBytes(); - break; - } - case 18: { - bitField0_ |= 0x00000002; - table_ = input.readBytes(); - break; - } - case 24: { - bitField0_ |= 0x00000004; - creationTime_ = input.readInt64(); - break; - } - case 32: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type value = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(4, rawValue); - } else { - bitField0_ |= 0x00000008; - type_ = value; - } - break; - } - case 40: { - bitField0_ |= 0x00000010; - version_ = input.readInt32(); - break; - } - case 50: { - bitField0_ |= 0x00000020; - owner_ = input.readBytes(); - break; - } - case 58: { - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder subBuilder = null; - if (((bitField0_ & 0x00000040) == 0x00000040)) { - subBuilder = usersAndPermissions_.toBuilder(); - } - usersAndPermissions_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(usersAndPermissions_); - usersAndPermissions_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000040; - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder.class); - } - - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SnapshotDescription parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SnapshotDescription(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - /** - * Protobuf enum {@code hbase.pb.SnapshotDescription.Type} - */ - public enum Type - implements com.google.protobuf.ProtocolMessageEnum { - /** - * DISABLED = 0; - */ - DISABLED(0, 0), - /** - * FLUSH = 1; - */ - FLUSH(1, 1), - /** - * SKIPFLUSH = 2; - */ - SKIPFLUSH(2, 2), - ; - - /** - * DISABLED = 0; - */ - public static final int DISABLED_VALUE = 0; - /** - * FLUSH = 1; - */ - public static final int FLUSH_VALUE = 1; - /** - * SKIPFLUSH = 2; - */ - public static final int SKIPFLUSH_VALUE = 2; - - - public final int getNumber() { return value; } - - public static Type valueOf(int value) { - switch (value) { - case 0: return DISABLED; - case 1: return FLUSH; - case 2: return SKIPFLUSH; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap() { - public Type findValueByNumber(int number) { - return Type.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDescriptor().getEnumTypes().get(0); - } - - private static final Type[] VALUES = values(); - - public static Type valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private Type(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:hbase.pb.SnapshotDescription.Type) - } - - private int bitField0_; - // required string name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private java.lang.Object name_; - /** - * required string name = 1; - */ - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required string name = 1; - */ - public java.lang.String getName() { - java.lang.Object ref = name_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - name_ = s; - } - return s; - } - } - /** - * required string name = 1; - */ - public com.google.protobuf.ByteString - getNameBytes() { - java.lang.Object ref = name_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - name_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional string table = 2; - public static final int TABLE_FIELD_NUMBER = 2; - private java.lang.Object table_; - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public boolean hasTable() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public java.lang.String getTable() { - java.lang.Object ref = table_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - table_ = s; - } - return s; - } - } - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public com.google.protobuf.ByteString - getTableBytes() { - java.lang.Object ref = table_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - table_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional int64 creation_time = 3 [default = 0]; - public static final int CREATION_TIME_FIELD_NUMBER = 3; - private long creationTime_; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public boolean hasCreationTime() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public long getCreationTime() { - return creationTime_; - } - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - public static final int TYPE_FIELD_NUMBER = 4; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type type_; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public boolean hasType() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType() { - return type_; - } - - // optional int32 version = 5; - public static final int VERSION_FIELD_NUMBER = 5; - private int version_; - /** - * optional int32 version = 5; - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional int32 version = 5; - */ - public int getVersion() { - return version_; - } - - // optional string owner = 6; - public static final int OWNER_FIELD_NUMBER = 6; - private java.lang.Object owner_; - /** - * optional string owner = 6; - */ - public boolean hasOwner() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * optional string owner = 6; - */ - public java.lang.String getOwner() { - java.lang.Object ref = owner_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - owner_ = s; - } - return s; - } - } - /** - * optional string owner = 6; - */ - public com.google.protobuf.ByteString - getOwnerBytes() { - java.lang.Object ref = owner_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - owner_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - public static final int USERS_AND_PERMISSIONS_FIELD_NUMBER = 7; - private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public boolean hasUsersAndPermissions() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { - return usersAndPermissions_; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { - return usersAndPermissions_; - } - - private void initFields() { - name_ = ""; - table_ = ""; - creationTime_ = 0L; - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - version_ = 0; - owner_ = ""; - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasName()) { - memoizedIsInitialized = 0; - return false; - } - if (hasUsersAndPermissions()) { - if (!getUsersAndPermissions().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getTableBytes()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt64(3, creationTime_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeEnum(4, type_.getNumber()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeInt32(5, version_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, getOwnerBytes()); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeMessage(7, usersAndPermissions_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getTableBytes()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeInt64Size(3, creationTime_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeEnumSize(4, type_.getNumber()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(5, version_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, getOwnerBytes()); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(7, usersAndPermissions_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription other = (org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) obj; - - boolean result = true; - result = result && (hasName() == other.hasName()); - if (hasName()) { - result = result && getName() - .equals(other.getName()); - } - result = result && (hasTable() == other.hasTable()); - if (hasTable()) { - result = result && getTable() - .equals(other.getTable()); - } - result = result && (hasCreationTime() == other.hasCreationTime()); - if (hasCreationTime()) { - result = result && (getCreationTime() - == other.getCreationTime()); - } - result = result && (hasType() == other.hasType()); - if (hasType()) { - result = result && - (getType() == other.getType()); - } - result = result && (hasVersion() == other.hasVersion()); - if (hasVersion()) { - result = result && (getVersion() - == other.getVersion()); - } - result = result && (hasOwner() == other.hasOwner()); - if (hasOwner()) { - result = result && getOwner() - .equals(other.getOwner()); - } - result = result && (hasUsersAndPermissions() == other.hasUsersAndPermissions()); - if (hasUsersAndPermissions()) { - result = result && getUsersAndPermissions() - .equals(other.getUsersAndPermissions()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasName()) { - hash = (37 * hash) + NAME_FIELD_NUMBER; - hash = (53 * hash) + getName().hashCode(); - } - if (hasTable()) { - hash = (37 * hash) + TABLE_FIELD_NUMBER; - hash = (53 * hash) + getTable().hashCode(); - } - if (hasCreationTime()) { - hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getCreationTime()); - } - if (hasType()) { - hash = (37 * hash) + TYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getType()); - } - if (hasVersion()) { - hash = (37 * hash) + VERSION_FIELD_NUMBER; - hash = (53 * hash) + getVersion(); - } - if (hasOwner()) { - hash = (37 * hash) + OWNER_FIELD_NUMBER; - hash = (53 * hash) + getOwner().hashCode(); - } - if (hasUsersAndPermissions()) { - hash = (37 * hash) + USERS_AND_PERMISSIONS_FIELD_NUMBER; - hash = (53 * hash) + getUsersAndPermissions().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.SnapshotDescription} - * - *
-     **
-     * Description of the snapshot to take
-     * 
- */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getUsersAndPermissionsFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - name_ = ""; - bitField0_ = (bitField0_ & ~0x00000001); - table_ = ""; - bitField0_ = (bitField0_ & ~0x00000002); - creationTime_ = 0L; - bitField0_ = (bitField0_ & ~0x00000004); - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - bitField0_ = (bitField0_ & ~0x00000008); - version_ = 0; - bitField0_ = (bitField0_ & ~0x00000010); - owner_ = ""; - bitField0_ = (bitField0_ & ~0x00000020); - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - } else { - usersAndPermissionsBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000040); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription build() { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription result = new org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.name_ = name_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.table_ = table_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.creationTime_ = creationTime_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.type_ = type_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - result.version_ = version_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; - } - result.owner_ = owner_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { - to_bitField0_ |= 0x00000040; - } - if (usersAndPermissionsBuilder_ == null) { - result.usersAndPermissions_ = usersAndPermissions_; - } else { - result.usersAndPermissions_ = usersAndPermissionsBuilder_.build(); - } - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) return this; - if (other.hasName()) { - bitField0_ |= 0x00000001; - name_ = other.name_; - onChanged(); - } - if (other.hasTable()) { - bitField0_ |= 0x00000002; - table_ = other.table_; - onChanged(); - } - if (other.hasCreationTime()) { - setCreationTime(other.getCreationTime()); - } - if (other.hasType()) { - setType(other.getType()); - } - if (other.hasVersion()) { - setVersion(other.getVersion()); - } - if (other.hasOwner()) { - bitField0_ |= 0x00000020; - owner_ = other.owner_; - onChanged(); - } - if (other.hasUsersAndPermissions()) { - mergeUsersAndPermissions(other.getUsersAndPermissions()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasName()) { - - return false; - } - if (hasUsersAndPermissions()) { - if (!getUsersAndPermissions().isInitialized()) { - - return false; - } - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // required string name = 1; - private java.lang.Object name_ = ""; - /** - * required string name = 1; - */ - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required string name = 1; - */ - public java.lang.String getName() { - java.lang.Object ref = name_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - name_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * required string name = 1; - */ - public com.google.protobuf.ByteString - getNameBytes() { - java.lang.Object ref = name_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - name_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * required string name = 1; - */ - public Builder setName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - /** - * required string name = 1; - */ - public Builder clearName() { - bitField0_ = (bitField0_ & ~0x00000001); - name_ = getDefaultInstance().getName(); - onChanged(); - return this; - } - /** - * required string name = 1; - */ - public Builder setNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - - // optional string table = 2; - private java.lang.Object table_ = ""; - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public boolean hasTable() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public java.lang.String getTable() { - java.lang.Object ref = table_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - table_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public com.google.protobuf.ByteString - getTableBytes() { - java.lang.Object ref = table_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - table_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder setTable( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - table_ = value; - onChanged(); - return this; - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder clearTable() { - bitField0_ = (bitField0_ & ~0x00000002); - table_ = getDefaultInstance().getTable(); - onChanged(); - return this; - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder setTableBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - table_ = value; - onChanged(); - return this; - } - - // optional int64 creation_time = 3 [default = 0]; - private long creationTime_ ; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public boolean hasCreationTime() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public long getCreationTime() { - return creationTime_; - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public Builder setCreationTime(long value) { - bitField0_ |= 0x00000004; - creationTime_ = value; - onChanged(); - return this; - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public Builder clearCreationTime() { - bitField0_ = (bitField0_ & ~0x00000004); - creationTime_ = 0L; - onChanged(); - return this; - } - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public boolean hasType() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType() { - return type_; - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public Builder setType(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000008; - type_ = value; - onChanged(); - return this; - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public Builder clearType() { - bitField0_ = (bitField0_ & ~0x00000008); - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - onChanged(); - return this; - } - - // optional int32 version = 5; - private int version_ ; - /** - * optional int32 version = 5; - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional int32 version = 5; - */ - public int getVersion() { - return version_; - } - /** - * optional int32 version = 5; - */ - public Builder setVersion(int value) { - bitField0_ |= 0x00000010; - version_ = value; - onChanged(); - return this; - } - /** - * optional int32 version = 5; - */ - public Builder clearVersion() { - bitField0_ = (bitField0_ & ~0x00000010); - version_ = 0; - onChanged(); - return this; - } - - // optional string owner = 6; - private java.lang.Object owner_ = ""; - /** - * optional string owner = 6; - */ - public boolean hasOwner() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * optional string owner = 6; - */ - public java.lang.String getOwner() { - java.lang.Object ref = owner_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - owner_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * optional string owner = 6; - */ - public com.google.protobuf.ByteString - getOwnerBytes() { - java.lang.Object ref = owner_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - owner_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * optional string owner = 6; - */ - public Builder setOwner( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - owner_ = value; - onChanged(); - return this; - } - /** - * optional string owner = 6; - */ - public Builder clearOwner() { - bitField0_ = (bitField0_ & ~0x00000020); - owner_ = getDefaultInstance().getOwner(); - onChanged(); - return this; - } - /** - * optional string owner = 6; - */ - public Builder setOwnerBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - owner_ = value; - onChanged(); - return this; - } - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> usersAndPermissionsBuilder_; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public boolean hasUsersAndPermissions() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { - if (usersAndPermissionsBuilder_ == null) { - return usersAndPermissions_; - } else { - return usersAndPermissionsBuilder_.getMessage(); - } - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder setUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { - if (usersAndPermissionsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - usersAndPermissions_ = value; - onChanged(); - } else { - usersAndPermissionsBuilder_.setMessage(value); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder setUsersAndPermissions( - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder builderForValue) { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = builderForValue.build(); - onChanged(); - } else { - usersAndPermissionsBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder mergeUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { - if (usersAndPermissionsBuilder_ == null) { - if (((bitField0_ & 0x00000040) == 0x00000040) && - usersAndPermissions_ != org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance()) { - usersAndPermissions_ = - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.newBuilder(usersAndPermissions_).mergeFrom(value).buildPartial(); - } else { - usersAndPermissions_ = value; - } - onChanged(); - } else { - usersAndPermissionsBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder clearUsersAndPermissions() { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - onChanged(); - } else { - usersAndPermissionsBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000040); - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder getUsersAndPermissionsBuilder() { - bitField0_ |= 0x00000040; - onChanged(); - return getUsersAndPermissionsFieldBuilder().getBuilder(); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { - if (usersAndPermissionsBuilder_ != null) { - return usersAndPermissionsBuilder_.getMessageOrBuilder(); - } else { - return usersAndPermissions_; - } - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> - getUsersAndPermissionsFieldBuilder() { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissionsBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder>( - usersAndPermissions_, - getParentForChildren(), - isClean()); - usersAndPermissions_ = null; - } - return usersAndPermissionsBuilder_; - } - - // @@protoc_insertion_point(builder_scope:hbase.pb.SnapshotDescription) - } - - static { - defaultInstance = new SnapshotDescription(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotDescription) - } - public interface SnapshotFileInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -6173,11 +4685,6 @@ public final class SnapshotProtos { // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotDataManifest) } - private static com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_SnapshotDescription_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SnapshotFileInfo_descriptor; private static @@ -6212,52 +4719,39 @@ public final class SnapshotProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\016Snapshot.proto\022\010hbase.pb\032\023AccessContro" + - "l.proto\032\010FS.proto\032\013HBase.proto\"\223\002\n\023Snaps" + - "hotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002 " + - "\001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\0227\n\004type\030\004 " + - "\001(\0162\".hbase.pb.SnapshotDescription.Type:" + - "\005FLUSH\022\017\n\007version\030\005 \001(\005\022\r\n\005owner\030\006 \001(\t\022<" + - "\n\025users_and_permissions\030\007 \001(\0132\035.hbase.pb" + - ".UsersAndPermissions\".\n\004Type\022\014\n\010DISABLED" + - "\020\000\022\t\n\005FLUSH\020\001\022\r\n\tSKIPFLUSH\020\002\"\222\001\n\020Snapsho" + - "tFileInfo\022-\n\004type\030\001 \002(\0162\037.hbase.pb.Snaps", - "hotFileInfo.Type\022\r\n\005hfile\030\003 \001(\t\022\022\n\nwal_s" + - "erver\030\004 \001(\t\022\020\n\010wal_name\030\005 \001(\t\"\032\n\004Type\022\t\n" + - "\005HFILE\020\001\022\007\n\003WAL\020\002\"\323\002\n\026SnapshotRegionMani" + - "fest\022\017\n\007version\030\001 \001(\005\022)\n\013region_info\030\002 \002" + - "(\0132\024.hbase.pb.RegionInfo\022B\n\014family_files" + - "\030\003 \003(\0132,.hbase.pb.SnapshotRegionManifest" + - ".FamilyFiles\032T\n\tStoreFile\022\014\n\004name\030\001 \002(\t\022" + - "&\n\treference\030\002 \001(\0132\023.hbase.pb.Reference\022" + - "\021\n\tfile_size\030\003 \001(\004\032c\n\013FamilyFiles\022\023\n\013fam" + - "ily_name\030\001 \002(\014\022?\n\013store_files\030\002 \003(\0132*.hb", - "ase.pb.SnapshotRegionManifest.StoreFile\"" + - "\177\n\024SnapshotDataManifest\022+\n\014table_schema\030" + - "\001 \002(\0132\025.hbase.pb.TableSchema\022:\n\020region_m" + - "anifests\030\002 \003(\0132 .hbase.pb.SnapshotRegion" + - "ManifestBD\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\016SnapshotProtosH\001\210\001\001\240\001\001" + "\n\016Snapshot.proto\022\010hbase.pb\032\010FS.proto\032\013HB" + + "ase.proto\"\222\001\n\020SnapshotFileInfo\022-\n\004type\030\001" + + " \002(\0162\037.hbase.pb.SnapshotFileInfo.Type\022\r\n" + + "\005hfile\030\003 \001(\t\022\022\n\nwal_server\030\004 \001(\t\022\020\n\010wal_" + + "name\030\005 \001(\t\"\032\n\004Type\022\t\n\005HFILE\020\001\022\007\n\003WAL\020\002\"\323" + + "\002\n\026SnapshotRegionManifest\022\017\n\007version\030\001 \001" + + "(\005\022)\n\013region_info\030\002 \002(\0132\024.hbase.pb.Regio" + + "nInfo\022B\n\014family_files\030\003 \003(\0132,.hbase.pb.S" + + "napshotRegionManifest.FamilyFiles\032T\n\tSto" + + "reFile\022\014\n\004name\030\001 \002(\t\022&\n\treference\030\002 \001(\0132", + "\023.hbase.pb.Reference\022\021\n\tfile_size\030\003 \001(\004\032" + + "c\n\013FamilyFiles\022\023\n\013family_name\030\001 \002(\014\022?\n\013s" + + "tore_files\030\002 \003(\0132*.hbase.pb.SnapshotRegi" + + "onManifest.StoreFile\"\177\n\024SnapshotDataMani" + + "fest\022+\n\014table_schema\030\001 \002(\0132\025.hbase.pb.Ta" + + "bleSchema\022:\n\020region_manifests\030\002 \003(\0132 .hb" + + "ase.pb.SnapshotRegionManifestBD\n*org.apa" + + "che.hadoop.hbase.protobuf.generatedB\016Sna" + + "pshotProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_hbase_pb_SnapshotDescription_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_hbase_pb_SnapshotDescription_descriptor, - new java.lang.String[] { "Name", "Table", "CreationTime", "Type", "Version", "Owner", "UsersAndPermissions", }); internal_static_hbase_pb_SnapshotFileInfo_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotFileInfo_descriptor, new java.lang.String[] { "Type", "Hfile", "WalServer", "WalName", }); internal_static_hbase_pb_SnapshotRegionManifest_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_SnapshotRegionManifest_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotRegionManifest_descriptor, @@ -6275,7 +4769,7 @@ public final class SnapshotProtos { internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor, new java.lang.String[] { "FamilyName", "StoreFiles", }); internal_static_hbase_pb_SnapshotDataManifest_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(2); internal_static_hbase_pb_SnapshotDataManifest_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotDataManifest_descriptor, @@ -6286,7 +4780,6 @@ public final class SnapshotProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.FSProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), }, assigner); diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java new file mode 100644 index 00000000000..9507d013131 --- /dev/null +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TableProtos.java @@ -0,0 +1,607 @@ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: Table.proto + +package org.apache.hadoop.hbase.protobuf.generated; + +public final class TableProtos { + private TableProtos() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistry registry) { + } + public interface TableNameOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required bytes namespace = 1; + /** + * required bytes namespace = 1; + */ + boolean hasNamespace(); + /** + * required bytes namespace = 1; + */ + com.google.protobuf.ByteString getNamespace(); + + // required bytes qualifier = 2; + /** + * required bytes qualifier = 2; + */ + boolean hasQualifier(); + /** + * required bytes qualifier = 2; + */ + com.google.protobuf.ByteString getQualifier(); + } + /** + * Protobuf type {@code hbase.pb.TableName} + * + *
+   **
+   * Table Name
+   * 
+ */ + public static final class TableName extends + com.google.protobuf.GeneratedMessage + implements TableNameOrBuilder { + // Use TableName.newBuilder() to construct. + private TableName(com.google.protobuf.GeneratedMessage.Builder builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private TableName(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final TableName defaultInstance; + public static TableName getDefaultInstance() { + return defaultInstance; + } + + public TableName getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private TableName( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + namespace_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + qualifier_ = input.readBytes(); + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class); + } + + public static com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + public TableName parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TableName(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + private int bitField0_; + // required bytes namespace = 1; + public static final int NAMESPACE_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString namespace_; + /** + * required bytes namespace = 1; + */ + public boolean hasNamespace() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * required bytes namespace = 1; + */ + public com.google.protobuf.ByteString getNamespace() { + return namespace_; + } + + // required bytes qualifier = 2; + public static final int QUALIFIER_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString qualifier_; + /** + * required bytes qualifier = 2; + */ + public boolean hasQualifier() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * required bytes qualifier = 2; + */ + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + + private void initFields() { + namespace_ = com.google.protobuf.ByteString.EMPTY; + qualifier_ = com.google.protobuf.ByteString.EMPTY; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasNamespace()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasQualifier()) { + memoizedIsInitialized = 0; + return false; + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, namespace_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, qualifier_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, namespace_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, qualifier_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)) { + return super.equals(obj); + } + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) obj; + + boolean result = true; + result = result && (hasNamespace() == other.hasNamespace()); + if (hasNamespace()) { + result = result && getNamespace() + .equals(other.getNamespace()); + } + result = result && (hasQualifier() == other.hasQualifier()); + if (hasQualifier()) { + result = result && getQualifier() + .equals(other.getQualifier()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasNamespace()) { + hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; + hash = (53 * hash) + getNamespace().hashCode(); + } + if (hasQualifier()) { + hash = (37 * hash) + QUALIFIER_FIELD_NUMBER; + hash = (53 * hash) + getQualifier().hashCode(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code hbase.pb.TableName} + * + *
+     **
+     * Table Name
+     * 
+ */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder + implements org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.class, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder.class); + } + + // Construct using org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + namespace_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + qualifier_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000002); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.internal_static_hbase_pb_TableName_descriptor; + } + + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName build() { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName result = new org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.namespace_ = namespace_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.qualifier_ = qualifier_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) return this; + if (other.hasNamespace()) { + setNamespace(other.getNamespace()); + } + if (other.hasQualifier()) { + setQualifier(other.getQualifier()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + if (!hasNamespace()) { + + return false; + } + if (!hasQualifier()) { + + return false; + } + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + private int bitField0_; + + // required bytes namespace = 1; + private com.google.protobuf.ByteString namespace_ = com.google.protobuf.ByteString.EMPTY; + /** + * required bytes namespace = 1; + */ + public boolean hasNamespace() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * required bytes namespace = 1; + */ + public com.google.protobuf.ByteString getNamespace() { + return namespace_; + } + /** + * required bytes namespace = 1; + */ + public Builder setNamespace(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + namespace_ = value; + onChanged(); + return this; + } + /** + * required bytes namespace = 1; + */ + public Builder clearNamespace() { + bitField0_ = (bitField0_ & ~0x00000001); + namespace_ = getDefaultInstance().getNamespace(); + onChanged(); + return this; + } + + // required bytes qualifier = 2; + private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY; + /** + * required bytes qualifier = 2; + */ + public boolean hasQualifier() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * required bytes qualifier = 2; + */ + public com.google.protobuf.ByteString getQualifier() { + return qualifier_; + } + /** + * required bytes qualifier = 2; + */ + public Builder setQualifier(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + qualifier_ = value; + onChanged(); + return this; + } + /** + * required bytes qualifier = 2; + */ + public Builder clearQualifier() { + bitField0_ = (bitField0_ & ~0x00000002); + qualifier_ = getDefaultInstance().getQualifier(); + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:hbase.pb.TableName) + } + + static { + defaultInstance = new TableName(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:hbase.pb.TableName) + } + + private static com.google.protobuf.Descriptors.Descriptor + internal_static_hbase_pb_TableName_descriptor; + private static + com.google.protobuf.GeneratedMessage.FieldAccessorTable + internal_static_hbase_pb_TableName_fieldAccessorTable; + + public static com.google.protobuf.Descriptors.FileDescriptor + getDescriptor() { + return descriptor; + } + private static com.google.protobuf.Descriptors.FileDescriptor + descriptor; + static { + java.lang.String[] descriptorData = { + "\n\013Table.proto\022\010hbase.pb\"1\n\tTableName\022\021\n\t" + + "namespace\030\001 \002(\014\022\021\n\tqualifier\030\002 \002(\014B>\n*or" + + "g.apache.hadoop.hbase.protobuf.generated" + + "B\013TableProtosH\001\240\001\001" + }; + com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = + new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { + public com.google.protobuf.ExtensionRegistry assignDescriptors( + com.google.protobuf.Descriptors.FileDescriptor root) { + descriptor = root; + internal_static_hbase_pb_TableName_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_hbase_pb_TableName_fieldAccessorTable = new + com.google.protobuf.GeneratedMessage.FieldAccessorTable( + internal_static_hbase_pb_TableName_descriptor, + new java.lang.String[] { "Namespace", "Qualifier", }); + return null; + } + }; + com.google.protobuf.Descriptors.FileDescriptor + .internalBuildGeneratedFileFrom(descriptorData, + new com.google.protobuf.Descriptors.FileDescriptor[] { + }, assigner); + } + + // @@protoc_insertion_point(outer_class_scope) +} diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java index 9bae06fb8cb..a466e6cb7cd 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java @@ -8813,11 +8813,11 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // required bytes encoded_region_name = 2; /** @@ -8921,11 +8921,11 @@ public final class WALProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -8996,7 +8996,7 @@ public final class WALProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -9006,13 +9006,13 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -9085,7 +9085,7 @@ public final class WALProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); encodedRegionName_ = com.google.protobuf.ByteString.EMPTY; stores_ = java.util.Collections.emptyList(); bulkloadSeqNum_ = 0L; @@ -9347,7 +9347,7 @@ public final class WALProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -9516,9 +9516,9 @@ public final class WALProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -9528,7 +9528,7 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -9538,7 +9538,7 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -9555,7 +9555,7 @@ public final class WALProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -9568,12 +9568,12 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -9589,7 +9589,7 @@ public final class WALProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -9600,7 +9600,7 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -9608,7 +9608,7 @@ public final class WALProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -9619,11 +9619,11 @@ public final class WALProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -11978,56 +11978,56 @@ public final class WALProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\tWAL.proto\022\010hbase.pb\032\013HBase.proto\032\014Clie" + - "nt.proto\"\217\001\n\tWALHeader\022\027\n\017has_compressio" + - "n\030\001 \001(\010\022\026\n\016encryption_key\030\002 \001(\014\022\033\n\023has_t" + - "ag_compression\030\003 \001(\010\022\027\n\017writer_cls_name\030" + - "\004 \001(\t\022\033\n\023cell_codec_cls_name\030\005 \001(\t\"\273\002\n\006W" + - "ALKey\022\033\n\023encoded_region_name\030\001 \002(\014\022\022\n\nta" + - "ble_name\030\002 \002(\014\022\033\n\023log_sequence_number\030\003 " + - "\002(\004\022\022\n\nwrite_time\030\004 \002(\004\022&\n\ncluster_id\030\005 " + - "\001(\0132\016.hbase.pb.UUIDB\002\030\001\022%\n\006scopes\030\006 \003(\0132" + - "\025.hbase.pb.FamilyScope\022\032\n\022following_kv_c", - "ount\030\007 \001(\r\022#\n\013cluster_ids\030\010 \003(\0132\016.hbase." + - "pb.UUID\022\022\n\nnonceGroup\030\t \001(\004\022\r\n\005nonce\030\n \001" + - "(\004\022\034\n\024orig_sequence_number\030\013 \001(\004\"F\n\013Fami" + - "lyScope\022\016\n\006family\030\001 \002(\014\022\'\n\nscope_type\030\002 " + - "\002(\0162\023.hbase.pb.ScopeType\"\276\001\n\024CompactionD" + - "escriptor\022\022\n\ntable_name\030\001 \002(\014\022\033\n\023encoded" + - "_region_name\030\002 \002(\014\022\023\n\013family_name\030\003 \002(\014\022" + - "\030\n\020compaction_input\030\004 \003(\t\022\031\n\021compaction_" + - "output\030\005 \003(\t\022\026\n\016store_home_dir\030\006 \002(\t\022\023\n\013" + - "region_name\030\007 \001(\014\"\244\003\n\017FlushDescriptor\0225\n", - "\006action\030\001 \002(\0162%.hbase.pb.FlushDescriptor" + - ".FlushAction\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023enco" + - "ded_region_name\030\003 \002(\014\022\035\n\025flush_sequence_" + - "number\030\004 \001(\004\022E\n\rstore_flushes\030\005 \003(\0132..hb" + - "ase.pb.FlushDescriptor.StoreFlushDescrip" + - "tor\022\023\n\013region_name\030\006 \001(\014\032Y\n\024StoreFlushDe" + - "scriptor\022\023\n\013family_name\030\001 \002(\014\022\026\n\016store_h" + - "ome_dir\030\002 \002(\t\022\024\n\014flush_output\030\003 \003(\t\"S\n\013F" + - "lushAction\022\017\n\013START_FLUSH\020\000\022\020\n\014COMMIT_FL" + - "USH\020\001\022\017\n\013ABORT_FLUSH\020\002\022\020\n\014CANNOT_FLUSH\020\003", - "\"q\n\017StoreDescriptor\022\023\n\013family_name\030\001 \002(\014" + - "\022\026\n\016store_home_dir\030\002 \002(\t\022\022\n\nstore_file\030\003" + - " \003(\t\022\035\n\025store_file_size_bytes\030\004 \001(\004\"\237\001\n\022" + - "BulkLoadDescriptor\022\'\n\ntable_name\030\001 \002(\0132\023" + - ".hbase.pb.TableName\022\033\n\023encoded_region_na" + - "me\030\002 \002(\014\022)\n\006stores\030\003 \003(\0132\031.hbase.pb.Stor" + - "eDescriptor\022\030\n\020bulkload_seq_num\030\004 \002(\003\"\272\002" + - "\n\025RegionEventDescriptor\022=\n\nevent_type\030\001 " + - "\002(\0162).hbase.pb.RegionEventDescriptor.Eve" + - "ntType\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023encoded_re", - "gion_name\030\003 \002(\014\022\033\n\023log_sequence_number\030\004" + - " \001(\004\022)\n\006stores\030\005 \003(\0132\031.hbase.pb.StoreDes" + - "criptor\022$\n\006server\030\006 \001(\0132\024.hbase.pb.Serve" + - "rName\022\023\n\013region_name\030\007 \001(\014\".\n\tEventType\022" + - "\017\n\013REGION_OPEN\020\000\022\020\n\014REGION_CLOSE\020\001\"\014\n\nWA" + - "LTrailer*d\n\tScopeType\022\033\n\027REPLICATION_SCO" + - "PE_LOCAL\020\000\022\034\n\030REPLICATION_SCOPE_GLOBAL\020\001" + - "\022\034\n\030REPLICATION_SCOPE_SERIAL\020\002B?\n*org.ap" + - "ache.hadoop.hbase.protobuf.generatedB\tWA" + - "LProtosH\001\210\001\000\240\001\001" + "\n\tWAL.proto\022\010hbase.pb\032\013Table.proto\032\013HBas" + + "e.proto\032\014Client.proto\"\217\001\n\tWALHeader\022\027\n\017h" + + "as_compression\030\001 \001(\010\022\026\n\016encryption_key\030\002" + + " \001(\014\022\033\n\023has_tag_compression\030\003 \001(\010\022\027\n\017wri" + + "ter_cls_name\030\004 \001(\t\022\033\n\023cell_codec_cls_nam" + + "e\030\005 \001(\t\"\273\002\n\006WALKey\022\033\n\023encoded_region_nam" + + "e\030\001 \002(\014\022\022\n\ntable_name\030\002 \002(\014\022\033\n\023log_seque" + + "nce_number\030\003 \002(\004\022\022\n\nwrite_time\030\004 \002(\004\022&\n\n" + + "cluster_id\030\005 \001(\0132\016.hbase.pb.UUIDB\002\030\001\022%\n\006" + + "scopes\030\006 \003(\0132\025.hbase.pb.FamilyScope\022\032\n\022f", + "ollowing_kv_count\030\007 \001(\r\022#\n\013cluster_ids\030\010" + + " \003(\0132\016.hbase.pb.UUID\022\022\n\nnonceGroup\030\t \001(\004" + + "\022\r\n\005nonce\030\n \001(\004\022\034\n\024orig_sequence_number\030" + + "\013 \001(\004\"F\n\013FamilyScope\022\016\n\006family\030\001 \002(\014\022\'\n\n" + + "scope_type\030\002 \002(\0162\023.hbase.pb.ScopeType\"\276\001" + + "\n\024CompactionDescriptor\022\022\n\ntable_name\030\001 \002" + + "(\014\022\033\n\023encoded_region_name\030\002 \002(\014\022\023\n\013famil" + + "y_name\030\003 \002(\014\022\030\n\020compaction_input\030\004 \003(\t\022\031" + + "\n\021compaction_output\030\005 \003(\t\022\026\n\016store_home_" + + "dir\030\006 \002(\t\022\023\n\013region_name\030\007 \001(\014\"\244\003\n\017Flush", + "Descriptor\0225\n\006action\030\001 \002(\0162%.hbase.pb.Fl" + + "ushDescriptor.FlushAction\022\022\n\ntable_name\030" + + "\002 \002(\014\022\033\n\023encoded_region_name\030\003 \002(\014\022\035\n\025fl" + + "ush_sequence_number\030\004 \001(\004\022E\n\rstore_flush" + + "es\030\005 \003(\0132..hbase.pb.FlushDescriptor.Stor" + + "eFlushDescriptor\022\023\n\013region_name\030\006 \001(\014\032Y\n" + + "\024StoreFlushDescriptor\022\023\n\013family_name\030\001 \002" + + "(\014\022\026\n\016store_home_dir\030\002 \002(\t\022\024\n\014flush_outp" + + "ut\030\003 \003(\t\"S\n\013FlushAction\022\017\n\013START_FLUSH\020\000" + + "\022\020\n\014COMMIT_FLUSH\020\001\022\017\n\013ABORT_FLUSH\020\002\022\020\n\014C", + "ANNOT_FLUSH\020\003\"q\n\017StoreDescriptor\022\023\n\013fami" + + "ly_name\030\001 \002(\014\022\026\n\016store_home_dir\030\002 \002(\t\022\022\n" + + "\nstore_file\030\003 \003(\t\022\035\n\025store_file_size_byt" + + "es\030\004 \001(\004\"\237\001\n\022BulkLoadDescriptor\022\'\n\ntable" + + "_name\030\001 \002(\0132\023.hbase.pb.TableName\022\033\n\023enco" + + "ded_region_name\030\002 \002(\014\022)\n\006stores\030\003 \003(\0132\031." + + "hbase.pb.StoreDescriptor\022\030\n\020bulkload_seq" + + "_num\030\004 \002(\003\"\272\002\n\025RegionEventDescriptor\022=\n\n" + + "event_type\030\001 \002(\0162).hbase.pb.RegionEventD" + + "escriptor.EventType\022\022\n\ntable_name\030\002 \002(\014\022", + "\033\n\023encoded_region_name\030\003 \002(\014\022\033\n\023log_sequ" + + "ence_number\030\004 \001(\004\022)\n\006stores\030\005 \003(\0132\031.hbas" + + "e.pb.StoreDescriptor\022$\n\006server\030\006 \001(\0132\024.h" + + "base.pb.ServerName\022\023\n\013region_name\030\007 \001(\014\"" + + ".\n\tEventType\022\017\n\013REGION_OPEN\020\000\022\020\n\014REGION_" + + "CLOSE\020\001\"\014\n\nWALTrailer*d\n\tScopeType\022\033\n\027RE" + + "PLICATION_SCOPE_LOCAL\020\000\022\034\n\030REPLICATION_S" + + "COPE_GLOBAL\020\001\022\034\n\030REPLICATION_SCOPE_SERIA" + + "L\020\002B?\n*org.apache.hadoop.hbase.protobuf." + + "generatedB\tWALProtosH\001\210\001\000\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -12100,6 +12100,7 @@ public final class WALProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(), }, assigner); diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java index fb06a78333e..fc181a8bc18 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java @@ -5043,11 +5043,11 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // repeated bytes families = 2; /** @@ -5115,11 +5115,11 @@ public final class ZooKeeperProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -5180,7 +5180,7 @@ public final class ZooKeeperProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -5190,13 +5190,13 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -5224,7 +5224,7 @@ public final class ZooKeeperProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); families_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; @@ -5435,7 +5435,7 @@ public final class ZooKeeperProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -5546,9 +5546,9 @@ public final class ZooKeeperProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -5558,7 +5558,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -5568,7 +5568,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -5585,7 +5585,7 @@ public final class ZooKeeperProtos { * optional .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -5598,12 +5598,12 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -5619,7 +5619,7 @@ public final class ZooKeeperProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -5630,7 +5630,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -5638,7 +5638,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -5649,11 +5649,11 @@ public final class ZooKeeperProtos { * optional .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -9260,11 +9260,11 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional .hbase.pb.ServerName lock_owner = 2; /** @@ -9382,11 +9382,11 @@ public final class ZooKeeperProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -9469,7 +9469,7 @@ public final class ZooKeeperProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -9479,13 +9479,13 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } @@ -9603,7 +9603,7 @@ public final class ZooKeeperProtos { } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); threadId_ = 0L; isShared_ = false; @@ -9892,7 +9892,7 @@ public final class ZooKeeperProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -10047,9 +10047,9 @@ public final class ZooKeeperProtos { private int bitField0_; // optional .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 1; */ @@ -10059,7 +10059,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -10069,7 +10069,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -10086,7 +10086,7 @@ public final class ZooKeeperProtos { * optional .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -10099,12 +10099,12 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -10120,7 +10120,7 @@ public final class ZooKeeperProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -10131,7 +10131,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -10139,7 +10139,7 @@ public final class ZooKeeperProtos { /** * optional .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -10150,11 +10150,11 @@ public final class ZooKeeperProtos { * optional .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -10982,47 +10982,47 @@ public final class ZooKeeperProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\017ZooKeeper.proto\022\010hbase.pb\032\013HBase.proto" + - "\032\023ClusterStatus.proto\"y\n\020MetaRegionServe" + - "r\022$\n\006server\030\001 \002(\0132\024.hbase.pb.ServerName\022" + - "\023\n\013rpc_version\030\002 \001(\r\022*\n\005state\030\003 \001(\0162\033.hb" + - "ase.pb.RegionState.State\"V\n\006Master\022$\n\006ma" + - "ster\030\001 \002(\0132\024.hbase.pb.ServerName\022\023\n\013rpc_" + - "version\030\002 \001(\r\022\021\n\tinfo_port\030\003 \001(\r\"\037\n\tClus" + - "terUp\022\022\n\nstart_date\030\001 \002(\t\"\221\001\n\020RegionTran" + - "sition\022\027\n\017event_type_code\030\001 \002(\r\022\023\n\013regio" + - "n_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(\004\022)\n\013ser", - "ver_name\030\004 \002(\0132\024.hbase.pb.ServerName\022\017\n\007" + - "payload\030\005 \001(\014\"\247\002\n\014SplitLogTask\022+\n\005state\030" + - "\001 \002(\0162\034.hbase.pb.SplitLogTask.State\022)\n\013s" + - "erver_name\030\002 \002(\0132\024.hbase.pb.ServerName\022:" + - "\n\004mode\030\003 \001(\0162#.hbase.pb.SplitLogTask.Rec" + - "overyMode:\007UNKNOWN\"C\n\005State\022\016\n\nUNASSIGNE" + - "D\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004DONE\020\003\022\007" + - "\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007UNKNOWN\020\000\022\021\n\r" + - "LOG_SPLITTING\020\001\022\016\n\nLOG_REPLAY\020\002\"w\n\005Table" + - "\022-\n\005state\030\001 \002(\0162\025.hbase.pb.Table.State:\007", - "ENABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLE" + - "D\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"D\n\007Tabl" + - "eCF\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Table" + - "Name\022\020\n\010families\030\002 \003(\014\"\330\001\n\017ReplicationPe" + - "er\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027replicationEnd" + - "pointImpl\030\002 \001(\t\022&\n\004data\030\003 \003(\0132\030.hbase.pb" + - ".BytesBytesPair\022/\n\rconfiguration\030\004 \003(\0132\030" + - ".hbase.pb.NameStringPair\022$\n\ttable_cfs\030\005 " + - "\003(\0132\021.hbase.pb.TableCF\022\021\n\tbandwidth\030\006 \001(" + - "\003\"g\n\020ReplicationState\022/\n\005state\030\001 \002(\0162 .h", - "base.pb.ReplicationState.State\"\"\n\005State\022" + - "\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020\001\"+\n\027Replicatio" + - "nHLogPosition\022\020\n\010position\030\001 \002(\003\"%\n\017Repli" + - "cationLock\022\022\n\nlock_owner\030\001 \002(\t\"\252\001\n\tTable" + - "Lock\022\'\n\ntable_name\030\001 \001(\0132\023.hbase.pb.Tabl" + - "eName\022(\n\nlock_owner\030\002 \001(\0132\024.hbase.pb.Ser" + - "verName\022\021\n\tthread_id\030\003 \001(\003\022\021\n\tis_shared\030" + - "\004 \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013create_time\030\006 " + - "\001(\003\"\036\n\013SwitchState\022\017\n\007enabled\030\001 \001(\010BE\n*o" + - "rg.apache.hadoop.hbase.protobuf.generate", - "dB\017ZooKeeperProtosH\001\210\001\001\240\001\001" + "\n\017ZooKeeper.proto\022\010hbase.pb\032\013Table.proto" + + "\032\013HBase.proto\032\023ClusterStatus.proto\"y\n\020Me" + + "taRegionServer\022$\n\006server\030\001 \002(\0132\024.hbase.p" + + "b.ServerName\022\023\n\013rpc_version\030\002 \001(\r\022*\n\005sta" + + "te\030\003 \001(\0162\033.hbase.pb.RegionState.State\"V\n" + + "\006Master\022$\n\006master\030\001 \002(\0132\024.hbase.pb.Serve" + + "rName\022\023\n\013rpc_version\030\002 \001(\r\022\021\n\tinfo_port\030" + + "\003 \001(\r\"\037\n\tClusterUp\022\022\n\nstart_date\030\001 \002(\t\"\221" + + "\001\n\020RegionTransition\022\027\n\017event_type_code\030\001" + + " \002(\r\022\023\n\013region_name\030\002 \002(\014\022\023\n\013create_time", + "\030\003 \002(\004\022)\n\013server_name\030\004 \002(\0132\024.hbase.pb.S" + + "erverName\022\017\n\007payload\030\005 \001(\014\"\247\002\n\014SplitLogT" + + "ask\022+\n\005state\030\001 \002(\0162\034.hbase.pb.SplitLogTa" + + "sk.State\022)\n\013server_name\030\002 \002(\0132\024.hbase.pb" + + ".ServerName\022:\n\004mode\030\003 \001(\0162#.hbase.pb.Spl" + + "itLogTask.RecoveryMode:\007UNKNOWN\"C\n\005State" + + "\022\016\n\nUNASSIGNED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020" + + "\002\022\010\n\004DONE\020\003\022\007\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007" + + "UNKNOWN\020\000\022\021\n\rLOG_SPLITTING\020\001\022\016\n\nLOG_REPL" + + "AY\020\002\"w\n\005Table\022-\n\005state\030\001 \002(\0162\025.hbase.pb.", + "Table.State:\007ENABLED\"?\n\005State\022\013\n\007ENABLED" + + "\020\000\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABL" + + "ING\020\003\"D\n\007TableCF\022\'\n\ntable_name\030\001 \001(\0132\023.h" + + "base.pb.TableName\022\020\n\010families\030\002 \003(\014\"\330\001\n\017" + + "ReplicationPeer\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027r" + + "eplicationEndpointImpl\030\002 \001(\t\022&\n\004data\030\003 \003" + + "(\0132\030.hbase.pb.BytesBytesPair\022/\n\rconfigur" + + "ation\030\004 \003(\0132\030.hbase.pb.NameStringPair\022$\n" + + "\ttable_cfs\030\005 \003(\0132\021.hbase.pb.TableCF\022\021\n\tb" + + "andwidth\030\006 \001(\003\"g\n\020ReplicationState\022/\n\005st", + "ate\030\001 \002(\0162 .hbase.pb.ReplicationState.St" + + "ate\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020\001\"" + + "+\n\027ReplicationHLogPosition\022\020\n\010position\030\001" + + " \002(\003\"%\n\017ReplicationLock\022\022\n\nlock_owner\030\001 " + + "\002(\t\"\252\001\n\tTableLock\022\'\n\ntable_name\030\001 \001(\0132\023." + + "hbase.pb.TableName\022(\n\nlock_owner\030\002 \001(\0132\024" + + ".hbase.pb.ServerName\022\021\n\tthread_id\030\003 \001(\003\022" + + "\021\n\tis_shared\030\004 \001(\010\022\017\n\007purpose\030\005 \001(\t\022\023\n\013c" + + "reate_time\030\006 \001(\003\"\036\n\013SwitchState\022\017\n\007enabl" + + "ed\030\001 \001(\010BE\n*org.apache.hadoop.hbase.prot", + "obuf.generatedB\017ZooKeeperProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -11113,6 +11113,7 @@ public final class ZooKeeperProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(), }, assigner); diff --git a/hbase-protocol/src/main/protobuf/AccessControl.proto b/hbase-protocol/src/main/protobuf/AccessControl.proto index cc0d4a53933..02e3a5b056e 100644 --- a/hbase-protocol/src/main/protobuf/AccessControl.proto +++ b/hbase-protocol/src/main/protobuf/AccessControl.proto @@ -23,7 +23,7 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; -import "HBase.proto"; +import "Table.proto"; message Permission { enum Action { diff --git a/hbase-protocol/src/main/protobuf/HBase.proto b/hbase-protocol/src/main/protobuf/HBase.proto index 279eb39d143..67645d4c9f4 100644 --- a/hbase-protocol/src/main/protobuf/HBase.proto +++ b/hbase-protocol/src/main/protobuf/HBase.proto @@ -25,14 +25,8 @@ option java_generate_equals_and_hash = true; option optimize_for = SPEED; import "Cell.proto"; - -/** - * Table Name - */ -message TableName { - required bytes namespace = 1; - required bytes qualifier = 2; -} +import "Table.proto"; +import "AccessControl.proto"; /** * Table Schema @@ -221,3 +215,21 @@ message RegionServerInfo { optional int32 infoPort = 1; optional VersionInfo version_info = 2; } + +/** + * Description of the snapshot to take + */ +message SnapshotDescription { + required string name = 1; + optional string table = 2; // not needed for delete, but checked for in taking snapshot + optional int64 creation_time = 3 [default = 0]; + enum Type { + DISABLED = 0; + FLUSH = 1; + SKIPFLUSH = 2; + } + optional Type type = 4 [default = FLUSH]; + optional int32 version = 5; + optional string owner = 6; + optional UsersAndPermissions users_and_permissions = 7; +} diff --git a/hbase-protocol/src/main/protobuf/Master.proto b/hbase-protocol/src/main/protobuf/Master.proto index d19856bc595..4db87791349 100644 --- a/hbase-protocol/src/main/protobuf/Master.proto +++ b/hbase-protocol/src/main/protobuf/Master.proto @@ -26,6 +26,7 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; +import "Table.proto"; import "HBase.proto"; import "Client.proto"; import "ClusterStatus.proto"; diff --git a/hbase-protocol/src/main/protobuf/MasterProcedure.proto b/hbase-protocol/src/main/protobuf/MasterProcedure.proto index 2d2aff4f248..c1d37892b56 100644 --- a/hbase-protocol/src/main/protobuf/MasterProcedure.proto +++ b/hbase-protocol/src/main/protobuf/MasterProcedure.proto @@ -23,6 +23,7 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; +import "Table.proto"; import "HBase.proto"; import "RPC.proto"; diff --git a/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto b/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto index 814735b2892..12e7cf766c5 100644 --- a/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto +++ b/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto @@ -23,6 +23,7 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; +import 'Table.proto'; import 'HBase.proto'; import 'Client.proto'; diff --git a/hbase-protocol/src/main/protobuf/Snapshot.proto b/hbase-protocol/src/main/protobuf/Snapshot.proto index 015787df955..ae1a1e680ac 100644 --- a/hbase-protocol/src/main/protobuf/Snapshot.proto +++ b/hbase-protocol/src/main/protobuf/Snapshot.proto @@ -23,28 +23,9 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; -import "AccessControl.proto"; import "FS.proto"; import "HBase.proto"; -/** - * Description of the snapshot to take - */ -message SnapshotDescription { - required string name = 1; - optional string table = 2; // not needed for delete, but checked for in taking snapshot - optional int64 creation_time = 3 [default = 0]; - enum Type { - DISABLED = 0; - FLUSH = 1; - SKIPFLUSH = 2; - } - optional Type type = 4 [default = FLUSH]; - optional int32 version = 5; - optional string owner = 6; - optional UsersAndPermissions users_and_permissions = 7; -} - message SnapshotFileInfo { enum Type { HFILE = 1; diff --git a/hbase-protocol/src/main/protobuf/Table.proto b/hbase-protocol/src/main/protobuf/Table.proto new file mode 100644 index 00000000000..4452eb2de30 --- /dev/null +++ b/hbase-protocol/src/main/protobuf/Table.proto @@ -0,0 +1,33 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +// This file contains protocol buffers that are shared throughout HBase +package hbase.pb; + +option java_package = "org.apache.hadoop.hbase.protobuf.generated"; +option java_outer_classname = "TableProtos"; +option java_generate_equals_and_hash = true; +option optimize_for = SPEED; + +/** + * Table Name + */ +message TableName { + required bytes namespace = 1; + required bytes qualifier = 2; +} diff --git a/hbase-protocol/src/main/protobuf/WAL.proto b/hbase-protocol/src/main/protobuf/WAL.proto index 5d91c4c9855..08925f82152 100644 --- a/hbase-protocol/src/main/protobuf/WAL.proto +++ b/hbase-protocol/src/main/protobuf/WAL.proto @@ -23,6 +23,7 @@ option java_generic_services = false; option java_generate_equals_and_hash = true; option optimize_for = SPEED; +import "Table.proto"; import "HBase.proto"; import "Client.proto"; diff --git a/hbase-protocol/src/main/protobuf/ZooKeeper.proto b/hbase-protocol/src/main/protobuf/ZooKeeper.proto index a632552daf7..1638bf70714 100644 --- a/hbase-protocol/src/main/protobuf/ZooKeeper.proto +++ b/hbase-protocol/src/main/protobuf/ZooKeeper.proto @@ -26,6 +26,7 @@ option java_generic_services = true; option java_generate_equals_and_hash = true; option optimize_for = SPEED; +import "Table.proto"; import "HBase.proto"; import "ClusterStatus.proto"; diff --git a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.jamon b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.jamon index 58032976c77..0ecc131b709 100644 --- a/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.jamon +++ b/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.jamon @@ -49,7 +49,7 @@ org.apache.hadoop.hbase.HTableDescriptor; org.apache.hadoop.hbase.HBaseConfiguration; org.apache.hadoop.hbase.TableName; org.apache.hadoop.hbase.tool.Canary; -org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; org.apache.hadoop.hbase.master.DeadServer; org.apache.hadoop.hbase.protobuf.ProtobufUtil; org.apache.hadoop.hbase.security.visibility.VisibilityConstants; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java index b2f76d2dd5d..2d8fdbacc5c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterAndRegionObserver.java @@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.master.RegionPlan; import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import java.io.IOException; import java.util.List; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java index f747599c61c..5fe80c0deb4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java @@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.master.RegionPlan; import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import java.io.IOException; import java.util.List; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java index d6afec055eb..3fb0858c494 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java @@ -45,6 +45,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper; import org.apache.hadoop.hbase.io.Reference; import org.apache.hadoop.hbase.io.hfile.CacheConfig; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.KeyValueScanner; import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; @@ -56,7 +57,6 @@ import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.regionserver.StoreFile.Reader; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; import org.apache.hadoop.hbase.wal.WALKey; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java index 75581479b9d..24c62b2dfea 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java @@ -37,8 +37,8 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.master.RegionPlan; import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; /** * Defines coprocessor hooks for interacting with operations on the diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java index ad562314a75..ad16b977264 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java @@ -45,6 +45,7 @@ import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper; import org.apache.hadoop.hbase.io.Reference; import org.apache.hadoop.hbase.io.hfile.CacheConfig; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.Region.Operation; import org.apache.hadoop.hbase.regionserver.InternalScanner; @@ -55,7 +56,6 @@ import org.apache.hadoop.hbase.regionserver.ScanType; import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; import org.apache.hadoop.hbase.wal.WALKey; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java index 9f52850d959..a8d387ac8c0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java @@ -36,8 +36,8 @@ import org.apache.hadoop.hbase.client.IsolationLevel; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos.TableSnapshotRegionSplit; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java index c7dd2823237..226ba069c95 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java @@ -45,8 +45,8 @@ import org.apache.hadoop.hbase.coprocessor.ObserverContext; import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; import org.apache.hadoop.hbase.metrics.MetricRegistry; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; /** * Provides the coprocessor framework and environment for master oriented diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java index c216995f253..32f62eea90b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java @@ -60,6 +60,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionStor import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AbortProcedureResponse; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest; @@ -185,7 +186,6 @@ import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.Repor import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRSFatalErrorResponse; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest; import org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionResponse; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.RSRpcServices; import org.apache.hadoop.hbase.security.AccessDeniedException; import org.apache.hadoop.hbase.security.User; @@ -903,7 +903,7 @@ public class MasterRpcServices extends RSRpcServices List tableNameList = null; if (req.getTableNamesCount() > 0) { tableNameList = new ArrayList(req.getTableNamesCount()); - for (HBaseProtos.TableName tableNamePB: req.getTableNamesList()) { + for (TableProtos.TableName tableNamePB: req.getTableNamesList()) { tableNameList.add(ProtobufUtil.toTableName(tableNamePB)); } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java index 0f1f495b09d..2f769f3c3c2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotSentinel.java @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.master; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.errorhandling.ForeignException; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; /** * Watch the current snapshot under process diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/CloneSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/CloneSnapshotHandler.java index ff59ea12d8b..6f8bcd4db87 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/CloneSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/CloneSnapshotHandler.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.master.SnapshotSentinel; import org.apache.hadoop.hbase.master.handler.CreateTableHandler; import org.apache.hadoop.hbase.monitoring.MonitoredTask; import org.apache.hadoop.hbase.monitoring.TaskMonitor; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java index 8e40a7d0f3d..b87c8269739 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.client.RegionReplicaUtil; import org.apache.hadoop.hbase.errorhandling.ForeignException; import org.apache.hadoop.hbase.master.MasterServices; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.SnapshotManifest; import org.apache.hadoop.hbase.util.FSUtils; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java index fa4245a3c93..81a3bb4daf4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.errorhandling.ForeignException; import org.apache.hadoop.hbase.master.MasterServices; import org.apache.hadoop.hbase.procedure.Procedure; import org.apache.hadoop.hbase.procedure.ProcedureCoordinator; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException; import org.apache.hadoop.hbase.util.Pair; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/MasterSnapshotVerifier.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/MasterSnapshotVerifier.java index 73b21989ae1..bb54fc30596 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/MasterSnapshotVerifier.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/MasterSnapshotVerifier.java @@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.RegionReplicaUtil; import org.apache.hadoop.hbase.MetaTableAccessor; import org.apache.hadoop.hbase.master.MasterServices; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.CorruptedSnapshotException; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/RestoreSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/RestoreSnapshotHandler.java index 80f151d74b6..1185073babd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/RestoreSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/RestoreSnapshotHandler.java @@ -46,7 +46,7 @@ import org.apache.hadoop.hbase.master.SnapshotSentinel; import org.apache.hadoop.hbase.master.handler.TableEventHandler; import org.apache.hadoop.hbase.monitoring.MonitoredTask; import org.apache.hadoop.hbase.monitoring.TaskMonitor; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotHelper; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java index 4e0181f7663..3efe984bdd8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java @@ -62,8 +62,8 @@ import org.apache.hadoop.hbase.procedure.ProcedureCoordinatorRpcs; import org.apache.hadoop.hbase.procedure.ZKProcedureCoordinatorRpcs; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ProcedureDescription; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type; import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos; import org.apache.hadoop.hbase.quotas.QuotaExceededException; import org.apache.hadoop.hbase.security.AccessDeniedException; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java index 7b45610c8c4..39387cbebeb 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java @@ -47,7 +47,7 @@ import org.apache.hadoop.hbase.master.TableLockManager; import org.apache.hadoop.hbase.master.TableLockManager.TableLock; import org.apache.hadoop.hbase.monitoring.MonitoredTask; import org.apache.hadoop.hbase.monitoring.TaskMonitor; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DeleteTracker.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java similarity index 98% rename from hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DeleteTracker.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java index 4e1ba4ec2e1..6ebc912092b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DeleteTracker.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java @@ -15,7 +15,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package org.apache.hadoop.hbase.regionserver.querymatcher; +package org.apache.hadoop.hbase.regionserver; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.Cell; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 1fac6833553..71c7892b2bb 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -156,7 +156,7 @@ import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall; import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad; import org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.StoreSequenceId; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.WALProtos; import org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor; import org.apache.hadoop.hbase.protobuf.generated.WALProtos.FlushDescriptor; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 96855b88a47..dbe8521ef1f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -73,9 +73,9 @@ import org.apache.hadoop.hbase.io.Reference; import org.apache.hadoop.hbase.io.hfile.CacheConfig; import org.apache.hadoop.hbase.ipc.RpcServer; import org.apache.hadoop.hbase.metrics.MetricRegistry; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.Region.Operation; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; import org.apache.hadoop.hbase.wal.WALKey; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/CompactionScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/CompactionScanQueryMatcher.java index 49e5f42deab..0d083b4f9f9 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/CompactionScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/CompactionScanQueryMatcher.java @@ -24,6 +24,7 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeepDeletedCells; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.filter.Filter; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; import org.apache.hadoop.hbase.regionserver.ScanInfo; import org.apache.hadoop.hbase.regionserver.ScanType; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DropDeletesCompactionScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DropDeletesCompactionScanQueryMatcher.java index b20abeefa6b..ad946e00407 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DropDeletesCompactionScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/DropDeletesCompactionScanQueryMatcher.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver.querymatcher; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.KeepDeletedCells; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.ScanInfo; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/LegacyScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/LegacyScanQueryMatcher.java index 207445033e9..b253a524a44 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/LegacyScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/LegacyScanQueryMatcher.java @@ -32,10 +32,11 @@ import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.filter.Filter.ReturnCode; import org.apache.hadoop.hbase.io.TimeRange; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; +import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult; import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; import org.apache.hadoop.hbase.regionserver.ScanInfo; import org.apache.hadoop.hbase.regionserver.ScanType; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MajorCompactionScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MajorCompactionScanQueryMatcher.java index 9830faeca5f..5c8db051a1a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MajorCompactionScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MajorCompactionScanQueryMatcher.java @@ -22,6 +22,7 @@ import java.io.IOException; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.ScanInfo; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MinorCompactionScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MinorCompactionScanQueryMatcher.java index 385fc188b91..4aacbaa1810 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MinorCompactionScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/MinorCompactionScanQueryMatcher.java @@ -22,6 +22,7 @@ import java.io.IOException; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.ScanInfo; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.java index 8f5059fe4f0..512f1a84bf0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.java @@ -24,6 +24,7 @@ import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.KeepDeletedCells; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; import org.apache.hadoop.hbase.regionserver.ScanInfo; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanDeleteTracker.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanDeleteTracker.java index f3552c384e6..f51404f487b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanDeleteTracker.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanDeleteTracker.java @@ -22,6 +22,7 @@ import java.util.SortedSet; import java.util.TreeSet; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.java index 76dcad15fd8..45ca5ac6388 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.java @@ -26,10 +26,11 @@ import org.apache.hadoop.hbase.KeyValue.Type; import org.apache.hadoop.hbase.KeyValueUtil; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.filter.Filter; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; +import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult; import org.apache.hadoop.hbase.regionserver.HStore; import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; import org.apache.hadoop.hbase.regionserver.ScanInfo; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult; /** * A query matcher that is specifically designed for the scan case. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.java index 9ae3975a237..b78ee05a6db 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/querymatcher/StripeCompactionScanQueryMatcher.java @@ -22,6 +22,7 @@ import java.io.IOException; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.ScanInfo; /** diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/FlushSnapshotSubprocedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/FlushSnapshotSubprocedure.java index b1179b541b6..6b33e751448 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/FlushSnapshotSubprocedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/FlushSnapshotSubprocedure.java @@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.errorhandling.ForeignException; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; import org.apache.hadoop.hbase.procedure.ProcedureMember; import org.apache.hadoop.hbase.procedure.Subprocedure; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.Region.FlushResult; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java index 99a0261ec68..0f9ac195b92 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java @@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.procedure.RegionServerProcedureManager; import org.apache.hadoop.hbase.procedure.Subprocedure; import org.apache.hadoop.hbase.procedure.SubprocedureFactory; import org.apache.hadoop.hbase.procedure.ZKProcedureMemberRpcs; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.RegionServerServices; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java index 722d9ebc0d5..dd15c3cc1fd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java @@ -91,10 +91,10 @@ import org.apache.hadoop.hbase.protobuf.ResponseConverter; import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos; import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.CleanupBulkLoadRequest; import org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.PrepareBulkLoadRequest; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; import org.apache.hadoop.hbase.regionserver.Region; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java index ac9dc79795b..ed4137108d2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java @@ -90,13 +90,13 @@ import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.Visibil import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsResponse; import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsService; import org.apache.hadoop.hbase.regionserver.BloomType; +import org.apache.hadoop.hbase.regionserver.DeleteTracker; import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; import org.apache.hadoop.hbase.regionserver.OperationStatus; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.RegionScanner; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker; import org.apache.hadoop.hbase.replication.ReplicationEndpoint; import org.apache.hadoop.hbase.security.AccessDeniedException; import org.apache.hadoop.hbase.security.Superusers; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java index e9ca7dc7de5..c2da8e00216 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java @@ -23,7 +23,7 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.util.AbstractHBaseTool; import java.util.Arrays; import java.util.Locale; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java index 6ec0b9020a3..97d06030d66 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java @@ -53,7 +53,7 @@ import org.apache.hadoop.hbase.io.FileLink; import org.apache.hadoop.hbase.io.HFileLink; import org.apache.hadoop.hbase.io.WALLink; import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.util.FSUtils; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java index 39887c6bd7e..47b3c3474e8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java @@ -54,7 +54,7 @@ import org.apache.hadoop.hbase.io.Reference; import org.apache.hadoop.hbase.monitoring.MonitoredTask; import org.apache.hadoop.hbase.monitoring.TaskMonitor; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.java index b735d6491ed..cdc1b1fde7f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.java @@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.security.User; import org.apache.hadoop.hbase.security.access.AccessControlLists; import org.apache.hadoop.hbase.security.access.TablePermission; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java index bd3d0c2a06f..e4c4c5964c3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java @@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java index 265ff5bed29..7822fa436f0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java @@ -43,8 +43,8 @@ import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.regionserver.HRegion; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java index 89d0ba22130..a5afb919692 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java @@ -36,7 +36,7 @@ import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HRegionInfo; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; import org.apache.hadoop.hbase.regionserver.StoreFileInfo; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java index 82ec2826a04..3b918a0a81d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java @@ -40,7 +40,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.PathFilter; import org.apache.hadoop.hbase.HRegionInfo; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.regionserver.StoreFileInfo; import org.apache.hadoop.hbase.util.ByteStringer; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java index 0f09f066e4e..56699d61365 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java @@ -39,7 +39,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.io.HFileLink; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.regionserver.StoreFileInfo; diff --git a/hbase-server/src/main/resources/hbase-webapps/master/snapshot.jsp b/hbase-server/src/main/resources/hbase-webapps/master/snapshot.jsp index f5a0d9e66b2..6099a2181bb 100644 --- a/hbase-server/src/main/resources/hbase-webapps/master/snapshot.jsp +++ b/hbase-server/src/main/resources/hbase-webapps/master/snapshot.jsp @@ -24,7 +24,7 @@ import="org.apache.hadoop.hbase.client.HConnectionManager" import="org.apache.hadoop.hbase.master.HMaster" import="org.apache.hadoop.hbase.snapshot.SnapshotInfo" - import="org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription" + import="org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription" import="org.apache.hadoop.util.StringUtils" import="org.apache.hadoop.hbase.TableName" import="org.apache.hadoop.hbase.HBaseConfiguration" %> diff --git a/hbase-server/src/main/resources/hbase-webapps/master/snapshotsStats.jsp b/hbase-server/src/main/resources/hbase-webapps/master/snapshotsStats.jsp index 289a72cb201..dceed8ee1d8 100644 --- a/hbase-server/src/main/resources/hbase-webapps/master/snapshotsStats.jsp +++ b/hbase-server/src/main/resources/hbase-webapps/master/snapshotsStats.jsp @@ -27,7 +27,7 @@ import="org.apache.hadoop.hbase.HBaseConfiguration" import="org.apache.hadoop.hbase.client.Admin" import="org.apache.hadoop.hbase.master.HMaster" - import="org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription" + import="org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription" import="org.apache.hadoop.hbase.snapshot.SnapshotInfo" import="org.apache.hadoop.hbase.TableName" import="org.apache.hadoop.util.StringUtils" %> diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java index 85c3ddf8e3f..541cb8935bc 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java @@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java index 0851267c4cd..54745f53c0e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java @@ -59,10 +59,10 @@ import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.RequestConverter; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableDescriptorsRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableNamesRequest; import org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java index 4514ae2b5e3..7b5c4156611 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java @@ -34,7 +34,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java index e29bef857d9..7d1595e55e0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionInfo.java @@ -34,8 +34,8 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.exceptions.DeserializationException; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionInfo; +import org.apache.hadoop.hbase.protobuf.generated.TableProtos; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.FSTableDescriptors; import org.apache.hadoop.hbase.util.MD5Hash; @@ -240,7 +240,7 @@ public class TestHRegionInfo { // test convert RegionInfo without replicaId RegionInfo info = RegionInfo.newBuilder() - .setTableName(HBaseProtos.TableName.newBuilder() + .setTableName(TableProtos.TableName.newBuilder() .setQualifier(ByteString.copyFrom(tableName.getQualifier())) .setNamespace(ByteString.copyFrom(tableName.getNamespace())) .build()) diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanDeleteTracker.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanDeleteTracker.java index fce35bdec67..8c93520462a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanDeleteTracker.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanDeleteTracker.java @@ -22,7 +22,7 @@ package org.apache.hadoop.hbase.regionserver.querymatcher; import static org.junit.Assert.*; import org.apache.hadoop.hbase.KeyValue; -import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult; +import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java index c5c0470b83d..eff0e687fe8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java @@ -52,7 +52,7 @@ import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.ProcedureInfo; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.security.Superusers; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.MiniHBaseCluster; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java index 23e5f3faf00..e64c6c8c37e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java @@ -57,7 +57,7 @@ import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter; import org.apache.hadoop.hbase.master.MasterCoprocessorHost; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java index 06ec4a0f11f..9847bc428e2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java @@ -59,7 +59,7 @@ import org.apache.hadoop.hbase.io.HFileLink; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.MasterFileSystem; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java index c2139d963bd..a3f5382c5b2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java @@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotFileInfo; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils.SnapshotMock; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java index 9289a020d8f..667c0156370 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java @@ -43,7 +43,7 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java index 8ebf26c6772..403b1e6abbf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java @@ -27,6 +27,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.HRegion; @@ -108,10 +109,9 @@ public class TestRegionSnapshotTask { List hRegions = TEST_UTIL.getHBaseCluster().getRegions(tableName); - final SnapshotProtos.SnapshotDescription snapshot = - SnapshotProtos.SnapshotDescription.newBuilder() + final HBaseProtos.SnapshotDescription snapshot = HBaseProtos.SnapshotDescription.newBuilder() .setTable(tableName.getNameAsString()) - .setType(SnapshotProtos.SnapshotDescription.Type.FLUSH) + .setType(HBaseProtos.SnapshotDescription.Type.FLUSH) .setName("test_table_snapshot") .setVersion(SnapshotManifestV2.DESCRIPTOR_VERSION) .build(); @@ -161,7 +161,7 @@ public class TestRegionSnapshotTask { SnapshotReferenceUtil.verifySnapshot(conf, fs, manifest); } - private void addRegionToSnapshot(SnapshotProtos.SnapshotDescription snapshot, + private void addRegionToSnapshot(HBaseProtos.SnapshotDescription snapshot, HRegion region, SnapshotManifest manifest) throws Exception { LOG.info("Adding region to snapshot: " + region.getRegionInfo().getRegionNameAsString()); Path workingDir = SnapshotDescriptionUtils.getWorkingSnapshotDir(snapshot, rootDir); @@ -171,7 +171,7 @@ public class TestRegionSnapshotTask { } private SnapshotManifest.RegionVisitor createRegionVisitorWithDelay( - SnapshotProtos.SnapshotDescription desc, Path workingDir) { + HBaseProtos.SnapshotDescription desc, Path workingDir) { return new SnapshotManifestV2.ManifestBuilder(conf, fs, workingDir) { @Override public void storeFile(final SnapshotProtos.SnapshotRegionManifest.Builder region, diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java index 442ac0dd7fe..c9c4202ca39 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java @@ -23,7 +23,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java index 8746ea3e39f..2a5bd766d8f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java @@ -31,7 +31,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; import org.apache.hadoop.hbase.io.HFileLink; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotClientRetries.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotClientRetries.java index 01452ebaab1..8ebeb97ed5e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotClientRetries.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotClientRetries.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.coprocessor.BaseMasterObserver; import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.coprocessor.MasterCoprocessorEnvironment; import org.apache.hadoop.hbase.coprocessor.ObserverContext; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.TestTableName; import org.junit.After; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotDescriptionUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotDescriptionUtils.java index 68377de876b..6583b2ce96f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotDescriptionUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotDescriptionUtils.java @@ -29,7 +29,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper; import org.junit.After; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java index 461ea98b9c8..076ce40db89 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestSnapshotManifest.java @@ -27,8 +27,8 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.HBaseTestingUtility; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDataManifest; -import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.SmallTests;