HBASE-26096 Cleanup the deprecated methods in HBTU related classes and format code (#3503)
Signed-off-by: Xiaolin Ha <haxiaolin@apache.org> Signed-off-by: Yulin Niu <niuyulin@apache.org>
This commit is contained in:
parent
332d9d13b1
commit
5f0950558f
|
@ -49,26 +49,19 @@ public class HBaseCommonTestingUtil {
|
|||
* Compression algorithms to use in parameterized JUnit 4 tests
|
||||
*/
|
||||
public static final List<Object[]> COMPRESSION_ALGORITHMS_PARAMETERIZED =
|
||||
Arrays.asList(new Object[][] {
|
||||
{ Compression.Algorithm.NONE },
|
||||
{ Compression.Algorithm.GZ }
|
||||
});
|
||||
Arrays.asList(new Object[][] { { Compression.Algorithm.NONE }, { Compression.Algorithm.GZ } });
|
||||
|
||||
/**
|
||||
* This is for unit tests parameterized with a two booleans.
|
||||
*/
|
||||
public static final List<Object[]> BOOLEAN_PARAMETERIZED =
|
||||
Arrays.asList(new Object[][] {
|
||||
{false},
|
||||
{true}
|
||||
});
|
||||
Arrays.asList(new Object[][] { { false }, { true } });
|
||||
|
||||
/**
|
||||
* Compression algorithms to use in testing
|
||||
*/
|
||||
public static final Compression.Algorithm[] COMPRESSION_ALGORITHMS = {
|
||||
Compression.Algorithm.NONE, Compression.Algorithm.GZ
|
||||
};
|
||||
public static final Compression.Algorithm[] COMPRESSION_ALGORITHMS =
|
||||
{ Compression.Algorithm.NONE, Compression.Algorithm.GZ };
|
||||
|
||||
protected final Configuration conf;
|
||||
|
||||
|
@ -82,7 +75,6 @@ public class HBaseCommonTestingUtil {
|
|||
|
||||
/**
|
||||
* Returns this classes's instance of {@link Configuration}.
|
||||
*
|
||||
* @return Instance of Configuration.
|
||||
*/
|
||||
public Configuration getConfiguration() {
|
||||
|
@ -92,8 +84,7 @@ public class HBaseCommonTestingUtil {
|
|||
/**
|
||||
* System property key to get base test directory value
|
||||
*/
|
||||
public static final String BASE_TEST_DIRECTORY_KEY =
|
||||
"test.build.data.basedirectory";
|
||||
public static final String BASE_TEST_DIRECTORY_KEY = "test.build.data.basedirectory";
|
||||
|
||||
/**
|
||||
* Default base directory for test output.
|
||||
|
@ -127,13 +118,11 @@ public class HBaseCommonTestingUtil {
|
|||
|
||||
/**
|
||||
* Sets up a directory for a test to use.
|
||||
*
|
||||
* @return New directory path, if created.
|
||||
*/
|
||||
protected Path setupDataTestDir() {
|
||||
if (this.dataTestDir != null) {
|
||||
LOG.warn("Data test dir already setup in " +
|
||||
dataTestDir.getAbsolutePath());
|
||||
LOG.warn("Data test dir already setup in " + dataTestDir.getAbsolutePath());
|
||||
return null;
|
||||
}
|
||||
Path testPath = getRandomDir();
|
||||
|
@ -151,7 +140,7 @@ public class HBaseCommonTestingUtil {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns A dir with a random (uuid) name under the test dir
|
||||
* Returns a dir with a random (uuid) name under the test dir
|
||||
* @see #getBaseTestDir()
|
||||
*/
|
||||
public Path getRandomDir() {
|
||||
|
@ -159,8 +148,7 @@ public class HBaseCommonTestingUtil {
|
|||
}
|
||||
|
||||
public static UUID getRandomUUID() {
|
||||
return new UUID(ThreadLocalRandom.current().nextLong(),
|
||||
ThreadLocalRandom.current().nextLong());
|
||||
return new UUID(ThreadLocalRandom.current().nextLong(), ThreadLocalRandom.current().nextLong());
|
||||
}
|
||||
|
||||
protected void createSubDir(String propertyName, Path parent, String subDirName) {
|
||||
|
@ -212,8 +200,7 @@ public class HBaseCommonTestingUtil {
|
|||
* @see #setupDataTestDir()
|
||||
*/
|
||||
private Path getBaseTestDir() {
|
||||
String PathName = System.getProperty(
|
||||
BASE_TEST_DIRECTORY_KEY, DEFAULT_BASE_TEST_DIRECTORY);
|
||||
String PathName = System.getProperty(BASE_TEST_DIRECTORY_KEY, DEFAULT_BASE_TEST_DIRECTORY);
|
||||
|
||||
return new Path(PathName);
|
||||
}
|
||||
|
@ -248,8 +235,7 @@ public class HBaseCommonTestingUtil {
|
|||
/**
|
||||
* Wrapper method for {@link Waiter#waitFor(Configuration, long, Predicate)}.
|
||||
*/
|
||||
public <E extends Exception> long waitFor(long timeout, Predicate<E> predicate)
|
||||
throws E {
|
||||
public <E extends Exception> long waitFor(long timeout, Predicate<E> predicate) throws E {
|
||||
return Waiter.waitFor(this.conf, timeout, predicate);
|
||||
}
|
||||
|
||||
|
@ -257,15 +243,15 @@ public class HBaseCommonTestingUtil {
|
|||
* Wrapper method for {@link Waiter#waitFor(Configuration, long, long, Predicate)}.
|
||||
*/
|
||||
public <E extends Exception> long waitFor(long timeout, long interval, Predicate<E> predicate)
|
||||
throws E {
|
||||
throws E {
|
||||
return Waiter.waitFor(this.conf, timeout, interval, predicate);
|
||||
}
|
||||
|
||||
/**
|
||||
* Wrapper method for {@link Waiter#waitFor(Configuration, long, long, boolean, Predicate)}.
|
||||
*/
|
||||
public <E extends Exception> long waitFor(long timeout, long interval,
|
||||
boolean failIfTimeout, Predicate<E> predicate) throws E {
|
||||
public <E extends Exception> long waitFor(long timeout, long interval, boolean failIfTimeout,
|
||||
Predicate<E> predicate) throws E {
|
||||
return Waiter.waitFor(this.conf, timeout, interval, failIfTimeout, predicate);
|
||||
}
|
||||
|
||||
|
@ -331,12 +317,11 @@ public class HBaseCommonTestingUtil {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns a random port. These ports cannot be registered with IANA and are
|
||||
* intended for dynamic allocation (see http://bit.ly/dynports).
|
||||
* Returns a random port. These ports cannot be registered with IANA and are intended for
|
||||
* dynamic allocation (see http://bit.ly/dynports).
|
||||
*/
|
||||
private int randomPort() {
|
||||
return MIN_RANDOM_PORT
|
||||
+ random.nextInt(MAX_RANDOM_PORT - MIN_RANDOM_PORT);
|
||||
return MIN_RANDOM_PORT + random.nextInt(MAX_RANDOM_PORT - MIN_RANDOM_PORT);
|
||||
}
|
||||
|
||||
interface AvailablePortChecker {
|
||||
|
|
|
@ -211,7 +211,7 @@ public class IntegrationTestBulkLoad extends IntegrationTestBase {
|
|||
TableDescriptor desc = admin.getDescriptor(t);
|
||||
TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(desc);
|
||||
builder.setCoprocessor(SlowMeCoproScanOperations.class.getName());
|
||||
HBaseTestingUtil.modifyTableSync(admin, builder.build());
|
||||
admin.modifyTable(builder.build());
|
||||
}
|
||||
|
||||
@Test
|
||||
|
|
|
@ -37,6 +37,7 @@ import org.apache.hadoop.hbase.client.Result;
|
|||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.client.Table;
|
||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||
import org.apache.hadoop.hbase.logging.Log4jUtils;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.io.NullWritable;
|
||||
import org.apache.hadoop.mapreduce.Job;
|
||||
|
@ -73,7 +74,7 @@ public abstract class MultiTableInputFormatTestBase {
|
|||
@BeforeClass
|
||||
public static void setUpBeforeClass() throws Exception {
|
||||
// switch TIF to log at DEBUG level
|
||||
TEST_UTIL.enableDebug(MultiTableInputFormatBase.class);
|
||||
Log4jUtils.enableDebug(MultiTableInputFormatBase.class);
|
||||
// start mini hbase cluster
|
||||
TEST_UTIL.startMiniCluster(3);
|
||||
// create and fill table
|
||||
|
|
|
@ -22,6 +22,7 @@ import java.util.List;
|
|||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||
import org.apache.hadoop.hbase.logging.Log4jUtils;
|
||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||
import org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests;
|
||||
import org.apache.hadoop.mapreduce.Job;
|
||||
|
@ -43,8 +44,8 @@ public class TestMultiTableInputFormat extends MultiTableInputFormatTestBase {
|
|||
|
||||
@BeforeClass
|
||||
public static void setupLogging() {
|
||||
TEST_UTIL.enableDebug(MultiTableInputFormat.class);
|
||||
}
|
||||
Log4jUtils.enableDebug(MultiTableInputFormat.class);
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void initJob(List<Scan> scans, Job job) throws IOException {
|
||||
|
|
|
@ -27,6 +27,7 @@ import org.apache.hadoop.hbase.HBaseClassTestRule;
|
|||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||
import org.apache.hadoop.hbase.logging.Log4jUtils;
|
||||
import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils;
|
||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||
import org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests;
|
||||
|
@ -53,8 +54,8 @@ public class TestMultiTableSnapshotInputFormat extends MultiTableInputFormatTest
|
|||
|
||||
@BeforeClass
|
||||
public static void setUpSnapshots() throws Exception {
|
||||
TEST_UTIL.enableDebug(MultiTableSnapshotInputFormat.class);
|
||||
TEST_UTIL.enableDebug(MultiTableSnapshotInputFormatImpl.class);
|
||||
Log4jUtils.enableDebug(MultiTableSnapshotInputFormat.class);
|
||||
Log4jUtils.enableDebug(MultiTableSnapshotInputFormatImpl.class);
|
||||
|
||||
// take a snapshot of every table we have.
|
||||
for (String tableName : TABLES) {
|
||||
|
|
|
@ -28,28 +28,27 @@ import org.apache.yetus.audience.InterfaceAudience;
|
|||
import org.apache.yetus.audience.InterfaceStability;
|
||||
|
||||
/**
|
||||
* This class defines methods that can help with managing HBase clusters
|
||||
* from unit tests and system tests. There are 3 types of cluster deployments:
|
||||
* This class defines methods that can help with managing HBase clusters from unit tests and system
|
||||
* tests. There are 3 types of cluster deployments:
|
||||
* <ul>
|
||||
* <li><b>SingleProcessHBaseCluster:</b> each server is run in the same JVM in separate threads,
|
||||
* used by unit tests</li>
|
||||
* <li><b>DistributedHBaseCluster:</b> the cluster is pre-deployed, system and integration tests can
|
||||
* interact with the cluster. </li>
|
||||
* <li><b>ProcessBasedLocalHBaseCluster:</b> each server is deployed locally but in separate
|
||||
* JVMs. </li>
|
||||
* interact with the cluster.</li>
|
||||
* <li><b>ProcessBasedLocalHBaseCluster:</b> each server is deployed locally but in separate JVMs.
|
||||
* </li>
|
||||
* </ul>
|
||||
* <p>
|
||||
* HBaseCluster unifies the way tests interact with the cluster, so that the same test can
|
||||
* be run against a mini-cluster during unit test execution, or a distributed cluster having
|
||||
* tens/hundreds of nodes during execution of integration tests.
|
||||
*
|
||||
* HBaseCluster unifies the way tests interact with the cluster, so that the same test can be run
|
||||
* against a mini-cluster during unit test execution, or a distributed cluster having tens/hundreds
|
||||
* of nodes during execution of integration tests.
|
||||
* <p>
|
||||
* HBaseCluster exposes client-side public interfaces to tests, so that tests does not assume
|
||||
* running in a particular mode. Not all the tests are suitable to be run on an actual cluster,
|
||||
* and some tests will still need to mock stuff and introspect internal state. For those use
|
||||
* cases from unit tests, or if more control is needed, you can use the subclasses directly.
|
||||
* In that sense, this class does not abstract away <strong>every</strong> interface that
|
||||
* SingleProcessHBaseCluster or DistributedHBaseCluster provide.
|
||||
* running in a particular mode. Not all the tests are suitable to be run on an actual cluster, and
|
||||
* some tests will still need to mock stuff and introspect internal state. For those use cases from
|
||||
* unit tests, or if more control is needed, you can use the subclasses directly. In that sense,
|
||||
* this class does not abstract away <strong>every</strong> interface that SingleProcessHBaseCluster
|
||||
* or DistributedHBaseCluster provide.
|
||||
*/
|
||||
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.PHOENIX)
|
||||
@InterfaceStability.Evolving
|
||||
|
@ -85,24 +84,23 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
public abstract ClusterMetrics getClusterMetrics() throws IOException;
|
||||
|
||||
/**
|
||||
* Returns a ClusterStatus for this HBase cluster as observed at the
|
||||
* starting of the HBaseCluster
|
||||
* Returns a ClusterStatus for this HBase cluster as observed at the starting of the HBaseCluster
|
||||
*/
|
||||
public ClusterMetrics getInitialClusterMetrics() throws IOException {
|
||||
return initialClusterStatus;
|
||||
}
|
||||
|
||||
/**
|
||||
* Starts a new region server on the given hostname or if this is a mini/local cluster,
|
||||
* starts a region server locally.
|
||||
* Starts a new region server on the given hostname or if this is a mini/local cluster, starts a
|
||||
* region server locally.
|
||||
* @param hostname the hostname to start the regionserver on
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void startRegionServer(String hostname, int port) throws IOException;
|
||||
|
||||
/**
|
||||
* Kills the region server process if this is a distributed cluster, otherwise
|
||||
* this causes the region server to exit doing basic clean up only.
|
||||
* Kills the region server process if this is a distributed cluster, otherwise this causes the
|
||||
* region server to exit doing basic clean up only.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void killRegionServer(ServerName serverName) throws IOException;
|
||||
|
@ -110,9 +108,9 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
/**
|
||||
* Keeping track of killed servers and being able to check if a particular server was killed makes
|
||||
* it possible to do fault tolerance testing for dead servers in a deterministic way. A concrete
|
||||
* example of such case is - killing servers and waiting for all regions of a particular table
|
||||
* to be assigned. We can check for server column in META table and that its value is not one
|
||||
* of the killed servers.
|
||||
* example of such case is - killing servers and waiting for all regions of a particular table to
|
||||
* be assigned. We can check for server column in META table and that its value is not one of the
|
||||
* killed servers.
|
||||
*/
|
||||
public abstract boolean isKilledRS(ServerName serverName);
|
||||
|
||||
|
@ -127,7 +125,7 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public void waitForRegionServerToStart(String hostname, int port, long timeout)
|
||||
throws IOException {
|
||||
throws IOException {
|
||||
long start = EnvironmentEdgeManager.currentTime();
|
||||
while ((EnvironmentEdgeManager.currentTime() - start) < timeout) {
|
||||
for (ServerName server : getClusterMetrics().getLiveServerMetrics().keySet()) {
|
||||
|
@ -137,8 +135,8 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
}
|
||||
Threads.sleep(100);
|
||||
}
|
||||
throw new IOException("did timeout " + timeout + "ms waiting for region server to start: "
|
||||
+ hostname);
|
||||
throw new IOException(
|
||||
"did timeout " + timeout + "ms waiting for region server to start: " + hostname);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -146,7 +144,7 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForRegionServerToStop(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
throws IOException;
|
||||
|
||||
/**
|
||||
* Suspend the region server
|
||||
|
@ -163,23 +161,23 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
public abstract void resumeRegionServer(ServerName serverName) throws IOException;
|
||||
|
||||
/**
|
||||
* Starts a new zookeeper node on the given hostname or if this is a mini/local cluster,
|
||||
* silently logs warning message.
|
||||
* Starts a new zookeeper node on the given hostname or if this is a mini/local cluster, silently
|
||||
* logs warning message.
|
||||
* @param hostname the hostname to start the regionserver on
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void startZkNode(String hostname, int port) throws IOException;
|
||||
|
||||
/**
|
||||
* Kills the zookeeper node process if this is a distributed cluster, otherwise,
|
||||
* this causes master to exit doing basic clean up only.
|
||||
* Kills the zookeeper node process if this is a distributed cluster, otherwise, this causes
|
||||
* master to exit doing basic clean up only.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void killZkNode(ServerName serverName) throws IOException;
|
||||
|
||||
/**
|
||||
* Stops the region zookeeper if this is a distributed cluster, otherwise
|
||||
* silently logs warning message.
|
||||
* Stops the region zookeeper if this is a distributed cluster, otherwise silently logs warning
|
||||
* message.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void stopZkNode(ServerName serverName) throws IOException;
|
||||
|
@ -188,33 +186,30 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* Wait for the specified zookeeper node to join the cluster
|
||||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForZkNodeToStart(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
public abstract void waitForZkNodeToStart(ServerName serverName, long timeout) throws IOException;
|
||||
|
||||
/**
|
||||
* Wait for the specified zookeeper node to stop the thread / process.
|
||||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForZkNodeToStop(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
public abstract void waitForZkNodeToStop(ServerName serverName, long timeout) throws IOException;
|
||||
|
||||
/**
|
||||
* Starts a new datanode on the given hostname or if this is a mini/local cluster,
|
||||
* silently logs warning message.
|
||||
* Starts a new datanode on the given hostname or if this is a mini/local cluster, silently logs
|
||||
* warning message.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void startDataNode(ServerName serverName) throws IOException;
|
||||
|
||||
/**
|
||||
* Kills the datanode process if this is a distributed cluster, otherwise,
|
||||
* this causes master to exit doing basic clean up only.
|
||||
* Kills the datanode process if this is a distributed cluster, otherwise, this causes master to
|
||||
* exit doing basic clean up only.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void killDataNode(ServerName serverName) throws IOException;
|
||||
|
||||
/**
|
||||
* Stops the datanode if this is a distributed cluster, otherwise
|
||||
* silently logs warning message.
|
||||
* Stops the datanode if this is a distributed cluster, otherwise silently logs warning message.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void stopDataNode(ServerName serverName) throws IOException;
|
||||
|
@ -258,26 +253,26 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForNameNodeToStart(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
throws IOException;
|
||||
|
||||
/**
|
||||
* Wait for the specified namenode to stop
|
||||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForNameNodeToStop(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
throws IOException;
|
||||
|
||||
/**
|
||||
* Starts a new master on the given hostname or if this is a mini/local cluster,
|
||||
* starts a master locally.
|
||||
* Starts a new master on the given hostname or if this is a mini/local cluster, starts a master
|
||||
* locally.
|
||||
* @param hostname the hostname to start the master on
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void startMaster(String hostname, int port) throws IOException;
|
||||
|
||||
/**
|
||||
* Kills the master process if this is a distributed cluster, otherwise,
|
||||
* this causes master to exit doing basic clean up only.
|
||||
* Kills the master process if this is a distributed cluster, otherwise, this causes master to
|
||||
* exit doing basic clean up only.
|
||||
* @throws IOException if something goes wrong
|
||||
*/
|
||||
public abstract void killMaster(ServerName serverName) throws IOException;
|
||||
|
@ -292,31 +287,23 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* Wait for the specified master to stop the thread / process.
|
||||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public abstract void waitForMasterToStop(ServerName serverName, long timeout)
|
||||
throws IOException;
|
||||
public abstract void waitForMasterToStop(ServerName serverName, long timeout) throws IOException;
|
||||
|
||||
/**
|
||||
* Blocks until there is an active master and that master has completed
|
||||
* initialization.
|
||||
*
|
||||
* @return true if an active master becomes available. false if there are no
|
||||
* masters left.
|
||||
* Blocks until there is an active master and that master has completed initialization.
|
||||
* @return true if an active master becomes available. false if there are no masters left.
|
||||
* @throws IOException if something goes wrong or timeout occurs
|
||||
*/
|
||||
public boolean waitForActiveAndReadyMaster()
|
||||
throws IOException {
|
||||
public boolean waitForActiveAndReadyMaster() throws IOException {
|
||||
return waitForActiveAndReadyMaster(Long.MAX_VALUE);
|
||||
}
|
||||
|
||||
/**
|
||||
* Blocks until there is an active master and that master has completed
|
||||
* initialization.
|
||||
* Blocks until there is an active master and that master has completed initialization.
|
||||
* @param timeout the timeout limit in ms
|
||||
* @return true if an active master becomes available. false if there are no
|
||||
* masters left.
|
||||
* @return true if an active master becomes available. false if there are no masters left.
|
||||
*/
|
||||
public abstract boolean waitForActiveAndReadyMaster(long timeout)
|
||||
throws IOException;
|
||||
public abstract boolean waitForActiveAndReadyMaster(long timeout) throws IOException;
|
||||
|
||||
/**
|
||||
* Wait for HBase Cluster to shut down.
|
||||
|
@ -329,10 +316,9 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
public abstract void shutdown() throws IOException;
|
||||
|
||||
/**
|
||||
* Restores the cluster to it's initial state if this is a real cluster,
|
||||
* otherwise does nothing.
|
||||
* This is a best effort restore. If the servers are not reachable, or insufficient
|
||||
* permissions, etc. restoration might be partial.
|
||||
* Restores the cluster to it's initial state if this is a real cluster, otherwise does nothing.
|
||||
* This is a best effort restore. If the servers are not reachable, or insufficient permissions,
|
||||
* etc. restoration might be partial.
|
||||
* @return whether restoration is complete
|
||||
*/
|
||||
public boolean restoreInitialStatus() throws IOException {
|
||||
|
@ -340,10 +326,9 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
}
|
||||
|
||||
/**
|
||||
* Restores the cluster to given state if this is a real cluster,
|
||||
* otherwise does nothing.
|
||||
* This is a best effort restore. If the servers are not reachable, or insufficient
|
||||
* permissions, etc. restoration might be partial.
|
||||
* Restores the cluster to given state if this is a real cluster, otherwise does nothing. This is
|
||||
* a best effort restore. If the servers are not reachable, or insufficient permissions, etc.
|
||||
* restoration might be partial.
|
||||
* @return whether restoration is complete
|
||||
*/
|
||||
public boolean restoreClusterMetrics(ClusterMetrics desiredStatus) throws IOException {
|
||||
|
@ -365,19 +350,19 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
* @return ServerName that hosts the region or null
|
||||
*/
|
||||
public abstract ServerName getServerHoldingRegion(final TableName tn, byte[] regionName)
|
||||
throws IOException;
|
||||
throws IOException;
|
||||
|
||||
/**
|
||||
* @return whether we are interacting with a distributed cluster as opposed to an
|
||||
* in-process mini/local cluster.
|
||||
* @return whether we are interacting with a distributed cluster as opposed to an in-process
|
||||
* mini/local cluster.
|
||||
*/
|
||||
public boolean isDistributedCluster() {
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* Closes all the resources held open for this cluster. Note that this call does not shutdown
|
||||
* the cluster.
|
||||
* Closes all the resources held open for this cluster. Note that this call does not shutdown the
|
||||
* cluster.
|
||||
* @see #shutdown()
|
||||
*/
|
||||
@Override
|
||||
|
@ -385,8 +370,6 @@ public abstract class HBaseClusterInterface implements Closeable, Configurable {
|
|||
|
||||
/**
|
||||
* Wait for the namenode.
|
||||
*
|
||||
* @throws InterruptedException
|
||||
*/
|
||||
public void waitForNamenodeAvailable() throws InterruptedException {
|
||||
}
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -97,7 +97,7 @@ public final class MiniClusterRule extends ExternalResource {
|
|||
}
|
||||
|
||||
/**
|
||||
* @return the underlying instance of {@link HBaseTestingUtil}
|
||||
* Returns the underlying instance of {@link HBaseTestingUtil}
|
||||
*/
|
||||
public HBaseTestingUtil getTestingUtility() {
|
||||
return testingUtility;
|
||||
|
|
|
@ -47,10 +47,9 @@ import org.slf4j.LoggerFactory;
|
|||
import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse;
|
||||
|
||||
/**
|
||||
* This class creates a single process HBase cluster.
|
||||
* each server. The master uses the 'default' FileSystem. The RegionServers,
|
||||
* if we are running on DistributedFilesystem, create a FileSystem instance
|
||||
* each and will close down their instance on the way out.
|
||||
* This class creates a single process HBase cluster. each server. The master uses the 'default'
|
||||
* FileSystem. The RegionServers, if we are running on DistributedFilesystem, create a FileSystem
|
||||
* instance each and will close down their instance on the way out.
|
||||
*/
|
||||
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.PHOENIX)
|
||||
@InterfaceStability.Evolving
|
||||
|
@ -101,16 +100,16 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* each cluster start.
|
||||
*/
|
||||
public SingleProcessHBaseCluster(Configuration conf, int numMasters, int numAlwaysStandByMasters,
|
||||
int numRegionServers, List<Integer> rsPorts, Class<? extends HMaster> masterClass,
|
||||
Class<? extends SingleProcessHBaseCluster.MiniHBaseClusterRegionServer> regionserverClass)
|
||||
throws IOException, InterruptedException {
|
||||
int numRegionServers, List<Integer> rsPorts, Class<? extends HMaster> masterClass,
|
||||
Class<? extends SingleProcessHBaseCluster.MiniHBaseClusterRegionServer> regionserverClass)
|
||||
throws IOException, InterruptedException {
|
||||
super(conf);
|
||||
|
||||
// Hadoop 2
|
||||
CompatibilityFactory.getInstance(MetricsAssertHelper.class).init();
|
||||
|
||||
init(numMasters, numAlwaysStandByMasters, numRegionServers, rsPorts, masterClass,
|
||||
regionserverClass);
|
||||
regionserverClass);
|
||||
this.initialClusterStatus = getClusterMetrics();
|
||||
}
|
||||
|
||||
|
@ -119,30 +118,29 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* Subclass so can get at protected methods (none at moment). Also, creates
|
||||
* a FileSystem instance per instantiation. Adds a shutdown own FileSystem
|
||||
* on the way out. Shuts down own Filesystem only, not All filesystems as
|
||||
* the FileSystem system exit hook does.
|
||||
* Subclass so can get at protected methods (none at moment). Also, creates a FileSystem instance
|
||||
* per instantiation. Adds a shutdown own FileSystem on the way out. Shuts down own Filesystem
|
||||
* only, not All filesystems as the FileSystem system exit hook does.
|
||||
*/
|
||||
public static class MiniHBaseClusterRegionServer extends HRegionServer {
|
||||
private Thread shutdownThread = null;
|
||||
private User user = null;
|
||||
/**
|
||||
* List of RegionServers killed so far. ServerName also comprises startCode of a server,
|
||||
* so any restarted instances of the same server will have different ServerName and will not
|
||||
* coincide with past dead ones. So there's no need to cleanup this list.
|
||||
* List of RegionServers killed so far. ServerName also comprises startCode of a server, so any
|
||||
* restarted instances of the same server will have different ServerName and will not coincide
|
||||
* with past dead ones. So there's no need to cleanup this list.
|
||||
*/
|
||||
static Set<ServerName> killedServers = new HashSet<>();
|
||||
|
||||
public MiniHBaseClusterRegionServer(Configuration conf)
|
||||
throws IOException, InterruptedException {
|
||||
throws IOException, InterruptedException {
|
||||
super(conf);
|
||||
this.user = User.getCurrent();
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void handleReportForDutyResponse(
|
||||
final RegionServerStartupResponse c) throws IOException {
|
||||
protected void handleReportForDutyResponse(final RegionServerStartupResponse c)
|
||||
throws IOException {
|
||||
super.handleReportForDutyResponse(c);
|
||||
// Run this thread to shutdown our filesystem on way out.
|
||||
this.shutdownThread = new SingleFileSystemShutdownThread(getFileSystem());
|
||||
|
@ -196,15 +194,17 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* Alternate shutdown hook.
|
||||
* Just shuts down the passed fs, not all as default filesystem hook does.
|
||||
* Alternate shutdown hook. Just shuts down the passed fs, not all as default filesystem hook
|
||||
* does.
|
||||
*/
|
||||
static class SingleFileSystemShutdownThread extends Thread {
|
||||
private final FileSystem fs;
|
||||
|
||||
SingleFileSystemShutdownThread(final FileSystem fs) {
|
||||
super("Shutdown of " + fs);
|
||||
this.fs = fs;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void run() {
|
||||
try {
|
||||
|
@ -232,7 +232,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
// start up a LocalHBaseCluster
|
||||
hbaseCluster = new LocalHBaseCluster(conf, nMasterNodes, numAlwaysStandByMasters, 0,
|
||||
masterClass, regionserverClass);
|
||||
masterClass, regionserverClass);
|
||||
|
||||
// manually add the regionservers as other users
|
||||
for (int i = 0; i < nRegionNodes; i++) {
|
||||
|
@ -240,8 +240,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
if (rsPorts != null) {
|
||||
rsConf.setInt(HConstants.REGIONSERVER_PORT, rsPorts.get(i));
|
||||
}
|
||||
User user = HBaseTestingUtil.getDifferentUser(rsConf,
|
||||
".hfs."+index++);
|
||||
User user = HBaseTestingUtil.getDifferentUser(rsConf, ".hfs." + index++);
|
||||
hbaseCluster.addRegionServer(rsConf, i, user);
|
||||
}
|
||||
|
||||
|
@ -296,7 +295,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
@Override
|
||||
public void waitForRegionServerToStop(ServerName serverName, long timeout) throws IOException {
|
||||
//ignore timeout for now
|
||||
// ignore timeout for now
|
||||
waitOnRegionServer(getRegionServerIndex(serverName));
|
||||
}
|
||||
|
||||
|
@ -392,7 +391,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
@Override
|
||||
public void waitForMasterToStop(ServerName serverName, long timeout) throws IOException {
|
||||
//ignore timeout for now
|
||||
// ignore timeout for now
|
||||
waitOnMaster(getMasterIndex(serverName));
|
||||
}
|
||||
|
||||
|
@ -400,20 +399,18 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* Starts a region server thread running
|
||||
* @return New RegionServerThread
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread startRegionServer()
|
||||
throws IOException {
|
||||
public JVMClusterUtil.RegionServerThread startRegionServer() throws IOException {
|
||||
final Configuration newConf = HBaseConfiguration.create(conf);
|
||||
return startRegionServer(newConf);
|
||||
}
|
||||
|
||||
private JVMClusterUtil.RegionServerThread startRegionServer(Configuration configuration)
|
||||
throws IOException {
|
||||
User rsUser =
|
||||
HBaseTestingUtil.getDifferentUser(configuration, ".hfs."+index++);
|
||||
JVMClusterUtil.RegionServerThread t = null;
|
||||
throws IOException {
|
||||
User rsUser = HBaseTestingUtil.getDifferentUser(configuration, ".hfs." + index++);
|
||||
JVMClusterUtil.RegionServerThread t = null;
|
||||
try {
|
||||
t = hbaseCluster.addRegionServer(
|
||||
configuration, hbaseCluster.getRegionServers().size(), rsUser);
|
||||
t =
|
||||
hbaseCluster.addRegionServer(configuration, hbaseCluster.getRegionServers().size(), rsUser);
|
||||
t.start();
|
||||
t.waitForServerOnline();
|
||||
} catch (InterruptedException ie) {
|
||||
|
@ -423,16 +420,15 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* Starts a region server thread and waits until its processed by master. Throws an exception
|
||||
* when it can't start a region server or when the region server is not processed by master
|
||||
* within the timeout.
|
||||
*
|
||||
* Starts a region server thread and waits until its processed by master. Throws an exception when
|
||||
* it can't start a region server or when the region server is not processed by master within the
|
||||
* timeout.
|
||||
* @return New RegionServerThread
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread startRegionServerAndWait(long timeout)
|
||||
throws IOException {
|
||||
throws IOException {
|
||||
|
||||
JVMClusterUtil.RegionServerThread t = startRegionServer();
|
||||
JVMClusterUtil.RegionServerThread t = startRegionServer();
|
||||
ServerName rsServerName = t.getRegionServer().getServerName();
|
||||
|
||||
long start = EnvironmentEdgeManager.currentTime();
|
||||
|
@ -452,7 +448,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
/**
|
||||
* Cause a region server to exit doing basic clean up only on its way out.
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @param serverNumber Used as index into a list.
|
||||
*/
|
||||
public String abortRegionServer(int serverNumber) {
|
||||
HRegionServer server = getRegionServer(serverNumber);
|
||||
|
@ -463,8 +459,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
/**
|
||||
* Shut down the specified region server cleanly
|
||||
*
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @return the region server that was stopped
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread stopRegionServer(int serverNumber) {
|
||||
|
@ -480,9 +475,8 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* @return the region server that was stopped
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread stopRegionServer(int serverNumber,
|
||||
final boolean shutdownFS) {
|
||||
JVMClusterUtil.RegionServerThread server =
|
||||
hbaseCluster.getRegionServers().get(serverNumber);
|
||||
final boolean shutdownFS) {
|
||||
JVMClusterUtil.RegionServerThread server = hbaseCluster.getRegionServers().get(serverNumber);
|
||||
LOG.info("Stopping " + server.toString());
|
||||
server.getRegionServer().stop("Stopping rs " + serverNumber);
|
||||
return server;
|
||||
|
@ -493,8 +487,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* @param serverNumber Used as index into a list.
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread suspendRegionServer(int serverNumber) {
|
||||
JVMClusterUtil.RegionServerThread server =
|
||||
hbaseCluster.getRegionServers().get(serverNumber);
|
||||
JVMClusterUtil.RegionServerThread server = hbaseCluster.getRegionServers().get(serverNumber);
|
||||
LOG.info("Suspending {}", server.toString());
|
||||
server.suspend();
|
||||
return server;
|
||||
|
@ -505,8 +498,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* @param serverNumber Used as index into a list.
|
||||
*/
|
||||
public JVMClusterUtil.RegionServerThread resumeRegionServer(int serverNumber) {
|
||||
JVMClusterUtil.RegionServerThread server =
|
||||
hbaseCluster.getRegionServers().get(serverNumber);
|
||||
JVMClusterUtil.RegionServerThread server = hbaseCluster.getRegionServers().get(serverNumber);
|
||||
LOG.info("Resuming {}", server.toString());
|
||||
server.resume();
|
||||
return server;
|
||||
|
@ -520,16 +512,13 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
return this.hbaseCluster.waitOnRegionServer(serverNumber);
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Starts a master thread running
|
||||
*
|
||||
* @return New RegionServerThread
|
||||
*/
|
||||
public JVMClusterUtil.MasterThread startMaster() throws IOException {
|
||||
Configuration c = HBaseConfiguration.create(conf);
|
||||
User user =
|
||||
HBaseTestingUtil.getDifferentUser(c, ".hfs."+index++);
|
||||
User user = HBaseTestingUtil.getDifferentUser(c, ".hfs." + index++);
|
||||
|
||||
JVMClusterUtil.MasterThread t = null;
|
||||
try {
|
||||
|
@ -539,7 +528,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
throw new IOException("Interrupted adding master to cluster", ie);
|
||||
}
|
||||
conf.set(HConstants.MASTER_ADDRS_KEY,
|
||||
hbaseCluster.getConfiguration().get(HConstants.MASTER_ADDRS_KEY));
|
||||
hbaseCluster.getConfiguration().get(HConstants.MASTER_ADDRS_KEY));
|
||||
return t;
|
||||
}
|
||||
|
||||
|
@ -556,7 +545,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* @return the active MasterThread, null if none is active.
|
||||
*/
|
||||
public MasterThread getMasterThread() {
|
||||
for (MasterThread mt: hbaseCluster.getLiveMasters()) {
|
||||
for (MasterThread mt : hbaseCluster.getLiveMasters()) {
|
||||
if (mt.getMaster().isActiveMaster()) {
|
||||
return mt;
|
||||
}
|
||||
|
@ -574,7 +563,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
/**
|
||||
* Cause a master to exit without shutting down entire cluster.
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @param serverNumber Used as index into a list.
|
||||
*/
|
||||
public String abortMaster(int serverNumber) {
|
||||
HMaster server = getMaster(serverNumber);
|
||||
|
@ -585,8 +574,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
|
||||
/**
|
||||
* Shut down the specified master cleanly
|
||||
*
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @param serverNumber Used as index into a list.
|
||||
* @return the region server that was stopped
|
||||
*/
|
||||
public JVMClusterUtil.MasterThread stopMaster(int serverNumber) {
|
||||
|
@ -601,10 +589,8 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* test and you shut down one before end of the test.
|
||||
* @return the master that was stopped
|
||||
*/
|
||||
public JVMClusterUtil.MasterThread stopMaster(int serverNumber,
|
||||
final boolean shutdownFS) {
|
||||
JVMClusterUtil.MasterThread server =
|
||||
hbaseCluster.getMasters().get(serverNumber);
|
||||
public JVMClusterUtil.MasterThread stopMaster(int serverNumber, final boolean shutdownFS) {
|
||||
JVMClusterUtil.MasterThread server = hbaseCluster.getMasters().get(serverNumber);
|
||||
LOG.info("Stopping " + server.toString());
|
||||
server.getMaster().stop("Stopping master " + serverNumber);
|
||||
return server;
|
||||
|
@ -619,24 +605,18 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* Blocks until there is an active master and that master has completed
|
||||
* initialization.
|
||||
*
|
||||
* @return true if an active master becomes available. false if there are no
|
||||
* masters left.
|
||||
* Blocks until there is an active master and that master has completed initialization.
|
||||
* @return true if an active master becomes available. false if there are no masters left.
|
||||
*/
|
||||
@Override
|
||||
public boolean waitForActiveAndReadyMaster(long timeout) throws IOException {
|
||||
List<JVMClusterUtil.MasterThread> mts;
|
||||
long start = EnvironmentEdgeManager.currentTime();
|
||||
while (!(mts = getMasterThreads()).isEmpty()
|
||||
&& (EnvironmentEdgeManager.currentTime() - start) < timeout) {
|
||||
for (JVMClusterUtil.MasterThread mt : mts) {
|
||||
while (EnvironmentEdgeManager.currentTime() - start < timeout) {
|
||||
for (JVMClusterUtil.MasterThread mt : getMasterThreads()) {
|
||||
if (mt.getMaster().isActiveMaster() && mt.getMaster().isInitialized()) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
Threads.sleep(100);
|
||||
}
|
||||
return false;
|
||||
|
@ -722,9 +702,8 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* Call flushCache on all regions on all participating regionservers.
|
||||
*/
|
||||
public void compact(boolean major) throws IOException {
|
||||
for (JVMClusterUtil.RegionServerThread t:
|
||||
this.hbaseCluster.getRegionServers()) {
|
||||
for(HRegion r: t.getRegionServer().getOnlineRegionsLocalContext()) {
|
||||
for (JVMClusterUtil.RegionServerThread t : this.hbaseCluster.getRegionServers()) {
|
||||
for (HRegion r : t.getRegionServer().getOnlineRegionsLocalContext()) {
|
||||
r.compact(major);
|
||||
}
|
||||
}
|
||||
|
@ -734,10 +713,9 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
* Call flushCache on all regions of the specified table.
|
||||
*/
|
||||
public void compact(TableName tableName, boolean major) throws IOException {
|
||||
for (JVMClusterUtil.RegionServerThread t:
|
||||
this.hbaseCluster.getRegionServers()) {
|
||||
for(HRegion r: t.getRegionServer().getOnlineRegionsLocalContext()) {
|
||||
if(r.getTableDescriptor().getTableName().equals(tableName)) {
|
||||
for (JVMClusterUtil.RegionServerThread t : this.hbaseCluster.getRegionServers()) {
|
||||
for (HRegion r : t.getRegionServer().getOnlineRegionsLocalContext()) {
|
||||
if (r.getTableDescriptor().getTableName().equals(tableName)) {
|
||||
r.compact(major);
|
||||
}
|
||||
}
|
||||
|
@ -760,7 +738,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* @return List of live region server threads (skips the aborted and the killed)
|
||||
* Returns List of live region server threads (skips the aborted and the killed)
|
||||
*/
|
||||
public List<JVMClusterUtil.RegionServerThread> getLiveRegionServerThreads() {
|
||||
return this.hbaseCluster.getLiveRegionServers();
|
||||
|
@ -775,10 +753,8 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
public HRegionServer getRegionServer(ServerName serverName) {
|
||||
return hbaseCluster.getRegionServers().stream()
|
||||
.map(t -> t.getRegionServer())
|
||||
.filter(r -> r.getServerName().equals(serverName))
|
||||
.findFirst().orElse(null);
|
||||
return hbaseCluster.getRegionServers().stream().map(t -> t.getRegionServer())
|
||||
.filter(r -> r.getServerName().equals(serverName)).findFirst().orElse(null);
|
||||
}
|
||||
|
||||
public List<HRegion> getRegions(byte[] tableName) {
|
||||
|
@ -791,7 +767,7 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
HRegionServer hrs = rst.getRegionServer();
|
||||
for (Region region : hrs.getOnlineRegionsLocalContext()) {
|
||||
if (region.getTableDescriptor().getTableName().equals(tableName)) {
|
||||
ret.add((HRegion)region);
|
||||
ret.add((HRegion) region);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -809,12 +785,12 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
/**
|
||||
* Get the location of the specified region
|
||||
* @param regionName Name of the region in bytes
|
||||
* @return Index into List of {@link SingleProcessHBaseCluster#getRegionServerThreads()}
|
||||
* of HRS carrying hbase:meta. Returns -1 if none found.
|
||||
* @return Index into List of {@link SingleProcessHBaseCluster#getRegionServerThreads()} of HRS
|
||||
* carrying hbase:meta. Returns -1 if none found.
|
||||
*/
|
||||
public int getServerWith(byte[] regionName) {
|
||||
int index = 0;
|
||||
for (JVMClusterUtil.RegionServerThread rst: getRegionServerThreads()) {
|
||||
for (JVMClusterUtil.RegionServerThread rst : getRegionServerThreads()) {
|
||||
HRegionServer hrs = rst.getRegionServer();
|
||||
if (!hrs.isStopped()) {
|
||||
Region region = hrs.getOnlineRegion(regionName);
|
||||
|
@ -864,8 +840,8 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
}
|
||||
|
||||
/**
|
||||
* Do a simulated kill all masters and regionservers. Useful when it is
|
||||
* impossible to bring the mini-cluster back for clean shutdown.
|
||||
* Do a simulated kill all masters and regionservers. Useful when it is impossible to bring the
|
||||
* mini-cluster back for clean shutdown.
|
||||
*/
|
||||
public void killAll() {
|
||||
// Do backups first.
|
||||
|
@ -897,18 +873,17 @@ public class SingleProcessHBaseCluster extends HBaseClusterInterface {
|
|||
HRegionServer hrs = rst.getRegionServer();
|
||||
for (Region region : hrs.getRegions(tableName)) {
|
||||
if (region.getTableDescriptor().getTableName().equals(tableName)) {
|
||||
ret.add((HRegion)region);
|
||||
ret.add((HRegion) region);
|
||||
}
|
||||
}
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
protected int getRegionServerIndex(ServerName serverName) {
|
||||
//we have a small number of region servers, this should be fine for now.
|
||||
// we have a small number of region servers, this should be fine for now.
|
||||
List<RegionServerThread> servers = getRegionServerThreads();
|
||||
for (int i=0; i < servers.size(); i++) {
|
||||
for (int i = 0; i < servers.size(); i++) {
|
||||
if (servers.get(i).getRegionServer().getServerName().equals(serverName)) {
|
||||
return i;
|
||||
}
|
||||
|
|
|
@ -45,8 +45,8 @@ import org.apache.yetus.audience.InterfaceStability;
|
|||
@InterfaceStability.Evolving
|
||||
public final class StartTestingClusterOption {
|
||||
/**
|
||||
* Number of masters to start up. We'll start this many hbase masters. If numMasters > 1, you
|
||||
* can find the active/primary master with {@link SingleProcessHBaseCluster#getMaster()}.
|
||||
* Number of masters to start up. We'll start this many hbase masters. If numMasters > 1, you can
|
||||
* find the active/primary master with {@link SingleProcessHBaseCluster#getMaster()}.
|
||||
*/
|
||||
private final int numMasters;
|
||||
|
||||
|
@ -63,9 +63,8 @@ public final class StartTestingClusterOption {
|
|||
private final Class<? extends HMaster> masterClass;
|
||||
|
||||
/**
|
||||
* Number of region servers to start up.
|
||||
* If this value is > 1, then make sure config "hbase.regionserver.info.port" is -1
|
||||
* (i.e. no ui per regionserver) otherwise bind errors.
|
||||
* Number of region servers to start up. If this value is > 1, then make sure config
|
||||
* "hbase.regionserver.info.port" is -1 (i.e. no ui per regionserver) otherwise bind errors.
|
||||
*/
|
||||
private final int numRegionServers;
|
||||
/**
|
||||
|
@ -96,13 +95,13 @@ public final class StartTestingClusterOption {
|
|||
private final int numZkServers;
|
||||
|
||||
/**
|
||||
* Whether to create a new root or data directory path. If true, the newly created data directory
|
||||
* will be configured as HBase rootdir. This will overwrite existing root directory config.
|
||||
* Whether to create a new root or data directory path. If true, the newly created data directory
|
||||
* will be configured as HBase rootdir. This will overwrite existing root directory config.
|
||||
*/
|
||||
private final boolean createRootDir;
|
||||
|
||||
/**
|
||||
* Whether to create a new WAL directory. If true, the newly created directory will be configured
|
||||
* Whether to create a new WAL directory. If true, the newly created directory will be configured
|
||||
* as HBase wal.dir which is separate from HBase rootdir.
|
||||
*/
|
||||
private final boolean createWALDir;
|
||||
|
@ -174,15 +173,15 @@ public final class StartTestingClusterOption {
|
|||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "StartMiniClusterOption{" + "numMasters=" + numMasters + ", masterClass=" + masterClass
|
||||
+ ", numRegionServers=" + numRegionServers + ", rsPorts=" + StringUtils.join(rsPorts)
|
||||
+ ", rsClass=" + rsClass + ", numDataNodes=" + numDataNodes
|
||||
+ ", dataNodeHosts=" + Arrays.toString(dataNodeHosts) + ", numZkServers=" + numZkServers
|
||||
+ ", createRootDir=" + createRootDir + ", createWALDir=" + createWALDir + '}';
|
||||
return "StartMiniClusterOption{" + "numMasters=" + numMasters + ", masterClass=" + masterClass +
|
||||
", numRegionServers=" + numRegionServers + ", rsPorts=" + StringUtils.join(rsPorts) +
|
||||
", rsClass=" + rsClass + ", numDataNodes=" + numDataNodes + ", dataNodeHosts=" +
|
||||
Arrays.toString(dataNodeHosts) + ", numZkServers=" + numZkServers + ", createRootDir=" +
|
||||
createRootDir + ", createWALDir=" + createWALDir + '}';
|
||||
}
|
||||
|
||||
/**
|
||||
* @return a new builder.
|
||||
* Returns a new builder.
|
||||
*/
|
||||
public static Builder builder() {
|
||||
return new Builder();
|
||||
|
@ -190,7 +189,7 @@ public final class StartTestingClusterOption {
|
|||
|
||||
/**
|
||||
* Builder pattern for creating an {@link StartTestingClusterOption}.
|
||||
*
|
||||
* <p/>
|
||||
* The default values of its fields should be considered public and constant. Changing the default
|
||||
* values may cause other tests fail.
|
||||
*/
|
||||
|
@ -214,9 +213,9 @@ public final class StartTestingClusterOption {
|
|||
if (dataNodeHosts != null && dataNodeHosts.length != 0) {
|
||||
numDataNodes = dataNodeHosts.length;
|
||||
}
|
||||
return new StartTestingClusterOption(numMasters,numAlwaysStandByMasters, masterClass,
|
||||
numRegionServers, rsPorts, rsClass, numDataNodes, dataNodeHosts, numZkServers,
|
||||
createRootDir, createWALDir);
|
||||
return new StartTestingClusterOption(numMasters, numAlwaysStandByMasters, masterClass,
|
||||
numRegionServers, rsPorts, rsClass, numDataNodes, dataNodeHosts, numZkServers,
|
||||
createRootDir, createWALDir);
|
||||
}
|
||||
|
||||
public Builder numMasters(int numMasters) {
|
||||
|
|
|
@ -41,6 +41,7 @@ import org.apache.hadoop.hbase.ClusterMetrics.Option;
|
|||
import org.apache.hadoop.hbase.HBaseTestingUtil;
|
||||
import org.apache.hadoop.hbase.HConstants;
|
||||
import org.apache.hadoop.hbase.HRegionLocation;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.ServerName;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
|
@ -108,9 +109,10 @@ public class BaseTestHBaseFsck {
|
|||
* Debugging method to dump the contents of meta.
|
||||
*/
|
||||
protected void dumpMeta(TableName tableName) throws IOException {
|
||||
List<byte[]> metaRows = TEST_UTIL.getMetaTableRows(tableName);
|
||||
for (byte[] row : metaRows) {
|
||||
LOG.info(Bytes.toString(row));
|
||||
List<RegionInfo> regions =
|
||||
MetaTableAccessor.getTableRegions(TEST_UTIL.getConnection(), tableName);
|
||||
for (RegionInfo region : regions) {
|
||||
LOG.info(region.getRegionNameAsString());
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -210,7 +212,6 @@ public class BaseTestHBaseFsck {
|
|||
LOG.info(hri.toString() + hsa.toString());
|
||||
}
|
||||
|
||||
TEST_UTIL.getMetaTableRows(htd.getTableName());
|
||||
LOG.info("*** After delete:");
|
||||
dumpMeta(htd.getTableName());
|
||||
}
|
||||
|
|
|
@ -54,8 +54,8 @@ public class HBaseZKTestingUtil extends HBaseCommonTestingUtil {
|
|||
}
|
||||
|
||||
/**
|
||||
* @return Where the cluster will write data on the local subsystem. Creates it if it does not
|
||||
* exist already. A subdir of {@code HBaseCommonTestingUtility#getBaseTestDir()}
|
||||
* Returns Where the cluster will write data on the local subsystem. Creates it if it does not
|
||||
* exist already. A subdir of {@code HBaseCommonTestingUtility#getBaseTestDir()}
|
||||
*/
|
||||
Path getClusterTestDir() {
|
||||
if (clusterTestDir == null) {
|
||||
|
@ -99,7 +99,7 @@ public class HBaseZKTestingUtil extends HBaseCommonTestingUtil {
|
|||
* @return zk cluster started.
|
||||
*/
|
||||
public MiniZooKeeperCluster startMiniZKCluster(int zooKeeperServerNum, int... clientPortList)
|
||||
throws Exception {
|
||||
throws Exception {
|
||||
setupClusterTestDir();
|
||||
return startMiniZKCluster(clusterTestDir, zooKeeperServerNum, clientPortList);
|
||||
}
|
||||
|
@ -109,7 +109,7 @@ public class HBaseZKTestingUtil extends HBaseCommonTestingUtil {
|
|||
* port mentioned is used as the default port for ZooKeeper.
|
||||
*/
|
||||
private MiniZooKeeperCluster startMiniZKCluster(File dir, int zooKeeperServerNum,
|
||||
int[] clientPortList) throws Exception {
|
||||
int[] clientPortList) throws Exception {
|
||||
if (this.zkCluster != null) {
|
||||
throw new IOException("Cluster already running at " + dir);
|
||||
}
|
||||
|
@ -159,7 +159,7 @@ public class HBaseZKTestingUtil extends HBaseCommonTestingUtil {
|
|||
* users. Don't close it, it will be closed automatically when the cluster shutdowns
|
||||
* @return The ZKWatcher instance.
|
||||
*/
|
||||
public synchronized ZKWatcher getZooKeeperWatcher() throws IOException {
|
||||
public ZKWatcher getZooKeeperWatcher() throws IOException {
|
||||
if (zooKeeperWatcher == null) {
|
||||
zooKeeperWatcher = new ZKWatcher(conf, "testing utility", new Abortable() {
|
||||
@Override
|
||||
|
@ -177,27 +177,7 @@ public class HBaseZKTestingUtil extends HBaseCommonTestingUtil {
|
|||
}
|
||||
|
||||
/**
|
||||
* Gets a ZKWatcher.
|
||||
*/
|
||||
public static ZKWatcher getZooKeeperWatcher(HBaseZKTestingUtil testUtil) throws IOException {
|
||||
return new ZKWatcher(testUtil.getConfiguration(), "unittest", new Abortable() {
|
||||
boolean aborted = false;
|
||||
|
||||
@Override
|
||||
public void abort(String why, Throwable e) {
|
||||
aborted = true;
|
||||
throw new RuntimeException("Fatal ZK error, why=" + why, e);
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean isAborted() {
|
||||
return aborted;
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* @return True if we removed the test dirs
|
||||
* Returns true if we removed the test dirs
|
||||
*/
|
||||
@Override
|
||||
public boolean cleanupTestDir() {
|
||||
|
|
Loading…
Reference in New Issue