HBASE-15568 Procedure V2 - Remove CreateTableHandler in HBase Apache 2.0 release (Stephen Yuan Jiang)
This commit is contained in:
parent
2d8e0a0477
commit
89d750166d
|
@ -1,311 +0,0 @@
|
|||
/**
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.master.handler;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.InterruptedIOException;
|
||||
import java.security.PrivilegedExceptionAction;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.CoordinatedStateException;
|
||||
import org.apache.hadoop.hbase.HRegionInfo;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.NotAllMetaRegionsOnlineException;
|
||||
import org.apache.hadoop.hbase.Server;
|
||||
import org.apache.hadoop.hbase.TableDescriptor;
|
||||
import org.apache.hadoop.hbase.TableExistsException;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||
import org.apache.hadoop.hbase.client.TableState;
|
||||
import org.apache.hadoop.hbase.executor.EventHandler;
|
||||
import org.apache.hadoop.hbase.executor.EventType;
|
||||
import org.apache.hadoop.hbase.ipc.RpcServer;
|
||||
import org.apache.hadoop.hbase.master.AssignmentManager;
|
||||
import org.apache.hadoop.hbase.master.HMaster;
|
||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
||||
import org.apache.hadoop.hbase.master.MasterServices;
|
||||
import org.apache.hadoop.hbase.master.TableLockManager;
|
||||
import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
|
||||
import org.apache.hadoop.hbase.security.User;
|
||||
import org.apache.hadoop.hbase.security.UserProvider;
|
||||
import org.apache.hadoop.hbase.util.FSTableDescriptors;
|
||||
import org.apache.hadoop.hbase.util.FSUtils;
|
||||
import org.apache.hadoop.hbase.util.ModifyRegionUtils;
|
||||
import org.apache.hadoop.hbase.util.ServerRegionReplicaUtil;
|
||||
|
||||
/**
|
||||
* Handler to create a table.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
public class CreateTableHandler extends EventHandler {
|
||||
private static final Log LOG = LogFactory.getLog(CreateTableHandler.class);
|
||||
protected final MasterFileSystem fileSystemManager;
|
||||
protected final HTableDescriptor hTableDescriptor;
|
||||
protected final Configuration conf;
|
||||
private final AssignmentManager assignmentManager;
|
||||
private final TableLockManager tableLockManager;
|
||||
private final HRegionInfo [] newRegions;
|
||||
private final MasterServices masterServices;
|
||||
private final TableLock tableLock;
|
||||
private User activeUser;
|
||||
|
||||
public CreateTableHandler(Server server, MasterFileSystem fileSystemManager,
|
||||
HTableDescriptor hTableDescriptor, Configuration conf, HRegionInfo [] newRegions,
|
||||
MasterServices masterServices) {
|
||||
super(server, EventType.C_M_CREATE_TABLE);
|
||||
|
||||
this.fileSystemManager = fileSystemManager;
|
||||
this.hTableDescriptor = hTableDescriptor;
|
||||
this.conf = conf;
|
||||
this.newRegions = newRegions;
|
||||
this.masterServices = masterServices;
|
||||
this.assignmentManager = masterServices.getAssignmentManager();
|
||||
this.tableLockManager = masterServices.getTableLockManager();
|
||||
|
||||
this.tableLock = this.tableLockManager.writeLock(this.hTableDescriptor.getTableName()
|
||||
, EventType.C_M_CREATE_TABLE.toString());
|
||||
}
|
||||
|
||||
@Override
|
||||
public CreateTableHandler prepare()
|
||||
throws NotAllMetaRegionsOnlineException, TableExistsException, IOException {
|
||||
int timeout = conf.getInt("hbase.client.catalog.timeout", 10000);
|
||||
// Need hbase:meta availability to create a table
|
||||
try {
|
||||
if (server.getMetaTableLocator().waitMetaRegionLocation(
|
||||
server.getZooKeeper(), timeout) == null) {
|
||||
throw new NotAllMetaRegionsOnlineException();
|
||||
}
|
||||
// If we are creating the table in service to an RPC request, record the
|
||||
// active user for later, so proper permissions will be applied to the
|
||||
// new table by the AccessController if it is active
|
||||
this.activeUser = RpcServer.getRequestUser();
|
||||
if (this.activeUser == null) {
|
||||
this.activeUser = UserProvider.instantiate(conf).getCurrent();
|
||||
}
|
||||
} catch (InterruptedException e) {
|
||||
LOG.warn("Interrupted waiting for meta availability", e);
|
||||
InterruptedIOException ie = new InterruptedIOException(e.getMessage());
|
||||
ie.initCause(e);
|
||||
throw ie;
|
||||
}
|
||||
|
||||
//acquire the table write lock, blocking. Make sure that it is released.
|
||||
this.tableLock.acquire();
|
||||
boolean success = false;
|
||||
try {
|
||||
TableName tableName = this.hTableDescriptor.getTableName();
|
||||
if (MetaTableAccessor.tableExists(this.server.getConnection(), tableName)) {
|
||||
throw new TableExistsException(tableName);
|
||||
}
|
||||
success = true;
|
||||
} finally {
|
||||
if (!success) {
|
||||
releaseTableLock();
|
||||
}
|
||||
}
|
||||
return this;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
String name = "UnknownServerName";
|
||||
if(server != null && server.getServerName() != null) {
|
||||
name = server.getServerName().toString();
|
||||
}
|
||||
return getClass().getSimpleName() + "-" + name + "-" + getSeqid() + "-" +
|
||||
this.hTableDescriptor.getTableName();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void process() {
|
||||
TableName tableName = this.hTableDescriptor.getTableName();
|
||||
LOG.info("Create table " + tableName);
|
||||
HMaster master = ((HMaster) this.server);
|
||||
try {
|
||||
final MasterCoprocessorHost cpHost = master.getMasterCoprocessorHost();
|
||||
if (cpHost != null) {
|
||||
cpHost.preCreateTableHandler(this.hTableDescriptor, this.newRegions);
|
||||
}
|
||||
handleCreateTable(tableName);
|
||||
completed(null);
|
||||
if (cpHost != null) {
|
||||
this.activeUser.runAs(new PrivilegedExceptionAction<Void>() {
|
||||
@Override
|
||||
public Void run() throws Exception {
|
||||
cpHost.postCreateTableHandler(hTableDescriptor, newRegions);
|
||||
return null;
|
||||
}
|
||||
});
|
||||
}
|
||||
} catch (Throwable e) {
|
||||
LOG.error("Error trying to create the table " + tableName, e);
|
||||
if (master.isInitialized()) {
|
||||
try {
|
||||
((HMaster) this.server).getMasterQuotaManager().removeTableFromNamespaceQuota(
|
||||
hTableDescriptor.getTableName());
|
||||
} catch (IOException e1) {
|
||||
LOG.error("Error trying to update namespace quota " + e1);
|
||||
}
|
||||
}
|
||||
completed(e);
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Called after that process() is completed.
|
||||
* @param exception null if process() is successful or not null if something has failed.
|
||||
*/
|
||||
protected void completed(final Throwable exception) {
|
||||
releaseTableLock();
|
||||
LOG.info("Table, " + this.hTableDescriptor.getTableName() + ", creation " +
|
||||
(exception == null ? "successful" : "failed. " + exception));
|
||||
}
|
||||
|
||||
/**
|
||||
* Responsible of table creation (on-disk and META) and assignment.
|
||||
* - Create the table directory and descriptor (temp folder)
|
||||
* - Create the on-disk regions (temp folder)
|
||||
* [If something fails here: we've just some trash in temp]
|
||||
* - Move the table from temp to the root directory
|
||||
* [If something fails here: we've the table in place but some of the rows required
|
||||
* present in META. (hbck needed)]
|
||||
* - Add regions to META
|
||||
* [If something fails here: we don't have regions assigned: table disabled]
|
||||
* - Assign regions to Region Servers
|
||||
* [If something fails here: we still have the table in disabled state]
|
||||
* - Update ZooKeeper with the enabled state
|
||||
*/
|
||||
private void handleCreateTable(TableName tableName)
|
||||
throws IOException, CoordinatedStateException {
|
||||
Path tempdir = fileSystemManager.getTempDir();
|
||||
FileSystem fs = fileSystemManager.getFileSystem();
|
||||
|
||||
// 1. Create Table Descriptor
|
||||
// using a copy of descriptor, table will be created enabling first
|
||||
TableDescriptor underConstruction = new TableDescriptor(
|
||||
this.hTableDescriptor);
|
||||
Path tempTableDir = FSUtils.getTableDir(tempdir, tableName);
|
||||
((FSTableDescriptors)(masterServices.getTableDescriptors()))
|
||||
.createTableDescriptorForTableDirectory(
|
||||
tempTableDir, underConstruction, false);
|
||||
Path tableDir = FSUtils.getTableDir(fileSystemManager.getRootDir(), tableName);
|
||||
|
||||
// 2. Create Regions
|
||||
List<HRegionInfo> regionInfos = handleCreateHdfsRegions(tempdir, tableName);
|
||||
// 3. Move Table temp directory to the hbase root location
|
||||
if (!fs.rename(tempTableDir, tableDir)) {
|
||||
throw new IOException("Unable to move table from temp=" + tempTableDir +
|
||||
" to hbase root=" + tableDir);
|
||||
}
|
||||
|
||||
// populate descriptors cache to be visible in getAll
|
||||
masterServices.getTableDescriptors().get(tableName);
|
||||
|
||||
MetaTableAccessor.updateTableState(this.server.getConnection(), hTableDescriptor.getTableName(),
|
||||
TableState.State.ENABLING);
|
||||
|
||||
if (regionInfos != null && regionInfos.size() > 0) {
|
||||
// 4. Add regions to META
|
||||
addRegionsToMeta(regionInfos, hTableDescriptor.getRegionReplication());
|
||||
// 5. Add replicas if needed
|
||||
regionInfos = addReplicas(hTableDescriptor, regionInfos);
|
||||
|
||||
// 6. Setup replication for region replicas if needed
|
||||
if (hTableDescriptor.getRegionReplication() > 1) {
|
||||
ServerRegionReplicaUtil.setupRegionReplicaReplication(conf);
|
||||
}
|
||||
|
||||
// 7. Trigger immediate assignment of the regions in round-robin fashion
|
||||
ModifyRegionUtils.assignRegions(assignmentManager, regionInfos);
|
||||
}
|
||||
|
||||
// 8. Enable table
|
||||
assignmentManager.getTableStateManager().setTableState(tableName,
|
||||
TableState.State.ENABLED);
|
||||
|
||||
// 9. Update the tabledescriptor cache.
|
||||
((HMaster) this.server).getTableDescriptors().get(tableName);
|
||||
}
|
||||
|
||||
/**
|
||||
* Create any replicas for the regions (the default replicas that was
|
||||
* already created is passed to the method)
|
||||
* @param hTableDescriptor descriptor to use
|
||||
* @param regions default replicas
|
||||
* @return the combined list of default and non-default replicas
|
||||
*/
|
||||
protected List<HRegionInfo> addReplicas(HTableDescriptor hTableDescriptor,
|
||||
List<HRegionInfo> regions) {
|
||||
int numRegionReplicas = hTableDescriptor.getRegionReplication() - 1;
|
||||
if (numRegionReplicas <= 0) {
|
||||
return regions;
|
||||
}
|
||||
List<HRegionInfo> hRegionInfos =
|
||||
new ArrayList<HRegionInfo>((numRegionReplicas+1)*regions.size());
|
||||
for (int i = 0; i < regions.size(); i++) {
|
||||
for (int j = 1; j <= numRegionReplicas; j++) {
|
||||
hRegionInfos.add(RegionReplicaUtil.getRegionInfoForReplica(regions.get(i), j));
|
||||
}
|
||||
}
|
||||
hRegionInfos.addAll(regions);
|
||||
return hRegionInfos;
|
||||
}
|
||||
|
||||
private void releaseTableLock() {
|
||||
if (this.tableLock != null) {
|
||||
try {
|
||||
this.tableLock.release();
|
||||
} catch (IOException ex) {
|
||||
LOG.warn("Could not release the table lock", ex);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Create the on-disk structure for the table, and returns the regions info.
|
||||
* @param tableRootDir directory where the table is being created
|
||||
* @param tableName name of the table under construction
|
||||
* @return the list of regions created
|
||||
*/
|
||||
protected List<HRegionInfo> handleCreateHdfsRegions(final Path tableRootDir,
|
||||
final TableName tableName)
|
||||
throws IOException {
|
||||
return ModifyRegionUtils.createRegions(conf, tableRootDir,
|
||||
hTableDescriptor, newRegions, null);
|
||||
}
|
||||
|
||||
/**
|
||||
* Add the specified set of regions to the hbase:meta table.
|
||||
*/
|
||||
protected void addRegionsToMeta(final List<HRegionInfo> regionInfos, int regionReplication)
|
||||
throws IOException {
|
||||
MetaTableAccessor.addRegionsToMeta(this.server.getConnection(), regionInfos, regionReplication);
|
||||
}
|
||||
}
|
|
@ -1,178 +0,0 @@
|
|||
/**
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.master.handler;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||
import org.apache.hadoop.hbase.HRegionInfo;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||
import org.apache.hadoop.hbase.Server;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.master.HMaster;
|
||||
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
||||
import org.apache.hadoop.hbase.master.MasterServices;
|
||||
import org.apache.hadoop.hbase.master.RegionState.State;
|
||||
import org.apache.hadoop.hbase.master.RegionStates;
|
||||
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.hbase.util.FSUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.experimental.categories.Category;
|
||||
|
||||
@Category({MasterTests.class, MediumTests.class})
|
||||
public class TestCreateTableHandler {
|
||||
private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
|
||||
private static final Log LOG = LogFactory.getLog(TestCreateTableHandler.class);
|
||||
private static final byte[] FAMILYNAME = Bytes.toBytes("fam");
|
||||
private static boolean throwException = false;
|
||||
|
||||
|
||||
@Before
|
||||
public void setUp() throws Exception {
|
||||
TEST_UTIL.startMiniCluster(1);
|
||||
}
|
||||
|
||||
@After
|
||||
public void tearDown() throws Exception {
|
||||
TEST_UTIL.shutdownMiniCluster();
|
||||
throwException = false;
|
||||
}
|
||||
|
||||
@Test (timeout=300000)
|
||||
public void testCreateTableCalledTwiceAndFirstOneInProgress() throws Exception {
|
||||
final TableName tableName = TableName.valueOf("testCreateTableCalledTwiceAndFirstOneInProgress");
|
||||
final MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
|
||||
final HMaster m = cluster.getMaster();
|
||||
final HTableDescriptor desc = new HTableDescriptor(tableName);
|
||||
desc.addFamily(new HColumnDescriptor(FAMILYNAME));
|
||||
final HRegionInfo[] hRegionInfos = new HRegionInfo[] { new HRegionInfo(desc.getTableName(), null,
|
||||
null) };
|
||||
CustomCreateTableHandler handler = new CustomCreateTableHandler(m, m.getMasterFileSystem(),
|
||||
desc, cluster.getConfiguration(), hRegionInfos, m);
|
||||
handler.prepare();
|
||||
throwException = true;
|
||||
handler.process();
|
||||
throwException = false;
|
||||
CustomCreateTableHandler handler1 = new CustomCreateTableHandler(m, m.getMasterFileSystem(),
|
||||
desc, cluster.getConfiguration(), hRegionInfos, m);
|
||||
handler1.prepare();
|
||||
handler1.process();
|
||||
for (int i = 0; i < 100; i++) {
|
||||
if (!TEST_UTIL.getHBaseAdmin().isTableAvailable(tableName)) {
|
||||
Thread.sleep(200);
|
||||
}
|
||||
}
|
||||
assertTrue(TEST_UTIL.getHBaseAdmin().isTableEnabled(tableName));
|
||||
}
|
||||
|
||||
@Test (timeout=300000)
|
||||
public void testCreateTableWithSplitRegion() throws Exception {
|
||||
final TableName tableName = TableName.valueOf("testCreateTableWithSplitRegion");
|
||||
final MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
|
||||
final HMaster m = cluster.getMaster();
|
||||
final HTableDescriptor desc = new HTableDescriptor(tableName);
|
||||
desc.addFamily(new HColumnDescriptor(FAMILYNAME));
|
||||
byte[] splitPoint = Bytes.toBytes("split-point");
|
||||
long ts = System.currentTimeMillis();
|
||||
HRegionInfo d1 = new HRegionInfo(desc.getTableName(), null, splitPoint, false, ts);
|
||||
HRegionInfo d2 = new HRegionInfo(desc.getTableName(), splitPoint, null, false, ts + 1);
|
||||
HRegionInfo parent = new HRegionInfo(desc.getTableName(), null, null, true, ts + 2);
|
||||
parent.setOffline(true);
|
||||
|
||||
Path tempdir = m.getMasterFileSystem().getTempDir();
|
||||
FileSystem fs = m.getMasterFileSystem().getFileSystem();
|
||||
Path tempTableDir = FSUtils.getTableDir(tempdir, desc.getTableName());
|
||||
fs.delete(tempTableDir, true); // Clean up temp table dir if exists
|
||||
|
||||
final HRegionInfo[] hRegionInfos = new HRegionInfo[] {d1, d2, parent};
|
||||
CreateTableHandler handler = new CreateTableHandler(m, m.getMasterFileSystem(),
|
||||
desc, cluster.getConfiguration(), hRegionInfos, m);
|
||||
handler.prepare();
|
||||
handler.process();
|
||||
for (int i = 0; i < 100; i++) {
|
||||
if (!TEST_UTIL.getHBaseAdmin().isTableAvailable(tableName)) {
|
||||
Thread.sleep(300);
|
||||
}
|
||||
}
|
||||
assertTrue(TEST_UTIL.getHBaseAdmin().isTableEnabled(tableName));
|
||||
assertTrue(TEST_UTIL.getHBaseAdmin().isTableAvailable(tableName));
|
||||
assertTrue(TEST_UTIL.getHBaseAdmin().isTableAvailable(tableName, new byte[][] { splitPoint }));
|
||||
RegionStates regionStates = m.getAssignmentManager().getRegionStates();
|
||||
assertTrue("Parent should be in SPLIT state",
|
||||
regionStates.isRegionInState(parent, State.SPLIT));
|
||||
}
|
||||
|
||||
@Test (timeout=60000)
|
||||
public void testMasterRestartAfterEnablingNodeIsCreated() throws Exception {
|
||||
byte[] tableName = Bytes.toBytes("testMasterRestartAfterEnablingNodeIsCreated");
|
||||
final MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
|
||||
final HMaster m = cluster.getMaster();
|
||||
final HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
|
||||
desc.addFamily(new HColumnDescriptor(FAMILYNAME));
|
||||
final HRegionInfo[] hRegionInfos = new HRegionInfo[] { new HRegionInfo(desc.getTableName(), null,
|
||||
null) };
|
||||
CustomCreateTableHandler handler = new CustomCreateTableHandler(m, m.getMasterFileSystem(),
|
||||
desc, cluster.getConfiguration(), hRegionInfos, m);
|
||||
handler.prepare();
|
||||
throwException = true;
|
||||
handler.process();
|
||||
abortAndStartNewMaster(cluster);
|
||||
assertTrue(cluster.getLiveMasterThreads().size() == 1);
|
||||
}
|
||||
|
||||
private void abortAndStartNewMaster(final MiniHBaseCluster cluster) throws IOException {
|
||||
cluster.abortMaster(0);
|
||||
cluster.waitOnMaster(0);
|
||||
LOG.info("Starting new master");
|
||||
cluster.startMaster();
|
||||
LOG.info("Waiting for master to become active.");
|
||||
cluster.waitForActiveAndReadyMaster();
|
||||
}
|
||||
|
||||
private static class CustomCreateTableHandler extends CreateTableHandler {
|
||||
public CustomCreateTableHandler(Server server, MasterFileSystem fileSystemManager,
|
||||
HTableDescriptor hTableDescriptor, Configuration conf, HRegionInfo[] newRegions,
|
||||
MasterServices masterServices) {
|
||||
super(server, fileSystemManager, hTableDescriptor, conf, newRegions, masterServices);
|
||||
}
|
||||
|
||||
@Override
|
||||
protected List<HRegionInfo> handleCreateHdfsRegions(Path tableRootDir,
|
||||
TableName tableName) throws IOException {
|
||||
if (throwException) {
|
||||
throw new IOException("Test throws exceptions.");
|
||||
}
|
||||
return super.handleCreateHdfsRegions(tableRootDir, tableName);
|
||||
}
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue