HBASE-24788: Fix the connection leaks on getting hbase admin from unclosed connection (#2162)
Signed-off-by: Bharath Vissapragada <bharathv@apache.org> Signed-off-by: Viraj Jasani <vjasani@apache.org>
This commit is contained in:
parent
5f27a009ab
commit
d65fb87a2e
|
@ -19,9 +19,6 @@
|
|||
package org.apache.hadoop.hbase.mapreduce;
|
||||
|
||||
import java.io.IOException;
|
||||
import org.apache.yetus.audience.InterfaceAudience;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.apache.hadoop.conf.Configurable;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||
|
@ -41,6 +38,9 @@ import org.apache.hadoop.mapreduce.OutputCommitter;
|
|||
import org.apache.hadoop.mapreduce.OutputFormat;
|
||||
import org.apache.hadoop.mapreduce.RecordWriter;
|
||||
import org.apache.hadoop.mapreduce.TaskAttemptContext;
|
||||
import org.apache.yetus.audience.InterfaceAudience;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Convert Map/Reduce output and write it to an HBase table. The KEY is ignored
|
||||
|
@ -155,11 +155,10 @@ implements Configurable {
|
|||
* @param context The current task context.
|
||||
* @return The newly created writer instance.
|
||||
* @throws IOException When creating the writer fails.
|
||||
* @throws InterruptedException When the jobs is cancelled.
|
||||
*/
|
||||
@Override
|
||||
public RecordWriter<KEY, Mutation> getRecordWriter(TaskAttemptContext context)
|
||||
throws IOException, InterruptedException {
|
||||
throws IOException {
|
||||
return new TableRecordWriter();
|
||||
}
|
||||
|
||||
|
@ -168,18 +167,18 @@ implements Configurable {
|
|||
*
|
||||
* @param context The current context.
|
||||
* @throws IOException When the check fails.
|
||||
* @throws InterruptedException When the job is aborted.
|
||||
* @see OutputFormat#checkOutputSpecs(JobContext)
|
||||
*/
|
||||
@Override
|
||||
public void checkOutputSpecs(JobContext context) throws IOException,
|
||||
InterruptedException {
|
||||
public void checkOutputSpecs(JobContext context)
|
||||
throws IOException {
|
||||
Configuration hConf = getConf();
|
||||
if (hConf == null) {
|
||||
hConf = context.getConfiguration();
|
||||
}
|
||||
|
||||
try (Admin admin = ConnectionFactory.createConnection(hConf).getAdmin()) {
|
||||
try (Connection connection = ConnectionFactory.createConnection(hConf);
|
||||
Admin admin = connection.getAdmin()) {
|
||||
TableName tableName = TableName.valueOf(hConf.get(OUTPUT_TABLE));
|
||||
if (!admin.tableExists(tableName)) {
|
||||
throw new TableNotFoundException("Can't write, table does not exist:" +
|
||||
|
|
|
@ -24,6 +24,7 @@ import org.apache.hadoop.fs.FileSystem;
|
|||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.ReplicationPeerNotFoundException;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.RegionInfo;
|
||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||
|
@ -161,15 +162,18 @@ public class ServerRegionReplicaUtil extends RegionReplicaUtil {
|
|||
if (!isRegionReplicaReplicationEnabled(conf)) {
|
||||
return;
|
||||
}
|
||||
Admin admin = ConnectionFactory.createConnection(conf).getAdmin();
|
||||
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Admin admin = connection.getAdmin()) {
|
||||
ReplicationPeerConfig peerConfig = null;
|
||||
try {
|
||||
peerConfig = admin.getReplicationPeerConfig(REGION_REPLICA_REPLICATION_PEER);
|
||||
} catch (ReplicationPeerNotFoundException e) {
|
||||
LOG.warn("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER
|
||||
+ " not exist", e);
|
||||
LOG.warn(
|
||||
"Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER + " not exist",
|
||||
e);
|
||||
}
|
||||
try {
|
||||
|
||||
if (peerConfig == null) {
|
||||
LOG.info("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER
|
||||
+ " not exist. Creating...");
|
||||
|
@ -178,8 +182,6 @@ public class ServerRegionReplicaUtil extends RegionReplicaUtil {
|
|||
peerConfig.setReplicationEndpointImpl(RegionReplicaReplicationEndpoint.class.getName());
|
||||
admin.addReplicationPeer(REGION_REPLICA_REPLICATION_PEER, peerConfig);
|
||||
}
|
||||
} finally {
|
||||
admin.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -23,6 +23,7 @@ import static org.junit.Assert.assertFalse;
|
|||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Collection;
|
||||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
|
@ -41,6 +42,7 @@ import org.apache.hadoop.hbase.testclassification.ClientTests;
|
|||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.ClassRule;
|
||||
import org.junit.Test;
|
||||
|
@ -64,6 +66,7 @@ public class TestAsyncReplicationAdminApiWithClusters extends TestAsyncAdminBase
|
|||
private static HBaseTestingUtility TEST_UTIL2;
|
||||
private static Configuration conf2;
|
||||
private static AsyncAdmin admin2;
|
||||
private static AsyncConnection connection;
|
||||
|
||||
@BeforeClass
|
||||
public static void setUpBeforeClass() throws Exception {
|
||||
|
@ -78,14 +81,21 @@ public class TestAsyncReplicationAdminApiWithClusters extends TestAsyncAdminBase
|
|||
conf2.set(HConstants.ZOOKEEPER_ZNODE_PARENT, "/2");
|
||||
TEST_UTIL2 = new HBaseTestingUtility(conf2);
|
||||
TEST_UTIL2.startMiniCluster();
|
||||
admin2 =
|
||||
ConnectionFactory.createAsyncConnection(TEST_UTIL2.getConfiguration()).get().getAdmin();
|
||||
|
||||
connection =
|
||||
ConnectionFactory.createAsyncConnection(TEST_UTIL2.getConfiguration()).get();
|
||||
admin2 = connection.getAdmin();
|
||||
|
||||
ReplicationPeerConfig rpc = new ReplicationPeerConfig();
|
||||
rpc.setClusterKey(TEST_UTIL2.getClusterKey());
|
||||
ASYNC_CONN.getAdmin().addReplicationPeer(ID_SECOND, rpc).join();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void clearUp() throws IOException {
|
||||
connection.close();
|
||||
}
|
||||
|
||||
@Override
|
||||
@After
|
||||
public void tearDown() throws Exception {
|
||||
|
|
|
@ -396,12 +396,12 @@ public class TestReplicaWithCluster {
|
|||
LOG.info("Setup second Zk");
|
||||
HTU2.getAdmin().createTable(tableDescriptor, HBaseTestingUtility.KEYS_FOR_HBA_CREATE_TABLE);
|
||||
|
||||
Admin admin = ConnectionFactory.createConnection(HTU.getConfiguration()).getAdmin();
|
||||
|
||||
try (Connection connection = ConnectionFactory.createConnection(HTU.getConfiguration());
|
||||
Admin admin = connection.getAdmin()) {
|
||||
ReplicationPeerConfig rpc = new ReplicationPeerConfig();
|
||||
rpc.setClusterKey(HTU2.getClusterKey());
|
||||
admin.addReplicationPeer("2", rpc);
|
||||
admin.close();
|
||||
}
|
||||
|
||||
Put p = new Put(row);
|
||||
p.addColumn(row, row, row);
|
||||
|
|
|
@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster;
|
|||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.AsyncAdmin;
|
||||
import org.apache.hadoop.hbase.client.AsyncConnection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.client.Table;
|
||||
|
@ -149,8 +150,9 @@ public class TestClearRegionBlockCache {
|
|||
|
||||
@Test
|
||||
public void testClearBlockCacheFromAsyncAdmin() throws Exception {
|
||||
AsyncAdmin admin =
|
||||
ConnectionFactory.createAsyncConnection(HTU.getConfiguration()).get().getAdmin();
|
||||
try (AsyncConnection conn = ConnectionFactory.createAsyncConnection(HTU.getConfiguration())
|
||||
.get()) {
|
||||
AsyncAdmin admin = conn.getAdmin();
|
||||
|
||||
BlockCache blockCache1 = rs1.getBlockCache().get();
|
||||
BlockCache blockCache2 = rs2.getBlockCache().get();
|
||||
|
@ -171,6 +173,7 @@ public class TestClearRegionBlockCache {
|
|||
assertEquals(initialBlockCount1, blockCache1.getBlockCount());
|
||||
assertEquals(initialBlockCount2, blockCache2.getBlockCount());
|
||||
}
|
||||
}
|
||||
|
||||
private void scanAllRegionsForRS(HRegionServer rs) throws IOException {
|
||||
for (Region region : rs.getRegions(TABLE_NAME)) {
|
||||
|
|
|
@ -43,6 +43,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster;
|
|||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Delete;
|
||||
import org.apache.hadoop.hbase.client.Durability;
|
||||
|
@ -486,8 +487,8 @@ public class TestMasterReplication {
|
|||
|
||||
private void addPeer(String id, int masterClusterNumber,
|
||||
int slaveClusterNumber) throws Exception {
|
||||
try (Admin admin = ConnectionFactory.createConnection(configurations[masterClusterNumber])
|
||||
.getAdmin()) {
|
||||
try (Connection conn = ConnectionFactory.createConnection(configurations[masterClusterNumber]);
|
||||
Admin admin = conn.getAdmin()) {
|
||||
admin.addReplicationPeer(id,
|
||||
new ReplicationPeerConfig().setClusterKey(utilities[slaveClusterNumber].getClusterKey()));
|
||||
}
|
||||
|
@ -495,8 +496,8 @@ public class TestMasterReplication {
|
|||
|
||||
private void addPeer(String id, int masterClusterNumber, int slaveClusterNumber, String tableCfs)
|
||||
throws Exception {
|
||||
try (Admin admin =
|
||||
ConnectionFactory.createConnection(configurations[masterClusterNumber]).getAdmin()) {
|
||||
try (Connection conn = ConnectionFactory.createConnection(configurations[masterClusterNumber]);
|
||||
Admin admin = conn.getAdmin()) {
|
||||
admin.addReplicationPeer(
|
||||
id,
|
||||
new ReplicationPeerConfig().setClusterKey(utilities[slaveClusterNumber].getClusterKey())
|
||||
|
@ -506,15 +507,15 @@ public class TestMasterReplication {
|
|||
}
|
||||
|
||||
private void disablePeer(String id, int masterClusterNumber) throws Exception {
|
||||
try (Admin admin = ConnectionFactory.createConnection(configurations[masterClusterNumber])
|
||||
.getAdmin()) {
|
||||
try (Connection conn = ConnectionFactory.createConnection(configurations[masterClusterNumber]);
|
||||
Admin admin = conn.getAdmin()) {
|
||||
admin.disableReplicationPeer(id);
|
||||
}
|
||||
}
|
||||
|
||||
private void enablePeer(String id, int masterClusterNumber) throws Exception {
|
||||
try (Admin admin = ConnectionFactory.createConnection(configurations[masterClusterNumber])
|
||||
.getAdmin()) {
|
||||
try (Connection conn = ConnectionFactory.createConnection(configurations[masterClusterNumber]);
|
||||
Admin admin = conn.getAdmin()) {
|
||||
admin.enableReplicationPeer(id);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -35,6 +35,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster;
|
|||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Delete;
|
||||
import org.apache.hadoop.hbase.client.Get;
|
||||
|
@ -138,8 +139,8 @@ public class TestMultiSlaveReplication {
|
|||
MiniHBaseCluster master = utility1.startMiniCluster();
|
||||
utility2.startMiniCluster();
|
||||
utility3.startMiniCluster();
|
||||
Admin admin1 = ConnectionFactory.createConnection(conf1).getAdmin();
|
||||
|
||||
try (Connection conn = ConnectionFactory.createConnection(conf1);
|
||||
Admin admin1 = conn.getAdmin()) {
|
||||
utility1.getAdmin().createTable(table);
|
||||
utility2.getAdmin().createTable(table);
|
||||
utility3.getAdmin().createTable(table);
|
||||
|
@ -155,7 +156,7 @@ public class TestMultiSlaveReplication {
|
|||
putAndWait(row, famName, htable1, htable2);
|
||||
deleteAndWait(row, htable1, htable2);
|
||||
// check it wasn't replication to cluster 3
|
||||
checkRow(row,0,htable3);
|
||||
checkRow(row, 0, htable3);
|
||||
|
||||
putAndWait(row2, famName, htable1, htable2);
|
||||
|
||||
|
@ -176,11 +177,11 @@ public class TestMultiSlaveReplication {
|
|||
|
||||
// make sure row2 did not get replicated after
|
||||
// cluster 3 was added
|
||||
checkRow(row2,0,htable3);
|
||||
checkRow(row2, 0, htable3);
|
||||
|
||||
// row3 will get replicated, because it was in the
|
||||
// latest log
|
||||
checkRow(row3,1,htable3);
|
||||
checkRow(row3, 1, htable3);
|
||||
|
||||
Put p = new Put(row);
|
||||
p.addColumn(famName, row, row);
|
||||
|
@ -206,6 +207,7 @@ public class TestMultiSlaveReplication {
|
|||
utility2.shutdownMiniCluster();
|
||||
utility1.shutdownMiniCluster();
|
||||
}
|
||||
}
|
||||
|
||||
private void rollWALAndWait(final HBaseTestingUtility utility, final TableName table,
|
||||
final byte[] row) throws IOException {
|
||||
|
|
|
@ -17,8 +17,13 @@
|
|||
*/
|
||||
package org.apache.hadoop.hbase.replication;
|
||||
|
||||
import static org.junit.Assert.*;
|
||||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashMap;
|
||||
|
@ -56,7 +61,6 @@ import org.junit.experimental.categories.Category;
|
|||
import org.junit.rules.TestName;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos;
|
||||
|
||||
@Category({FlakeyTests.class, LargeTests.class})
|
||||
|
@ -375,14 +379,13 @@ public class TestPerTableCFReplication {
|
|||
@Test
|
||||
public void testPerTableCFReplication() throws Exception {
|
||||
LOG.info("testPerTableCFReplication");
|
||||
Admin replicationAdmin = ConnectionFactory.createConnection(conf1).getAdmin();
|
||||
Connection connection1 = ConnectionFactory.createConnection(conf1);
|
||||
try (Connection connection1 = ConnectionFactory.createConnection(conf1);
|
||||
Connection connection2 = ConnectionFactory.createConnection(conf2);
|
||||
Connection connection3 = ConnectionFactory.createConnection(conf3);
|
||||
try {
|
||||
Admin admin1 = connection1.getAdmin();
|
||||
Admin admin2 = connection2.getAdmin();
|
||||
Admin admin3 = connection3.getAdmin();
|
||||
Admin replicationAdmin = connection1.getAdmin()) {
|
||||
|
||||
admin1.createTable(tabA);
|
||||
admin1.createTable(tabB);
|
||||
|
@ -524,10 +527,6 @@ public class TestPerTableCFReplication {
|
|||
// cf 'f3' of tableC can replicated to cluster2 and cluster3
|
||||
putAndWaitWithFamily(row2, f3Name, htab1C, htab2C, htab3C);
|
||||
deleteAndWaitWithFamily(row2, f3Name, htab1C, htab2C, htab3C);
|
||||
} finally {
|
||||
connection1.close();
|
||||
connection2.close();
|
||||
connection3.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.hadoop.hbase.replication;
|
|||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
@ -54,7 +53,6 @@ import org.junit.Before;
|
|||
import org.junit.BeforeClass;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableList;
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableMap;
|
||||
|
||||
|
@ -66,7 +64,8 @@ import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableMap;
|
|||
*/
|
||||
public class TestReplicationBase {
|
||||
private static final Logger LOG = LoggerFactory.getLogger(TestReplicationBase.class);
|
||||
|
||||
private static Connection connection1;
|
||||
private static Connection connection2;
|
||||
protected static Configuration CONF_WITH_LOCALFS;
|
||||
|
||||
protected static Admin hbaseAdmin;
|
||||
|
@ -244,27 +243,28 @@ public class TestReplicationBase {
|
|||
// as a component in deciding maximum number of parallel batches to send to the peer cluster.
|
||||
UTIL2.startMiniCluster(NUM_SLAVES2);
|
||||
|
||||
hbaseAdmin = ConnectionFactory.createConnection(CONF1).getAdmin();
|
||||
connection1 = ConnectionFactory.createConnection(CONF1);
|
||||
connection2 = ConnectionFactory.createConnection(CONF2);
|
||||
hbaseAdmin = connection1.getAdmin();
|
||||
|
||||
TableDescriptor table = TableDescriptorBuilder.newBuilder(tableName)
|
||||
.setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(famName).setMaxVersions(100)
|
||||
.setScope(HConstants.REPLICATION_SCOPE_GLOBAL).build())
|
||||
.setColumnFamily(ColumnFamilyDescriptorBuilder.of(noRepfamName)).build();
|
||||
|
||||
Connection connection1 = ConnectionFactory.createConnection(CONF1);
|
||||
Connection connection2 = ConnectionFactory.createConnection(CONF2);
|
||||
try (Admin admin1 = connection1.getAdmin()) {
|
||||
try (
|
||||
Admin admin1 = connection1.getAdmin();
|
||||
Admin admin2 = connection2.getAdmin()) {
|
||||
admin1.createTable(table, HBaseTestingUtility.KEYS_FOR_HBA_CREATE_TABLE);
|
||||
}
|
||||
try (Admin admin2 = connection2.getAdmin()) {
|
||||
admin2.createTable(table, HBaseTestingUtility.KEYS_FOR_HBA_CREATE_TABLE);
|
||||
}
|
||||
UTIL1.waitUntilAllRegionsAssigned(tableName);
|
||||
UTIL2.waitUntilAllRegionsAssigned(tableName);
|
||||
htable1 = connection1.getTable(tableName);
|
||||
UTIL2.waitUntilAllRegionsAssigned(tableName);
|
||||
htable2 = connection2.getTable(tableName);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
public static void setUpBeforeClass() throws Exception {
|
||||
configureClusters(UTIL1, UTIL2);
|
||||
|
@ -366,6 +366,13 @@ public class TestReplicationBase {
|
|||
if (hbaseAdmin != null) {
|
||||
hbaseAdmin.close();
|
||||
}
|
||||
|
||||
if (connection2 != null) {
|
||||
connection2.close();
|
||||
}
|
||||
if (connection1 != null) {
|
||||
connection1.close();
|
||||
}
|
||||
UTIL2.shutdownMiniCluster();
|
||||
UTIL1.shutdownMiniCluster();
|
||||
}
|
||||
|
|
|
@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.HTestConst;
|
|||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Put;
|
||||
import org.apache.hadoop.hbase.client.Result;
|
||||
|
@ -105,12 +106,14 @@ public class TestGlobalReplicationThrottler {
|
|||
utility1.startMiniCluster();
|
||||
utility2.startMiniCluster();
|
||||
|
||||
Admin admin1 = ConnectionFactory.createConnection(conf1).getAdmin();
|
||||
try (Connection connection = ConnectionFactory.createConnection(utility1.getConfiguration());
|
||||
Admin admin1 = connection.getAdmin()) {
|
||||
admin1.addReplicationPeer("peer1", rpc);
|
||||
admin1.addReplicationPeer("peer2", rpc);
|
||||
admin1.addReplicationPeer("peer3", rpc);
|
||||
numOfPeer = admin1.listReplicationPeers().size();
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDownAfterClass() throws Exception {
|
||||
|
|
|
@ -53,7 +53,6 @@ import org.apache.hadoop.hbase.regionserver.HRegion;
|
|||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||
import org.apache.hadoop.hbase.regionserver.Region;
|
||||
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
||||
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
||||
import org.apache.hadoop.hbase.testclassification.FlakeyTests;
|
||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||
|
@ -122,10 +121,11 @@ public class TestRegionReplicaReplicationEndpoint {
|
|||
}
|
||||
|
||||
@Test
|
||||
public void testRegionReplicaReplicationPeerIsCreated() throws IOException, ReplicationException {
|
||||
public void testRegionReplicaReplicationPeerIsCreated() throws IOException {
|
||||
// create a table with region replicas. Check whether the replication peer is created
|
||||
// and replication started.
|
||||
Admin admin = ConnectionFactory.createConnection(HTU.getConfiguration()).getAdmin();
|
||||
try (Connection connection = ConnectionFactory.createConnection(HTU.getConfiguration());
|
||||
Admin admin = connection.getAdmin()) {
|
||||
String peerId = "region_replica_replication";
|
||||
|
||||
ReplicationPeerConfig peerConfig = null;
|
||||
|
@ -140,8 +140,8 @@ public class TestRegionReplicaReplicationEndpoint {
|
|||
peerConfig = null;
|
||||
}
|
||||
|
||||
HTableDescriptor htd = HTU.createTableDescriptor(
|
||||
TableName.valueOf("testReplicationPeerIsCreated_no_region_replicas"),
|
||||
HTableDescriptor htd = HTU
|
||||
.createTableDescriptor(TableName.valueOf("testReplicationPeerIsCreated_no_region_replicas"),
|
||||
HColumnDescriptor.DEFAULT_MIN_VERSIONS, 3, HConstants.FOREVER,
|
||||
HColumnDescriptor.DEFAULT_KEEP_DELETED);
|
||||
HTU.getAdmin().createTable(htd);
|
||||
|
@ -162,18 +162,19 @@ public class TestRegionReplicaReplicationEndpoint {
|
|||
// assert peer configuration is correct
|
||||
peerConfig = admin.getReplicationPeerConfig(peerId);
|
||||
assertNotNull(peerConfig);
|
||||
assertEquals(peerConfig.getClusterKey(), ZKConfig.getZooKeeperClusterKey(
|
||||
HTU.getConfiguration()));
|
||||
assertEquals(peerConfig.getClusterKey(),
|
||||
ZKConfig.getZooKeeperClusterKey(HTU.getConfiguration()));
|
||||
assertEquals(RegionReplicaReplicationEndpoint.class.getName(),
|
||||
peerConfig.getReplicationEndpointImpl());
|
||||
admin.close();
|
||||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testRegionReplicaReplicationPeerIsCreatedForModifyTable() throws Exception {
|
||||
// modify a table by adding region replicas. Check whether the replication peer is created
|
||||
// and replication started.
|
||||
Admin admin = ConnectionFactory.createConnection(HTU.getConfiguration()).getAdmin();
|
||||
try (Connection connection = ConnectionFactory.createConnection(HTU.getConfiguration());
|
||||
Admin admin = connection.getAdmin()) {
|
||||
String peerId = "region_replica_replication";
|
||||
|
||||
ReplicationPeerConfig peerConfig = null;
|
||||
|
@ -211,11 +212,11 @@ public class TestRegionReplicaReplicationEndpoint {
|
|||
// assert peer configuration is correct
|
||||
peerConfig = admin.getReplicationPeerConfig(peerId);
|
||||
assertNotNull(peerConfig);
|
||||
assertEquals(peerConfig.getClusterKey(), ZKConfig.getZooKeeperClusterKey(
|
||||
HTU.getConfiguration()));
|
||||
assertEquals(peerConfig.getClusterKey(),
|
||||
ZKConfig.getZooKeeperClusterKey(HTU.getConfiguration()));
|
||||
assertEquals(RegionReplicaReplicationEndpoint.class.getName(),
|
||||
peerConfig.getReplicationEndpointImpl());
|
||||
admin.close();
|
||||
}
|
||||
}
|
||||
|
||||
public void testRegionReplicaReplication(int regionReplication) throws Exception {
|
||||
|
|
Loading…
Reference in New Issue