HBASE-19546 TestMasterReplication.testCyclicReplication2 uses wrong assertion

This commit is contained in:
zhangduo 2017-12-18 16:38:10 +08:00
parent 2a5e5e9e2c
commit 62b591bdc7
1 changed files with 17 additions and 20 deletions

View File

@ -1,4 +1,4 @@
/* /**
* Licensed to the Apache Software Foundation (ASF) under one * Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file * or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information * distributed with this work for additional information
@ -38,19 +38,18 @@ import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.ClusterStatus;
import org.apache.hadoop.hbase.ClusterStatus.Option;
import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HBaseTestingUtility;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.MiniHBaseCluster;
import org.apache.hadoop.hbase.ServerLoad; import org.apache.hadoop.hbase.ServerLoad;
import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.ServerName;
import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.Waiter; import org.apache.hadoop.hbase.Waiter;
import org.apache.hadoop.hbase.ClusterStatus.Option;
import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Admin;
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.ConnectionFactory;
import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Delete;
import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Durability;
@ -58,6 +57,8 @@ import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.client.Table;
import org.apache.hadoop.hbase.client.TableDescriptor;
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil; import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
import org.apache.hadoop.hbase.coprocessor.ObserverContext; import org.apache.hadoop.hbase.coprocessor.ObserverContext;
@ -66,13 +67,13 @@ import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
import org.apache.hadoop.hbase.coprocessor.RegionObserver; import org.apache.hadoop.hbase.coprocessor.RegionObserver;
import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener; import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
import org.apache.hadoop.hbase.wal.WALEdit;
import org.apache.hadoop.hbase.replication.regionserver.TestSourceFSConfigurationProvider; import org.apache.hadoop.hbase.replication.regionserver.TestSourceFSConfigurationProvider;
import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.LargeTests;
import org.apache.hadoop.hbase.testclassification.ReplicationTests; import org.apache.hadoop.hbase.testclassification.ReplicationTests;
import org.apache.hadoop.hbase.tool.LoadIncrementalHFiles; import org.apache.hadoop.hbase.tool.LoadIncrementalHFiles;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.HFileTestUtil; import org.apache.hadoop.hbase.util.HFileTestUtil;
import org.apache.hadoop.hbase.wal.WALEdit;
import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster; import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;
import org.apache.hadoop.hbase.zookeeper.ZKUtil; import org.apache.hadoop.hbase.zookeeper.ZKUtil;
import org.apache.hadoop.hbase.zookeeper.ZKWatcher; import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
@ -110,7 +111,7 @@ public class TestMasterReplication {
private static final byte[] put = Bytes.toBytes("put"); private static final byte[] put = Bytes.toBytes("put");
private static final byte[] delete = Bytes.toBytes("delete"); private static final byte[] delete = Bytes.toBytes("delete");
private HTableDescriptor table; private TableDescriptor table;
@Before @Before
public void setUp() throws Exception { public void setUp() throws Exception {
@ -130,16 +131,12 @@ public class TestMasterReplication {
baseConfiguration.setStrings( baseConfiguration.setStrings(
CoprocessorHost.USER_REGION_COPROCESSOR_CONF_KEY, CoprocessorHost.USER_REGION_COPROCESSOR_CONF_KEY,
CoprocessorCounter.class.getName()); CoprocessorCounter.class.getName());
table = TableDescriptorBuilder.newBuilder(tableName)
table = new HTableDescriptor(tableName); .addColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(famName)
HColumnDescriptor fam = new HColumnDescriptor(famName); .setScope(HConstants.REPLICATION_SCOPE_GLOBAL).build())
fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL); .addColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(famName1)
table.addFamily(fam); .setScope(HConstants.REPLICATION_SCOPE_GLOBAL).build())
fam = new HColumnDescriptor(famName1); .addColumnFamily(ColumnFamilyDescriptorBuilder.of(noRepfamName)).build();
fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
table.addFamily(fam);
fam = new HColumnDescriptor(noRepfamName);
table.addFamily(fam);
} }
/** /**
@ -273,7 +270,7 @@ public class TestMasterReplication {
*/ */
@Test(timeout = 300000) @Test(timeout = 300000)
public void testCyclicReplication2() throws Exception { public void testCyclicReplication2() throws Exception {
LOG.info("testCyclicReplication1"); LOG.info("testCyclicReplication2");
int numClusters = 3; int numClusters = 3;
Table[] htables = null; Table[] htables = null;
try { try {
@ -312,7 +309,7 @@ public class TestMasterReplication {
// without HBASE-9158 the edit for row4 would have been marked with // without HBASE-9158 the edit for row4 would have been marked with
// cluster 0's id // cluster 0's id
// and hence not replicated to cluster 0 // and hence not replicated to cluster 0
wait(row4, htables[0], true); wait(row4, htables[0], false);
} finally { } finally {
close(htables); close(htables);
shutDownMiniClusters(); shutDownMiniClusters();
@ -507,7 +504,7 @@ public class TestMasterReplication {
miniZK.shutdown(); miniZK.shutdown();
} }
private void createTableOnClusters(HTableDescriptor table) throws Exception { private void createTableOnClusters(TableDescriptor table) throws Exception {
for (HBaseTestingUtility utility : utilities) { for (HBaseTestingUtility utility : utilities) {
utility.getAdmin().createTable(table); utility.getAdmin().createTable(table);
} }
@ -608,7 +605,7 @@ public class TestMasterReplication {
Path dir = util.getDataTestDirOnTestFS(testName); Path dir = util.getDataTestDirOnTestFS(testName);
FileSystem fs = util.getTestFileSystem(); FileSystem fs = util.getTestFileSystem();
dir = dir.makeQualified(fs); dir = dir.makeQualified(fs.getUri(), fs.getWorkingDirectory());
Path familyDir = new Path(dir, Bytes.toString(fam)); Path familyDir = new Path(dir, Bytes.toString(fam));
int hfileIdx = 0; int hfileIdx = 0;