diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestNameNodePrunesMissingStorages.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestNameNodePrunesMissingStorages.java index 948a8fb2386..96d227d1779 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestNameNodePrunesMissingStorages.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestNameNodePrunesMissingStorages.java @@ -383,51 +383,60 @@ public void testNameNodePrunesUnreportedStorages() throws Exception { .Builder(conf).numDataNodes(1) .storagesPerDatanode(2) .build(); - // Create two files to ensure each storage has a block - DFSTestUtil.createFile(cluster.getFileSystem(), new Path("file1"), - 102400, 102400, 102400, (short)1, - 0x1BAD5EE); - DFSTestUtil.createFile(cluster.getFileSystem(), new Path("file2"), - 102400, 102400, 102400, (short)1, - 0x1BAD5EED); - // Get the datanode storages and data directories - DataNode dn = cluster.getDataNodes().get(0); - BlockManager bm = cluster.getNameNode().getNamesystem().getBlockManager(); - DatanodeDescriptor dnDescriptor = bm.getDatanodeManager(). - getDatanode(cluster.getDataNodes().get(0).getDatanodeUuid()); - DatanodeStorageInfo[] dnStoragesInfosBeforeRestart = - dnDescriptor.getStorageInfos(); - Collection oldDirs = new ArrayList(dn.getConf(). - getTrimmedStringCollection(DFSConfigKeys.DFS_DATANODE_DATA_DIR_KEY)); - // Keep the first data directory and remove the second. - String newDirs = oldDirs.iterator().next(); - conf.set(DFSConfigKeys.DFS_DATANODE_DATA_DIR_KEY, newDirs); - // Restart the datanode with the new conf - cluster.stopDataNode(0); - cluster.startDataNodes(conf, 1, false, null, null); - dn = cluster.getDataNodes().get(0); - cluster.waitActive(); - // Assert that the dnDescriptor has both the storages after restart - assertArrayEquals(dnStoragesInfosBeforeRestart, - dnDescriptor.getStorageInfos()); - // Assert that the removed storage is marked as FAILED - // when DN heartbeats to the NN - int numFailedStoragesWithBlocks = 0; - DatanodeStorageInfo failedStorageInfo = null; - for (DatanodeStorageInfo dnStorageInfo: dnDescriptor.getStorageInfos()) { - if (dnStorageInfo.areBlocksOnFailedStorage()) { - numFailedStoragesWithBlocks++; - failedStorageInfo = dnStorageInfo; + try { + cluster.waitActive(); + // Create two files to ensure each storage has a block + DFSTestUtil.createFile(cluster.getFileSystem(), new Path("file1"), + 102400, 102400, 102400, (short)1, + 0x1BAD5EE); + DFSTestUtil.createFile(cluster.getFileSystem(), new Path("file2"), + 102400, 102400, 102400, (short)1, + 0x1BAD5EED); + // Get the datanode storages and data directories + DataNode dn = cluster.getDataNodes().get(0); + BlockManager bm = + cluster.getNameNode().getNamesystem().getBlockManager(); + DatanodeDescriptor dnDescriptor = bm.getDatanodeManager(). + getDatanode(cluster.getDataNodes().get(0).getDatanodeUuid()); + DatanodeStorageInfo[] dnStoragesInfosBeforeRestart = + dnDescriptor.getStorageInfos(); + Collection oldDirs = new ArrayList(dn.getConf(). + getTrimmedStringCollection(DFSConfigKeys.DFS_DATANODE_DATA_DIR_KEY)); + // Keep the first data directory and remove the second. + String newDirs = oldDirs.iterator().next(); + conf.set(DFSConfigKeys.DFS_DATANODE_DATA_DIR_KEY, newDirs); + // Restart the datanode with the new conf + cluster.stopDataNode(0); + cluster.startDataNodes(conf, 1, false, null, null); + dn = cluster.getDataNodes().get(0); + cluster.waitActive(); + // Assert that the dnDescriptor has both the storages after restart + assertArrayEquals(dnStoragesInfosBeforeRestart, + dnDescriptor.getStorageInfos()); + // Assert that the removed storage is marked as FAILED + // when DN heartbeats to the NN + int numFailedStoragesWithBlocks = 0; + DatanodeStorageInfo failedStorageInfo = null; + for (DatanodeStorageInfo dnStorageInfo: dnDescriptor.getStorageInfos()) { + if (dnStorageInfo.areBlocksOnFailedStorage()) { + numFailedStoragesWithBlocks++; + failedStorageInfo = dnStorageInfo; + } + } + assertEquals(1, numFailedStoragesWithBlocks); + // Heartbeat manager removes the blocks associated with this failed + // storage + bm.getDatanodeManager().getHeartbeatManager().heartbeatCheck(); + assertTrue(!failedStorageInfo.areBlocksOnFailedStorage()); + // pruneStorageMap removes the unreported storage + cluster.triggerHeartbeats(); + // Assert that the unreported storage is pruned + assertEquals(DataNode.getStorageLocations(dn.getConf()).size(), + dnDescriptor.getStorageInfos().length); + } finally { + if (cluster != null) { + cluster.shutdown(); } } - assertEquals(1, numFailedStoragesWithBlocks); - // Heartbeat manager removes the blocks associated with this failed storage - bm.getDatanodeManager().getHeartbeatManager().heartbeatCheck(); - assertTrue(!failedStorageInfo.areBlocksOnFailedStorage()); - // pruneStorageMap removes the unreported storage - cluster.triggerHeartbeats(); - // Assert that the unreported storage is pruned - assertEquals(DataNode.getStorageLocations(dn.getConf()).size(), - dnDescriptor.getStorageInfos().length); } }