diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index d857c58e79d..04bb1e15fe3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1657,6 +1657,9 @@ Release 2.8.0 - UNRELEASED HDFS-9252. Change TestFileTruncate to use FsDatasetTestUtils to get block file size and genstamp. (Lei (Eddy) Xu via cmccabe) + HDFS-8056. Decommissioned dead nodes should continue to be counted as dead + after NN restart. (mingma) + OPTIMIZATIONS HDFS-8026. Trace FSOutputSummer#writeChecksumChunks rather than diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java index 3406cf4c06e..d35b23761de 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java @@ -1272,7 +1272,7 @@ public List getDatanodeListForReport( if (listDeadNodes) { for (InetSocketAddress addr : includedNodes) { - if (foundNodes.matchedBy(addr) || excludedNodes.match(addr)) { + if (foundNodes.matchedBy(addr)) { continue; } // The remaining nodes are ones that are referenced by the hosts @@ -1289,6 +1289,9 @@ public List getDatanodeListForReport( addr.getPort() == 0 ? defaultXferPort : addr.getPort(), defaultInfoPort, defaultInfoSecurePort, defaultIpcPort)); setDatanodeDead(dn); + if (excludedNodes.match(addr)) { + dn.setDecommissioned(); + } nodes.add(dn); } } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java index d648bca4630..0b70e24aeb0 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java @@ -923,6 +923,41 @@ public void testDecommissionWithNamenodeRestart()throws IOException, Interrupted cluster.shutdown(); } + /** + * Tests dead node count after restart of namenode + **/ + @Test(timeout=360000) + public void testDeadNodeCountAfterNamenodeRestart()throws Exception { + LOG.info("Starting test testDeadNodeCountAfterNamenodeRestart"); + int numNamenodes = 1; + int numDatanodes = 2; + + startCluster(numNamenodes, numDatanodes, conf); + + DFSClient client = getDfsClient(cluster.getNameNode(), conf); + DatanodeInfo[] info = client.datanodeReport(DatanodeReportType.LIVE); + DatanodeInfo excludedDatanode = info[0]; + String excludedDatanodeName = info[0].getXferAddr(); + + writeConfigFile(hostsFile, new ArrayList(Arrays.asList( + excludedDatanodeName, info[1].getXferAddr()))); + decommissionNode(0, excludedDatanode.getDatanodeUuid(), null, + AdminStates.DECOMMISSIONED); + + cluster.stopDataNode(excludedDatanodeName); + DFSTestUtil.waitForDatanodeState( + cluster, excludedDatanode.getDatanodeUuid(), false, 20000); + + //Restart the namenode + cluster.restartNameNode(); + + assertEquals("There should be one node alive", 1, + client.datanodeReport(DatanodeReportType.LIVE).length); + assertEquals("There should be one node dead", 1, + client.datanodeReport(DatanodeReportType.DEAD).length); + cluster.shutdown(); + } + /** * Test using a "registration name" in a host include file. * diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHostFileManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHostFileManager.java index c65b5804396..6f17040ef68 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHostFileManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHostFileManager.java @@ -151,7 +151,7 @@ public void testIncludeExcludeLists() throws IOException { Assert.assertEquals(1, dm.getDatanodeListForReport(HdfsConstants .DatanodeReportType.DEAD).size()); excludedNodes.add(entry("127.0.0.3")); - Assert.assertEquals(0, dm.getDatanodeListForReport(HdfsConstants + Assert.assertEquals(1, dm.getDatanodeListForReport(HdfsConstants .DatanodeReportType.DEAD).size()); } }