diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 8b5faaba5df..1f26752ca2c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -151,6 +151,9 @@ Trunk (unreleased changes) HDFS-3368. Missing blocks due to bad DataNodes coming up and down. (shv) + HDFS-3462. TestDFSClientRetries.busyTest() should restore default + xceiver count in the config. (Madhukara Phatak via harsh) + Release 2.0.1-alpha - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java index 91f2b97b800..ffe187be1de 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java @@ -526,7 +526,7 @@ public class TestDFSClientRetries extends TestCase { short replicationFactor = 1; long blockSize = 128*1024*1024; // DFS block size int bufferSize = 4096; - + int originalXcievers = conf.getInt(DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY,0); conf.setInt(DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY, xcievers); conf.setInt(DFSConfigKeys.DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_KEY, retries); @@ -605,6 +605,7 @@ public class TestDFSClientRetries extends TestCase { e.printStackTrace(); ret = false; } finally { + conf.setInt(DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY,originalXcievers); fs.delete(file1, false); cluster.shutdown(); }