diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 31d8db3d4a1..c2268e63e58 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -232,6 +232,9 @@ Release 2.0.1-alpha - UNRELEASED HDFS-3524. Update TestFileLengthOnClusterRestart for HDFS-3522. (Brandon Li via szetszwo) + HDFS-3559. DFSTestUtil: use Builder class to construct DFSTestUtil + instances. (Colin Patrick McCabe via atm) + BREAKDOWN OF HDFS-3042 SUBTASKS HDFS-2185. HDFS portion of ZK-based FailoverController (todd) diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java index 1ed534c0065..5b416d6a82e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java @@ -55,6 +55,7 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem.Statistics; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hdfs.DFSTestUtil.Builder; import org.apache.hadoop.hdfs.MiniDFSCluster.NameNodeInfo; import org.apache.hadoop.hdfs.client.HdfsDataInputStream; import org.apache.hadoop.hdfs.protocol.DatanodeID; @@ -95,10 +96,10 @@ public class DFSTestUtil { "zero", "one", "two", "three", "four", "five", "six", "seven", "eight", "nine" }; - private int maxLevels;// = 3; - private int maxSize;// = 8*1024; - private int minSize = 1; - private int nFiles; + private final int maxLevels; + private final int maxSize; + private final int minSize; + private final int nFiles; private MyFile[] files; /** Creates a new instance of DFSTestUtil @@ -108,10 +109,11 @@ public class DFSTestUtil { * @param maxLevels Maximum number of directory levels * @param maxSize Maximum size for file */ - public DFSTestUtil(String testName, int nFiles, int maxLevels, int maxSize) { + private DFSTestUtil(int nFiles, int maxLevels, int maxSize, int minSize) { this.nFiles = nFiles; this.maxLevels = maxLevels; this.maxSize = maxSize; + this.minSize = minSize; } /** @@ -778,4 +780,42 @@ public static DatanodeRegistration getLocalDatanodeRegistration() { return new DatanodeRegistration(getLocalDatanodeID(), new StorageInfo(), new ExportedBlockKeys(), VersionInfo.getVersion()); } + + public static class Builder { + private int maxLevels = 3; + private int maxSize = 8*1024; + private int minSize = 1; + private int nFiles = 1; + + public Builder() { + } + + public Builder setName(String string) { + return this; + } + + public Builder setNumFiles(int nFiles) { + this.nFiles = nFiles; + return this; + } + + public Builder setMaxLevels(int maxLevels) { + this.maxLevels = maxLevels; + return this; + } + + public Builder setMaxSize(int maxSize) { + this.maxSize = maxSize; + return this; + } + + public Builder setMinSize(int minSize) { + this.minSize = minSize; + return this; + } + + public DFSTestUtil build() { + return new DFSTestUtil(nFiles, maxLevels, maxSize, minSize); + } + } } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestCrcCorruption.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestCrcCorruption.java index 482f12b00dd..010d3ff9f6f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestCrcCorruption.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestCrcCorruption.java @@ -206,7 +206,8 @@ public void testCrcCorruption() throws Exception { System.out.println("TestCrcCorruption with default parameters"); Configuration conf1 = new HdfsConfiguration(); conf1.setInt(DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, 3 * 1000); - DFSTestUtil util1 = new DFSTestUtil("TestCrcCorruption", 40, 3, 8*1024); + DFSTestUtil util1 = new DFSTestUtil.Builder().setName("TestCrcCorruption"). + setNumFiles(40).build(); thistest(conf1, util1); // @@ -216,7 +217,8 @@ public void testCrcCorruption() throws Exception { Configuration conf2 = new HdfsConfiguration(); conf2.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, 17); conf2.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 34); - DFSTestUtil util2 = new DFSTestUtil("TestCrcCorruption", 40, 3, 400); + DFSTestUtil util2 = new DFSTestUtil.Builder().setName("TestCrcCorruption"). + setNumFiles(40).setMaxSize(400).build(); thistest(conf2, util2); } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileCorruption.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileCorruption.java index d3df0c0c049..a8624400aa4 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileCorruption.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileCorruption.java @@ -58,7 +58,8 @@ public class TestFileCorruption extends TestCase { /** check if DFS can handle corrupted blocks properly */ public void testFileCorruption() throws Exception { MiniDFSCluster cluster = null; - DFSTestUtil util = new DFSTestUtil("TestFileCorruption", 20, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder().setName("TestFileCorruption"). + setNumFiles(20).build(); try { Configuration conf = new HdfsConfiguration(); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build(); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRestartDFS.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRestartDFS.java index 4dfdb5d1e8f..7e130c98525 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRestartDFS.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRestartDFS.java @@ -31,7 +31,8 @@ public class TestRestartDFS extends TestCase { public void runTests(Configuration conf, boolean serviceTest) throws Exception { MiniDFSCluster cluster = null; - DFSTestUtil files = new DFSTestUtil("TestRestartDFS", 20, 3, 8*1024); + DFSTestUtil files = new DFSTestUtil.Builder().setName("TestRestartDFS"). + setNumFiles(20).build(); final String dir = "/srcdat"; final Path rootpath = new Path("/"); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestDatanodeRestart.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestDatanodeRestart.java index 773864e2f3c..9b4e0d44aed 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestDatanodeRestart.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestDatanodeRestart.java @@ -59,7 +59,8 @@ public class TestDatanodeRestart { try { // test finalized replicas final String TopDir = "/test"; - DFSTestUtil util = new DFSTestUtil("TestCrcCorruption", 2, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("TestDatanodeRestart").setNumFiles(2).build(); util.createFiles(fs, TopDir, (short)3); util.waitReplication(fs, TopDir, (short)3); util.checkFiles(fs, TopDir); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java index 694d84f4d1b..d7bb0f7cbc7 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java @@ -80,7 +80,8 @@ public void setupCluster() throws Exception { final long precision = 1L; conf.setLong(DFSConfigKeys.DFS_NAMENODE_ACCESSTIME_PRECISION_KEY, precision); conf.setLong(DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, 10000L); - util = new DFSTestUtil("TestAuditAllowed", 20, 3, 8*1024); + util = new DFSTestUtil.Builder().setName("TestAuditAllowed"). + setNumFiles(20).build(); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(4).build(); fs = cluster.getFileSystem(); util.createFiles(fs, fileName); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java index 531dc879d8d..e228019c499 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java @@ -101,7 +101,8 @@ static String runFsck(Configuration conf, int expectedErrCode, /** do fsck */ @Test public void testFsck() throws Exception { - DFSTestUtil util = new DFSTestUtil("TestFsck", 20, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder().setName("TestFsck"). + setNumFiles(20).build(); MiniDFSCluster cluster = null; FileSystem fs = null; try { @@ -174,7 +175,8 @@ private void verifyAuditLogs() throws IOException { @Test public void testFsckNonExistent() throws Exception { - DFSTestUtil util = new DFSTestUtil("TestFsck", 20, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder().setName("TestFsck"). + setNumFiles(20).build(); MiniDFSCluster cluster = null; FileSystem fs = null; try { @@ -197,7 +199,8 @@ public void testFsckNonExistent() throws Exception { /** Test fsck with permission set on inodes */ @Test public void testFsckPermission() throws Exception { - final DFSTestUtil util = new DFSTestUtil(getClass().getSimpleName(), 20, 3, 8*1024); + final DFSTestUtil util = new DFSTestUtil.Builder(). + setName(getClass().getSimpleName()).setNumFiles(20).build(); final Configuration conf = new HdfsConfiguration(); conf.setLong(DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, 10000L); @@ -246,7 +249,8 @@ public Object run() throws Exception { @Test public void testFsckMoveAndDelete() throws Exception { final int MAX_MOVE_TRIES = 5; - DFSTestUtil util = new DFSTestUtil("TestFsck", 5, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("TestFsckMoveAndDelete").setNumFiles(5).build(); MiniDFSCluster cluster = null; FileSystem fs = null; try { @@ -319,7 +323,8 @@ public void testFsckMoveAndDelete() throws Exception { @Test public void testFsckOpenFiles() throws Exception { - DFSTestUtil util = new DFSTestUtil("TestFsck", 4, 3, 8*1024); + DFSTestUtil util = new DFSTestUtil.Builder().setName("TestFsck"). + setNumFiles(4).build(); MiniDFSCluster cluster = null; FileSystem fs = null; try { @@ -492,7 +497,9 @@ public void testFsckListCorruptFilesBlocks() throws Exception { cluster = new MiniDFSCluster.Builder(conf).build(); cluster.waitActive(); fs = cluster.getFileSystem(); - DFSTestUtil util = new DFSTestUtil("testGetCorruptFiles", 3, 1, 1024); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("testGetCorruptFiles").setNumFiles(3).setMaxLevels(1). + setMaxSize(1024).build(); util.createFiles(fs, "/corruptData", (short) 1); util.waitReplication(fs, "/corruptData", (short) 1); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestListCorruptFileBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestListCorruptFileBlocks.java index 79b0bf41526..3443fa814f4 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestListCorruptFileBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestListCorruptFileBlocks.java @@ -68,7 +68,9 @@ public void testListCorruptFilesCorruptedBlock() throws Exception { FileSystem fs = cluster.getFileSystem(); // create two files with one block each - DFSTestUtil util = new DFSTestUtil("testCorruptFilesCorruptedBlock", 2, 1, 512); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("testCorruptFilesCorruptedBlock").setNumFiles(2). + setMaxLevels(1).setMaxSize(512).build(); util.createFiles(fs, "/srcdat10"); // fetch bad file list from namenode. There should be none. @@ -152,8 +154,9 @@ public void testListCorruptFileBlocksInSafeMode() throws Exception { FileSystem fs = cluster.getFileSystem(); // create two files with one block each - DFSTestUtil util = new DFSTestUtil("testListCorruptFileBlocksInSafeMode", - 2, 1, 512); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("testListCorruptFileBlocksInSafeMode").setNumFiles(2). + setMaxLevels(1).setMaxSize(512).build(); util.createFiles(fs, "/srcdat10"); // fetch bad file list from namenode. There should be none. @@ -272,7 +275,9 @@ public void testlistCorruptFileBlocks() throws Exception { cluster = new MiniDFSCluster.Builder(conf).build(); cluster.waitActive(); fs = cluster.getFileSystem(); - DFSTestUtil util = new DFSTestUtil("testGetCorruptFiles", 3, 1, 1024); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("testGetCorruptFiles").setNumFiles(3).setMaxLevels(1). + setMaxSize(1024).build(); util.createFiles(fs, "/corruptData"); final NameNode namenode = cluster.getNameNode(); @@ -381,7 +386,9 @@ public void testlistCorruptFileBlocksDFS() throws Exception { cluster.waitActive(); fs = cluster.getFileSystem(); DistributedFileSystem dfs = (DistributedFileSystem) fs; - DFSTestUtil util = new DFSTestUtil("testGetCorruptFiles", 3, 1, 1024); + DFSTestUtil util = new DFSTestUtil.Builder(). + setName("testGetCorruptFiles").setNumFiles(3). + setMaxLevels(1).setMaxSize(1024).build(); util.createFiles(fs, "/corruptData"); RemoteIterator corruptFileBlocks = @@ -451,8 +458,9 @@ public void testMaxCorruptFiles() throws Exception { FSNamesystem.DEFAULT_MAX_CORRUPT_FILEBLOCKS_RETURNED; // create 110 files with one block each - DFSTestUtil util = new DFSTestUtil("testMaxCorruptFiles", - maxCorruptFileBlocks * 3, 1, 512); + DFSTestUtil util = new DFSTestUtil.Builder().setName("testMaxCorruptFiles"). + setNumFiles(maxCorruptFileBlocks * 3).setMaxLevels(1).setMaxSize(512). + build(); util.createFiles(fs, "/srcdat2", (short) 1); util.waitReplication(fs, "/srcdat2", (short) 1); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestParallelImageWrite.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestParallelImageWrite.java index b62dcc1bd4c..4b8409535c1 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestParallelImageWrite.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestParallelImageWrite.java @@ -49,7 +49,8 @@ public void testRestartDFS() throws Exception { MiniDFSCluster cluster = null; FSNamesystem fsn = null; int numNamenodeDirs; - DFSTestUtil files = new DFSTestUtil("TestRestartDFS", 200, 3, 8*1024); + DFSTestUtil files = new DFSTestUtil.Builder().setName("TestRestartDFS"). + setNumFiles(200).build(); final String dir = "/srcdat"; final Path rootpath = new Path("/");