MAPREDUCE-6520. Migrate MR Client test cases part 1.

This commit is contained in:
Akira Ajisaka 2016-03-11 22:20:55 +09:00
parent 682adc6ba9
commit ef68b44195
25 changed files with 284 additions and 67 deletions

View File

@ -17,17 +17,30 @@
*/ */
package org.apache.hadoop.conf; package org.apache.hadoop.conf;
import org.junit.Assert;
import org.apache.hadoop.mapred.*;
import org.apache.hadoop.mapreduce.MRConfig;
import org.apache.hadoop.mapreduce.server.jobtracker.JTConfig;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.FileInputFormat;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;
import org.apache.hadoop.mapred.Utils;
import org.junit.Test;
import java.io.*; import java.io.BufferedReader;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.io.OutputStream;
import java.io.OutputStreamWriter;
import java.io.Writer;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
/** /**
* This testcase tests that a JobConf without default values submits jobs * This testcase tests that a JobConf without default values submits jobs
@ -40,6 +53,7 @@ public class TestNoDefaultsJobConf extends HadoopTestCase {
super(HadoopTestCase.CLUSTER_MR, HadoopTestCase.DFS_FS, 1, 1); super(HadoopTestCase.CLUSTER_MR, HadoopTestCase.DFS_FS, 1, 1);
} }
@Test
public void testNoDefaults() throws Exception { public void testNoDefaults() throws Exception {
JobConf configuration = new JobConf(); JobConf configuration = new JobConf();
assertTrue(configuration.get("hadoop.tmp.dir", null) != null); assertTrue(configuration.get("hadoop.tmp.dir", null) != null);

View File

@ -18,13 +18,12 @@
package org.apache.hadoop.mapred; package org.apache.hadoop.mapred;
import junit.framework.TestCase;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.mapreduce.MRConfig; import org.apache.hadoop.mapreduce.MRConfig;
import org.junit.After;
import org.junit.Before;
import java.io.File;
import java.io.IOException; import java.io.IOException;
/** /**
@ -38,7 +37,7 @@ import java.io.IOException;
* Job Configurations should be created using a configuration returned by the * Job Configurations should be created using a configuration returned by the
* 'createJobConf()' method. * 'createJobConf()' method.
*/ */
public abstract class HadoopTestCase extends TestCase { public abstract class HadoopTestCase {
public static final int LOCAL_MR = 1; public static final int LOCAL_MR = 1;
public static final int CLUSTER_MR = 2; public static final int CLUSTER_MR = 2;
public static final int LOCAL_FS = 4; public static final int LOCAL_FS = 4;
@ -140,8 +139,8 @@ public abstract class HadoopTestCase extends TestCase {
* *
* @throws Exception * @throws Exception
*/ */
protected void setUp() throws Exception { @Before
super.setUp(); public void setUp() throws Exception {
if (localFS) { if (localFS) {
fileSystem = FileSystem.getLocal(new JobConf()); fileSystem = FileSystem.getLocal(new JobConf());
} }
@ -164,7 +163,8 @@ public abstract class HadoopTestCase extends TestCase {
* *
* @throws Exception * @throws Exception
*/ */
protected void tearDown() throws Exception { @After
public void tearDown() throws Exception {
try { try {
if (mrCluster != null) { if (mrCluster != null) {
mrCluster.shutdown(); mrCluster.shutdown();
@ -181,7 +181,6 @@ public abstract class HadoopTestCase extends TestCase {
catch (Exception ex) { catch (Exception ex) {
System.out.println(ex); System.out.println(ex);
} }
super.tearDown();
} }
/** /**

View File

@ -34,6 +34,13 @@ import javax.servlet.ServletException;
import java.io.IOException; import java.io.IOException;
import java.io.DataOutputStream; import java.io.DataOutputStream;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertEquals;
import org.junit.Before;
import org.junit.After;
import org.junit.Test;
/** /**
* Base class to test Job end notification in local and cluster mode. * Base class to test Job end notification in local and cluster mode.
* *
@ -140,17 +147,19 @@ public abstract class NotificationTestCase extends HadoopTestCase {
return conf; return conf;
} }
@Before
protected void setUp() throws Exception { public void setUp() throws Exception {
super.setUp(); super.setUp();
startHttpServer(); startHttpServer();
} }
protected void tearDown() throws Exception { @After
public void tearDown() throws Exception {
stopHttpServer(); stopHttpServer();
super.tearDown(); super.tearDown();
} }
@Test
public void testMR() throws Exception { public void testMR() throws Exception {
System.out.println(launchWordCount(this.createJobConf(), System.out.println(launchWordCount(this.createJobConf(),

View File

@ -30,12 +30,16 @@ import java.io.IOException;
import java.io.OutputStream; import java.io.OutputStream;
import java.util.Iterator; import java.util.Iterator;
import org.junit.Test;
import static org.junit.Assert.assertTrue;
public class TestFileOutputFormat extends HadoopTestCase { public class TestFileOutputFormat extends HadoopTestCase {
public TestFileOutputFormat() throws IOException { public TestFileOutputFormat() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testCustomFile() throws Exception { public void testCustomFile() throws Exception {
Path inDir = new Path("testing/fileoutputformat/input"); Path inDir = new Path("testing/fileoutputformat/input");
Path outDir = new Path("testing/fileoutputformat/output"); Path outDir = new Path("testing/fileoutputformat/output");

View File

@ -17,9 +17,6 @@
*/ */
package org.apache.hadoop.mapred; package org.apache.hadoop.mapred;
import java.io.File;
import java.io.IOException;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -27,9 +24,18 @@ import org.apache.hadoop.io.Text;
import org.apache.hadoop.ipc.ProtocolSignature; import org.apache.hadoop.ipc.ProtocolSignature;
import org.apache.hadoop.mapred.SortedRanges.Range; import org.apache.hadoop.mapred.SortedRanges.Range;
import org.apache.hadoop.mapreduce.TaskType; import org.apache.hadoop.mapreduce.TaskType;
import org.apache.hadoop.mapreduce.checkpoint.CheckpointID;
import org.apache.hadoop.mapreduce.checkpoint.FSCheckpointID;
import org.apache.hadoop.mapreduce.checkpoint.TaskCheckpointID; import org.apache.hadoop.mapreduce.checkpoint.TaskCheckpointID;
import org.junit.After;
import org.junit.Test;
import java.io.File;
import java.io.IOException;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
public class TestTaskCommit extends HadoopTestCase { public class TestTaskCommit extends HadoopTestCase {
@ -80,12 +86,13 @@ public class TestTaskCommit extends HadoopTestCase {
super(LOCAL_MR, LOCAL_FS, 1, 1); super(LOCAL_MR, LOCAL_FS, 1, 1);
} }
@Override @After
public void tearDown() throws Exception { public void tearDown() throws Exception {
super.tearDown(); super.tearDown();
FileUtil.fullyDelete(new File(rootDir.toString())); FileUtil.fullyDelete(new File(rootDir.toString()));
} }
@Test
public void testCommitFail() throws IOException { public void testCommitFail() throws IOException {
final Path inDir = new Path(rootDir, "./input"); final Path inDir = new Path(rootDir, "./input");
final Path outDir = new Path(rootDir, "./output"); final Path outDir = new Path(rootDir, "./output");
@ -199,6 +206,7 @@ public class TestTaskCommit extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testTaskCleanupDoesNotCommit() throws Exception { public void testTaskCleanupDoesNotCommit() throws Exception {
// Mimic a job with a special committer that does not cleanup // Mimic a job with a special committer that does not cleanup
// files when a task fails. // files when a task fails.
@ -245,16 +253,19 @@ public class TestTaskCommit extends HadoopTestCase {
assertTrue("Task did not succeed", umbilical.taskDone); assertTrue("Task did not succeed", umbilical.taskDone);
} }
@Test
public void testCommitRequiredForMapTask() throws Exception { public void testCommitRequiredForMapTask() throws Exception {
Task testTask = createDummyTask(TaskType.MAP); Task testTask = createDummyTask(TaskType.MAP);
assertTrue("MapTask should need commit", testTask.isCommitRequired()); assertTrue("MapTask should need commit", testTask.isCommitRequired());
} }
@Test
public void testCommitRequiredForReduceTask() throws Exception { public void testCommitRequiredForReduceTask() throws Exception {
Task testTask = createDummyTask(TaskType.REDUCE); Task testTask = createDummyTask(TaskType.REDUCE);
assertTrue("ReduceTask should need commit", testTask.isCommitRequired()); assertTrue("ReduceTask should need commit", testTask.isCommitRequired());
} }
@Test
public void testCommitNotRequiredForJobSetup() throws Exception { public void testCommitNotRequiredForJobSetup() throws Exception {
Task testTask = createDummyTask(TaskType.MAP); Task testTask = createDummyTask(TaskType.MAP);
testTask.setJobSetupTask(); testTask.setJobSetupTask();
@ -262,6 +273,7 @@ public class TestTaskCommit extends HadoopTestCase {
testTask.isCommitRequired()); testTask.isCommitRequired());
} }
@Test
public void testCommitNotRequiredForJobCleanup() throws Exception { public void testCommitNotRequiredForJobCleanup() throws Exception {
Task testTask = createDummyTask(TaskType.MAP); Task testTask = createDummyTask(TaskType.MAP);
testTask.setJobCleanupTask(); testTask.setJobCleanupTask();
@ -269,6 +281,7 @@ public class TestTaskCommit extends HadoopTestCase {
testTask.isCommitRequired()); testTask.isCommitRequired());
} }
@Test
public void testCommitNotRequiredForTaskCleanup() throws Exception { public void testCommitNotRequiredForTaskCleanup() throws Exception {
Task testTask = createDummyTask(TaskType.REDUCE); Task testTask = createDummyTask(TaskType.REDUCE);
testTask.setTaskCleanupTask(); testTask.setTaskCleanupTask();

View File

@ -27,6 +27,8 @@ import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.JobConf;
import org.junit.Test;
import static org.junit.Assert.assertEquals;
/** /**
* HadoopTestCase that tests the local job runner. * HadoopTestCase that tests the local job runner.
@ -59,6 +61,7 @@ public class TestLocalJobControl extends HadoopTestCase {
* object. Finally, it creates a thread to run the JobControl object and * object. Finally, it creates a thread to run the JobControl object and
* monitors/reports the job states. * monitors/reports the job states.
*/ */
@Test
public void testLocalJobControlDataCopy() throws Exception { public void testLocalJobControlDataCopy() throws Exception {
FileSystem fs = FileSystem.get(createJobConf()); FileSystem fs = FileSystem.get(createJobConf());

View File

@ -21,12 +21,29 @@ import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.*; import org.apache.hadoop.mapred.FileInputFormat;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.RunningJob;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;
import org.junit.Test;
import java.io.DataOutputStream; import java.io.DataOutputStream;
import java.io.IOException; import java.io.IOException;
import java.util.Iterator; import java.util.Iterator;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotSame;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
public class TestChainMapReduce extends HadoopTestCase { public class TestChainMapReduce extends HadoopTestCase {
private static Path getFlagDir(boolean local) { private static Path getFlagDir(boolean local) {
@ -67,6 +84,7 @@ public class TestChainMapReduce extends HadoopTestCase {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testChain() throws Exception { public void testChain() throws Exception {
Path inDir = new Path("testing/chain/input"); Path inDir = new Path("testing/chain/input");
Path outDir = new Path("testing/chain/output"); Path outDir = new Path("testing/chain/output");

View File

@ -36,6 +36,8 @@ import org.apache.hadoop.mapred.TextOutputFormat;
import org.apache.hadoop.mapred.Utils; import org.apache.hadoop.mapred.Utils;
import org.junit.After; import org.junit.After;
import org.junit.Test; import org.junit.Test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
import java.io.BufferedReader; import java.io.BufferedReader;
import java.io.File; import java.io.File;

View File

@ -17,11 +17,6 @@
*/ */
package org.apache.hadoop.mapred.lib; package org.apache.hadoop.mapred.lib;
import java.io.IOException;
import java.util.Map;
import junit.framework.TestCase;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapred.InputFormat; import org.apache.hadoop.mapred.InputFormat;
import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.JobConf;
@ -30,12 +25,19 @@ import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector; import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reporter; import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.TextInputFormat; import org.apache.hadoop.mapred.TextInputFormat;
import org.junit.Test;
import java.io.IOException;
import java.util.Map;
import static org.junit.Assert.assertEquals;
/** /**
* @see TestDelegatingInputFormat * @see TestDelegatingInputFormat
*/ */
public class TestMultipleInputs extends TestCase { public class TestMultipleInputs {
@Test
public void testAddInputPathWithFormat() { public void testAddInputPathWithFormat() {
final JobConf conf = new JobConf(); final JobConf conf = new JobConf();
MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class); MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class);
@ -48,6 +50,7 @@ public class TestMultipleInputs extends TestCase {
.getClass()); .getClass());
} }
@Test
public void testAddInputPathWithMapper() { public void testAddInputPathWithMapper() {
final JobConf conf = new JobConf(); final JobConf conf = new JobConf();
MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class, MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class,

View File

@ -24,7 +24,23 @@ import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.SequenceFile; import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.io.serializer.JavaSerializationComparator; import org.apache.hadoop.io.serializer.JavaSerializationComparator;
import org.apache.hadoop.mapred.*; import org.apache.hadoop.mapred.Counters;
import org.apache.hadoop.mapred.FileInputFormat;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.RunningJob;
import org.apache.hadoop.mapred.SequenceFileOutputFormat;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import java.io.BufferedReader; import java.io.BufferedReader;
import java.io.DataOutputStream; import java.io.DataOutputStream;
@ -32,17 +48,23 @@ import java.io.IOException;
import java.io.InputStreamReader; import java.io.InputStreamReader;
import java.util.Iterator; import java.util.Iterator;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
public class TestMultipleOutputs extends HadoopTestCase { public class TestMultipleOutputs extends HadoopTestCase {
public TestMultipleOutputs() throws IOException { public TestMultipleOutputs() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testWithoutCounters() throws Exception { public void testWithoutCounters() throws Exception {
_testMultipleOutputs(false); _testMultipleOutputs(false);
_testMOWithJavaSerialization(false); _testMOWithJavaSerialization(false);
} }
@Test
public void testWithCounters() throws Exception { public void testWithCounters() throws Exception {
_testMultipleOutputs(true); _testMultipleOutputs(true);
_testMOWithJavaSerialization(true); _testMOWithJavaSerialization(true);
@ -62,6 +84,7 @@ public class TestMultipleOutputs extends HadoopTestCase {
return dir; return dir;
} }
@Before
public void setUp() throws Exception { public void setUp() throws Exception {
super.setUp(); super.setUp();
Path rootDir = getDir(ROOT_DIR); Path rootDir = getDir(ROOT_DIR);
@ -75,6 +98,7 @@ public class TestMultipleOutputs extends HadoopTestCase {
} }
} }
@After
public void tearDown() throws Exception { public void tearDown() throws Exception {
Path rootDir = getDir(ROOT_DIR); Path rootDir = getDir(ROOT_DIR);

View File

@ -22,26 +22,45 @@ import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.*; import org.apache.hadoop.mapred.FileInputFormat;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.RunningJob;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;
import org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper; import org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper;
import org.junit.Test;
import java.io.DataOutputStream; import java.io.DataOutputStream;
import java.io.IOException; import java.io.IOException;
import java.util.Iterator; import java.util.Iterator;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
public class TestMultithreadedMapRunner extends HadoopTestCase { public class TestMultithreadedMapRunner extends HadoopTestCase {
public TestMultithreadedMapRunner() throws IOException { public TestMultithreadedMapRunner() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testOKRun() throws Exception { public void testOKRun() throws Exception {
run(false, false); run(false, false);
} }
@Test
public void testIOExRun() throws Exception { public void testIOExRun() throws Exception {
run(true, false); run(true, false);
} }
@Test
public void testRuntimeExRun() throws Exception { public void testRuntimeExRun() throws Exception {
run(false, true); run(false, true);
} }

View File

@ -30,6 +30,13 @@ import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.JobConf;
import org.apache.log4j.Level; import org.apache.log4j.Level;
import org.junit.Before;
import org.junit.Test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
public class TestChild extends HadoopTestCase { public class TestChild extends HadoopTestCase {
private static String TEST_ROOT_DIR = private static String TEST_ROOT_DIR =
@ -146,6 +153,7 @@ public class TestChild extends HadoopTestCase {
return job; return job;
} }
@Test
public void testChild() throws Exception { public void testChild() throws Exception {
try { try {
submitAndValidateJob(createJobConf(), 1, 1, true); submitAndValidateJob(createJobConf(), 1, 1, true);

View File

@ -30,6 +30,9 @@ import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl; import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
import org.junit.Ignore; import org.junit.Ignore;
import org.junit.Test;
import static org.junit.Assert.assertTrue;
@Ignore @Ignore
public class TestNoJobSetupCleanup extends HadoopTestCase { public class TestNoJobSetupCleanup extends HadoopTestCase {
private static String TEST_ROOT_DIR = private static String TEST_ROOT_DIR =
@ -69,6 +72,7 @@ public class TestNoJobSetupCleanup extends HadoopTestCase {
return job; return job;
} }
@Test
public void testNoJobSetupCleanup() throws Exception { public void testNoJobSetupCleanup() throws Exception {
try { try {
Configuration conf = createJobConf(); Configuration conf = createJobConf();

View File

@ -33,6 +33,8 @@ import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Ignore; import org.junit.Ignore;
import org.junit.Test; import org.junit.Test;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
/** /**
* Tests context api and {@link StatusReporter#getProgress()} via * Tests context api and {@link StatusReporter#getProgress()} via

View File

@ -29,6 +29,10 @@ import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil; import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.Mapper; import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer; import org.apache.hadoop.mapreduce.Reducer;
import org.junit.Test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertEquals;
/** /**
* Tests error conditions in ChainMapper/ChainReducer. * Tests error conditions in ChainMapper/ChainReducer.
@ -51,6 +55,7 @@ public class TestChainErrors extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testChainSubmission() throws Exception { public void testChainSubmission() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
@ -89,6 +94,7 @@ public class TestChainErrors extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testChainFail() throws Exception { public void testChainFail() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
@ -114,6 +120,7 @@ public class TestChainErrors extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testReducerFail() throws Exception { public void testReducerFail() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
@ -139,6 +146,7 @@ public class TestChainErrors extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testChainMapNoOuptut() throws Exception { public void testChainMapNoOuptut() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
String expectedOutput = ""; String expectedOutput = "";
@ -163,6 +171,7 @@ public class TestChainErrors extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testChainReduceNoOuptut() throws Exception { public void testChainReduceNoOuptut() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
String expectedOutput = ""; String expectedOutput = "";

View File

@ -30,6 +30,11 @@ import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil; import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.Mapper; import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer; import org.apache.hadoop.mapreduce.Reducer;
import org.junit.Test;
import static org.junit.Assert.fail;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
public class TestMapReduceChain extends HadoopTestCase { public class TestMapReduceChain extends HadoopTestCase {
@ -63,6 +68,7 @@ public class TestMapReduceChain extends HadoopTestCase {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testChain() throws Exception { public void testChain() throws Exception {
Path inDir = new Path(localPathRoot, "testing/chain/input"); Path inDir = new Path(localPathRoot, "testing/chain/input");
Path outDir = new Path(localPathRoot, "testing/chain/output"); Path outDir = new Path(localPathRoot, "testing/chain/output");

View File

@ -26,6 +26,9 @@ import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil; import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper; import org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper;
import org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer; import org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer;
import org.junit.Test;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
import java.io.IOException; import java.io.IOException;
@ -42,6 +45,7 @@ public class TestSingleElementChain extends HadoopTestCase {
} }
// test chain mapper and reducer by adding single mapper and reducer to chain // test chain mapper and reducer by adding single mapper and reducer to chain
@Test
public void testNoChain() throws Exception { public void testNoChain() throws Exception {
Path inDir = new Path(localPathRoot, "testing/chain/input"); Path inDir = new Path(localPathRoot, "testing/chain/input");
Path outDir = new Path(localPathRoot, "testing/chain/output"); Path outDir = new Path(localPathRoot, "testing/chain/output");

View File

@ -18,25 +18,40 @@
package org.apache.hadoop.mapreduce.lib.db; package org.apache.hadoop.mapreduce.lib.db;
import java.sql.*;
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
//import org.apache.hadoop.examples.DBCountPageView;
import org.apache.hadoop.fs.*;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.WritableComparable; import org.apache.hadoop.io.WritableComparable;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.*; import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.db.*; import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.lib.input.*; import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.output.*; import org.apache.hadoop.mapreduce.TaskCounter;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.StringUtils;
import org.hsqldb.server.Server; import org.hsqldb.server.Server;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
import java.sql.Connection;
import java.sql.Date;
import java.sql.DriverManager;
import java.sql.PreparedStatement;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.sql.Statement;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
//import org.apache.hadoop.examples.DBCountPageView;
/** /**
* Test aspects of DataDrivenDBInputFormat * Test aspects of DataDrivenDBInputFormat
@ -109,11 +124,13 @@ public class TestDataDrivenDBInputFormat extends HadoopTestCase {
createConnection(driverClassName, url); createConnection(driverClassName, url);
} }
@Before
public void setUp() throws Exception { public void setUp() throws Exception {
initialize(DRIVER_CLASS, DB_URL); initialize(DRIVER_CLASS, DB_URL);
super.setUp(); super.setUp();
} }
@After
public void tearDown() throws Exception { public void tearDown() throws Exception {
super.tearDown(); super.tearDown();
shutdown(); shutdown();
@ -170,6 +187,7 @@ public class TestDataDrivenDBInputFormat extends HadoopTestCase {
} }
} }
@Test
public void testDateSplits() throws Exception { public void testDateSplits() throws Exception {
Statement s = connection.createStatement(); Statement s = connection.createStatement();
final String DATE_TABLE = "datetable"; final String DATE_TABLE = "datetable";

View File

@ -38,6 +38,8 @@ import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer; import org.apache.hadoop.mapreduce.Reducer;
import org.junit.Before; import org.junit.Before;
import org.junit.Test; import org.junit.Test;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
/** /**
* @see TestDelegatingInputFormat * @see TestDelegatingInputFormat
@ -139,7 +141,7 @@ public class TestMultipleInputs extends HadoopTestCase {
assertTrue(output.readLine().equals("e 2")); assertTrue(output.readLine().equals("e 2"));
} }
@SuppressWarnings("unchecked") @Test
public void testAddInputPathWithFormat() throws IOException { public void testAddInputPathWithFormat() throws IOException {
final Job conf = Job.getInstance(); final Job conf = Job.getInstance();
MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class); MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class);
@ -152,7 +154,7 @@ public class TestMultipleInputs extends HadoopTestCase {
.getClass()); .getClass());
} }
@SuppressWarnings("unchecked") @Test
public void testAddInputPathWithMapper() throws IOException { public void testAddInputPathWithMapper() throws IOException {
final Job conf = Job.getInstance(); final Job conf = Job.getInstance();
MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class, MultipleInputs.addInputPath(conf, new Path("/foo"), TextInputFormat.class,

View File

@ -33,6 +33,9 @@ import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil; import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.junit.Test; import org.junit.Test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
/** /**
* This class performs unit test for Job/JobControl classes. * This class performs unit test for Job/JobControl classes.
@ -121,6 +124,7 @@ public class TestMapReduceJobControl extends HadoopTestCase {
} }
} }
@Test
public void testJobControlWithFailJob() throws Exception { public void testJobControlWithFailJob() throws Exception {
LOG.info("Starting testJobControlWithFailJob"); LOG.info("Starting testJobControlWithFailJob");
Configuration conf = createJobConf(); Configuration conf = createJobConf();
@ -144,6 +148,7 @@ public class TestMapReduceJobControl extends HadoopTestCase {
theControl.stop(); theControl.stop();
} }
@Test
public void testJobControlWithKillJob() throws Exception { public void testJobControlWithKillJob() throws Exception {
LOG.info("Starting testJobControlWithKillJob"); LOG.info("Starting testJobControlWithKillJob");
@ -182,6 +187,7 @@ public class TestMapReduceJobControl extends HadoopTestCase {
theControl.stop(); theControl.stop();
} }
@Test
public void testJobControl() throws Exception { public void testJobControl() throws Exception {
LOG.info("Starting testJobControl"); LOG.info("Starting testJobControl");

View File

@ -23,23 +23,33 @@ import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.*; import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.junit.Test;
import java.io.IOException; import java.io.IOException;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
public class TestMultithreadedMapper extends HadoopTestCase { public class TestMultithreadedMapper extends HadoopTestCase {
public TestMultithreadedMapper() throws IOException { public TestMultithreadedMapper() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testOKRun() throws Exception { public void testOKRun() throws Exception {
run(false, false); run(false, false);
} }
@Test
public void testIOExRun() throws Exception { public void testIOExRun() throws Exception {
run(true, false); run(true, false);
} }
@Test
public void testRuntimeExRun() throws Exception { public void testRuntimeExRun() throws Exception {
run(false, true); run(false, true);
} }

View File

@ -33,6 +33,11 @@ import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.OutputCommitter; import org.apache.hadoop.mapreduce.OutputCommitter;
import org.apache.hadoop.mapreduce.OutputFormat; import org.apache.hadoop.mapreduce.OutputFormat;
import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hadoop.mapreduce.TaskAttemptContext;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertFalse;
/** /**
* A JUnit test to test Map-Reduce job committer. * A JUnit test to test Map-Reduce job committer.
@ -54,15 +59,15 @@ public class TestJobOutputCommitter extends HadoopTestCase {
private FileSystem fs; private FileSystem fs;
private Configuration conf = null; private Configuration conf = null;
@Override @Before
protected void setUp() throws Exception { public void setUp() throws Exception {
super.setUp(); super.setUp();
conf = createJobConf(); conf = createJobConf();
fs = getFileSystem(); fs = getFileSystem();
} }
@Override @After
protected void tearDown() throws Exception { public void tearDown() throws Exception {
fs.delete(new Path(TEST_ROOT_DIR), true); fs.delete(new Path(TEST_ROOT_DIR), true);
super.tearDown(); super.tearDown();
} }
@ -219,6 +224,7 @@ public class TestJobOutputCommitter extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testDefaultCleanupAndAbort() throws Exception { public void testDefaultCleanupAndAbort() throws Exception {
// check with a successful job // check with a successful job
testSuccessfulJob(FileOutputCommitter.SUCCEEDED_FILE_NAME, testSuccessfulJob(FileOutputCommitter.SUCCEEDED_FILE_NAME,
@ -238,6 +244,7 @@ public class TestJobOutputCommitter extends HadoopTestCase {
* *
* @throws Exception * @throws Exception
*/ */
@Test
public void testCustomAbort() throws Exception { public void testCustomAbort() throws Exception {
// check with a successful job // check with a successful job
testSuccessfulJob(FileOutputCommitter.SUCCEEDED_FILE_NAME, testSuccessfulJob(FileOutputCommitter.SUCCEEDED_FILE_NAME,
@ -264,6 +271,7 @@ public class TestJobOutputCommitter extends HadoopTestCase {
* compatibility testing. * compatibility testing.
* @throws Exception * @throws Exception
*/ */
@Test
public void testCustomCleanup() throws Exception { public void testCustomCleanup() throws Exception {
// check with a successful job // check with a successful job
testSuccessfulJob(CUSTOM_CLEANUP_FILE_NAME, testSuccessfulJob(CUSTOM_CLEANUP_FILE_NAME,

View File

@ -27,23 +27,36 @@ import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
import org.apache.hadoop.io.serializer.JavaSerializationComparator; import org.apache.hadoop.io.serializer.JavaSerializationComparator;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.mapreduce.*; import org.apache.hadoop.mapreduce.CounterGroup;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import java.io.BufferedReader; import java.io.BufferedReader;
import java.io.IOException; import java.io.IOException;
import java.io.InputStreamReader; import java.io.InputStreamReader;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
public class TestMRMultipleOutputs extends HadoopTestCase { public class TestMRMultipleOutputs extends HadoopTestCase {
public TestMRMultipleOutputs() throws IOException { public TestMRMultipleOutputs() throws IOException {
super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1); super(HadoopTestCase.LOCAL_MR, HadoopTestCase.LOCAL_FS, 1, 1);
} }
@Test
public void testWithoutCounters() throws Exception { public void testWithoutCounters() throws Exception {
_testMultipleOutputs(false); _testMultipleOutputs(false);
_testMOWithJavaSerialization(false); _testMOWithJavaSerialization(false);
} }
@Test
public void testWithCounters() throws Exception { public void testWithCounters() throws Exception {
_testMultipleOutputs(true); _testMultipleOutputs(true);
_testMOWithJavaSerialization(true); _testMOWithJavaSerialization(true);
@ -57,6 +70,7 @@ public class TestMRMultipleOutputs extends HadoopTestCase {
private static String TEXT = "text"; private static String TEXT = "text";
private static String SEQUENCE = "sequence"; private static String SEQUENCE = "sequence";
@Before
public void setUp() throws Exception { public void setUp() throws Exception {
super.setUp(); super.setUp();
Configuration conf = createJobConf(); Configuration conf = createJobConf();
@ -64,6 +78,7 @@ public class TestMRMultipleOutputs extends HadoopTestCase {
fs.delete(ROOT_DIR, true); fs.delete(ROOT_DIR, true);
} }
@After
public void tearDown() throws Exception { public void tearDown() throws Exception {
Configuration conf = createJobConf(); Configuration conf = createJobConf();
FileSystem fs = FileSystem.get(conf); FileSystem fs = FileSystem.get(conf);

View File

@ -18,8 +18,6 @@
package org.apache.hadoop.mapreduce.lib.partition; package org.apache.hadoop.mapreduce.lib.partition;
import java.io.*;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -32,6 +30,15 @@ import org.apache.hadoop.mapreduce.MRJobConfig;
import org.apache.hadoop.mapreduce.MapReduceTestUtil; import org.apache.hadoop.mapreduce.MapReduceTestUtil;
import org.apache.hadoop.mapreduce.Reducer; import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.map.InverseMapper; import org.apache.hadoop.mapreduce.lib.map.InverseMapper;
import org.junit.Test;
import java.io.BufferedReader;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
public class TestMRKeyFieldBasedComparator extends HadoopTestCase { public class TestMRKeyFieldBasedComparator extends HadoopTestCase {
@ -94,6 +101,7 @@ public class TestMRKeyFieldBasedComparator extends HadoopTestCase {
} }
} }
@Test
public void testBasicUnixComparator() throws Exception { public void testBasicUnixComparator() throws Exception {
testComparator("-k1,1n", 1); testComparator("-k1,1n", 1);
testComparator("-k2,2n", 1); testComparator("-k2,2n", 1);
@ -117,7 +125,7 @@ public class TestMRKeyFieldBasedComparator extends HadoopTestCase {
byte[] line1_bytes = line1.getBytes(); byte[] line1_bytes = line1.getBytes();
byte[] line2_bytes = line2.getBytes(); byte[] line2_bytes = line2.getBytes();
public void testWithoutMRJob(String keySpec, int expect) throws Exception { private void testWithoutMRJob(String keySpec, int expect) throws Exception {
KeyFieldBasedComparator<Void, Void> keyFieldCmp = KeyFieldBasedComparator<Void, Void> keyFieldCmp =
new KeyFieldBasedComparator<Void, Void>(); new KeyFieldBasedComparator<Void, Void>();
conf.set("mapreduce.partition.keycomparator.options", keySpec); conf.set("mapreduce.partition.keycomparator.options", keySpec);

View File

@ -27,6 +27,12 @@ import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapred.FileAlreadyExistsException; import org.apache.hadoop.mapred.FileAlreadyExistsException;
import org.apache.hadoop.mapred.HadoopTestCase; import org.apache.hadoop.mapred.HadoopTestCase;
import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.util.ToolRunner;
import org.junit.After;
import org.junit.Test;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.fail;
public class TestTeraSort extends HadoopTestCase { public class TestTeraSort extends HadoopTestCase {
private static Log LOG = LogFactory.getLog(TestTeraSort.class); private static Log LOG = LogFactory.getLog(TestTeraSort.class);
@ -35,7 +41,8 @@ public class TestTeraSort extends HadoopTestCase {
super(LOCAL_MR, LOCAL_FS, 1, 1); super(LOCAL_MR, LOCAL_FS, 1, 1);
} }
protected void tearDown() throws Exception { @After
public void tearDown() throws Exception {
getFileSystem().delete(new Path(TEST_DIR), true); getFileSystem().delete(new Path(TEST_DIR), true);
super.tearDown(); super.tearDown();
} }
@ -76,6 +83,7 @@ public class TestTeraSort extends HadoopTestCase {
assertEquals(ToolRunner.run(job, new TeraValidate(), svArgs), 0); assertEquals(ToolRunner.run(job, new TeraValidate(), svArgs), 0);
} }
@Test
public void testTeraSort() throws Exception { public void testTeraSort() throws Exception {
// Run TeraGen to generate input for 'terasort' // Run TeraGen to generate input for 'terasort'
runTeraGen(createJobConf(), SORT_INPUT_PATH); runTeraGen(createJobConf(), SORT_INPUT_PATH);
@ -104,6 +112,7 @@ public class TestTeraSort extends HadoopTestCase {
TERA_OUTPUT_PATH); TERA_OUTPUT_PATH);
} }
@Test
public void testTeraSortWithLessThanTwoArgs() throws Exception { public void testTeraSortWithLessThanTwoArgs() throws Exception {
String[] args = new String[1]; String[] args = new String[1];
assertEquals(new TeraSort().run(args), 2); assertEquals(new TeraSort().run(args), 2);