Revert HBASE-8453, HBASE-8454.
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1477448 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
0e9fbc727c
commit
fa3918d300
|
@ -1997,10 +1997,6 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
|
|||
return dfsCluster;
|
||||
}
|
||||
|
||||
public MiniMRCluster getMRCluster() {
|
||||
return mrCluster;
|
||||
}
|
||||
|
||||
public void setDFSCluster(MiniDFSCluster cluster) throws IOException {
|
||||
if (dfsCluster != null && dfsCluster.isClusterUp()) {
|
||||
throw new IOException("DFSCluster is already running! Shut it down first.");
|
||||
|
|
|
@ -26,7 +26,6 @@ import java.io.IOException;
|
|||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
|
@ -42,7 +41,6 @@ import org.apache.hadoop.hbase.client.Scan;
|
|||
import org.apache.hadoop.hbase.filter.Filter;
|
||||
import org.apache.hadoop.hbase.filter.PrefixFilter;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.mapred.JobConf;
|
||||
import org.apache.hadoop.mapreduce.Job;
|
||||
import org.apache.hadoop.util.GenericOptionsParser;
|
||||
import org.junit.After;
|
||||
|
@ -105,9 +103,7 @@ public class TestImportExport {
|
|||
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
||||
Configuration conf = opts.getConfiguration();
|
||||
args = opts.getRemainingArgs();
|
||||
JobConf jobConf = UTIL.getMRCluster().createJobConf();
|
||||
HBaseConfiguration.merge(jobConf, conf);
|
||||
Job job = Export.createSubmittableJob(jobConf, args);
|
||||
Job job = Export.createSubmittableJob(conf, args);
|
||||
job.waitForCompletion(false);
|
||||
return job.isSuccessful();
|
||||
}
|
||||
|
@ -125,9 +121,7 @@ public class TestImportExport {
|
|||
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
||||
Configuration conf = opts.getConfiguration();
|
||||
args = opts.getRemainingArgs();
|
||||
JobConf jobConf = UTIL.getMRCluster().createJobConf();
|
||||
HBaseConfiguration.merge(jobConf, conf);
|
||||
Job job = Import.createSubmittableJob(jobConf, args);
|
||||
Job job = Import.createSubmittableJob(conf, args);
|
||||
job.waitForCompletion(false);
|
||||
return job.isSuccessful();
|
||||
}
|
||||
|
|
|
@ -38,7 +38,6 @@ import org.apache.hadoop.fs.FSDataOutputStream;
|
|||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.KeyValue;
|
||||
import org.apache.hadoop.hbase.LargeTests;
|
||||
|
@ -47,7 +46,6 @@ import org.apache.hadoop.hbase.client.Result;
|
|||
import org.apache.hadoop.hbase.client.ResultScanner;
|
||||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.mapred.JobConf;
|
||||
import org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter;
|
||||
import org.apache.hadoop.util.Tool;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
|
@ -222,9 +220,7 @@ public class TestImportTsv implements Configurable {
|
|||
argv.add(inputPath.toString());
|
||||
Tool tool = new ImportTsv();
|
||||
LOG.debug("Running ImportTsv with arguments: " + argv);
|
||||
JobConf jobConf = util.getMRCluster().createJobConf();
|
||||
HBaseConfiguration.merge(jobConf, conf);
|
||||
assertEquals(0, ToolRunner.run(jobConf, tool, argv.toArray(args)));
|
||||
assertEquals(0, ToolRunner.run(conf, tool, argv.toArray(args)));
|
||||
|
||||
// Perform basic validation. If the input args did not include
|
||||
// ImportTsv.BULK_OUTPUT_CONF_KEY then validate data in the table.
|
||||
|
|
Loading…
Reference in New Issue