Revert HBASE-8453, HBASE-8454.
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1477448 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
0e9fbc727c
commit
fa3918d300
|
@ -1997,10 +1997,6 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
|
||||||
return dfsCluster;
|
return dfsCluster;
|
||||||
}
|
}
|
||||||
|
|
||||||
public MiniMRCluster getMRCluster() {
|
|
||||||
return mrCluster;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void setDFSCluster(MiniDFSCluster cluster) throws IOException {
|
public void setDFSCluster(MiniDFSCluster cluster) throws IOException {
|
||||||
if (dfsCluster != null && dfsCluster.isClusterUp()) {
|
if (dfsCluster != null && dfsCluster.isClusterUp()) {
|
||||||
throw new IOException("DFSCluster is already running! Shut it down first.");
|
throw new IOException("DFSCluster is already running! Shut it down first.");
|
||||||
|
|
|
@ -26,7 +26,6 @@ import java.io.IOException;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
|
@ -42,7 +41,6 @@ import org.apache.hadoop.hbase.client.Scan;
|
||||||
import org.apache.hadoop.hbase.filter.Filter;
|
import org.apache.hadoop.hbase.filter.Filter;
|
||||||
import org.apache.hadoop.hbase.filter.PrefixFilter;
|
import org.apache.hadoop.hbase.filter.PrefixFilter;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.mapred.JobConf;
|
|
||||||
import org.apache.hadoop.mapreduce.Job;
|
import org.apache.hadoop.mapreduce.Job;
|
||||||
import org.apache.hadoop.util.GenericOptionsParser;
|
import org.apache.hadoop.util.GenericOptionsParser;
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
|
@ -105,9 +103,7 @@ public class TestImportExport {
|
||||||
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
||||||
Configuration conf = opts.getConfiguration();
|
Configuration conf = opts.getConfiguration();
|
||||||
args = opts.getRemainingArgs();
|
args = opts.getRemainingArgs();
|
||||||
JobConf jobConf = UTIL.getMRCluster().createJobConf();
|
Job job = Export.createSubmittableJob(conf, args);
|
||||||
HBaseConfiguration.merge(jobConf, conf);
|
|
||||||
Job job = Export.createSubmittableJob(jobConf, args);
|
|
||||||
job.waitForCompletion(false);
|
job.waitForCompletion(false);
|
||||||
return job.isSuccessful();
|
return job.isSuccessful();
|
||||||
}
|
}
|
||||||
|
@ -125,9 +121,7 @@ public class TestImportExport {
|
||||||
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
|
||||||
Configuration conf = opts.getConfiguration();
|
Configuration conf = opts.getConfiguration();
|
||||||
args = opts.getRemainingArgs();
|
args = opts.getRemainingArgs();
|
||||||
JobConf jobConf = UTIL.getMRCluster().createJobConf();
|
Job job = Import.createSubmittableJob(conf, args);
|
||||||
HBaseConfiguration.merge(jobConf, conf);
|
|
||||||
Job job = Import.createSubmittableJob(jobConf, args);
|
|
||||||
job.waitForCompletion(false);
|
job.waitForCompletion(false);
|
||||||
return job.isSuccessful();
|
return job.isSuccessful();
|
||||||
}
|
}
|
||||||
|
|
|
@ -38,7 +38,6 @@ import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.KeyValue;
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
import org.apache.hadoop.hbase.LargeTests;
|
import org.apache.hadoop.hbase.LargeTests;
|
||||||
|
@ -47,7 +46,6 @@ import org.apache.hadoop.hbase.client.Result;
|
||||||
import org.apache.hadoop.hbase.client.ResultScanner;
|
import org.apache.hadoop.hbase.client.ResultScanner;
|
||||||
import org.apache.hadoop.hbase.client.Scan;
|
import org.apache.hadoop.hbase.client.Scan;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.mapred.JobConf;
|
|
||||||
import org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter;
|
import org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter;
|
||||||
import org.apache.hadoop.util.Tool;
|
import org.apache.hadoop.util.Tool;
|
||||||
import org.apache.hadoop.util.ToolRunner;
|
import org.apache.hadoop.util.ToolRunner;
|
||||||
|
@ -222,9 +220,7 @@ public class TestImportTsv implements Configurable {
|
||||||
argv.add(inputPath.toString());
|
argv.add(inputPath.toString());
|
||||||
Tool tool = new ImportTsv();
|
Tool tool = new ImportTsv();
|
||||||
LOG.debug("Running ImportTsv with arguments: " + argv);
|
LOG.debug("Running ImportTsv with arguments: " + argv);
|
||||||
JobConf jobConf = util.getMRCluster().createJobConf();
|
assertEquals(0, ToolRunner.run(conf, tool, argv.toArray(args)));
|
||||||
HBaseConfiguration.merge(jobConf, conf);
|
|
||||||
assertEquals(0, ToolRunner.run(jobConf, tool, argv.toArray(args)));
|
|
||||||
|
|
||||||
// Perform basic validation. If the input args did not include
|
// Perform basic validation. If the input args did not include
|
||||||
// ImportTsv.BULK_OUTPUT_CONF_KEY then validate data in the table.
|
// ImportTsv.BULK_OUTPUT_CONF_KEY then validate data in the table.
|
||||||
|
|
Loading…
Reference in New Issue