HBASE-8534 addendum removes TestDriver (Nick Dimiduk)
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1489878 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
f5ecf2c83c
commit
45da73e18e
|
@ -1,71 +0,0 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more contributor license
|
||||
* agreements. See the NOTICE file distributed with this work for additional information regarding
|
||||
* copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with the License. You may
|
||||
* obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software distributed under the
|
||||
* License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
|
||||
* either express or implied. See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.hbase.mapreduce;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.PrintStream;
|
||||
import java.lang.reflect.InvocationTargetException;
|
||||
|
||||
import org.apache.hadoop.hbase.SmallTests;
|
||||
import org.apache.hadoop.hbase.util.LauncherSecurityManager;
|
||||
import org.junit.Test;
|
||||
import org.junit.experimental.categories.Category;
|
||||
|
||||
@Category(SmallTests.class)
|
||||
public class TestDriver {
|
||||
/**
|
||||
* Test main method of Driver class
|
||||
*/
|
||||
@Test
|
||||
public void testDriver() throws Throwable {
|
||||
|
||||
PrintStream oldPrintStream = System.out;
|
||||
SecurityManager SECURITY_MANAGER = System.getSecurityManager();
|
||||
LauncherSecurityManager newSecurityManager= new LauncherSecurityManager();
|
||||
System.setSecurityManager(newSecurityManager);
|
||||
ByteArrayOutputStream data = new ByteArrayOutputStream();
|
||||
String[] args = {};
|
||||
System.setOut(new PrintStream(data));
|
||||
try {
|
||||
System.setOut(new PrintStream(data));
|
||||
|
||||
try {
|
||||
Driver.main(args);
|
||||
fail("should be SecurityException");
|
||||
} catch (InvocationTargetException e) {
|
||||
assertEquals(-1, newSecurityManager.getExitCode());
|
||||
assertTrue(data.toString().contains(
|
||||
"An example program must be given as the first argument."));
|
||||
assertTrue(data.toString().contains("CellCounter: Count cells in HBase table"));
|
||||
assertTrue(data.toString().contains("completebulkload: Complete a bulk data load."));
|
||||
assertTrue(data.toString().contains(
|
||||
"copytable: Export a table from local cluster to peer cluster"));
|
||||
assertTrue(data.toString().contains("export: Write table data to HDFS."));
|
||||
assertTrue(data.toString().contains("import: Import data written by Export."));
|
||||
assertTrue(data.toString().contains("importtsv: Import data in TSV format."));
|
||||
assertTrue(data.toString().contains("rowcounter: Count rows in HBase table"));
|
||||
}
|
||||
} finally {
|
||||
System.setOut(oldPrintStream);
|
||||
System.setSecurityManager(SECURITY_MANAGER);
|
||||
}
|
||||
|
||||
}
|
||||
}
|
|
@ -49,12 +49,6 @@ public class TestTableMapReduceUtil {
|
|||
assertEquals(Text.class, job.getOutputValueClass());
|
||||
assertNull(job.getCombinerClass());
|
||||
assertEquals("Table", job.getConfiguration().get(TableInputFormat.INPUT_TABLE));
|
||||
assertEquals(
|
||||
"org.apache.hadoop.io.serializer.WritableSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.MutationSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.ResultSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.KeyValueSerialization",
|
||||
job.getConfiguration().get("io.serializations"));
|
||||
|
||||
configuration = new Configuration();
|
||||
job = new Job(configuration, "tableName");
|
||||
|
@ -66,12 +60,6 @@ public class TestTableMapReduceUtil {
|
|||
assertEquals(Text.class, job.getOutputValueClass());
|
||||
assertNull(job.getCombinerClass());
|
||||
assertEquals("Table", job.getConfiguration().get(TableInputFormat.INPUT_TABLE));
|
||||
assertEquals(
|
||||
"org.apache.hadoop.io.serializer.WritableSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.MutationSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.ResultSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.KeyValueSerialization",
|
||||
job.getConfiguration().get("io.serializations"));
|
||||
|
||||
configuration = new Configuration();
|
||||
job = new Job(configuration, "tableName");
|
||||
|
@ -83,12 +71,6 @@ public class TestTableMapReduceUtil {
|
|||
assertEquals(Text.class, job.getOutputValueClass());
|
||||
assertNull(job.getCombinerClass());
|
||||
assertEquals("Table", job.getConfiguration().get(TableInputFormat.INPUT_TABLE));
|
||||
assertEquals(
|
||||
"org.apache.hadoop.io.serializer.WritableSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.MutationSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.ResultSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.KeyValueSerialization",
|
||||
job.getConfiguration().get("io.serializations"));
|
||||
|
||||
configuration = new Configuration();
|
||||
job = new Job(configuration, "tableName");
|
||||
|
@ -100,11 +82,5 @@ public class TestTableMapReduceUtil {
|
|||
assertEquals(Text.class, job.getOutputValueClass());
|
||||
assertNull(job.getCombinerClass());
|
||||
assertEquals("Table", job.getConfiguration().get(TableInputFormat.INPUT_TABLE));
|
||||
assertEquals(
|
||||
"org.apache.hadoop.io.serializer.WritableSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.MutationSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.ResultSerialization," +
|
||||
"org.apache.hadoop.hbase.mapreduce.KeyValueSerialization",
|
||||
job.getConfiguration().get("io.serializations"));
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue