svn merge -c 1530816 FIXES: MAPREDUCE-5102. fix coverage org.apache.hadoop.mapreduce.lib.db and org.apache.hadoop.mapred.lib.db. Contributed by Aleksey Gorshkov, Andrey Klochkov, and Nathan Roberts
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1530818 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
5f2d5faa7f
commit
9b5a023fc2
|
@ -50,6 +50,10 @@ Release 2.3.0 - UNRELEASED
|
||||||
|
|
||||||
MAPREDUCE-5514. Fix TestRMContainerAllocator. (Zhijie Shen via acmurthy)
|
MAPREDUCE-5514. Fix TestRMContainerAllocator. (Zhijie Shen via acmurthy)
|
||||||
|
|
||||||
|
MAPREDUCE-5102. fix coverage org.apache.hadoop.mapreduce.lib.db and
|
||||||
|
org.apache.hadoop.mapred.lib.db (Aleksey Gorshkov, Andrey Klochkov, and
|
||||||
|
Nathan Roberts via jlowe)
|
||||||
|
|
||||||
Release 2.2.1 - UNRELEASED
|
Release 2.2.1 - UNRELEASED
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -0,0 +1,157 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.mapred.lib.db;
|
||||||
|
|
||||||
|
import java.sql.DriverManager;
|
||||||
|
|
||||||
|
import org.apache.hadoop.io.LongWritable;
|
||||||
|
import org.apache.hadoop.mapred.JobConf;
|
||||||
|
import org.apache.hadoop.mapred.RecordReader;
|
||||||
|
import org.apache.hadoop.mapred.Reporter;
|
||||||
|
import org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit;
|
||||||
|
import org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader;
|
||||||
|
import org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable;
|
||||||
|
import org.apache.hadoop.mapred.InputSplit;
|
||||||
|
import org.apache.hadoop.mapreduce.MRJobConfig;
|
||||||
|
import org.apache.hadoop.mapred.lib.db.DBConfiguration;
|
||||||
|
import org.apache.hadoop.mapreduce.lib.db.DriverForTest;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import static org.junit.Assert.*;
|
||||||
|
import static org.mockito.Mockito.*;
|
||||||
|
|
||||||
|
public class TestDBInputFormat {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* test DBInputFormat class. Class should split result for chunks
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test(timeout = 10000)
|
||||||
|
public void testDBInputFormat() throws Exception {
|
||||||
|
JobConf configuration = new JobConf();
|
||||||
|
setupDriver(configuration);
|
||||||
|
|
||||||
|
DBInputFormat<NullDBWritable> format = new DBInputFormat<NullDBWritable>();
|
||||||
|
format.setConf(configuration);
|
||||||
|
format.setConf(configuration);
|
||||||
|
DBInputFormat.DBInputSplit splitter = new DBInputFormat.DBInputSplit(1, 10);
|
||||||
|
Reporter reporter = mock(Reporter.class);
|
||||||
|
RecordReader<LongWritable, NullDBWritable> reader = format.getRecordReader(
|
||||||
|
splitter, configuration, reporter);
|
||||||
|
|
||||||
|
configuration.setInt(MRJobConfig.NUM_MAPS, 3);
|
||||||
|
InputSplit[] lSplits = format.getSplits(configuration, 3);
|
||||||
|
assertEquals(5, lSplits[0].getLength());
|
||||||
|
assertEquals(3, lSplits.length);
|
||||||
|
|
||||||
|
// test reader .Some simple tests
|
||||||
|
assertEquals(LongWritable.class, reader.createKey().getClass());
|
||||||
|
assertEquals(0, reader.getPos());
|
||||||
|
assertEquals(0, reader.getProgress(), 0.001);
|
||||||
|
reader.close();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* test configuration for db. should works DBConfiguration.* parameters.
|
||||||
|
*/
|
||||||
|
@Test (timeout = 5000)
|
||||||
|
public void testSetInput() {
|
||||||
|
JobConf configuration = new JobConf();
|
||||||
|
|
||||||
|
String[] fieldNames = { "field1", "field2" };
|
||||||
|
DBInputFormat.setInput(configuration, NullDBWritable.class, "table",
|
||||||
|
"conditions", "orderBy", fieldNames);
|
||||||
|
assertEquals(
|
||||||
|
"org.apache.hadoop.mapred.lib.db.DBInputFormat$NullDBWritable",
|
||||||
|
configuration.getClass(DBConfiguration.INPUT_CLASS_PROPERTY, null)
|
||||||
|
.getName());
|
||||||
|
assertEquals("table",
|
||||||
|
configuration.get(DBConfiguration.INPUT_TABLE_NAME_PROPERTY, null));
|
||||||
|
|
||||||
|
String[] fields = configuration
|
||||||
|
.getStrings(DBConfiguration.INPUT_FIELD_NAMES_PROPERTY);
|
||||||
|
assertEquals("field1", fields[0]);
|
||||||
|
assertEquals("field2", fields[1]);
|
||||||
|
|
||||||
|
assertEquals("conditions",
|
||||||
|
configuration.get(DBConfiguration.INPUT_CONDITIONS_PROPERTY, null));
|
||||||
|
assertEquals("orderBy",
|
||||||
|
configuration.get(DBConfiguration.INPUT_ORDER_BY_PROPERTY, null));
|
||||||
|
|
||||||
|
configuration = new JobConf();
|
||||||
|
|
||||||
|
DBInputFormat.setInput(configuration, NullDBWritable.class, "query",
|
||||||
|
"countQuery");
|
||||||
|
assertEquals("query", configuration.get(DBConfiguration.INPUT_QUERY, null));
|
||||||
|
assertEquals("countQuery",
|
||||||
|
configuration.get(DBConfiguration.INPUT_COUNT_QUERY, null));
|
||||||
|
|
||||||
|
JobConf jConfiguration = new JobConf();
|
||||||
|
DBConfiguration.configureDB(jConfiguration, "driverClass", "dbUrl", "user",
|
||||||
|
"password");
|
||||||
|
assertEquals("driverClass",
|
||||||
|
jConfiguration.get(DBConfiguration.DRIVER_CLASS_PROPERTY));
|
||||||
|
assertEquals("dbUrl", jConfiguration.get(DBConfiguration.URL_PROPERTY));
|
||||||
|
assertEquals("user", jConfiguration.get(DBConfiguration.USERNAME_PROPERTY));
|
||||||
|
assertEquals("password",
|
||||||
|
jConfiguration.get(DBConfiguration.PASSWORD_PROPERTY));
|
||||||
|
jConfiguration = new JobConf();
|
||||||
|
DBConfiguration.configureDB(jConfiguration, "driverClass", "dbUrl");
|
||||||
|
assertEquals("driverClass",
|
||||||
|
jConfiguration.get(DBConfiguration.DRIVER_CLASS_PROPERTY));
|
||||||
|
assertEquals("dbUrl", jConfiguration.get(DBConfiguration.URL_PROPERTY));
|
||||||
|
assertNull(jConfiguration.get(DBConfiguration.USERNAME_PROPERTY));
|
||||||
|
assertNull(jConfiguration.get(DBConfiguration.PASSWORD_PROPERTY));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* test DBRecordReader. This reader should creates keys, values, know about position..
|
||||||
|
*/
|
||||||
|
@SuppressWarnings("unchecked")
|
||||||
|
@Test (timeout = 5000)
|
||||||
|
public void testDBRecordReader() throws Exception {
|
||||||
|
|
||||||
|
JobConf job = mock(JobConf.class);
|
||||||
|
DBConfiguration dbConfig = mock(DBConfiguration.class);
|
||||||
|
String[] fields = { "field1", "filed2" };
|
||||||
|
|
||||||
|
@SuppressWarnings("rawtypes")
|
||||||
|
DBRecordReader reader = new DBInputFormat<NullDBWritable>().new DBRecordReader(
|
||||||
|
new DBInputSplit(), NullDBWritable.class, job,
|
||||||
|
DriverForTest.getConnection(), dbConfig, "condition", fields, "table");
|
||||||
|
LongWritable key = reader.createKey();
|
||||||
|
assertEquals(0, key.get());
|
||||||
|
DBWritable value = reader.createValue();
|
||||||
|
assertEquals(
|
||||||
|
"org.apache.hadoop.mapred.lib.db.DBInputFormat$NullDBWritable", value
|
||||||
|
.getClass().getName());
|
||||||
|
assertEquals(0, reader.getPos());
|
||||||
|
assertFalse(reader.next(key, value));
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
private void setupDriver(JobConf configuration) throws Exception {
|
||||||
|
configuration.set(DBConfiguration.URL_PROPERTY, "testUrl");
|
||||||
|
DriverManager.registerDriver(new DriverForTest());
|
||||||
|
configuration.set(DBConfiguration.DRIVER_CLASS_PROPERTY,
|
||||||
|
DriverForTest.class.getCanonicalName());
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
|
@ -0,0 +1,112 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.mapreduce.lib.db;
|
||||||
|
|
||||||
|
import java.sql.Connection;
|
||||||
|
import java.sql.DatabaseMetaData;
|
||||||
|
import java.sql.Driver;
|
||||||
|
import java.sql.DriverPropertyInfo;
|
||||||
|
import java.sql.PreparedStatement;
|
||||||
|
import java.sql.ResultSet;
|
||||||
|
import java.sql.SQLException;
|
||||||
|
import java.sql.SQLFeatureNotSupportedException;
|
||||||
|
import java.sql.Statement;
|
||||||
|
import java.util.Properties;
|
||||||
|
import java.util.logging.Logger;
|
||||||
|
|
||||||
|
import static org.mockito.Matchers.any;
|
||||||
|
import static org.mockito.Mockito.*;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* class emulates a connection to database
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
public class DriverForTest implements Driver {
|
||||||
|
|
||||||
|
public static Connection getConnection() {
|
||||||
|
Connection connection = mock(FakeConnection.class);
|
||||||
|
try {
|
||||||
|
Statement statement = mock(Statement.class);
|
||||||
|
ResultSet results = mock(ResultSet.class);
|
||||||
|
when(results.getLong(1)).thenReturn(15L);
|
||||||
|
when(statement.executeQuery(any(String.class))).thenReturn(results);
|
||||||
|
when(connection.createStatement()).thenReturn(statement);
|
||||||
|
|
||||||
|
DatabaseMetaData metadata = mock(DatabaseMetaData.class);
|
||||||
|
when(metadata.getDatabaseProductName()).thenReturn("Test");
|
||||||
|
when(connection.getMetaData()).thenReturn(metadata);
|
||||||
|
|
||||||
|
PreparedStatement reparedStatement0= mock(PreparedStatement.class);
|
||||||
|
when(connection.prepareStatement(anyString())).thenReturn(
|
||||||
|
reparedStatement0);
|
||||||
|
|
||||||
|
PreparedStatement preparedStatement = mock(PreparedStatement.class);
|
||||||
|
ResultSet resultSet = mock(ResultSet.class);
|
||||||
|
when(resultSet.next()).thenReturn(false);
|
||||||
|
when(preparedStatement.executeQuery()).thenReturn(resultSet);
|
||||||
|
|
||||||
|
when(connection.prepareStatement(anyString(), anyInt(), anyInt()))
|
||||||
|
.thenReturn(preparedStatement);
|
||||||
|
} catch (SQLException e) {
|
||||||
|
;
|
||||||
|
}
|
||||||
|
return connection;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean acceptsURL(String arg0) throws SQLException {
|
||||||
|
return "testUrl".equals(arg0);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Connection connect(String arg0, Properties arg1) throws SQLException {
|
||||||
|
|
||||||
|
return getConnection();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getMajorVersion() {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getMinorVersion() {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public DriverPropertyInfo[] getPropertyInfo(String arg0, Properties arg1)
|
||||||
|
throws SQLException {
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean jdbcCompliant() {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Logger getParentLogger() throws SQLFeatureNotSupportedException {
|
||||||
|
throw new SQLFeatureNotSupportedException();
|
||||||
|
}
|
||||||
|
|
||||||
|
private interface FakeConnection extends Connection{
|
||||||
|
public void setSessionTimeZone(String arg);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
|
@ -0,0 +1,170 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.mapreduce.lib.db;
|
||||||
|
|
||||||
|
import java.sql.Connection;
|
||||||
|
import java.sql.Types;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.mapreduce.Job;
|
||||||
|
import org.apache.hadoop.mapreduce.InputSplit;
|
||||||
|
import org.apache.hadoop.mapreduce.JobContext;
|
||||||
|
import org.apache.hadoop.mapreduce.MRJobConfig;
|
||||||
|
import org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit;
|
||||||
|
import org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable;
|
||||||
|
import org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import static org.junit.Assert.*;
|
||||||
|
import static org.mockito.Mockito.*;
|
||||||
|
|
||||||
|
public class TestDbClasses {
|
||||||
|
/**
|
||||||
|
* test splitters from DataDrivenDBInputFormat. For different data types may
|
||||||
|
* be different splitter
|
||||||
|
*/
|
||||||
|
@Test(timeout = 1000)
|
||||||
|
public void testDataDrivenDBInputFormatSplitter() {
|
||||||
|
DataDrivenDBInputFormat<NullDBWritable> format = new DataDrivenDBInputFormat<NullDBWritable>();
|
||||||
|
testCommonSplitterTypes(format);
|
||||||
|
assertEquals(DateSplitter.class, format.getSplitter(Types.TIMESTAMP)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(DateSplitter.class, format.getSplitter(Types.DATE).getClass());
|
||||||
|
assertEquals(DateSplitter.class, format.getSplitter(Types.TIME).getClass());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout = 1000)
|
||||||
|
public void testDataDrivenDBInputFormat() throws Exception {
|
||||||
|
JobContext jobContext = mock(JobContext.class);
|
||||||
|
Configuration configuration = new Configuration();
|
||||||
|
configuration.setInt(MRJobConfig.NUM_MAPS, 1);
|
||||||
|
|
||||||
|
when(jobContext.getConfiguration()).thenReturn(configuration);
|
||||||
|
DataDrivenDBInputFormat<NullDBWritable> format = new DataDrivenDBInputFormat<NullDBWritable>();
|
||||||
|
List<InputSplit> splits = format.getSplits(jobContext);
|
||||||
|
assertEquals(1, splits.size());
|
||||||
|
DataDrivenDBInputSplit split = (DataDrivenDBInputSplit) splits.get(0);
|
||||||
|
assertEquals("1=1", split.getLowerClause());
|
||||||
|
assertEquals("1=1", split.getUpperClause());
|
||||||
|
|
||||||
|
// 2
|
||||||
|
configuration.setInt(MRJobConfig.NUM_MAPS, 2);
|
||||||
|
|
||||||
|
DataDrivenDBInputFormat.setBoundingQuery(configuration, "query");
|
||||||
|
assertEquals("query",
|
||||||
|
configuration.get(DBConfiguration.INPUT_BOUNDING_QUERY));
|
||||||
|
|
||||||
|
Job job = mock(Job.class);
|
||||||
|
when(job.getConfiguration()).thenReturn(configuration);
|
||||||
|
DataDrivenDBInputFormat.setInput(job, NullDBWritable.class, "query",
|
||||||
|
"Bounding Query");
|
||||||
|
|
||||||
|
assertEquals("Bounding Query",
|
||||||
|
configuration.get(DBConfiguration.INPUT_BOUNDING_QUERY));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout = 1000)
|
||||||
|
public void testOracleDataDrivenDBInputFormat() throws Exception {
|
||||||
|
OracleDataDrivenDBInputFormat<NullDBWritable> format =
|
||||||
|
new OracleDataDrivenDBInputFormatForTest();
|
||||||
|
testCommonSplitterTypes(format);
|
||||||
|
assertEquals(OracleDateSplitter.class, format.getSplitter(Types.TIMESTAMP)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(OracleDateSplitter.class,
|
||||||
|
format.getSplitter(Types.DATE).getClass());
|
||||||
|
assertEquals(OracleDateSplitter.class,
|
||||||
|
format.getSplitter(Types.TIME).getClass());
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* test generate sql script for OracleDBRecordReader.
|
||||||
|
*/
|
||||||
|
|
||||||
|
@Test(timeout = 2000)
|
||||||
|
public void testOracleDBRecordReader() throws Exception {
|
||||||
|
DBInputSplit splitter = new DBInputSplit(1, 10);
|
||||||
|
Configuration configuration = new Configuration();
|
||||||
|
Connection connect = DriverForTest.getConnection();
|
||||||
|
|
||||||
|
DBConfiguration dbConfiguration = new DBConfiguration(configuration);
|
||||||
|
dbConfiguration.setInputOrderBy("Order");
|
||||||
|
String[] fields = { "f1", "f2" };
|
||||||
|
|
||||||
|
OracleDBRecordReader<NullDBWritable> recorder = new OracleDBRecordReader<NullDBWritable>(
|
||||||
|
splitter, NullDBWritable.class, configuration, connect,
|
||||||
|
dbConfiguration, "condition", fields, "table");
|
||||||
|
assertEquals(
|
||||||
|
"SELECT * FROM (SELECT a.*,ROWNUM dbif_rno FROM ( SELECT f1, f2 FROM table WHERE condition ORDER BY Order ) a WHERE rownum <= 1 + 9 ) WHERE dbif_rno >= 1",
|
||||||
|
recorder.getSelectQuery());
|
||||||
|
}
|
||||||
|
|
||||||
|
private void testCommonSplitterTypes(
|
||||||
|
DataDrivenDBInputFormat<NullDBWritable> format) {
|
||||||
|
assertEquals(BigDecimalSplitter.class, format.getSplitter(Types.DECIMAL)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(BigDecimalSplitter.class, format.getSplitter(Types.NUMERIC)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(BooleanSplitter.class, format.getSplitter(Types.BOOLEAN)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(BooleanSplitter.class, format.getSplitter(Types.BIT)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(IntegerSplitter.class, format.getSplitter(Types.BIGINT)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(IntegerSplitter.class, format.getSplitter(Types.TINYINT)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(IntegerSplitter.class, format.getSplitter(Types.SMALLINT)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(IntegerSplitter.class, format.getSplitter(Types.INTEGER)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(FloatSplitter.class, format.getSplitter(Types.DOUBLE)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(FloatSplitter.class, format.getSplitter(Types.REAL).getClass());
|
||||||
|
assertEquals(FloatSplitter.class, format.getSplitter(Types.FLOAT)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(TextSplitter.class, format.getSplitter(Types.LONGVARCHAR)
|
||||||
|
.getClass());
|
||||||
|
assertEquals(TextSplitter.class, format.getSplitter(Types.CHAR).getClass());
|
||||||
|
assertEquals(TextSplitter.class, format.getSplitter(Types.VARCHAR)
|
||||||
|
.getClass());
|
||||||
|
// if unknown data type splitter is null
|
||||||
|
assertNull(format.getSplitter(Types.BINARY));
|
||||||
|
}
|
||||||
|
|
||||||
|
private class OracleDataDrivenDBInputFormatForTest extends
|
||||||
|
OracleDataDrivenDBInputFormat<NullDBWritable> {
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public DBConfiguration getDBConf() {
|
||||||
|
|
||||||
|
String[] names = { "field1", "field2" };
|
||||||
|
DBConfiguration result = mock(DBConfiguration.class);
|
||||||
|
when(result.getInputConditions()).thenReturn("conditions");
|
||||||
|
when(result.getInputFieldNames()).thenReturn(names);
|
||||||
|
when(result.getInputTableName()).thenReturn("table");
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Connection getConnection() {
|
||||||
|
return DriverForTest.getConnection();
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
|
@ -0,0 +1,164 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.mapreduce.lib.db;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.math.BigDecimal;
|
||||||
|
import java.sql.ResultSet;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.mapreduce.InputSplit;
|
||||||
|
import org.apache.hadoop.mapreduce.MRJobConfig;
|
||||||
|
import org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test Splitters. Splitters should build parts of sql sentences for split result.
|
||||||
|
*/
|
||||||
|
public class TestSplitters {
|
||||||
|
|
||||||
|
private Configuration configuration;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
configuration = new Configuration();
|
||||||
|
configuration.setInt(MRJobConfig.NUM_MAPS, 2);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout=2000)
|
||||||
|
public void testBooleanSplitter() throws Exception{
|
||||||
|
BooleanSplitter splitter = new BooleanSplitter();
|
||||||
|
ResultSet result = mock(ResultSet.class);
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
|
||||||
|
List<InputSplit> splits=splitter.split(configuration, result, "column");
|
||||||
|
assertSplits(new String[] {"column = FALSE column = FALSE",
|
||||||
|
"column IS NULL column IS NULL"}, splits);
|
||||||
|
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
when(result.getString(2)).thenReturn("result2");
|
||||||
|
when(result.getBoolean(1)).thenReturn(true);
|
||||||
|
when(result.getBoolean(2)).thenReturn(false);
|
||||||
|
|
||||||
|
splits=splitter.split(configuration, result, "column");
|
||||||
|
assertEquals(0, splits.size());
|
||||||
|
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
when(result.getString(2)).thenReturn("result2");
|
||||||
|
when(result.getBoolean(1)).thenReturn(false);
|
||||||
|
when(result.getBoolean(2)).thenReturn(true);
|
||||||
|
|
||||||
|
splits = splitter.split(configuration, result, "column");
|
||||||
|
assertSplits(new String[] {
|
||||||
|
"column = FALSE column = FALSE", ".*column = TRUE"}, splits);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout=2000)
|
||||||
|
public void testFloatSplitter() throws Exception{
|
||||||
|
FloatSplitter splitter = new FloatSplitter();
|
||||||
|
|
||||||
|
ResultSet results = mock(ResultSet.class);
|
||||||
|
|
||||||
|
List<InputSplit> splits = splitter.split(configuration, results, "column");
|
||||||
|
assertSplits(new String[] {".*column IS NULL"}, splits);
|
||||||
|
|
||||||
|
when(results.getString(1)).thenReturn("result1");
|
||||||
|
when(results.getString(2)).thenReturn("result2");
|
||||||
|
when(results.getDouble(1)).thenReturn(5.0);
|
||||||
|
when(results.getDouble(2)).thenReturn(7.0);
|
||||||
|
|
||||||
|
splits = splitter.split(configuration, results, "column1");
|
||||||
|
assertSplits(new String[] {"column1 >= 5.0 column1 < 6.0",
|
||||||
|
"column1 >= 6.0 column1 <= 7.0"}, splits);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout=2000)
|
||||||
|
public void testBigDecimalSplitter() throws Exception{
|
||||||
|
BigDecimalSplitter splitter = new BigDecimalSplitter();
|
||||||
|
ResultSet result = mock(ResultSet.class);
|
||||||
|
|
||||||
|
List<InputSplit> splits = splitter.split(configuration, result, "column");
|
||||||
|
assertSplits(new String[] {".*column IS NULL"}, splits);
|
||||||
|
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
when(result.getString(2)).thenReturn("result2");
|
||||||
|
when(result.getBigDecimal(1)).thenReturn(new BigDecimal(10));
|
||||||
|
when(result.getBigDecimal(2)).thenReturn(new BigDecimal(12));
|
||||||
|
|
||||||
|
splits = splitter.split(configuration, result, "column1");
|
||||||
|
assertSplits(new String[] {"column1 >= 10 column1 < 11",
|
||||||
|
"column1 >= 11 column1 <= 12"}, splits);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout=2000)
|
||||||
|
public void testIntegerSplitter() throws Exception{
|
||||||
|
IntegerSplitter splitter = new IntegerSplitter();
|
||||||
|
ResultSet result = mock(ResultSet.class);
|
||||||
|
|
||||||
|
List<InputSplit> splits = splitter.split(configuration, result, "column");
|
||||||
|
assertSplits(new String[] {".*column IS NULL"}, splits);
|
||||||
|
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
when(result.getString(2)).thenReturn("result2");
|
||||||
|
when(result.getLong(1)).thenReturn(8L);
|
||||||
|
when(result.getLong(2)).thenReturn(19L);
|
||||||
|
|
||||||
|
splits = splitter.split(configuration, result, "column1");
|
||||||
|
assertSplits(new String[] {"column1 >= 8 column1 < 13",
|
||||||
|
"column1 >= 13 column1 < 18", "column1 >= 18 column1 <= 19"}, splits);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(timeout=2000)
|
||||||
|
public void testTextSplitter() throws Exception{
|
||||||
|
TextSplitter splitter = new TextSplitter();
|
||||||
|
ResultSet result = mock(ResultSet.class);
|
||||||
|
|
||||||
|
List<InputSplit> splits = splitter.split(configuration, result, "column");
|
||||||
|
assertSplits(new String[] {"column IS NULL column IS NULL"}, splits);
|
||||||
|
|
||||||
|
when(result.getString(1)).thenReturn("result1");
|
||||||
|
when(result.getString(2)).thenReturn("result2");
|
||||||
|
|
||||||
|
splits = splitter.split(configuration, result, "column1");
|
||||||
|
assertSplits(new String[] {"column1 >= 'result1' column1 < 'result1.'",
|
||||||
|
"column1 >= 'result1' column1 <= 'result2'"}, splits);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void assertSplits(String[] expectedSplitRE,
|
||||||
|
List<InputSplit> splits) throws IOException {
|
||||||
|
assertEquals(expectedSplitRE.length, splits.size());
|
||||||
|
for (int i = 0; i < expectedSplitRE.length; i++) {
|
||||||
|
DataDrivenDBInputSplit split = (DataDrivenDBInputSplit) splits.get(i);
|
||||||
|
String actualExpr = split.getLowerClause() + " " + split.getUpperClause();
|
||||||
|
assertTrue("Split #" + (i+1) + " expression is wrong."
|
||||||
|
+ " Expected " + expectedSplitRE[i]
|
||||||
|
+ " Actual " + actualExpr,
|
||||||
|
Pattern.matches(expectedSplitRE[i], actualExpr));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
Loading…
Reference in New Issue