HBASE-913 Classes using log4j directly
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@703031 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
99c892f030
commit
609ff27c83
@ -18,6 +18,7 @@ Release 0.19.0 - Unreleased
|
|||||||
HBASE-853 [shell] Cannot describe meta tables (Izaak Rubin via Stack)
|
HBASE-853 [shell] Cannot describe meta tables (Izaak Rubin via Stack)
|
||||||
HBASE-844 Can't pass script to hbase shell
|
HBASE-844 Can't pass script to hbase shell
|
||||||
HBASE-837 Add unit tests for ThriftServer.HBaseHandler (Izaak Rubin via Stack)
|
HBASE-837 Add unit tests for ThriftServer.HBaseHandler (Izaak Rubin via Stack)
|
||||||
|
HBASE-913 Classes using log4j directly
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
HBASE-901 Add a limit to key length, check key and value length on client side
|
HBASE-901 Add a limit to key length, check key and value length on client side
|
||||||
|
@ -23,7 +23,6 @@ import java.io.IOException;
|
|||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.HConstants; //TODO: remove
|
|
||||||
import org.apache.hadoop.hbase.io.Cell;
|
import org.apache.hadoop.hbase.io.Cell;
|
||||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||||
import org.apache.hadoop.hbase.io.RowResult;
|
import org.apache.hadoop.hbase.io.RowResult;
|
||||||
@ -33,9 +32,10 @@ import org.apache.hadoop.mapred.MapReduceBase;
|
|||||||
import org.apache.hadoop.mapred.OutputCollector;
|
import org.apache.hadoop.mapred.OutputCollector;
|
||||||
import org.apache.hadoop.mapred.Reducer;
|
import org.apache.hadoop.mapred.Reducer;
|
||||||
import org.apache.hadoop.mapred.Reporter;
|
import org.apache.hadoop.mapred.Reporter;
|
||||||
import org.apache.log4j.Logger;
|
|
||||||
import org.apache.lucene.document.Document;
|
import org.apache.lucene.document.Document;
|
||||||
import org.apache.lucene.document.Field;
|
import org.apache.lucene.document.Field;
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Construct a Lucene document per row, which is consumed by IndexOutputFormat
|
* Construct a Lucene document per row, which is consumed by IndexOutputFormat
|
||||||
@ -43,8 +43,7 @@ import org.apache.lucene.document.Field;
|
|||||||
*/
|
*/
|
||||||
public class IndexTableReduce extends MapReduceBase implements
|
public class IndexTableReduce extends MapReduceBase implements
|
||||||
Reducer<ImmutableBytesWritable, RowResult, ImmutableBytesWritable, LuceneDocumentWrapper> {
|
Reducer<ImmutableBytesWritable, RowResult, ImmutableBytesWritable, LuceneDocumentWrapper> {
|
||||||
private static final Logger LOG = Logger.getLogger(IndexTableReduce.class);
|
private static final Log LOG = LogFactory.getLog(IndexTableReduce.class);
|
||||||
|
|
||||||
private IndexConfiguration indexConf;
|
private IndexConfiguration indexConf;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -27,12 +27,13 @@ import org.apache.hadoop.hbase.io.BatchUpdate;
|
|||||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
/** Abstract base class for merge tests */
|
/** Abstract base class for merge tests */
|
||||||
public abstract class AbstractMergeTestBase extends HBaseClusterTestCase {
|
public abstract class AbstractMergeTestBase extends HBaseClusterTestCase {
|
||||||
static final Logger LOG =
|
static final Log LOG =
|
||||||
Logger.getLogger(AbstractMergeTestBase.class.getName());
|
LogFactory.getLog(AbstractMergeTestBase.class.getName());
|
||||||
protected static final byte [] COLUMN_NAME = Bytes.toBytes("contents:");
|
protected static final byte [] COLUMN_NAME = Bytes.toBytes("contents:");
|
||||||
protected final Random rand = new Random();
|
protected final Random rand = new Random();
|
||||||
protected HTableDescriptor desc;
|
protected HTableDescriptor desc;
|
||||||
|
@ -30,7 +30,8 @@ import org.apache.hadoop.fs.Path;
|
|||||||
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.io.MapFile;
|
import org.apache.hadoop.io.MapFile;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* <p>
|
* <p>
|
||||||
@ -42,8 +43,8 @@ public class MapFilePerformanceEvaluation {
|
|||||||
private static final int ROW_LENGTH = 1000;
|
private static final int ROW_LENGTH = 1000;
|
||||||
private static final int ROW_COUNT = 1000000;
|
private static final int ROW_COUNT = 1000000;
|
||||||
|
|
||||||
static final Logger LOG =
|
static final Log LOG =
|
||||||
Logger.getLogger(MapFilePerformanceEvaluation.class.getName());
|
LogFactory.getLog(MapFilePerformanceEvaluation.class.getName());
|
||||||
|
|
||||||
static ImmutableBytesWritable format(final int i, ImmutableBytesWritable w) {
|
static ImmutableBytesWritable format(final int i, ImmutableBytesWritable w) {
|
||||||
String v = Integer.toString(i);
|
String v = Integer.toString(i);
|
||||||
|
@ -22,7 +22,8 @@ package org.apache.hadoop.hbase;
|
|||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.master.HMaster;
|
import org.apache.hadoop.hbase.master.HMaster;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
@ -33,8 +34,7 @@ import org.apache.hadoop.hbase.regionserver.HRegion;
|
|||||||
* each server.
|
* each server.
|
||||||
*/
|
*/
|
||||||
public class MiniHBaseCluster implements HConstants {
|
public class MiniHBaseCluster implements HConstants {
|
||||||
static final Logger LOG =
|
static final Log LOG = LogFactory.getLog(MiniHBaseCluster.class.getName());
|
||||||
Logger.getLogger(MiniHBaseCluster.class.getName());
|
|
||||||
|
|
||||||
private HBaseConfiguration conf;
|
private HBaseConfiguration conf;
|
||||||
private LocalHBaseCluster hbaseCluster;
|
private LocalHBaseCluster hbaseCluster;
|
||||||
|
@ -27,8 +27,6 @@ import java.util.Random;
|
|||||||
import junit.framework.TestSuite;
|
import junit.framework.TestSuite;
|
||||||
import junit.textui.TestRunner;
|
import junit.textui.TestRunner;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
|
||||||
import org.apache.commons.logging.LogFactory;
|
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.FileUtil;
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
@ -47,8 +45,8 @@ import org.apache.hadoop.mapred.FileOutputFormat;
|
|||||||
import org.apache.hadoop.mapred.JobClient;
|
import org.apache.hadoop.mapred.JobClient;
|
||||||
import org.apache.hadoop.mapred.JobConf;
|
import org.apache.hadoop.mapred.JobConf;
|
||||||
import org.apache.hadoop.mapred.MiniMRCluster;
|
import org.apache.hadoop.mapred.MiniMRCluster;
|
||||||
import org.apache.log4j.Level;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.lucene.index.Term;
|
import org.apache.lucene.index.Term;
|
||||||
import org.apache.lucene.search.IndexSearcher;
|
import org.apache.lucene.search.IndexSearcher;
|
||||||
import org.apache.lucene.search.MultiSearcher;
|
import org.apache.lucene.search.MultiSearcher;
|
||||||
@ -82,8 +80,6 @@ public class TestTableIndex extends MultiRegionTable {
|
|||||||
desc = new HTableDescriptor(TABLE_NAME);
|
desc = new HTableDescriptor(TABLE_NAME);
|
||||||
desc.addFamily(new HColumnDescriptor(INPUT_COLUMN));
|
desc.addFamily(new HColumnDescriptor(INPUT_COLUMN));
|
||||||
desc.addFamily(new HColumnDescriptor(OUTPUT_COLUMN));
|
desc.addFamily(new HColumnDescriptor(OUTPUT_COLUMN));
|
||||||
// Enable DEBUG-level MR logging.
|
|
||||||
Logger.getLogger("org.apache.hadoop.mapred").setLevel(Level.DEBUG);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -29,7 +29,8 @@ import org.apache.hadoop.dfs.MiniDFSCluster;
|
|||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseTestCase;
|
import org.apache.hadoop.hbase.HBaseTestCase;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.hbase.HStoreKey;
|
import org.apache.hadoop.hbase.HStoreKey;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
@ -44,7 +45,7 @@ import org.apache.hadoop.hbase.util.Bytes;
|
|||||||
* HRegions or in the HBaseMaster, so only basic testing is possible.
|
* HRegions or in the HBaseMaster, so only basic testing is possible.
|
||||||
*/
|
*/
|
||||||
public class TestHRegion extends HBaseTestCase {
|
public class TestHRegion extends HBaseTestCase {
|
||||||
static final Logger LOG = Logger.getLogger(TestHRegion.class);
|
static final Log LOG = LogFactory.getLog(TestHRegion.class);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Since all the "tests" depend on the results of the previous test, they are
|
* Since all the "tests" depend on the results of the previous test, they are
|
||||||
|
@ -24,14 +24,11 @@ import java.util.TreeMap;
|
|||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.log4j.Level;
|
|
||||||
import org.apache.log4j.Logger;
|
|
||||||
import org.apache.hadoop.hbase.HStoreKey;
|
|
||||||
import org.apache.hadoop.hbase.HBaseClusterTestCase;
|
import org.apache.hadoop.hbase.HBaseClusterTestCase;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
|
||||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.HStoreKey;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
import org.apache.hadoop.hbase.NotServingRegionException;
|
|
||||||
import org.apache.hadoop.hbase.UnknownScannerException;
|
import org.apache.hadoop.hbase.UnknownScannerException;
|
||||||
import org.apache.hadoop.hbase.io.BatchUpdate;
|
import org.apache.hadoop.hbase.io.BatchUpdate;
|
||||||
import org.apache.hadoop.hbase.io.Cell;
|
import org.apache.hadoop.hbase.io.Cell;
|
||||||
@ -42,7 +39,6 @@ import org.apache.hadoop.hbase.util.Bytes;
|
|||||||
* split and manufactures odd-ball split scenarios.
|
* split and manufactures odd-ball split scenarios.
|
||||||
*/
|
*/
|
||||||
public class TestSplit extends HBaseClusterTestCase {
|
public class TestSplit extends HBaseClusterTestCase {
|
||||||
@SuppressWarnings("hiding")
|
|
||||||
static final Log LOG = LogFactory.getLog(TestSplit.class.getName());
|
static final Log LOG = LogFactory.getLog(TestSplit.class.getName());
|
||||||
|
|
||||||
/** constructor */
|
/** constructor */
|
||||||
@ -64,10 +60,6 @@ public class TestSplit extends HBaseClusterTestCase {
|
|||||||
// This size should make it so we always split using the addContent
|
// This size should make it so we always split using the addContent
|
||||||
// below. After adding all data, the first region is 1.3M
|
// below. After adding all data, the first region is 1.3M
|
||||||
conf.setLong("hbase.hregion.max.filesize", 1024 * 128);
|
conf.setLong("hbase.hregion.max.filesize", 1024 * 128);
|
||||||
|
|
||||||
Logger.getRootLogger().setLevel(Level.WARN);
|
|
||||||
Logger.getLogger(this.getClass().getPackage().getName()).
|
|
||||||
setLevel(Level.DEBUG);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
Loading…
x
Reference in New Issue
Block a user