diff --git a/build-webapps.xml b/build-webapps.xml index 22647781237..38443910bb3 100644 --- a/build-webapps.xml +++ b/build-webapps.xml @@ -41,20 +41,16 @@ the generated java classes and the web.xml. To run, do following: --> - - - - - + - + diff --git a/build.xml b/build.xml index 0480e6f8d5b..5444684ab6d 100644 --- a/build.xml +++ b/build.xml @@ -18,18 +18,60 @@ --> - + + + + - - - - - + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + - + + + + + - + @@ -63,8 +105,8 @@ - - + - - - + @@ -101,6 +142,7 @@ + @@ -110,7 +152,7 @@ - + @@ -138,8 +180,7 @@ - - + - - - - + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Tests failed! + + + + + + + + + diff --git a/conf/hbase-default.xml b/conf/hbase-default.xml index 5dfb0b5feee..ecb273a2304 100644 --- a/conf/hbase-default.xml +++ b/conf/hbase-default.xml @@ -221,12 +221,6 @@ such as hlog. - - hbase.hstore.blockCache.blockSize - 65536 - The size of each block in any block caches. - - diff --git a/lib/commons-cli-2.0-SNAPSHOT.jar b/lib/commons-cli-2.0-SNAPSHOT.jar new file mode 100644 index 00000000000..0b1d51072a7 Binary files /dev/null and b/lib/commons-cli-2.0-SNAPSHOT.jar differ diff --git a/lib/commons-logging-1.0.4.jar b/lib/commons-logging-1.0.4.jar new file mode 100644 index 00000000000..b73a80fab64 Binary files /dev/null and b/lib/commons-logging-1.0.4.jar differ diff --git a/lib/commons-logging-api-1.0.4.jar b/lib/commons-logging-api-1.0.4.jar new file mode 100644 index 00000000000..ade9a13c782 Binary files /dev/null and b/lib/commons-logging-api-1.0.4.jar differ diff --git a/lib/hadoop-0.16.0-core.jar b/lib/hadoop-0.16.0-core.jar new file mode 100644 index 00000000000..1ead3049445 Binary files /dev/null and b/lib/hadoop-0.16.0-core.jar differ diff --git a/lib/hadoop-0.16.0-test.jar b/lib/hadoop-0.16.0-test.jar new file mode 100644 index 00000000000..917411ee582 Binary files /dev/null and b/lib/hadoop-0.16.0-test.jar differ diff --git a/lib/jetty-5.1.4.jar b/lib/jetty-5.1.4.jar new file mode 100644 index 00000000000..dcbd99e5aa3 Binary files /dev/null and b/lib/jetty-5.1.4.jar differ diff --git a/lib/jetty-ext/commons-el.jar b/lib/jetty-ext/commons-el.jar new file mode 100644 index 00000000000..608ed796ca3 Binary files /dev/null and b/lib/jetty-ext/commons-el.jar differ diff --git a/lib/jetty-ext/jasper-compiler.jar b/lib/jetty-ext/jasper-compiler.jar new file mode 100644 index 00000000000..96ec53852d4 Binary files /dev/null and b/lib/jetty-ext/jasper-compiler.jar differ diff --git a/lib/jetty-ext/jasper-runtime.jar b/lib/jetty-ext/jasper-runtime.jar new file mode 100644 index 00000000000..5c3b4bfff7a Binary files /dev/null and b/lib/jetty-ext/jasper-runtime.jar differ diff --git a/lib/jetty-ext/jsp-api.jar b/lib/jetty-ext/jsp-api.jar new file mode 100644 index 00000000000..93f89576f0a Binary files /dev/null and b/lib/jetty-ext/jsp-api.jar differ diff --git a/lib/junit-3.8.1.jar b/lib/junit-3.8.1.jar new file mode 100644 index 00000000000..674d71e89ea Binary files /dev/null and b/lib/junit-3.8.1.jar differ diff --git a/lib/log4j-1.2.13.jar b/lib/log4j-1.2.13.jar new file mode 100644 index 00000000000..dde9972109d Binary files /dev/null and b/lib/log4j-1.2.13.jar differ diff --git a/lib/servlet-api.jar b/lib/servlet-api.jar new file mode 100644 index 00000000000..c9dab30f948 Binary files /dev/null and b/lib/servlet-api.jar differ diff --git a/lib/xmlenc-0.52.jar b/lib/xmlenc-0.52.jar new file mode 100644 index 00000000000..ec568b4c9e0 Binary files /dev/null and b/lib/xmlenc-0.52.jar differ diff --git a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 2e0dc31bed1..c5304305c07 100644 --- a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -42,7 +42,7 @@ import org.apache.hadoop.hbase.io.TextSequence; public class HColumnDescriptor implements WritableComparable { // For future backward compatibility - private static final byte COLUMN_DESCRIPTOR_VERSION = (byte)2; + private static final byte COLUMN_DESCRIPTOR_VERSION = (byte)1; /** Legal family names can only contain 'word characters' and end in a colon. */ public static final Pattern LEGAL_FAMILY_NAME = Pattern.compile("\\w+:"); @@ -76,11 +76,6 @@ public class HColumnDescriptor implements WritableComparable { */ public static final boolean DEFAULT_IN_MEMORY = false; - /** - * Default setting for whether to use a block cache or not. - */ - public static final boolean DEFAULT_BLOCK_CACHE_ENABLED = false; - /** * Default maximum length of cell contents. */ @@ -100,8 +95,6 @@ public class HColumnDescriptor implements WritableComparable { private CompressionType compressionType; // Serve reads from in-memory cache private boolean inMemory; - // Serve reads from in-memory block cache - private boolean blockCacheEnabled; // Maximum value size private int maxValueLength; // True if bloom filter was specified @@ -130,7 +123,6 @@ public class HColumnDescriptor implements WritableComparable { this(columnName == null || columnName.length() <= 0? new Text(): new Text(columnName), DEFAULT_N_VERSIONS, DEFAULT_COMPRESSION_TYPE, DEFAULT_IN_MEMORY, - DEFAULT_BLOCK_CACHE_ENABLED, Integer.MAX_VALUE, DEFAULT_BLOOM_FILTER_DESCRIPTOR); } @@ -142,7 +134,6 @@ public class HColumnDescriptor implements WritableComparable { * @param compression Compression type * @param inMemory If true, column data should be kept in an HRegionServer's * cache - * @param blockCacheEnabled If true, MapFile blocks should be cached * @param maxValueLength Restrict values to <= this value * @param bloomFilter Enable the specified bloom filter for this column * @@ -153,7 +144,6 @@ public class HColumnDescriptor implements WritableComparable { */ public HColumnDescriptor(final Text name, final int maxVersions, final CompressionType compression, final boolean inMemory, - final boolean blockCacheEnabled, final int maxValueLength, final BloomFilterDescriptor bloomFilter) { String familyStr = name.toString(); // Test name if not null (It can be null when deserializing after @@ -175,7 +165,6 @@ public class HColumnDescriptor implements WritableComparable { } this.maxVersions = maxVersions; this.inMemory = inMemory; - this.blockCacheEnabled = blockCacheEnabled; this.maxValueLength = maxValueLength; this.bloomFilter = bloomFilter; this.bloomFilterSpecified = this.bloomFilter == null ? false : true; @@ -223,13 +212,6 @@ public class HColumnDescriptor implements WritableComparable { return this.inMemory; } - /** - * @return True if MapFile blocks should be cached. - */ - public boolean isBlockCacheEnabled() { - return blockCacheEnabled; - } - /** * @return Maximum value length. */ @@ -252,7 +234,6 @@ public class HColumnDescriptor implements WritableComparable { return "{name: " + tmp.substring(0, tmp.length() - 1) + ", max versions: " + maxVersions + ", compression: " + this.compressionType + ", in memory: " + inMemory + - ", block cache enabled: " + blockCacheEnabled + ", max length: " + maxValueLength + ", bloom filter: " + (bloomFilterSpecified ? bloomFilter.toString() : "none") + "}"; } @@ -270,7 +251,6 @@ public class HColumnDescriptor implements WritableComparable { result ^= Integer.valueOf(this.maxVersions).hashCode(); result ^= this.compressionType.hashCode(); result ^= Boolean.valueOf(this.inMemory).hashCode(); - result ^= Boolean.valueOf(this.blockCacheEnabled).hashCode(); result ^= Integer.valueOf(this.maxValueLength).hashCode(); result ^= Boolean.valueOf(this.bloomFilterSpecified).hashCode(); result ^= Byte.valueOf(this.versionNumber).hashCode(); @@ -297,10 +277,6 @@ public class HColumnDescriptor implements WritableComparable { bloomFilter = new BloomFilterDescriptor(); bloomFilter.readFields(in); } - - if (this.versionNumber > 1) { - this.blockCacheEnabled = in.readBoolean(); - } } /** {@inheritDoc} */ @@ -316,8 +292,6 @@ public class HColumnDescriptor implements WritableComparable { if(bloomFilterSpecified) { bloomFilter.write(out); } - - out.writeBoolean(this.blockCacheEnabled); } // Comparable @@ -353,18 +327,6 @@ public class HColumnDescriptor implements WritableComparable { } } - if(result == 0) { - if(this.blockCacheEnabled == other.blockCacheEnabled) { - result = 0; - - } else if(this.blockCacheEnabled) { - result = -1; - - } else { - result = 1; - } - } - if(result == 0) { result = other.maxValueLength - this.maxValueLength; } diff --git a/src/java/org/apache/hadoop/hbase/HStore.java b/src/java/org/apache/hadoop/hbase/HStore.java index b768eec456e..1a66f45709e 100644 --- a/src/java/org/apache/hadoop/hbase/HStore.java +++ b/src/java/org/apache/hadoop/hbase/HStore.java @@ -741,19 +741,9 @@ public class HStore implements HConstants { // Finally, start up all the map readers! (There could be more than one // since we haven't compacted yet.) - boolean first = true; for(Map.Entry e: this.storefiles.entrySet()) { - if (first) { - // Use a block cache (if configured) for the first reader only - // so as to control memory usage. - this.readers.put(e.getKey(), - e.getValue().getReader(this.fs, this.bloomFilter, - family.isBlockCacheEnabled())); - first = false; - } else { - this.readers.put(e.getKey(), - e.getValue().getReader(this.fs, this.bloomFilter)); - } + this.readers.put(e.getKey(), + e.getValue().getReader(this.fs, this.bloomFilter)); } } @@ -1570,10 +1560,7 @@ public class HStore implements HConstants { // 6. Loading the new TreeMap. Long orderVal = Long.valueOf(finalCompactedFile.loadInfo(fs)); this.readers.put(orderVal, - // Use a block cache (if configured) for this reader since - // it is the only one. - finalCompactedFile.getReader(this.fs, this.bloomFilter, - family.isBlockCacheEnabled())); + finalCompactedFile.getReader(this.fs, this.bloomFilter)); this.storefiles.put(orderVal, finalCompactedFile); } catch (IOException e) { e = RemoteExceptionHandler.checkIOException(e); diff --git a/src/java/org/apache/hadoop/hbase/HStoreFile.java b/src/java/org/apache/hadoop/hbase/HStoreFile.java index 390d921deff..5b8d7cfbc40 100644 --- a/src/java/org/apache/hadoop/hbase/HStoreFile.java +++ b/src/java/org/apache/hadoop/hbase/HStoreFile.java @@ -31,11 +31,9 @@ import java.util.Random; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.io.BlockFSInputStream; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.util.Writables; import org.apache.hadoop.io.MapFile; @@ -415,37 +413,17 @@ public class HStoreFile implements HConstants { * @return MapFile.Reader * @throws IOException */ - public MapFile.Reader getReader(final FileSystem fs, - final Filter bloomFilter) - throws IOException { - return isReference()? - new HStoreFile.HalfMapFileReader(fs, getMapFilePath(reference).toString(), - conf, reference.getFileRegion(), reference.getMidkey(), bloomFilter): - new BloomFilterMapFile.Reader(fs, getMapFilePath().toString(), - conf, bloomFilter); - } - - /** - * Get reader for the store file map file. - * Client is responsible for closing file when done. - * @param fs - * @param bloomFilter If null, no filtering is done. - * @param blockCacheEnabled If true, MapFile blocks should be cached. - * @return MapFile.Reader - * @throws IOException - */ public synchronized MapFile.Reader getReader(final FileSystem fs, - final Filter bloomFilter, final boolean blockCacheEnabled) + final Filter bloomFilter) throws IOException { if (isReference()) { return new HStoreFile.HalfMapFileReader(fs, getMapFilePath(reference).toString(), conf, - reference.getFileRegion(), reference.getMidkey(), bloomFilter, - blockCacheEnabled); + reference.getFileRegion(), reference.getMidkey(), bloomFilter); } return new BloomFilterMapFile.Reader(fs, getMapFilePath().toString(), - conf, bloomFilter, blockCacheEnabled); + conf, bloomFilter); } /** @@ -606,13 +584,8 @@ public class HStoreFile implements HConstants { */ static class HbaseMapFile extends MapFile { - /** - * A reader capable of reading and caching blocks of the data file. - */ static class HbaseReader extends MapFile.Reader { - private final boolean blockCacheEnabled; - /** * @param fs * @param dirName @@ -621,23 +594,7 @@ public class HStoreFile implements HConstants { */ public HbaseReader(FileSystem fs, String dirName, Configuration conf) throws IOException { - this(fs, dirName, conf, false); - } - - /** - * @param fs - * @param dirName - * @param conf - * @param blockCacheEnabled - * @throws IOException - */ - public HbaseReader(FileSystem fs, String dirName, Configuration conf, - boolean blockCacheEnabled) - throws IOException { - super(fs, dirName, null, conf, false); // defer opening streams - this.blockCacheEnabled = blockCacheEnabled; - open(fs, dirName, null, conf); - + super(fs, dirName, conf); // Force reading of the mapfile index by calling midKey. // Reading the index will bring the index into memory over // here on the client and then close the index file freeing @@ -648,28 +605,6 @@ public class HStoreFile implements HConstants { // using up datanode resources. See HADOOP-2341. midKey(); } - - @Override - protected org.apache.hadoop.io.SequenceFile.Reader createDataFileReader( - FileSystem fs, Path dataFile, Configuration conf) - throws IOException { - if (!blockCacheEnabled) { - return super.createDataFileReader(fs, dataFile, conf); - } - LOG.info("Block Cache enabled"); - final int blockSize = conf.getInt("hbase.hstore.blockCache.blockSize", - 64 * 1024); - return new SequenceFile.Reader(fs, dataFile, conf) { - @Override - protected FSDataInputStream openFile(FileSystem fs, Path file, - int bufferSize, long length) throws IOException { - - return new FSDataInputStream(new BlockFSInputStream( - super.openFile(fs, file, bufferSize, length), length, - blockSize)); - } - }; - } } static class HbaseWriter extends MapFile.Writer { @@ -718,13 +653,6 @@ public class HStoreFile implements HConstants { bloomFilter = filter; } - public Reader(FileSystem fs, String dirName, Configuration conf, - final Filter filter, final boolean blockCacheEnabled) - throws IOException { - super(fs, dirName, conf, blockCacheEnabled); - bloomFilter = filter; - } - /** {@inheritDoc} */ @Override public Writable get(WritableComparable key, Writable val) @@ -817,7 +745,7 @@ public class HStoreFile implements HConstants { final Configuration conf, final Range r, final WritableComparable midKey) throws IOException { - this(fs, dirName, conf, r, midKey, null, false); + this(fs, dirName, conf, r, midKey, null); } HalfMapFileReader(final FileSystem fs, final String dirName, @@ -829,16 +757,6 @@ public class HStoreFile implements HConstants { midkey = midKey; } - HalfMapFileReader(final FileSystem fs, final String dirName, - final Configuration conf, final Range r, - final WritableComparable midKey, final Filter filter, - final boolean blockCacheEnabled) - throws IOException { - super(fs, dirName, conf, filter, blockCacheEnabled); - top = isTopFileRegion(r); - midkey = midKey; - } - @SuppressWarnings("unchecked") private void checkKey(final WritableComparable key) throws IOException { diff --git a/src/java/org/apache/hadoop/hbase/HTableDescriptor.java b/src/java/org/apache/hadoop/hbase/HTableDescriptor.java index 9b763f635cd..a77f745f250 100644 --- a/src/java/org/apache/hadoop/hbase/HTableDescriptor.java +++ b/src/java/org/apache/hadoop/hbase/HTableDescriptor.java @@ -43,15 +43,15 @@ public class HTableDescriptor implements WritableComparable { public static final HTableDescriptor rootTableDesc = new HTableDescriptor(HConstants.ROOT_TABLE_NAME, new HColumnDescriptor(HConstants.COLUMN_FAMILY, 1, - HColumnDescriptor.CompressionType.NONE, false, false, - Integer.MAX_VALUE, null)); + HColumnDescriptor.CompressionType.NONE, false, Integer.MAX_VALUE, + null)); /** table descriptor for meta table */ public static final HTableDescriptor metaTableDesc = new HTableDescriptor(HConstants.META_TABLE_NAME, new HColumnDescriptor(HConstants.COLUMN_FAMILY, 1, - HColumnDescriptor.CompressionType.NONE, false, false, - Integer.MAX_VALUE, null)); + HColumnDescriptor.CompressionType.NONE, false, Integer.MAX_VALUE, + null)); private boolean rootregion; private boolean metaregion; @@ -256,4 +256,4 @@ public class HTableDescriptor implements WritableComparable { public static Path getTableDir(Path rootdir, Text tableName) { return new Path(rootdir, tableName.toString()); } -} \ No newline at end of file +} diff --git a/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java b/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java index d3ad067437c..8a8355910aa 100644 --- a/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java +++ b/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java @@ -202,8 +202,6 @@ public class AlterCommand extends SchemaModificationCommand { .get(spec)).toUpperCase()); } else if (spec.equals("IN_MEMORY")) { inMemory = (Boolean) columnSpec.get(spec); - } else if (spec.equals("BLOCK_CACHE_ENABLED")) { - blockCacheEnabled = (Boolean) columnSpec.get(spec); } else if (spec.equals("BLOOMFILTER")) { bloomFilterType = BloomFilterType.valueOf(((String) columnSpec.get(spec)) .toUpperCase()); @@ -231,8 +229,7 @@ public class AlterCommand extends SchemaModificationCommand { column = appendDelimiter(column); HColumnDescriptor columnDesc = new HColumnDescriptor(new Text(column), - maxVersions, compression, inMemory, blockCacheEnabled, - maxLength, bloomFilterDesc); + maxVersions, compression, inMemory, maxLength, bloomFilterDesc); return columnDesc; } @@ -246,7 +243,6 @@ public class AlterCommand extends SchemaModificationCommand { maxLength = original.getMaxValueLength(); compression = original.getCompression(); inMemory = original.isInMemory(); - blockCacheEnabled = original.isBlockCacheEnabled(); bloomFilterDesc = original.getBloomFilter(); } } diff --git a/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java b/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java index b8384545413..30afcecc735 100644 --- a/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java +++ b/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java @@ -37,7 +37,6 @@ public abstract class SchemaModificationCommand extends BasicCommand { protected int maxLength; protected HColumnDescriptor.CompressionType compression; protected boolean inMemory; - protected boolean blockCacheEnabled; protected BloomFilterDescriptor bloomFilterDesc; protected BloomFilterType bloomFilterType; protected int vectorSize; @@ -53,7 +52,6 @@ public abstract class SchemaModificationCommand extends BasicCommand { maxLength = HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH; compression = HColumnDescriptor.DEFAULT_COMPRESSION_TYPE; inMemory = HColumnDescriptor.DEFAULT_IN_MEMORY; - blockCacheEnabled = HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED; bloomFilterDesc = HColumnDescriptor.DEFAULT_BLOOM_FILTER_DESCRIPTOR; } @@ -78,8 +76,6 @@ public abstract class SchemaModificationCommand extends BasicCommand { .valueOf(((String) columnSpec.get(spec)).toUpperCase()); } else if (spec.equals("IN_MEMORY")) { inMemory = (Boolean) columnSpec.get(spec); - } else if (spec.equals("BLOCK_CACHE_ENABLED")) { - blockCacheEnabled = (Boolean) columnSpec.get(spec); } else if (spec.equals("BLOOMFILTER")) { bloomFilterType = BloomFilterType.valueOf(((String) columnSpec.get(spec)) .toUpperCase()); @@ -107,8 +103,7 @@ public abstract class SchemaModificationCommand extends BasicCommand { column = appendDelimiter(column); HColumnDescriptor columnDesc = new HColumnDescriptor(new Text(column), - maxVersions, compression, inMemory, blockCacheEnabled, - maxLength, bloomFilterDesc); + maxVersions, compression, inMemory, maxLength, bloomFilterDesc); return columnDesc; } diff --git a/src/test/org/apache/hadoop/hbase/HBaseTestCase.java b/src/test/org/apache/hadoop/hbase/HBaseTestCase.java index 9571ecf4384..ce8df19b903 100644 --- a/src/test/org/apache/hadoop/hbase/HBaseTestCase.java +++ b/src/test/org/apache/hadoop/hbase/HBaseTestCase.java @@ -184,11 +184,11 @@ public abstract class HBaseTestCase extends TestCase { final int versions) { HTableDescriptor htd = new HTableDescriptor(name); htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME1), versions, - CompressionType.NONE, false, false, Integer.MAX_VALUE, null)); + CompressionType.NONE, false, Integer.MAX_VALUE, null)); htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME2), versions, - CompressionType.NONE, false, false, Integer.MAX_VALUE, null)); + CompressionType.NONE, false, Integer.MAX_VALUE, null)); htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME3), versions, - CompressionType.NONE, false, false, Integer.MAX_VALUE, null)); + CompressionType.NONE, false, Integer.MAX_VALUE, null)); return htd; } @@ -565,4 +565,4 @@ public abstract class HBaseTestCase extends TestCase { return this.table.get(row, column, ts, versions); } } -} \ No newline at end of file +} diff --git a/src/test/org/apache/hadoop/hbase/TestBloomFilters.java b/src/test/org/apache/hadoop/hbase/TestBloomFilters.java index c505c87647e..f0d19ff6bf1 100644 --- a/src/test/org/apache/hadoop/hbase/TestBloomFilters.java +++ b/src/test/org/apache/hadoop/hbase/TestBloomFilters.java @@ -169,7 +169,6 @@ public class TestBloomFilters extends HBaseClusterTestCase { 1, // Max versions HColumnDescriptor.CompressionType.NONE, // no compression HColumnDescriptor.DEFAULT_IN_MEMORY, // not in memory - HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED, HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH, bloomFilter ) @@ -235,7 +234,6 @@ public class TestBloomFilters extends HBaseClusterTestCase { 1, // Max versions HColumnDescriptor.CompressionType.NONE, // no compression HColumnDescriptor.DEFAULT_IN_MEMORY, // not in memory - HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED, HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH, bloomFilter ) diff --git a/src/test/org/apache/hadoop/hbase/TestTimestamp.java b/src/test/org/apache/hadoop/hbase/TestTimestamp.java index 483518f231b..35788dad225 100644 --- a/src/test/org/apache/hadoop/hbase/TestTimestamp.java +++ b/src/test/org/apache/hadoop/hbase/TestTimestamp.java @@ -337,7 +337,7 @@ public class TestTimestamp extends HBaseTestCase { private HRegion createRegion() throws IOException { HTableDescriptor htd = createTableDescriptor(getName()); htd.addFamily(new HColumnDescriptor(COLUMN, VERSIONS, - CompressionType.NONE, false, false, Integer.MAX_VALUE, null)); + CompressionType.NONE, false, Integer.MAX_VALUE, null)); return createNewHRegion(htd, null, null); } -} \ No newline at end of file +} diff --git a/src/test/org/apache/hadoop/hbase/TestToString.java b/src/test/org/apache/hadoop/hbase/TestToString.java index 4c1ea3da17b..99adc65764d 100644 --- a/src/test/org/apache/hadoop/hbase/TestToString.java +++ b/src/test/org/apache/hadoop/hbase/TestToString.java @@ -44,9 +44,8 @@ public class TestToString extends TestCase { HTableDescriptor htd = HTableDescriptor.rootTableDesc; System. out.println(htd.toString()); assertEquals("Table descriptor", "name: -ROOT-, families: {info:={name: " + - "info, max versions: 1, compression: NONE, in memory: false, " + - "block cache enabled: false, max length: 2147483647, " + - "bloom filter: none}}", htd.toString()); + "info, max versions: 1, compression: NONE, in memory: false, max " + + "length: 2147483647, bloom filter: none}}", htd.toString()); } /**