diff --git a/build-webapps.xml b/build-webapps.xml
index 22647781237..38443910bb3 100644
--- a/build-webapps.xml
+++ b/build-webapps.xml
@@ -41,20 +41,16 @@ the generated java classes and the web.xml. To run, do following:
-->
-
-
-
-
-
+
-
+
diff --git a/build.xml b/build.xml
index 0480e6f8d5b..5444684ab6d 100644
--- a/build.xml
+++ b/build.xml
@@ -18,18 +18,60 @@
-->
-
+
+
+
+
-
-
-
-
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
-
+
+
+
+
+
-
+
@@ -63,8 +105,8 @@
-
-
+
-
-
-
+
@@ -101,6 +142,7 @@
+
@@ -110,7 +152,7 @@
-
+
@@ -138,8 +180,7 @@
-
-
+
-
-
-
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+ Tests failed!
+
+
+
+
+
+
+
+
+
diff --git a/conf/hbase-default.xml b/conf/hbase-default.xml
index 5dfb0b5feee..ecb273a2304 100644
--- a/conf/hbase-default.xml
+++ b/conf/hbase-default.xml
@@ -221,12 +221,6 @@
such as hlog.
-
- hbase.hstore.blockCache.blockSize
- 65536
- The size of each block in any block caches.
-
-
diff --git a/lib/commons-cli-2.0-SNAPSHOT.jar b/lib/commons-cli-2.0-SNAPSHOT.jar
new file mode 100644
index 00000000000..0b1d51072a7
Binary files /dev/null and b/lib/commons-cli-2.0-SNAPSHOT.jar differ
diff --git a/lib/commons-logging-1.0.4.jar b/lib/commons-logging-1.0.4.jar
new file mode 100644
index 00000000000..b73a80fab64
Binary files /dev/null and b/lib/commons-logging-1.0.4.jar differ
diff --git a/lib/commons-logging-api-1.0.4.jar b/lib/commons-logging-api-1.0.4.jar
new file mode 100644
index 00000000000..ade9a13c782
Binary files /dev/null and b/lib/commons-logging-api-1.0.4.jar differ
diff --git a/lib/hadoop-0.16.0-core.jar b/lib/hadoop-0.16.0-core.jar
new file mode 100644
index 00000000000..1ead3049445
Binary files /dev/null and b/lib/hadoop-0.16.0-core.jar differ
diff --git a/lib/hadoop-0.16.0-test.jar b/lib/hadoop-0.16.0-test.jar
new file mode 100644
index 00000000000..917411ee582
Binary files /dev/null and b/lib/hadoop-0.16.0-test.jar differ
diff --git a/lib/jetty-5.1.4.jar b/lib/jetty-5.1.4.jar
new file mode 100644
index 00000000000..dcbd99e5aa3
Binary files /dev/null and b/lib/jetty-5.1.4.jar differ
diff --git a/lib/jetty-ext/commons-el.jar b/lib/jetty-ext/commons-el.jar
new file mode 100644
index 00000000000..608ed796ca3
Binary files /dev/null and b/lib/jetty-ext/commons-el.jar differ
diff --git a/lib/jetty-ext/jasper-compiler.jar b/lib/jetty-ext/jasper-compiler.jar
new file mode 100644
index 00000000000..96ec53852d4
Binary files /dev/null and b/lib/jetty-ext/jasper-compiler.jar differ
diff --git a/lib/jetty-ext/jasper-runtime.jar b/lib/jetty-ext/jasper-runtime.jar
new file mode 100644
index 00000000000..5c3b4bfff7a
Binary files /dev/null and b/lib/jetty-ext/jasper-runtime.jar differ
diff --git a/lib/jetty-ext/jsp-api.jar b/lib/jetty-ext/jsp-api.jar
new file mode 100644
index 00000000000..93f89576f0a
Binary files /dev/null and b/lib/jetty-ext/jsp-api.jar differ
diff --git a/lib/junit-3.8.1.jar b/lib/junit-3.8.1.jar
new file mode 100644
index 00000000000..674d71e89ea
Binary files /dev/null and b/lib/junit-3.8.1.jar differ
diff --git a/lib/log4j-1.2.13.jar b/lib/log4j-1.2.13.jar
new file mode 100644
index 00000000000..dde9972109d
Binary files /dev/null and b/lib/log4j-1.2.13.jar differ
diff --git a/lib/servlet-api.jar b/lib/servlet-api.jar
new file mode 100644
index 00000000000..c9dab30f948
Binary files /dev/null and b/lib/servlet-api.jar differ
diff --git a/lib/xmlenc-0.52.jar b/lib/xmlenc-0.52.jar
new file mode 100644
index 00000000000..ec568b4c9e0
Binary files /dev/null and b/lib/xmlenc-0.52.jar differ
diff --git a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index 2e0dc31bed1..c5304305c07 100644
--- a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -42,7 +42,7 @@ import org.apache.hadoop.hbase.io.TextSequence;
public class HColumnDescriptor implements WritableComparable {
// For future backward compatibility
- private static final byte COLUMN_DESCRIPTOR_VERSION = (byte)2;
+ private static final byte COLUMN_DESCRIPTOR_VERSION = (byte)1;
/** Legal family names can only contain 'word characters' and end in a colon. */
public static final Pattern LEGAL_FAMILY_NAME = Pattern.compile("\\w+:");
@@ -76,11 +76,6 @@ public class HColumnDescriptor implements WritableComparable {
*/
public static final boolean DEFAULT_IN_MEMORY = false;
- /**
- * Default setting for whether to use a block cache or not.
- */
- public static final boolean DEFAULT_BLOCK_CACHE_ENABLED = false;
-
/**
* Default maximum length of cell contents.
*/
@@ -100,8 +95,6 @@ public class HColumnDescriptor implements WritableComparable {
private CompressionType compressionType;
// Serve reads from in-memory cache
private boolean inMemory;
- // Serve reads from in-memory block cache
- private boolean blockCacheEnabled;
// Maximum value size
private int maxValueLength;
// True if bloom filter was specified
@@ -130,7 +123,6 @@ public class HColumnDescriptor implements WritableComparable {
this(columnName == null || columnName.length() <= 0?
new Text(): new Text(columnName),
DEFAULT_N_VERSIONS, DEFAULT_COMPRESSION_TYPE, DEFAULT_IN_MEMORY,
- DEFAULT_BLOCK_CACHE_ENABLED,
Integer.MAX_VALUE, DEFAULT_BLOOM_FILTER_DESCRIPTOR);
}
@@ -142,7 +134,6 @@ public class HColumnDescriptor implements WritableComparable {
* @param compression Compression type
* @param inMemory If true, column data should be kept in an HRegionServer's
* cache
- * @param blockCacheEnabled If true, MapFile blocks should be cached
* @param maxValueLength Restrict values to <= this value
* @param bloomFilter Enable the specified bloom filter for this column
*
@@ -153,7 +144,6 @@ public class HColumnDescriptor implements WritableComparable {
*/
public HColumnDescriptor(final Text name, final int maxVersions,
final CompressionType compression, final boolean inMemory,
- final boolean blockCacheEnabled,
final int maxValueLength, final BloomFilterDescriptor bloomFilter) {
String familyStr = name.toString();
// Test name if not null (It can be null when deserializing after
@@ -175,7 +165,6 @@ public class HColumnDescriptor implements WritableComparable {
}
this.maxVersions = maxVersions;
this.inMemory = inMemory;
- this.blockCacheEnabled = blockCacheEnabled;
this.maxValueLength = maxValueLength;
this.bloomFilter = bloomFilter;
this.bloomFilterSpecified = this.bloomFilter == null ? false : true;
@@ -223,13 +212,6 @@ public class HColumnDescriptor implements WritableComparable {
return this.inMemory;
}
- /**
- * @return True if MapFile blocks should be cached.
- */
- public boolean isBlockCacheEnabled() {
- return blockCacheEnabled;
- }
-
/**
* @return Maximum value length.
*/
@@ -252,7 +234,6 @@ public class HColumnDescriptor implements WritableComparable {
return "{name: " + tmp.substring(0, tmp.length() - 1) +
", max versions: " + maxVersions +
", compression: " + this.compressionType + ", in memory: " + inMemory +
- ", block cache enabled: " + blockCacheEnabled +
", max length: " + maxValueLength + ", bloom filter: " +
(bloomFilterSpecified ? bloomFilter.toString() : "none") + "}";
}
@@ -270,7 +251,6 @@ public class HColumnDescriptor implements WritableComparable {
result ^= Integer.valueOf(this.maxVersions).hashCode();
result ^= this.compressionType.hashCode();
result ^= Boolean.valueOf(this.inMemory).hashCode();
- result ^= Boolean.valueOf(this.blockCacheEnabled).hashCode();
result ^= Integer.valueOf(this.maxValueLength).hashCode();
result ^= Boolean.valueOf(this.bloomFilterSpecified).hashCode();
result ^= Byte.valueOf(this.versionNumber).hashCode();
@@ -297,10 +277,6 @@ public class HColumnDescriptor implements WritableComparable {
bloomFilter = new BloomFilterDescriptor();
bloomFilter.readFields(in);
}
-
- if (this.versionNumber > 1) {
- this.blockCacheEnabled = in.readBoolean();
- }
}
/** {@inheritDoc} */
@@ -316,8 +292,6 @@ public class HColumnDescriptor implements WritableComparable {
if(bloomFilterSpecified) {
bloomFilter.write(out);
}
-
- out.writeBoolean(this.blockCacheEnabled);
}
// Comparable
@@ -353,18 +327,6 @@ public class HColumnDescriptor implements WritableComparable {
}
}
- if(result == 0) {
- if(this.blockCacheEnabled == other.blockCacheEnabled) {
- result = 0;
-
- } else if(this.blockCacheEnabled) {
- result = -1;
-
- } else {
- result = 1;
- }
- }
-
if(result == 0) {
result = other.maxValueLength - this.maxValueLength;
}
diff --git a/src/java/org/apache/hadoop/hbase/HStore.java b/src/java/org/apache/hadoop/hbase/HStore.java
index b768eec456e..1a66f45709e 100644
--- a/src/java/org/apache/hadoop/hbase/HStore.java
+++ b/src/java/org/apache/hadoop/hbase/HStore.java
@@ -741,19 +741,9 @@ public class HStore implements HConstants {
// Finally, start up all the map readers! (There could be more than one
// since we haven't compacted yet.)
- boolean first = true;
for(Map.Entry e: this.storefiles.entrySet()) {
- if (first) {
- // Use a block cache (if configured) for the first reader only
- // so as to control memory usage.
- this.readers.put(e.getKey(),
- e.getValue().getReader(this.fs, this.bloomFilter,
- family.isBlockCacheEnabled()));
- first = false;
- } else {
- this.readers.put(e.getKey(),
- e.getValue().getReader(this.fs, this.bloomFilter));
- }
+ this.readers.put(e.getKey(),
+ e.getValue().getReader(this.fs, this.bloomFilter));
}
}
@@ -1570,10 +1560,7 @@ public class HStore implements HConstants {
// 6. Loading the new TreeMap.
Long orderVal = Long.valueOf(finalCompactedFile.loadInfo(fs));
this.readers.put(orderVal,
- // Use a block cache (if configured) for this reader since
- // it is the only one.
- finalCompactedFile.getReader(this.fs, this.bloomFilter,
- family.isBlockCacheEnabled()));
+ finalCompactedFile.getReader(this.fs, this.bloomFilter));
this.storefiles.put(orderVal, finalCompactedFile);
} catch (IOException e) {
e = RemoteExceptionHandler.checkIOException(e);
diff --git a/src/java/org/apache/hadoop/hbase/HStoreFile.java b/src/java/org/apache/hadoop/hbase/HStoreFile.java
index 390d921deff..5b8d7cfbc40 100644
--- a/src/java/org/apache/hadoop/hbase/HStoreFile.java
+++ b/src/java/org/apache/hadoop/hbase/HStoreFile.java
@@ -31,11 +31,9 @@ import java.util.Random;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.io.BlockFSInputStream;
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
import org.apache.hadoop.hbase.util.Writables;
import org.apache.hadoop.io.MapFile;
@@ -415,37 +413,17 @@ public class HStoreFile implements HConstants {
* @return MapFile.Reader
* @throws IOException
*/
- public MapFile.Reader getReader(final FileSystem fs,
- final Filter bloomFilter)
- throws IOException {
- return isReference()?
- new HStoreFile.HalfMapFileReader(fs, getMapFilePath(reference).toString(),
- conf, reference.getFileRegion(), reference.getMidkey(), bloomFilter):
- new BloomFilterMapFile.Reader(fs, getMapFilePath().toString(),
- conf, bloomFilter);
- }
-
- /**
- * Get reader for the store file map file.
- * Client is responsible for closing file when done.
- * @param fs
- * @param bloomFilter If null, no filtering is done.
- * @param blockCacheEnabled If true, MapFile blocks should be cached.
- * @return MapFile.Reader
- * @throws IOException
- */
public synchronized MapFile.Reader getReader(final FileSystem fs,
- final Filter bloomFilter, final boolean blockCacheEnabled)
+ final Filter bloomFilter)
throws IOException {
if (isReference()) {
return new HStoreFile.HalfMapFileReader(fs,
getMapFilePath(reference).toString(), conf,
- reference.getFileRegion(), reference.getMidkey(), bloomFilter,
- blockCacheEnabled);
+ reference.getFileRegion(), reference.getMidkey(), bloomFilter);
}
return new BloomFilterMapFile.Reader(fs, getMapFilePath().toString(),
- conf, bloomFilter, blockCacheEnabled);
+ conf, bloomFilter);
}
/**
@@ -606,13 +584,8 @@ public class HStoreFile implements HConstants {
*/
static class HbaseMapFile extends MapFile {
- /**
- * A reader capable of reading and caching blocks of the data file.
- */
static class HbaseReader extends MapFile.Reader {
- private final boolean blockCacheEnabled;
-
/**
* @param fs
* @param dirName
@@ -621,23 +594,7 @@ public class HStoreFile implements HConstants {
*/
public HbaseReader(FileSystem fs, String dirName, Configuration conf)
throws IOException {
- this(fs, dirName, conf, false);
- }
-
- /**
- * @param fs
- * @param dirName
- * @param conf
- * @param blockCacheEnabled
- * @throws IOException
- */
- public HbaseReader(FileSystem fs, String dirName, Configuration conf,
- boolean blockCacheEnabled)
- throws IOException {
- super(fs, dirName, null, conf, false); // defer opening streams
- this.blockCacheEnabled = blockCacheEnabled;
- open(fs, dirName, null, conf);
-
+ super(fs, dirName, conf);
// Force reading of the mapfile index by calling midKey.
// Reading the index will bring the index into memory over
// here on the client and then close the index file freeing
@@ -648,28 +605,6 @@ public class HStoreFile implements HConstants {
// using up datanode resources. See HADOOP-2341.
midKey();
}
-
- @Override
- protected org.apache.hadoop.io.SequenceFile.Reader createDataFileReader(
- FileSystem fs, Path dataFile, Configuration conf)
- throws IOException {
- if (!blockCacheEnabled) {
- return super.createDataFileReader(fs, dataFile, conf);
- }
- LOG.info("Block Cache enabled");
- final int blockSize = conf.getInt("hbase.hstore.blockCache.blockSize",
- 64 * 1024);
- return new SequenceFile.Reader(fs, dataFile, conf) {
- @Override
- protected FSDataInputStream openFile(FileSystem fs, Path file,
- int bufferSize, long length) throws IOException {
-
- return new FSDataInputStream(new BlockFSInputStream(
- super.openFile(fs, file, bufferSize, length), length,
- blockSize));
- }
- };
- }
}
static class HbaseWriter extends MapFile.Writer {
@@ -718,13 +653,6 @@ public class HStoreFile implements HConstants {
bloomFilter = filter;
}
- public Reader(FileSystem fs, String dirName, Configuration conf,
- final Filter filter, final boolean blockCacheEnabled)
- throws IOException {
- super(fs, dirName, conf, blockCacheEnabled);
- bloomFilter = filter;
- }
-
/** {@inheritDoc} */
@Override
public Writable get(WritableComparable key, Writable val)
@@ -817,7 +745,7 @@ public class HStoreFile implements HConstants {
final Configuration conf, final Range r,
final WritableComparable midKey)
throws IOException {
- this(fs, dirName, conf, r, midKey, null, false);
+ this(fs, dirName, conf, r, midKey, null);
}
HalfMapFileReader(final FileSystem fs, final String dirName,
@@ -829,16 +757,6 @@ public class HStoreFile implements HConstants {
midkey = midKey;
}
- HalfMapFileReader(final FileSystem fs, final String dirName,
- final Configuration conf, final Range r,
- final WritableComparable midKey, final Filter filter,
- final boolean blockCacheEnabled)
- throws IOException {
- super(fs, dirName, conf, filter, blockCacheEnabled);
- top = isTopFileRegion(r);
- midkey = midKey;
- }
-
@SuppressWarnings("unchecked")
private void checkKey(final WritableComparable key)
throws IOException {
diff --git a/src/java/org/apache/hadoop/hbase/HTableDescriptor.java b/src/java/org/apache/hadoop/hbase/HTableDescriptor.java
index 9b763f635cd..a77f745f250 100644
--- a/src/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/src/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -43,15 +43,15 @@ public class HTableDescriptor implements WritableComparable {
public static final HTableDescriptor rootTableDesc =
new HTableDescriptor(HConstants.ROOT_TABLE_NAME,
new HColumnDescriptor(HConstants.COLUMN_FAMILY, 1,
- HColumnDescriptor.CompressionType.NONE, false, false,
- Integer.MAX_VALUE, null));
+ HColumnDescriptor.CompressionType.NONE, false, Integer.MAX_VALUE,
+ null));
/** table descriptor for meta table */
public static final HTableDescriptor metaTableDesc =
new HTableDescriptor(HConstants.META_TABLE_NAME,
new HColumnDescriptor(HConstants.COLUMN_FAMILY, 1,
- HColumnDescriptor.CompressionType.NONE, false, false,
- Integer.MAX_VALUE, null));
+ HColumnDescriptor.CompressionType.NONE, false, Integer.MAX_VALUE,
+ null));
private boolean rootregion;
private boolean metaregion;
@@ -256,4 +256,4 @@ public class HTableDescriptor implements WritableComparable {
public static Path getTableDir(Path rootdir, Text tableName) {
return new Path(rootdir, tableName.toString());
}
-}
\ No newline at end of file
+}
diff --git a/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java b/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java
index d3ad067437c..8a8355910aa 100644
--- a/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java
+++ b/src/java/org/apache/hadoop/hbase/hql/AlterCommand.java
@@ -202,8 +202,6 @@ public class AlterCommand extends SchemaModificationCommand {
.get(spec)).toUpperCase());
} else if (spec.equals("IN_MEMORY")) {
inMemory = (Boolean) columnSpec.get(spec);
- } else if (spec.equals("BLOCK_CACHE_ENABLED")) {
- blockCacheEnabled = (Boolean) columnSpec.get(spec);
} else if (spec.equals("BLOOMFILTER")) {
bloomFilterType = BloomFilterType.valueOf(((String) columnSpec.get(spec))
.toUpperCase());
@@ -231,8 +229,7 @@ public class AlterCommand extends SchemaModificationCommand {
column = appendDelimiter(column);
HColumnDescriptor columnDesc = new HColumnDescriptor(new Text(column),
- maxVersions, compression, inMemory, blockCacheEnabled,
- maxLength, bloomFilterDesc);
+ maxVersions, compression, inMemory, maxLength, bloomFilterDesc);
return columnDesc;
}
@@ -246,7 +243,6 @@ public class AlterCommand extends SchemaModificationCommand {
maxLength = original.getMaxValueLength();
compression = original.getCompression();
inMemory = original.isInMemory();
- blockCacheEnabled = original.isBlockCacheEnabled();
bloomFilterDesc = original.getBloomFilter();
}
}
diff --git a/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java b/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java
index b8384545413..30afcecc735 100644
--- a/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java
+++ b/src/java/org/apache/hadoop/hbase/hql/SchemaModificationCommand.java
@@ -37,7 +37,6 @@ public abstract class SchemaModificationCommand extends BasicCommand {
protected int maxLength;
protected HColumnDescriptor.CompressionType compression;
protected boolean inMemory;
- protected boolean blockCacheEnabled;
protected BloomFilterDescriptor bloomFilterDesc;
protected BloomFilterType bloomFilterType;
protected int vectorSize;
@@ -53,7 +52,6 @@ public abstract class SchemaModificationCommand extends BasicCommand {
maxLength = HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH;
compression = HColumnDescriptor.DEFAULT_COMPRESSION_TYPE;
inMemory = HColumnDescriptor.DEFAULT_IN_MEMORY;
- blockCacheEnabled = HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED;
bloomFilterDesc = HColumnDescriptor.DEFAULT_BLOOM_FILTER_DESCRIPTOR;
}
@@ -78,8 +76,6 @@ public abstract class SchemaModificationCommand extends BasicCommand {
.valueOf(((String) columnSpec.get(spec)).toUpperCase());
} else if (spec.equals("IN_MEMORY")) {
inMemory = (Boolean) columnSpec.get(spec);
- } else if (spec.equals("BLOCK_CACHE_ENABLED")) {
- blockCacheEnabled = (Boolean) columnSpec.get(spec);
} else if (spec.equals("BLOOMFILTER")) {
bloomFilterType = BloomFilterType.valueOf(((String) columnSpec.get(spec))
.toUpperCase());
@@ -107,8 +103,7 @@ public abstract class SchemaModificationCommand extends BasicCommand {
column = appendDelimiter(column);
HColumnDescriptor columnDesc = new HColumnDescriptor(new Text(column),
- maxVersions, compression, inMemory, blockCacheEnabled,
- maxLength, bloomFilterDesc);
+ maxVersions, compression, inMemory, maxLength, bloomFilterDesc);
return columnDesc;
}
diff --git a/src/test/org/apache/hadoop/hbase/HBaseTestCase.java b/src/test/org/apache/hadoop/hbase/HBaseTestCase.java
index 9571ecf4384..ce8df19b903 100644
--- a/src/test/org/apache/hadoop/hbase/HBaseTestCase.java
+++ b/src/test/org/apache/hadoop/hbase/HBaseTestCase.java
@@ -184,11 +184,11 @@ public abstract class HBaseTestCase extends TestCase {
final int versions) {
HTableDescriptor htd = new HTableDescriptor(name);
htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME1), versions,
- CompressionType.NONE, false, false, Integer.MAX_VALUE, null));
+ CompressionType.NONE, false, Integer.MAX_VALUE, null));
htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME2), versions,
- CompressionType.NONE, false, false, Integer.MAX_VALUE, null));
+ CompressionType.NONE, false, Integer.MAX_VALUE, null));
htd.addFamily(new HColumnDescriptor(new Text(COLFAMILY_NAME3), versions,
- CompressionType.NONE, false, false, Integer.MAX_VALUE, null));
+ CompressionType.NONE, false, Integer.MAX_VALUE, null));
return htd;
}
@@ -565,4 +565,4 @@ public abstract class HBaseTestCase extends TestCase {
return this.table.get(row, column, ts, versions);
}
}
-}
\ No newline at end of file
+}
diff --git a/src/test/org/apache/hadoop/hbase/TestBloomFilters.java b/src/test/org/apache/hadoop/hbase/TestBloomFilters.java
index c505c87647e..f0d19ff6bf1 100644
--- a/src/test/org/apache/hadoop/hbase/TestBloomFilters.java
+++ b/src/test/org/apache/hadoop/hbase/TestBloomFilters.java
@@ -169,7 +169,6 @@ public class TestBloomFilters extends HBaseClusterTestCase {
1, // Max versions
HColumnDescriptor.CompressionType.NONE, // no compression
HColumnDescriptor.DEFAULT_IN_MEMORY, // not in memory
- HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED,
HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH,
bloomFilter
)
@@ -235,7 +234,6 @@ public class TestBloomFilters extends HBaseClusterTestCase {
1, // Max versions
HColumnDescriptor.CompressionType.NONE, // no compression
HColumnDescriptor.DEFAULT_IN_MEMORY, // not in memory
- HColumnDescriptor.DEFAULT_BLOCK_CACHE_ENABLED,
HColumnDescriptor.DEFAULT_MAX_VALUE_LENGTH,
bloomFilter
)
diff --git a/src/test/org/apache/hadoop/hbase/TestTimestamp.java b/src/test/org/apache/hadoop/hbase/TestTimestamp.java
index 483518f231b..35788dad225 100644
--- a/src/test/org/apache/hadoop/hbase/TestTimestamp.java
+++ b/src/test/org/apache/hadoop/hbase/TestTimestamp.java
@@ -337,7 +337,7 @@ public class TestTimestamp extends HBaseTestCase {
private HRegion createRegion() throws IOException {
HTableDescriptor htd = createTableDescriptor(getName());
htd.addFamily(new HColumnDescriptor(COLUMN, VERSIONS,
- CompressionType.NONE, false, false, Integer.MAX_VALUE, null));
+ CompressionType.NONE, false, Integer.MAX_VALUE, null));
return createNewHRegion(htd, null, null);
}
-}
\ No newline at end of file
+}
diff --git a/src/test/org/apache/hadoop/hbase/TestToString.java b/src/test/org/apache/hadoop/hbase/TestToString.java
index 4c1ea3da17b..99adc65764d 100644
--- a/src/test/org/apache/hadoop/hbase/TestToString.java
+++ b/src/test/org/apache/hadoop/hbase/TestToString.java
@@ -44,9 +44,8 @@ public class TestToString extends TestCase {
HTableDescriptor htd = HTableDescriptor.rootTableDesc;
System. out.println(htd.toString());
assertEquals("Table descriptor", "name: -ROOT-, families: {info:={name: " +
- "info, max versions: 1, compression: NONE, in memory: false, " +
- "block cache enabled: false, max length: 2147483647, " +
- "bloom filter: none}}", htd.toString());
+ "info, max versions: 1, compression: NONE, in memory: false, max " +
+ "length: 2147483647, bloom filter: none}}", htd.toString());
}
/**