HBASE-1943 Remove AgileJSON; unused
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@831155 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
47feb6be86
commit
a5248f95fc
|
@ -145,6 +145,7 @@ Release 0.21.0 - Unreleased
|
||||||
HBASE-1936 HLog group commit
|
HBASE-1936 HLog group commit
|
||||||
HBASE-1921 When the Master's session times out and there's only one, cluster is wedged
|
HBASE-1921 When the Master's session times out and there's only one, cluster is wedged
|
||||||
HBASE-1942 Update hadoop jars in trunk; update to r831142
|
HBASE-1942 Update hadoop jars in trunk; update to r831142
|
||||||
|
HBASE-1943 Remove AgileJSON; unused
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
HBASE-410 [testing] Speed up the test suite
|
HBASE-410 [testing] Speed up the test suite
|
||||||
|
|
Binary file not shown.
|
@ -33,8 +33,6 @@ import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.io.Text;
|
import org.apache.hadoop.io.Text;
|
||||||
import org.apache.hadoop.io.WritableComparable;
|
import org.apache.hadoop.io.WritableComparable;
|
||||||
|
|
||||||
import agilejson.TOJSON;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* An HColumnDescriptor contains information about a column family such as the
|
* An HColumnDescriptor contains information about a column family such as the
|
||||||
* number of versions, compression settings, etc.
|
* number of versions, compression settings, etc.
|
||||||
|
@ -278,7 +276,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return Name of this column family
|
* @return Name of this column family
|
||||||
*/
|
*/
|
||||||
@TOJSON(fieldName = "name", base64=true)
|
|
||||||
public byte [] getName() {
|
public byte [] getName() {
|
||||||
return name;
|
return name;
|
||||||
}
|
}
|
||||||
|
@ -344,14 +341,12 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @return compression type being used for the column family */
|
/** @return compression type being used for the column family */
|
||||||
@TOJSON
|
|
||||||
public Compression.Algorithm getCompression() {
|
public Compression.Algorithm getCompression() {
|
||||||
String n = getValue(COMPRESSION);
|
String n = getValue(COMPRESSION);
|
||||||
return Compression.Algorithm.valueOf(n.toUpperCase());
|
return Compression.Algorithm.valueOf(n.toUpperCase());
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @return maximum number of versions */
|
/** @return maximum number of versions */
|
||||||
@TOJSON
|
|
||||||
public synchronized int getMaxVersions() {
|
public synchronized int getMaxVersions() {
|
||||||
if (this.cachedMaxVersions == -1) {
|
if (this.cachedMaxVersions == -1) {
|
||||||
String value = getValue(HConstants.VERSIONS);
|
String value = getValue(HConstants.VERSIONS);
|
||||||
|
@ -371,7 +366,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return Blocksize.
|
* @return Blocksize.
|
||||||
*/
|
*/
|
||||||
@TOJSON
|
|
||||||
public synchronized int getBlocksize() {
|
public synchronized int getBlocksize() {
|
||||||
if (this.blocksize == null) {
|
if (this.blocksize == null) {
|
||||||
String value = getValue(BLOCKSIZE);
|
String value = getValue(BLOCKSIZE);
|
||||||
|
@ -392,7 +386,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return Compression type setting.
|
* @return Compression type setting.
|
||||||
*/
|
*/
|
||||||
@TOJSON
|
|
||||||
public Compression.Algorithm getCompressionType() {
|
public Compression.Algorithm getCompressionType() {
|
||||||
return getCompression();
|
return getCompression();
|
||||||
}
|
}
|
||||||
|
@ -417,7 +410,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return True if we are to keep all in use HRegionServer cache.
|
* @return True if we are to keep all in use HRegionServer cache.
|
||||||
*/
|
*/
|
||||||
@TOJSON(prefixLength = 2)
|
|
||||||
public boolean isInMemory() {
|
public boolean isInMemory() {
|
||||||
String value = getValue(HConstants.IN_MEMORY);
|
String value = getValue(HConstants.IN_MEMORY);
|
||||||
if (value != null)
|
if (value != null)
|
||||||
|
@ -436,7 +428,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return Time-to-live of cell contents, in seconds.
|
* @return Time-to-live of cell contents, in seconds.
|
||||||
*/
|
*/
|
||||||
@TOJSON
|
|
||||||
public int getTimeToLive() {
|
public int getTimeToLive() {
|
||||||
String value = getValue(TTL);
|
String value = getValue(TTL);
|
||||||
return (value != null)? Integer.valueOf(value).intValue(): DEFAULT_TTL;
|
return (value != null)? Integer.valueOf(value).intValue(): DEFAULT_TTL;
|
||||||
|
@ -452,7 +443,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return True if MapFile blocks should be cached.
|
* @return True if MapFile blocks should be cached.
|
||||||
*/
|
*/
|
||||||
@TOJSON(prefixLength = 2)
|
|
||||||
public boolean isBlockCacheEnabled() {
|
public boolean isBlockCacheEnabled() {
|
||||||
String value = getValue(BLOCKCACHE);
|
String value = getValue(BLOCKCACHE);
|
||||||
if (value != null)
|
if (value != null)
|
||||||
|
@ -470,7 +460,6 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
|
||||||
/**
|
/**
|
||||||
* @return true if a bloom filter is enabled
|
* @return true if a bloom filter is enabled
|
||||||
*/
|
*/
|
||||||
@TOJSON(prefixLength = 2)
|
|
||||||
public boolean isBloomfilter() {
|
public boolean isBloomfilter() {
|
||||||
String value = getValue(BLOOMFILTER);
|
String value = getValue(BLOOMFILTER);
|
||||||
if (value != null)
|
if (value != null)
|
||||||
|
|
|
@ -36,8 +36,6 @@ import org.apache.hadoop.hbase.io.hfile.Compression;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.io.WritableComparable;
|
import org.apache.hadoop.io.WritableComparable;
|
||||||
|
|
||||||
import agilejson.TOJSON;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* HTableDescriptor contains the name of an HTable, and its
|
* HTableDescriptor contains the name of an HTable, and its
|
||||||
* column families.
|
* column families.
|
||||||
|
@ -369,7 +367,6 @@ public class HTableDescriptor implements WritableComparable<HTableDescriptor> {
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @return name of table */
|
/** @return name of table */
|
||||||
@TOJSON
|
|
||||||
public byte [] getName() {
|
public byte [] getName() {
|
||||||
return name;
|
return name;
|
||||||
}
|
}
|
||||||
|
@ -603,7 +600,6 @@ public class HTableDescriptor implements WritableComparable<HTableDescriptor> {
|
||||||
return Collections.unmodifiableSet(this.families.keySet());
|
return Collections.unmodifiableSet(this.families.keySet());
|
||||||
}
|
}
|
||||||
|
|
||||||
@TOJSON(fieldName = "columns")
|
|
||||||
public HColumnDescriptor[] getColumnFamilies() {
|
public HColumnDescriptor[] getColumnFamilies() {
|
||||||
return getFamilies().toArray(new HColumnDescriptor[0]);
|
return getFamilies().toArray(new HColumnDescriptor[0]);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue