HBASE-4302 Only run Snappy compression tests if Snappy is available

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1166316 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Todd Lipcon 2011-09-07 18:57:51 +00:00
parent f86382fa16
commit 73661b9b5e
2 changed files with 55 additions and 1 deletions

View File

@ -241,6 +241,8 @@ Release 0.91.0 - Unreleased
while table is being disabled (Ming Ma) while table is being disabled (Ming Ma)
HBASE-4338 Package build for rpm and deb are broken (Eric Yang) HBASE-4338 Package build for rpm and deb are broken (Eric Yang)
HBASE-4309 slow query log metrics spewing warnings (Riley Patterson) HBASE-4309 slow query log metrics spewing warnings (Riley Patterson)
HBASE-4302 Only run Snappy compression tests if Snappy is available
(Alejandro Abdelnur via todd)
IMPROVEMENTS IMPROVEMENTS
HBASE-3290 Max Compaction Size (Nicolas Spiegelberg via Stack) HBASE-3290 Max Compaction Size (Nicolas Spiegelberg via Stack)

View File

@ -20,9 +20,17 @@
package org.apache.hadoop.hbase.util; package org.apache.hadoop.hbase.util;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.io.hfile.Compression; import org.apache.hadoop.hbase.io.hfile.Compression;
import org.apache.hadoop.io.DataOutputBuffer;
import org.apache.hadoop.io.compress.CompressionCodec;
import org.apache.hadoop.io.compress.CompressionOutputStream;
import org.apache.hadoop.util.NativeCodeLoader;
import org.apache.hadoop.util.ReflectionUtils;
import org.junit.Test; import org.junit.Test;
import java.io.BufferedOutputStream;
import java.io.DataOutputStream;
import java.io.IOException; import java.io.IOException;
import static org.junit.Assert.*; import static org.junit.Assert.*;
@ -56,9 +64,53 @@ public class TestCompressionTest {
assertTrue(CompressionTest.testCompression("GZ")); assertTrue(CompressionTest.testCompression("GZ"));
if (isCompressionAvailable("org.apache.hadoop.io.compress.SnappyCodec")) { if (isCompressionAvailable("org.apache.hadoop.io.compress.SnappyCodec")) {
if (NativeCodeLoader.isNativeCodeLoaded()) {
try {
System.loadLibrary("snappy");
try {
Configuration conf = new Configuration();
CompressionCodec codec = (CompressionCodec)
ReflectionUtils.newInstance(
conf.getClassByName("org.apache.hadoop.io.compress.SnappyCodec"), conf);
DataOutputBuffer compressedDataBuffer = new DataOutputBuffer();
CompressionOutputStream deflateFilter =
codec.createOutputStream(compressedDataBuffer);
byte[] data = new byte[1024];
DataOutputStream deflateOut = new DataOutputStream(
new BufferedOutputStream(deflateFilter));
deflateOut.write(data, 0, data.length);
deflateOut.flush();
deflateFilter.finish();
// Snappy Codec class, Snappy nativelib and Hadoop nativelib with
// Snappy JNIs are present
assertTrue(CompressionTest.testCompression("SNAPPY")); assertTrue(CompressionTest.testCompression("SNAPPY"));
} }
catch (UnsatisfiedLinkError ex) {
// Hadoop nativelib does not have Snappy JNIs
// cannot assert the codec here because the current logic of
// CompressionTest checks only classloading, not the codec
// usage.
}
catch (Exception ex) {
}
}
catch (UnsatisfiedLinkError ex) {
// Snappy nativelib is not available
assertFalse(CompressionTest.testCompression("SNAPPY"));
}
}
else { else {
// Hadoop nativelib is not available
assertFalse(CompressionTest.testCompression("SNAPPY"));
}
}
else {
// Snappy Codec class is not available
assertFalse(CompressionTest.testCompression("SNAPPY")); assertFalse(CompressionTest.testCompression("SNAPPY"));
} }
} }