Backport of HADOOP-8423 from trunk. svn merge -c 1359866 ../../trunk (harsh)

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1359869 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Harsh J 2012-07-10 19:59:44 +00:00
parent 1ed7052a82
commit 504353c259
3 changed files with 52 additions and 0 deletions

View File

@ -63,6 +63,10 @@ Release 2.0.1-alpha - UNRELEASED
HADOOP-8533. Remove parallel call ununsed capability in RPC. HADOOP-8533. Remove parallel call ununsed capability in RPC.
(Brandon Li via suresh) (Brandon Li via suresh)
HADOOP-8423. MapFile.Reader.get() crashes jvm or throws
EOFException on Snappy or LZO block-compressed data
(todd via harsh)
BUG FIXES BUG FIXES
HADOOP-8372. NetUtils.normalizeHostName() incorrectly handles hostname HADOOP-8372. NetUtils.normalizeHostName() incorrectly handles hostname

View File

@ -127,6 +127,8 @@ public class BlockDecompressorStream extends DecompressorStream {
} }
public void resetState() throws IOException { public void resetState() throws IOException {
originalBlockSize = 0;
noUncompressedBytes = 0;
super.resetState(); super.resetState();
} }

View File

@ -46,6 +46,7 @@ import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.DataInputBuffer; import org.apache.hadoop.io.DataInputBuffer;
import org.apache.hadoop.io.DataOutputBuffer; import org.apache.hadoop.io.DataOutputBuffer;
import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.io.MapFile;
import org.apache.hadoop.io.RandomDatum; import org.apache.hadoop.io.RandomDatum;
import org.apache.hadoop.io.SequenceFile; import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Text;
@ -68,6 +69,7 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.junit.Assert; import org.junit.Assert;
import org.junit.Assume;
import org.junit.Test; import org.junit.Test;
import static org.junit.Assert.*; import static org.junit.Assert.*;
@ -514,6 +516,50 @@ public class TestCodec {
LOG.info("SUCCESS! Completed SequenceFileCodecTest with codec \"" + codecClass + "\""); LOG.info("SUCCESS! Completed SequenceFileCodecTest with codec \"" + codecClass + "\"");
} }
/**
* Regression test for HADOOP-8423: seeking in a block-compressed
* stream would not properly reset the block decompressor state.
*/
@Test
public void testSnappyMapFile() throws Exception {
Assume.assumeTrue(SnappyCodec.isNativeCodeLoaded());
codecTestMapFile(SnappyCodec.class, CompressionType.BLOCK, 100);
}
private void codecTestMapFile(Class<? extends CompressionCodec> clazz,
CompressionType type, int records) throws Exception {
FileSystem fs = FileSystem.get(conf);
LOG.info("Creating MapFiles with " + records +
" records using codec " + clazz.getSimpleName());
Path path = new Path(new Path(
System.getProperty("test.build.data", "/tmp")),
clazz.getSimpleName() + "-" + type + "-" + records);
LOG.info("Writing " + path);
createMapFile(conf, fs, path, clazz.newInstance(), type, records);
MapFile.Reader reader = new MapFile.Reader(path, conf);
Text key1 = new Text("002");
assertNotNull(reader.get(key1, new Text()));
Text key2 = new Text("004");
assertNotNull(reader.get(key2, new Text()));
}
private static void createMapFile(Configuration conf, FileSystem fs, Path path,
CompressionCodec codec, CompressionType type, int records) throws IOException {
MapFile.Writer writer =
new MapFile.Writer(conf, path,
MapFile.Writer.keyClass(Text.class),
MapFile.Writer.valueClass(Text.class),
MapFile.Writer.compression(type, codec));
Text key = new Text();
for (int j = 0; j < records; j++) {
key.set(String.format("%03d", j));
writer.append(key, key);
}
writer.close();
}
public static void main(String[] args) throws IOException { public static void main(String[] args) throws IOException {
int count = 10000; int count = 10000;
String codecClass = "org.apache.hadoop.io.compress.DefaultCodec"; String codecClass = "org.apache.hadoop.io.compress.DefaultCodec";