HBASE-1784 Missing rows after medium intensity insert
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@810310 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
231e0c0043
commit
9c1869648c
|
@ -967,9 +967,8 @@ public class HFile {
|
||||||
private ByteBuffer decompress(final long offset, final int compressedSize,
|
private ByteBuffer decompress(final long offset, final int compressedSize,
|
||||||
final int decompressedSize)
|
final int decompressedSize)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
|
|
||||||
Decompressor decompressor = null;
|
Decompressor decompressor = null;
|
||||||
|
ByteBuffer buf = null;
|
||||||
try {
|
try {
|
||||||
decompressor = this.compressAlgo.getDecompressor();
|
decompressor = this.compressAlgo.getDecompressor();
|
||||||
// My guess is that the bounded range fis is needed to stop the
|
// My guess is that the bounded range fis is needed to stop the
|
||||||
|
@ -979,7 +978,7 @@ public class HFile {
|
||||||
InputStream is = this.compressAlgo.createDecompressionStream(
|
InputStream is = this.compressAlgo.createDecompressionStream(
|
||||||
new BoundedRangeFileInputStream(this.istream, offset, compressedSize),
|
new BoundedRangeFileInputStream(this.istream, offset, compressedSize),
|
||||||
decompressor, 0);
|
decompressor, 0);
|
||||||
ByteBuffer buf = ByteBuffer.allocate(decompressedSize);
|
buf = ByteBuffer.allocate(decompressedSize);
|
||||||
IOUtils.readFully(is, buf.array(), 0, buf.capacity());
|
IOUtils.readFully(is, buf.array(), 0, buf.capacity());
|
||||||
is.close();
|
is.close();
|
||||||
} finally {
|
} finally {
|
||||||
|
|
Loading…
Reference in New Issue