HBASE-19651 Remove LimitInputStream
This commit is contained in:
parent
2bf2b5b2c1
commit
a30d9fe8d9
|
@ -39,7 +39,6 @@ Licensed under the Apache License v2.0 as a part of the Bootstrap project.
|
||||||
|
|
||||||
--
|
--
|
||||||
This product includes portions of the Guava project v14 and v21, specifically
|
This product includes portions of the Guava project v14 and v21, specifically
|
||||||
'hbase-common/src/main/java/org/apache/hadoop/hbase/io/LimitInputStream.java'
|
|
||||||
'hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java'
|
'hbase-common/src/main/java/org/apache/hadoop/hbase/util/Bytes.java'
|
||||||
'hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java'
|
'hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java'
|
||||||
|
|
||||||
|
|
|
@ -85,7 +85,6 @@ import org.apache.hadoop.hbase.client.security.SecurityCapability;
|
||||||
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
||||||
import org.apache.hadoop.hbase.filter.ByteArrayComparable;
|
import org.apache.hadoop.hbase.filter.ByteArrayComparable;
|
||||||
import org.apache.hadoop.hbase.filter.Filter;
|
import org.apache.hadoop.hbase.filter.Filter;
|
||||||
import org.apache.hadoop.hbase.io.LimitInputStream;
|
|
||||||
import org.apache.hadoop.hbase.io.TimeRange;
|
import org.apache.hadoop.hbase.io.TimeRange;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufMagic;
|
import org.apache.hadoop.hbase.protobuf.ProtobufMagic;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufMessageConverter;
|
import org.apache.hadoop.hbase.protobuf.ProtobufMessageConverter;
|
||||||
|
@ -106,6 +105,7 @@ import org.apache.hadoop.hbase.util.VersionInfo;
|
||||||
import org.apache.hadoop.ipc.RemoteException;
|
import org.apache.hadoop.ipc.RemoteException;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
|
||||||
|
import org.apache.hbase.thirdparty.com.google.common.io.ByteStreams;
|
||||||
import org.apache.hbase.thirdparty.com.google.gson.JsonArray;
|
import org.apache.hbase.thirdparty.com.google.gson.JsonArray;
|
||||||
import org.apache.hbase.thirdparty.com.google.gson.JsonElement;
|
import org.apache.hbase.thirdparty.com.google.gson.JsonElement;
|
||||||
import org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
|
import org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
|
||||||
|
@ -2609,7 +2609,7 @@ public final class ProtobufUtil {
|
||||||
final int firstByte = in.read();
|
final int firstByte = in.read();
|
||||||
if (firstByte != -1) {
|
if (firstByte != -1) {
|
||||||
final int size = CodedInputStream.readRawVarint32(firstByte, in);
|
final int size = CodedInputStream.readRawVarint32(firstByte, in);
|
||||||
final InputStream limitedInput = new LimitInputStream(in, size);
|
final InputStream limitedInput = ByteStreams.limit(in, size);
|
||||||
final CodedInputStream codedInput = CodedInputStream.newInstance(limitedInput);
|
final CodedInputStream codedInput = CodedInputStream.newInstance(limitedInput);
|
||||||
codedInput.setSizeLimit(size);
|
codedInput.setSizeLimit(size);
|
||||||
builder.mergeFrom(codedInput);
|
builder.mergeFrom(codedInput);
|
||||||
|
|
|
@ -1,7 +0,0 @@
|
||||||
--
|
|
||||||
This product includes portions of the Guava project v14, specifically
|
|
||||||
'hbase-common/src/main/java/org/apache/hadoop/hbase/io/LimitInputStream.java'
|
|
||||||
|
|
||||||
Copyright (C) 2007 The Guava Authors
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0
|
|
|
@ -1,105 +0,0 @@
|
||||||
/*
|
|
||||||
* Copyright (C) 2007 The Guava Authors
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.apache.hadoop.hbase.io;
|
|
||||||
|
|
||||||
import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkArgument;
|
|
||||||
import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkNotNull;
|
|
||||||
|
|
||||||
import java.io.FilterInputStream;
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.io.InputStream;
|
|
||||||
|
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Copied from guava source code v15 (LimitedInputStream)
|
|
||||||
* Guava deprecated LimitInputStream in v14 and removed it in v15. Copying this class here
|
|
||||||
* allows to be compatible with guava 11 to 15+.
|
|
||||||
*/
|
|
||||||
@InterfaceAudience.Private
|
|
||||||
public final class LimitInputStream extends FilterInputStream {
|
|
||||||
private long left;
|
|
||||||
private long mark = -1;
|
|
||||||
|
|
||||||
public LimitInputStream(InputStream in, long limit) {
|
|
||||||
super(in);
|
|
||||||
checkNotNull(in);
|
|
||||||
checkArgument(limit >= 0, "limit must be non-negative");
|
|
||||||
left = limit;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int available() throws IOException {
|
|
||||||
return (int) Math.min(in.available(), left);
|
|
||||||
}
|
|
||||||
|
|
||||||
// it's okay to mark even if mark isn't supported, as reset won't work
|
|
||||||
@Override
|
|
||||||
public synchronized void mark(int readLimit) {
|
|
||||||
in.mark(readLimit);
|
|
||||||
mark = left;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int read() throws IOException {
|
|
||||||
if (left == 0) {
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
int result = in.read();
|
|
||||||
if (result != -1) {
|
|
||||||
--left;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int read(byte[] b, int off, int len) throws IOException {
|
|
||||||
if (left == 0) {
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
len = (int) Math.min(len, left);
|
|
||||||
int result = in.read(b, off, len);
|
|
||||||
if (result != -1) {
|
|
||||||
left -= result;
|
|
||||||
}
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public synchronized void reset() throws IOException {
|
|
||||||
if (!in.markSupported()) {
|
|
||||||
throw new IOException("Mark not supported");
|
|
||||||
}
|
|
||||||
if (mark == -1) {
|
|
||||||
throw new IOException("Mark not set");
|
|
||||||
}
|
|
||||||
|
|
||||||
in.reset();
|
|
||||||
left = mark;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public long skip(long n) throws IOException {
|
|
||||||
n = Math.min(n, left);
|
|
||||||
long skipped = in.skip(n);
|
|
||||||
left -= skipped;
|
|
||||||
return skipped;
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -26,15 +26,11 @@ import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.FSDataInputStream;
|
import org.apache.hadoop.fs.FSDataInputStream;
|
||||||
import org.apache.hadoop.hbase.codec.Codec;
|
import org.apache.hadoop.hbase.codec.Codec;
|
||||||
import org.apache.hadoop.hbase.io.LimitInputStream;
|
|
||||||
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
|
||||||
|
@ -43,7 +39,12 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALKey;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALTrailer;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.WALTrailer;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
||||||
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import org.apache.hbase.thirdparty.com.google.common.io.ByteStreams;
|
||||||
import org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream;
|
import org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream;
|
||||||
import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
|
import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
|
||||||
|
|
||||||
|
@ -350,7 +351,7 @@ public class ProtobufLogReader extends ReaderBase {
|
||||||
"inputStream.available()= " + this.inputStream.available() + ", " +
|
"inputStream.available()= " + this.inputStream.available() + ", " +
|
||||||
"entry size= " + size + " at offset = " + this.inputStream.getPos());
|
"entry size= " + size + " at offset = " + this.inputStream.getPos());
|
||||||
}
|
}
|
||||||
ProtobufUtil.mergeFrom(builder, new LimitInputStream(this.inputStream, size),
|
ProtobufUtil.mergeFrom(builder, ByteStreams.limit(this.inputStream, size),
|
||||||
(int)size);
|
(int)size);
|
||||||
} catch (InvalidProtocolBufferException ipbe) {
|
} catch (InvalidProtocolBufferException ipbe) {
|
||||||
throw (EOFException) new EOFException("Invalid PB, EOF? Ignoring; originalPosition=" +
|
throw (EOFException) new EOFException("Invalid PB, EOF? Ignoring; originalPosition=" +
|
||||||
|
|
Loading…
Reference in New Issue