HDFS-11006. Ozone: support setting chunk size in streaming API. Contributed by Yiqun Lin.
This commit is contained in:
parent
cd361fa490
commit
57b2a9b17d
|
@ -45,7 +45,11 @@ public final class ScmConfigKeys {
|
|||
= "GRPC";
|
||||
|
||||
// TODO : this is copied from OzoneConsts, may need to move to a better place
|
||||
public static final int CHUNK_SIZE = 1 * 1024 * 1024; // 1 MB
|
||||
public static final String OZONE_SCM_CHUNK_SIZE_KEY = "ozone.scm.chunk.size";
|
||||
// 1 MB by default
|
||||
public static final int OZONE_SCM_CHUNK_SIZE_DEFAULT = 1 * 1024 * 1024;
|
||||
public static final int OZONE_SCM_CHUNK_MAX_SIZE = 1 * 1024 * 1024;
|
||||
|
||||
|
||||
public static final int OZONE_SCM_CLIENT_PORT_DEFAULT = 9860;
|
||||
public static final int OZONE_SCM_DATANODE_PORT_DEFAULT = 9861;
|
||||
|
|
|
@ -31,7 +31,6 @@ import com.google.protobuf.ByteString;
|
|||
import org.apache.hadoop.hdfs.ozone.protocol.proto.ContainerProtos.ChunkInfo;
|
||||
import org.apache.hadoop.hdfs.ozone.protocol.proto.ContainerProtos.KeyData;
|
||||
import org.apache.hadoop.hdfs.ozone.protocol.proto.ContainerProtos.KeyValue;
|
||||
import org.apache.hadoop.scm.ScmConfigKeys;
|
||||
import org.apache.hadoop.scm.XceiverClientManager;
|
||||
import org.apache.hadoop.scm.XceiverClientSpi;
|
||||
|
||||
|
@ -62,6 +61,7 @@ public class ChunkOutputStream extends OutputStream {
|
|||
private ByteBuffer buffer;
|
||||
private final String streamId;
|
||||
private int chunkIndex;
|
||||
private int chunkSize;
|
||||
|
||||
/**
|
||||
* Creates a new ChunkOutputStream.
|
||||
|
@ -71,13 +71,15 @@ public class ChunkOutputStream extends OutputStream {
|
|||
* @param xceiverClientManager client manager that controls client
|
||||
* @param xceiverClient client to perform container calls
|
||||
* @param traceID container protocol call args
|
||||
* @param chunkSize chunk size
|
||||
*/
|
||||
public ChunkOutputStream(String containerKey, String key,
|
||||
XceiverClientManager xceiverClientManager, XceiverClientSpi xceiverClient,
|
||||
String traceID) {
|
||||
String traceID, int chunkSize) {
|
||||
this.containerKey = containerKey;
|
||||
this.key = key;
|
||||
this.traceID = traceID;
|
||||
this.chunkSize = chunkSize;
|
||||
KeyValue keyValue = KeyValue.newBuilder()
|
||||
.setKey("TYPE").setValue("KEY").build();
|
||||
this.containerKeyData = KeyData.newBuilder()
|
||||
|
@ -86,7 +88,7 @@ public class ChunkOutputStream extends OutputStream {
|
|||
.addMetadata(keyValue);
|
||||
this.xceiverClientManager = xceiverClientManager;
|
||||
this.xceiverClient = xceiverClient;
|
||||
this.buffer = ByteBuffer.allocate(ScmConfigKeys.CHUNK_SIZE);
|
||||
this.buffer = ByteBuffer.allocate(chunkSize);
|
||||
this.streamId = UUID.randomUUID().toString();
|
||||
this.chunkIndex = 0;
|
||||
}
|
||||
|
@ -97,7 +99,7 @@ public class ChunkOutputStream extends OutputStream {
|
|||
int rollbackPosition = buffer.position();
|
||||
int rollbackLimit = buffer.limit();
|
||||
buffer.put((byte)b);
|
||||
if (buffer.position() == ScmConfigKeys.CHUNK_SIZE) {
|
||||
if (buffer.position() == chunkSize) {
|
||||
flushBufferToChunk(rollbackPosition, rollbackLimit);
|
||||
}
|
||||
}
|
||||
|
@ -116,12 +118,11 @@ public class ChunkOutputStream extends OutputStream {
|
|||
}
|
||||
checkOpen();
|
||||
while (len > 0) {
|
||||
int writeLen = Math.min(
|
||||
ScmConfigKeys.CHUNK_SIZE - buffer.position(), len);
|
||||
int writeLen = Math.min(chunkSize - buffer.position(), len);
|
||||
int rollbackPosition = buffer.position();
|
||||
int rollbackLimit = buffer.limit();
|
||||
buffer.put(b, off, writeLen);
|
||||
if (buffer.position() == ScmConfigKeys.CHUNK_SIZE) {
|
||||
if (buffer.position() == chunkSize) {
|
||||
flushBufferToChunk(rollbackPosition, rollbackLimit);
|
||||
}
|
||||
off += writeLen;
|
||||
|
|
|
@ -26,6 +26,7 @@ import org.apache.hadoop.hdfs.server.datanode.fsdataset.LengthInputStream;
|
|||
import org.apache.hadoop.ozone.OzoneConfiguration;
|
||||
import org.apache.hadoop.ozone.OzoneConsts;
|
||||
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||
import org.apache.hadoop.scm.ScmConfigKeys;
|
||||
import org.apache.hadoop.scm.XceiverClientManager;
|
||||
import org.apache.hadoop.scm.protocol.LocatedContainer;
|
||||
import org.apache.hadoop.scm.protocolPB.StorageContainerLocationProtocolClientSideTranslatorPB;
|
||||
|
@ -40,6 +41,8 @@ import org.apache.hadoop.scm.XceiverClientSpi;
|
|||
import org.apache.hadoop.scm.storage.ChunkInputStream;
|
||||
import org.apache.hadoop.scm.storage.ChunkOutputStream;
|
||||
import org.apache.hadoop.util.StringUtils;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
|
@ -55,11 +58,15 @@ import static org.apache.hadoop.scm.storage.ContainerProtocolCalls.putKey;
|
|||
* across the nodes of an HDFS cluster.
|
||||
*/
|
||||
public final class DistributedStorageHandler implements StorageHandler {
|
||||
private static final Logger LOG =
|
||||
LoggerFactory.getLogger(DistributedStorageHandler.class);
|
||||
|
||||
private final StorageContainerLocationProtocolClientSideTranslatorPB
|
||||
storageContainerLocation;
|
||||
private final XceiverClientManager xceiverClientManager;
|
||||
|
||||
private int chunkSize;
|
||||
|
||||
/**
|
||||
* Creates a new DistributedStorageHandler.
|
||||
*
|
||||
|
@ -71,6 +78,16 @@ public final class DistributedStorageHandler implements StorageHandler {
|
|||
storageContainerLocation) {
|
||||
this.storageContainerLocation = storageContainerLocation;
|
||||
this.xceiverClientManager = new XceiverClientManager(conf);
|
||||
|
||||
chunkSize = conf.getInt(ScmConfigKeys.OZONE_SCM_CHUNK_SIZE_KEY,
|
||||
ScmConfigKeys.OZONE_SCM_CHUNK_SIZE_DEFAULT);
|
||||
if(chunkSize > ScmConfigKeys.OZONE_SCM_CHUNK_MAX_SIZE) {
|
||||
LOG.warn("The chunk size ({}) is not allowed to be more than"
|
||||
+ " the maximum size ({}),"
|
||||
+ " resetting to the maximum size.",
|
||||
chunkSize, ScmConfigKeys.OZONE_SCM_CHUNK_MAX_SIZE);
|
||||
chunkSize = ScmConfigKeys.OZONE_SCM_CHUNK_MAX_SIZE;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -227,7 +244,8 @@ public final class DistributedStorageHandler implements StorageHandler {
|
|||
key.setCreatedOn(dateToString(new Date()));
|
||||
XceiverClientSpi xceiverClient = acquireXceiverClient(containerKey);
|
||||
return new ChunkOutputStream(containerKey, key.getKeyName(),
|
||||
xceiverClientManager, xceiverClient, args.getRequestID());
|
||||
xceiverClientManager, xceiverClient, args.getRequestID(),
|
||||
chunkSize);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
|
|
@ -110,4 +110,16 @@
|
|||
The default is appropriate for small clusters (tens of nodes).
|
||||
</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>ozone.scm.chunk.size</name>
|
||||
<value>1048576</value>
|
||||
<description>
|
||||
The chunk size for read/write chunk operations in bytes.
|
||||
|
||||
The chunk size defaults to 1MB. If the value configured is more
|
||||
than the maximum size (1MB), it will be reset to the maximum
|
||||
size.
|
||||
</description>
|
||||
</property>
|
||||
</configuration>
|
||||
|
|
Loading…
Reference in New Issue