HDFS-9833. Erasure coding: recomputing block checksum on the fly by reconstructing the missed/corrupt block data. Contributed by Rakesh R.
This commit is contained in:
parent
8ceb06e239
commit
d749cf65e1
|
@ -460,7 +460,8 @@ final class FileChecksumHelper {
|
||||||
setRemaining(getRemaining() - block.getNumBytes());
|
setRemaining(getRemaining() - block.getNumBytes());
|
||||||
|
|
||||||
StripedBlockInfo stripedBlockInfo = new StripedBlockInfo(block,
|
StripedBlockInfo stripedBlockInfo = new StripedBlockInfo(block,
|
||||||
blockGroup.getLocations(), blockGroup.getBlockTokens(), ecPolicy);
|
blockGroup.getLocations(), blockGroup.getBlockTokens(),
|
||||||
|
blockGroup.getBlockIndices(), ecPolicy);
|
||||||
DatanodeInfo[] datanodes = blockGroup.getLocations();
|
DatanodeInfo[] datanodes = blockGroup.getLocations();
|
||||||
|
|
||||||
//try each datanode in the block group.
|
//try each datanode in the block group.
|
||||||
|
|
|
@ -32,14 +32,16 @@ public class StripedBlockInfo {
|
||||||
private final ExtendedBlock block;
|
private final ExtendedBlock block;
|
||||||
private final DatanodeInfo[] datanodes;
|
private final DatanodeInfo[] datanodes;
|
||||||
private final Token<BlockTokenIdentifier>[] blockTokens;
|
private final Token<BlockTokenIdentifier>[] blockTokens;
|
||||||
|
private final byte[] blockIndices;
|
||||||
private final ErasureCodingPolicy ecPolicy;
|
private final ErasureCodingPolicy ecPolicy;
|
||||||
|
|
||||||
public StripedBlockInfo(ExtendedBlock block, DatanodeInfo[] datanodes,
|
public StripedBlockInfo(ExtendedBlock block, DatanodeInfo[] datanodes,
|
||||||
Token<BlockTokenIdentifier>[] blockTokens,
|
Token<BlockTokenIdentifier>[] blockTokens, byte[] blockIndices,
|
||||||
ErasureCodingPolicy ecPolicy) {
|
ErasureCodingPolicy ecPolicy) {
|
||||||
this.block = block;
|
this.block = block;
|
||||||
this.datanodes = datanodes;
|
this.datanodes = datanodes;
|
||||||
this.blockTokens = blockTokens;
|
this.blockTokens = blockTokens;
|
||||||
|
this.blockIndices = blockIndices;
|
||||||
this.ecPolicy = ecPolicy;
|
this.ecPolicy = ecPolicy;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -55,6 +57,10 @@ public class StripedBlockInfo {
|
||||||
return blockTokens;
|
return blockTokens;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public byte[] getBlockIndices() {
|
||||||
|
return blockIndices;
|
||||||
|
}
|
||||||
|
|
||||||
public ErasureCodingPolicy getErasureCodingPolicy() {
|
public ErasureCodingPolicy getErasureCodingPolicy() {
|
||||||
return ecPolicy;
|
return ecPolicy;
|
||||||
}
|
}
|
||||||
|
|
|
@ -274,6 +274,8 @@ public class Sender implements DataTransferProtocol {
|
||||||
stripedBlockInfo.getDatanodes()))
|
stripedBlockInfo.getDatanodes()))
|
||||||
.addAllBlockTokens(PBHelperClient.convert(
|
.addAllBlockTokens(PBHelperClient.convert(
|
||||||
stripedBlockInfo.getBlockTokens()))
|
stripedBlockInfo.getBlockTokens()))
|
||||||
|
.addAllBlockIndices(PBHelperClient
|
||||||
|
.convertBlockIndices(stripedBlockInfo.getBlockIndices()))
|
||||||
.setEcPolicy(PBHelperClient.convertErasureCodingPolicy(
|
.setEcPolicy(PBHelperClient.convertErasureCodingPolicy(
|
||||||
stripedBlockInfo.getErasureCodingPolicy()))
|
stripedBlockInfo.getErasureCodingPolicy()))
|
||||||
.build();
|
.build();
|
||||||
|
|
|
@ -853,6 +853,22 @@ public class PBHelperClient {
|
||||||
return results;
|
return results;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static List<Integer> convertBlockIndices(byte[] blockIndices) {
|
||||||
|
List<Integer> results = new ArrayList<>(blockIndices.length);
|
||||||
|
for (byte bt : blockIndices) {
|
||||||
|
results.add(Integer.valueOf(bt));
|
||||||
|
}
|
||||||
|
return results;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static byte[] convertBlockIndices(List<Integer> blockIndices) {
|
||||||
|
byte[] blkIndices = new byte[blockIndices.size()];
|
||||||
|
for (int i = 0; i < blockIndices.size(); i++) {
|
||||||
|
blkIndices[i] = (byte) blockIndices.get(i).intValue();
|
||||||
|
}
|
||||||
|
return blkIndices;
|
||||||
|
}
|
||||||
|
|
||||||
public static BlockStoragePolicy convert(BlockStoragePolicyProto proto) {
|
public static BlockStoragePolicy convert(BlockStoragePolicyProto proto) {
|
||||||
List<StorageTypeProto> cList = proto.getCreationPolicy()
|
List<StorageTypeProto> cList = proto.getCreationPolicy()
|
||||||
.getStorageTypesList();
|
.getStorageTypesList();
|
||||||
|
|
|
@ -154,6 +154,7 @@ message OpBlockGroupChecksumProto {
|
||||||
// each internal block has a block token
|
// each internal block has a block token
|
||||||
repeated hadoop.common.TokenProto blockTokens = 3;
|
repeated hadoop.common.TokenProto blockTokens = 3;
|
||||||
required ErasureCodingPolicyProto ecPolicy = 4;
|
required ErasureCodingPolicyProto ecPolicy = 4;
|
||||||
|
repeated uint32 blockIndices = 5;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -306,6 +306,7 @@ public abstract class Receiver implements DataTransferProtocol {
|
||||||
PBHelperClient.convert(proto.getHeader().getBlock()),
|
PBHelperClient.convert(proto.getHeader().getBlock()),
|
||||||
PBHelperClient.convert(proto.getDatanodes()),
|
PBHelperClient.convert(proto.getDatanodes()),
|
||||||
PBHelperClient.convertTokens(proto.getBlockTokensList()),
|
PBHelperClient.convertTokens(proto.getBlockTokensList()),
|
||||||
|
PBHelperClient.convertBlockIndices(proto.getBlockIndicesList()),
|
||||||
PBHelperClient.convertErasureCodingPolicy(proto.getEcPolicy())
|
PBHelperClient.convertErasureCodingPolicy(proto.getEcPolicy())
|
||||||
);
|
);
|
||||||
|
|
||||||
|
|
|
@ -18,6 +18,8 @@
|
||||||
package org.apache.hadoop.hdfs.server.datanode;
|
package org.apache.hadoop.hdfs.server.datanode;
|
||||||
|
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hdfs.DFSUtilClient;
|
import org.apache.hadoop.hdfs.DFSUtilClient;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
||||||
|
@ -30,6 +32,8 @@ import org.apache.hadoop.hdfs.protocol.datatransfer.Sender;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos;
|
import org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos;
|
||||||
import org.apache.hadoop.hdfs.protocolPB.PBHelperClient;
|
import org.apache.hadoop.hdfs.protocolPB.PBHelperClient;
|
||||||
import org.apache.hadoop.hdfs.security.token.block.BlockTokenIdentifier;
|
import org.apache.hadoop.hdfs.security.token.block.BlockTokenIdentifier;
|
||||||
|
import org.apache.hadoop.hdfs.server.datanode.erasurecode.StripedBlockChecksumReconstructor;
|
||||||
|
import org.apache.hadoop.hdfs.server.datanode.erasurecode.StripedReconstructionInfo;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.fsdataset.LengthInputStream;
|
import org.apache.hadoop.hdfs.server.datanode.fsdataset.LengthInputStream;
|
||||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||||
import org.apache.hadoop.io.DataOutputBuffer;
|
import org.apache.hadoop.io.DataOutputBuffer;
|
||||||
|
@ -46,11 +50,14 @@ import java.io.DataOutputStream;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.InputStream;
|
import java.io.InputStream;
|
||||||
import java.security.MessageDigest;
|
import java.security.MessageDigest;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Utilities for Block checksum computing, for both replicated and striped
|
* Utilities for Block checksum computing, for both replicated and striped
|
||||||
* blocks.
|
* blocks.
|
||||||
*/
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
final class BlockChecksumHelper {
|
final class BlockChecksumHelper {
|
||||||
|
|
||||||
static final Logger LOG = LoggerFactory.getLogger(BlockChecksumHelper.class);
|
static final Logger LOG = LoggerFactory.getLogger(BlockChecksumHelper.class);
|
||||||
|
@ -327,6 +334,7 @@ final class BlockChecksumHelper {
|
||||||
private final ErasureCodingPolicy ecPolicy;
|
private final ErasureCodingPolicy ecPolicy;
|
||||||
private final DatanodeInfo[] datanodes;
|
private final DatanodeInfo[] datanodes;
|
||||||
private final Token<BlockTokenIdentifier>[] blockTokens;
|
private final Token<BlockTokenIdentifier>[] blockTokens;
|
||||||
|
private final byte[] blockIndices;
|
||||||
|
|
||||||
private final DataOutputBuffer md5writer = new DataOutputBuffer();
|
private final DataOutputBuffer md5writer = new DataOutputBuffer();
|
||||||
|
|
||||||
|
@ -338,17 +346,61 @@ final class BlockChecksumHelper {
|
||||||
this.ecPolicy = stripedBlockInfo.getErasureCodingPolicy();
|
this.ecPolicy = stripedBlockInfo.getErasureCodingPolicy();
|
||||||
this.datanodes = stripedBlockInfo.getDatanodes();
|
this.datanodes = stripedBlockInfo.getDatanodes();
|
||||||
this.blockTokens = stripedBlockInfo.getBlockTokens();
|
this.blockTokens = stripedBlockInfo.getBlockTokens();
|
||||||
|
this.blockIndices = stripedBlockInfo.getBlockIndices();
|
||||||
|
}
|
||||||
|
|
||||||
|
private static class LiveBlockInfo {
|
||||||
|
private final DatanodeInfo dn;
|
||||||
|
private final Token<BlockTokenIdentifier> token;
|
||||||
|
|
||||||
|
LiveBlockInfo(DatanodeInfo dn, Token<BlockTokenIdentifier> token) {
|
||||||
|
this.dn = dn;
|
||||||
|
this.token = token;
|
||||||
|
}
|
||||||
|
|
||||||
|
DatanodeInfo getDn() {
|
||||||
|
return dn;
|
||||||
|
}
|
||||||
|
|
||||||
|
Token<BlockTokenIdentifier> getToken() {
|
||||||
|
return token;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
void compute() throws IOException {
|
void compute() throws IOException {
|
||||||
for (int idx = 0; idx < ecPolicy.getNumDataUnits(); idx++) {
|
assert datanodes.length == blockIndices.length;
|
||||||
ExtendedBlock block =
|
|
||||||
StripedBlockUtil.constructInternalBlock(blockGroup,
|
Map<Byte, LiveBlockInfo> liveDns = new HashMap<>(datanodes.length);
|
||||||
ecPolicy.getCellSize(), ecPolicy.getNumDataUnits(), idx);
|
int blkIndxLen = blockIndices.length;
|
||||||
DatanodeInfo targetDatanode = datanodes[idx];
|
int numDataUnits = ecPolicy.getNumDataUnits();
|
||||||
Token<BlockTokenIdentifier> blockToken = blockTokens[idx];
|
// Prepare live datanode list. Missing data blocks will be reconstructed
|
||||||
checksumBlock(block, idx, blockToken, targetDatanode);
|
// and recalculate checksum.
|
||||||
|
for (int idx = 0; idx < blkIndxLen; idx++) {
|
||||||
|
liveDns.put(blockIndices[idx],
|
||||||
|
new LiveBlockInfo(datanodes[idx], blockTokens[idx]));
|
||||||
|
}
|
||||||
|
for (int idx = 0; idx < numDataUnits && idx < blkIndxLen; idx++) {
|
||||||
|
try {
|
||||||
|
LiveBlockInfo liveBlkInfo = liveDns.get((byte) idx);
|
||||||
|
if (liveBlkInfo == null) {
|
||||||
|
// reconstruct block and calculate checksum for missing node
|
||||||
|
recalculateChecksum(idx);
|
||||||
|
} else {
|
||||||
|
try {
|
||||||
|
ExtendedBlock block = StripedBlockUtil.constructInternalBlock(
|
||||||
|
blockGroup, ecPolicy.getCellSize(), numDataUnits, idx);
|
||||||
|
checksumBlock(block, idx, liveBlkInfo.getToken(),
|
||||||
|
liveBlkInfo.getDn());
|
||||||
|
} catch (IOException ioe) {
|
||||||
|
LOG.warn("Exception while reading checksum", ioe);
|
||||||
|
// reconstruct block and calculate checksum for the failed node
|
||||||
|
recalculateChecksum(idx);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.warn("Failed to get the checksum", e);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MD5Hash md5out = MD5Hash.digest(md5writer.getData());
|
MD5Hash md5out = MD5Hash.digest(md5writer.getData());
|
||||||
|
@ -379,8 +431,64 @@ final class BlockChecksumHelper {
|
||||||
DataTransferProtos.OpBlockChecksumResponseProto checksumData =
|
DataTransferProtos.OpBlockChecksumResponseProto checksumData =
|
||||||
reply.getChecksumResponse();
|
reply.getChecksumResponse();
|
||||||
|
|
||||||
|
// read crc-type
|
||||||
|
final DataChecksum.Type ct;
|
||||||
|
if (checksumData.hasCrcType()) {
|
||||||
|
ct = PBHelperClient.convert(checksumData.getCrcType());
|
||||||
|
} else {
|
||||||
|
LOG.debug("Retrieving checksum from an earlier-version DataNode: "
|
||||||
|
+ "inferring checksum by reading first byte");
|
||||||
|
ct = DataChecksum.Type.DEFAULT;
|
||||||
|
}
|
||||||
|
|
||||||
|
setOrVerifyChecksumProperties(blockIdx, checksumData.getBytesPerCrc(),
|
||||||
|
checksumData.getCrcPerBlock(), ct);
|
||||||
|
//read md5
|
||||||
|
final MD5Hash md5 = new MD5Hash(checksumData.getMd5().toByteArray());
|
||||||
|
md5.write(md5writer);
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("got reply from " + targetDatanode + ": md5=" + md5);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Reconstruct this data block and recalculate checksum.
|
||||||
|
*
|
||||||
|
* @param errBlkIndex
|
||||||
|
* error index to be reconstrcuted and recalculate checksum.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private void recalculateChecksum(int errBlkIndex) throws IOException {
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Recalculate checksum for the missing/failed block index "
|
||||||
|
+ errBlkIndex);
|
||||||
|
}
|
||||||
|
byte[] errIndices = new byte[1];
|
||||||
|
errIndices[0] = (byte) errBlkIndex;
|
||||||
|
StripedReconstructionInfo stripedReconInfo =
|
||||||
|
new StripedReconstructionInfo(
|
||||||
|
blockGroup, ecPolicy, blockIndices, datanodes, errIndices);
|
||||||
|
final StripedBlockChecksumReconstructor checksumRecon =
|
||||||
|
new StripedBlockChecksumReconstructor(
|
||||||
|
getDatanode().getErasureCodingWorker(), stripedReconInfo,
|
||||||
|
md5writer);
|
||||||
|
checksumRecon.reconstruct();
|
||||||
|
|
||||||
|
DataChecksum checksum = checksumRecon.getChecksum();
|
||||||
|
long crcPerBlock = checksum.getChecksumSize() <= 0 ? 0
|
||||||
|
: checksumRecon.getChecksumDataLen() / checksum.getChecksumSize();
|
||||||
|
setOrVerifyChecksumProperties(errBlkIndex, checksum.getBytesPerChecksum(),
|
||||||
|
crcPerBlock, checksum.getChecksumType());
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Recalculated checksum for the block index " + errBlkIndex
|
||||||
|
+ ": md5=" + checksumRecon.getMD5());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void setOrVerifyChecksumProperties(int blockIdx, int bpc,
|
||||||
|
final long cpb, DataChecksum.Type ct) throws IOException {
|
||||||
//read byte-per-checksum
|
//read byte-per-checksum
|
||||||
final int bpc = checksumData.getBytesPerCrc();
|
|
||||||
if (blockIdx == 0) { //first block
|
if (blockIdx == 0) { //first block
|
||||||
setBytesPerCRC(bpc);
|
setBytesPerCRC(bpc);
|
||||||
} else if (bpc != getBytesPerCRC()) {
|
} else if (bpc != getBytesPerCRC()) {
|
||||||
|
@ -389,26 +497,10 @@ final class BlockChecksumHelper {
|
||||||
}
|
}
|
||||||
|
|
||||||
//read crc-per-block
|
//read crc-per-block
|
||||||
final long cpb = checksumData.getCrcPerBlock();
|
|
||||||
if (blockIdx == 0) {
|
if (blockIdx == 0) {
|
||||||
setCrcPerBlock(cpb);
|
setCrcPerBlock(cpb);
|
||||||
}
|
}
|
||||||
|
|
||||||
//read md5
|
|
||||||
final MD5Hash md5 = new MD5Hash(
|
|
||||||
checksumData.getMd5().toByteArray());
|
|
||||||
md5.write(md5writer);
|
|
||||||
|
|
||||||
// read crc-type
|
|
||||||
final DataChecksum.Type ct;
|
|
||||||
if (checksumData.hasCrcType()) {
|
|
||||||
ct = PBHelperClient.convert(checksumData.getCrcType());
|
|
||||||
} else {
|
|
||||||
LOG.debug("Retrieving checksum from an earlier-version DataNode: " +
|
|
||||||
"inferring checksum by reading first byte");
|
|
||||||
ct = DataChecksum.Type.DEFAULT;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (blockIdx == 0) { // first block
|
if (blockIdx == 0) { // first block
|
||||||
setCrcType(ct);
|
setCrcType(ct);
|
||||||
} else if (getCrcType() != DataChecksum.Type.MIXED &&
|
} else if (getCrcType() != DataChecksum.Type.MIXED &&
|
||||||
|
@ -422,8 +514,6 @@ final class BlockChecksumHelper {
|
||||||
LOG.debug("set bytesPerCRC=" + getBytesPerCRC()
|
LOG.debug("set bytesPerCRC=" + getBytesPerCRC()
|
||||||
+ ", crcPerBlock=" + getCrcPerBlock());
|
+ ", crcPerBlock=" + getCrcPerBlock());
|
||||||
}
|
}
|
||||||
LOG.debug("got reply from " + targetDatanode + ": md5=" + md5);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -116,19 +116,24 @@ public final class ErasureCodingWorker {
|
||||||
*/
|
*/
|
||||||
public void processErasureCodingTasks(
|
public void processErasureCodingTasks(
|
||||||
Collection<BlockECReconstructionInfo> ecTasks) {
|
Collection<BlockECReconstructionInfo> ecTasks) {
|
||||||
for (BlockECReconstructionInfo reconstructionInfo : ecTasks) {
|
for (BlockECReconstructionInfo reconInfo : ecTasks) {
|
||||||
try {
|
try {
|
||||||
final StripedReconstructor task =
|
StripedReconstructionInfo stripedReconInfo =
|
||||||
new StripedReconstructor(this, reconstructionInfo);
|
new StripedReconstructionInfo(
|
||||||
|
reconInfo.getExtendedBlock(), reconInfo.getErasureCodingPolicy(),
|
||||||
|
reconInfo.getLiveBlockIndices(), reconInfo.getSourceDnInfos(),
|
||||||
|
reconInfo.getTargetDnInfos(), reconInfo.getTargetStorageTypes());
|
||||||
|
final StripedBlockReconstructor task =
|
||||||
|
new StripedBlockReconstructor(this, stripedReconInfo);
|
||||||
if (task.hasValidTargets()) {
|
if (task.hasValidTargets()) {
|
||||||
stripedReconstructionPool.submit(task);
|
stripedReconstructionPool.submit(task);
|
||||||
} else {
|
} else {
|
||||||
LOG.warn("No missing internal block. Skip reconstruction for task:{}",
|
LOG.warn("No missing internal block. Skip reconstruction for task:{}",
|
||||||
reconstructionInfo);
|
reconInfo);
|
||||||
}
|
}
|
||||||
} catch (Throwable e) {
|
} catch (Throwable e) {
|
||||||
LOG.warn("Failed to reconstruct striped block {}",
|
LOG.warn("Failed to reconstruct striped block {}",
|
||||||
reconstructionInfo.getExtendedBlock().getLocalBlock(), e);
|
reconInfo.getExtendedBlock().getLocalBlock(), e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,129 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.datanode.erasurecode;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.security.MessageDigest;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.io.DataOutputBuffer;
|
||||||
|
import org.apache.hadoop.io.MD5Hash;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* StripedBlockChecksumReconstructor reconstruct one or more missed striped
|
||||||
|
* block in the striped block group, the minimum number of live striped blocks
|
||||||
|
* should be no less than data block number. Then checksum will be recalculated
|
||||||
|
* using the newly reconstructed block.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
public class StripedBlockChecksumReconstructor extends StripedReconstructor {
|
||||||
|
|
||||||
|
private ByteBuffer targetBuffer;
|
||||||
|
private final byte[] targetIndices;
|
||||||
|
|
||||||
|
private byte[] checksumBuf;
|
||||||
|
private DataOutputBuffer checksumWriter;
|
||||||
|
private MD5Hash md5;
|
||||||
|
private long checksumDataLen;
|
||||||
|
|
||||||
|
public StripedBlockChecksumReconstructor(ErasureCodingWorker worker,
|
||||||
|
StripedReconstructionInfo stripedReconInfo,
|
||||||
|
DataOutputBuffer checksumWriter) throws IOException {
|
||||||
|
super(worker, stripedReconInfo);
|
||||||
|
this.targetIndices = stripedReconInfo.getTargetIndices();
|
||||||
|
assert targetIndices != null;
|
||||||
|
this.checksumWriter = checksumWriter;
|
||||||
|
init();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void init() throws IOException {
|
||||||
|
getStripedReader().init();
|
||||||
|
// allocate buffer to keep the reconstructed block data
|
||||||
|
targetBuffer = allocateBuffer(getBufferSize());
|
||||||
|
long maxTargetLen = 0L;
|
||||||
|
for (int targetIndex : targetIndices) {
|
||||||
|
maxTargetLen = Math.max(maxTargetLen, getBlockLen(targetIndex));
|
||||||
|
}
|
||||||
|
setMaxTargetLength(maxTargetLen);
|
||||||
|
int checksumSize = getChecksum().getChecksumSize();
|
||||||
|
int bytesPerChecksum = getChecksum().getBytesPerChecksum();
|
||||||
|
int tmpLen = checksumSize * (getBufferSize() / bytesPerChecksum);
|
||||||
|
checksumBuf = new byte[tmpLen];
|
||||||
|
}
|
||||||
|
|
||||||
|
public void reconstruct() throws IOException {
|
||||||
|
MessageDigest digester = MD5Hash.getDigester();
|
||||||
|
while (getPositionInBlock() < getMaxTargetLength()) {
|
||||||
|
long remaining = getMaxTargetLength() - getPositionInBlock();
|
||||||
|
final int toReconstructLen = (int) Math
|
||||||
|
.min(getStripedReader().getBufferSize(), remaining);
|
||||||
|
// step1: read from minimum source DNs required for reconstruction.
|
||||||
|
// The returned success list is the source DNs we do real read from
|
||||||
|
getStripedReader().readMinimumSources(toReconstructLen);
|
||||||
|
|
||||||
|
// step2: decode to reconstruct targets
|
||||||
|
reconstructTargets(toReconstructLen);
|
||||||
|
|
||||||
|
// step3: calculate checksum
|
||||||
|
getChecksum().calculateChunkedSums(targetBuffer.array(), 0,
|
||||||
|
targetBuffer.remaining(), checksumBuf, 0);
|
||||||
|
|
||||||
|
// step4: updates the digest using the checksum array of bytes
|
||||||
|
digester.update(checksumBuf, 0, checksumBuf.length);
|
||||||
|
checksumDataLen += checksumBuf.length;
|
||||||
|
updatePositionInBlock(toReconstructLen);
|
||||||
|
clearBuffers();
|
||||||
|
}
|
||||||
|
|
||||||
|
byte[] digest = digester.digest();
|
||||||
|
md5 = new MD5Hash(digest);
|
||||||
|
md5.write(checksumWriter);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void reconstructTargets(int toReconstructLen) {
|
||||||
|
initDecoderIfNecessary();
|
||||||
|
|
||||||
|
ByteBuffer[] inputs = getStripedReader().getInputBuffers(toReconstructLen);
|
||||||
|
|
||||||
|
ByteBuffer[] outputs = new ByteBuffer[1];
|
||||||
|
targetBuffer.limit(toReconstructLen);
|
||||||
|
outputs[0] = targetBuffer;
|
||||||
|
int[] tarIndices = new int[targetIndices.length];
|
||||||
|
for (int i = 0; i < targetIndices.length; i++) {
|
||||||
|
tarIndices[i] = targetIndices[i];
|
||||||
|
}
|
||||||
|
getDecoder().decode(inputs, tarIndices, outputs);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clear all associated buffers.
|
||||||
|
*/
|
||||||
|
private void clearBuffers() {
|
||||||
|
getStripedReader().clearBuffers();
|
||||||
|
targetBuffer.clear();
|
||||||
|
}
|
||||||
|
|
||||||
|
public MD5Hash getMD5() {
|
||||||
|
return md5;
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getChecksumDataLen() {
|
||||||
|
return checksumDataLen;
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,119 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.datanode.erasurecode;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* StripedBlockReconstructor reconstruct one or more missed striped block in
|
||||||
|
* the striped block group, the minimum number of live striped blocks should
|
||||||
|
* be no less than data block number.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
class StripedBlockReconstructor extends StripedReconstructor
|
||||||
|
implements Runnable {
|
||||||
|
|
||||||
|
private StripedWriter stripedWriter;
|
||||||
|
|
||||||
|
StripedBlockReconstructor(ErasureCodingWorker worker,
|
||||||
|
StripedReconstructionInfo stripedReconInfo) {
|
||||||
|
super(worker, stripedReconInfo);
|
||||||
|
|
||||||
|
stripedWriter = new StripedWriter(this, getDatanode(),
|
||||||
|
getConf(), stripedReconInfo);
|
||||||
|
}
|
||||||
|
|
||||||
|
boolean hasValidTargets() {
|
||||||
|
return stripedWriter.hasValidTargets();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void run() {
|
||||||
|
getDatanode().incrementXmitsInProgress();
|
||||||
|
try {
|
||||||
|
getStripedReader().init();
|
||||||
|
|
||||||
|
stripedWriter.init();
|
||||||
|
|
||||||
|
reconstruct();
|
||||||
|
|
||||||
|
stripedWriter.endTargetBlocks();
|
||||||
|
|
||||||
|
// Currently we don't check the acks for packets, this is similar as
|
||||||
|
// block replication.
|
||||||
|
} catch (Throwable e) {
|
||||||
|
LOG.warn("Failed to reconstruct striped block: {}", getBlockGroup(), e);
|
||||||
|
getDatanode().getMetrics().incrECFailedReconstructionTasks();
|
||||||
|
} finally {
|
||||||
|
getDatanode().decrementXmitsInProgress();
|
||||||
|
getDatanode().getMetrics().incrECReconstructionTasks();
|
||||||
|
getStripedReader().close();
|
||||||
|
stripedWriter.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void reconstruct() throws IOException {
|
||||||
|
while (getPositionInBlock() < getMaxTargetLength()) {
|
||||||
|
long remaining = getMaxTargetLength() - getPositionInBlock();
|
||||||
|
final int toReconstructLen =
|
||||||
|
(int) Math.min(getStripedReader().getBufferSize(), remaining);
|
||||||
|
|
||||||
|
// step1: read from minimum source DNs required for reconstruction.
|
||||||
|
// The returned success list is the source DNs we do real read from
|
||||||
|
getStripedReader().readMinimumSources(toReconstructLen);
|
||||||
|
|
||||||
|
// step2: decode to reconstruct targets
|
||||||
|
reconstructTargets(toReconstructLen);
|
||||||
|
|
||||||
|
// step3: transfer data
|
||||||
|
if (stripedWriter.transferData2Targets() == 0) {
|
||||||
|
String error = "Transfer failed for all targets.";
|
||||||
|
throw new IOException(error);
|
||||||
|
}
|
||||||
|
|
||||||
|
updatePositionInBlock(toReconstructLen);
|
||||||
|
|
||||||
|
clearBuffers();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void reconstructTargets(int toReconstructLen) {
|
||||||
|
initDecoderIfNecessary();
|
||||||
|
|
||||||
|
ByteBuffer[] inputs = getStripedReader().getInputBuffers(toReconstructLen);
|
||||||
|
|
||||||
|
int[] erasedIndices = stripedWriter.getRealTargetIndices();
|
||||||
|
ByteBuffer[] outputs = stripedWriter.getRealTargetBuffers(toReconstructLen);
|
||||||
|
|
||||||
|
getDecoder().decode(inputs, erasedIndices, outputs);
|
||||||
|
|
||||||
|
stripedWriter.updateRealTargetBuffers(toReconstructLen);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clear all associated buffers.
|
||||||
|
*/
|
||||||
|
private void clearBuffers() {
|
||||||
|
getStripedReader().clearBuffers();
|
||||||
|
|
||||||
|
stripedWriter.clearBuffers();
|
||||||
|
}
|
||||||
|
}
|
|
@ -23,11 +23,8 @@ import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.DFSUtilClient.CorruptedBlocks;
|
import org.apache.hadoop.hdfs.DFSUtilClient.CorruptedBlocks;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
|
||||||
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.BlockECReconstructionCommand.BlockECReconstructionInfo;
|
|
||||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil.StripingChunkReadResult;
|
import org.apache.hadoop.hdfs.util.StripedBlockUtil.StripingChunkReadResult;
|
||||||
import org.apache.hadoop.util.DataChecksum;
|
import org.apache.hadoop.util.DataChecksum;
|
||||||
|
@ -85,8 +82,7 @@ class StripedReader {
|
||||||
private final CompletionService<Void> readService;
|
private final CompletionService<Void> readService;
|
||||||
|
|
||||||
StripedReader(StripedReconstructor reconstructor, DataNode datanode,
|
StripedReader(StripedReconstructor reconstructor, DataNode datanode,
|
||||||
Configuration conf,
|
Configuration conf, StripedReconstructionInfo stripedReconInfo) {
|
||||||
BlockECReconstructionInfo reconstructionInfo) {
|
|
||||||
stripedReadTimeoutInMills = conf.getInt(
|
stripedReadTimeoutInMills = conf.getInt(
|
||||||
DFSConfigKeys.DFS_DN_EC_RECONSTRUCTION_STRIPED_READ_TIMEOUT_MILLIS_KEY,
|
DFSConfigKeys.DFS_DN_EC_RECONSTRUCTION_STRIPED_READ_TIMEOUT_MILLIS_KEY,
|
||||||
DFSConfigKeys.DFS_DN_EC_RECONSTRUCTION_STRIPED_READ_TIMEOUT_MILLIS_DEFAULT);
|
DFSConfigKeys.DFS_DN_EC_RECONSTRUCTION_STRIPED_READ_TIMEOUT_MILLIS_DEFAULT);
|
||||||
|
@ -98,13 +94,11 @@ class StripedReader {
|
||||||
this.datanode = datanode;
|
this.datanode = datanode;
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
|
|
||||||
ErasureCodingPolicy ecPolicy = reconstructionInfo.getErasureCodingPolicy();
|
dataBlkNum = stripedReconInfo.getEcPolicy().getNumDataUnits();
|
||||||
dataBlkNum = ecPolicy.getNumDataUnits();
|
parityBlkNum = stripedReconInfo.getEcPolicy().getNumParityUnits();
|
||||||
parityBlkNum = ecPolicy.getNumParityUnits();
|
|
||||||
|
|
||||||
ExtendedBlock blockGroup = reconstructionInfo.getExtendedBlock();
|
int cellsNum = (int) ((stripedReconInfo.getBlockGroup().getNumBytes() - 1)
|
||||||
int cellsNum = (int)((blockGroup.getNumBytes() - 1) / ecPolicy.getCellSize()
|
/ stripedReconInfo.getEcPolicy().getCellSize() + 1);
|
||||||
+ 1);
|
|
||||||
minRequiredSources = Math.min(cellsNum, dataBlkNum);
|
minRequiredSources = Math.min(cellsNum, dataBlkNum);
|
||||||
|
|
||||||
if (minRequiredSources < dataBlkNum) {
|
if (minRequiredSources < dataBlkNum) {
|
||||||
|
@ -113,8 +107,10 @@ class StripedReader {
|
||||||
zeroStripeIndices = new short[zeroStripNum];
|
zeroStripeIndices = new short[zeroStripNum];
|
||||||
}
|
}
|
||||||
|
|
||||||
liveIndices = reconstructionInfo.getLiveBlockIndices();
|
this.liveIndices = stripedReconInfo.getLiveIndices();
|
||||||
sources = reconstructionInfo.getSourceDnInfos();
|
assert liveIndices != null;
|
||||||
|
this.sources = stripedReconInfo.getSources();
|
||||||
|
assert sources != null;
|
||||||
|
|
||||||
readers = new ArrayList<>(sources.length);
|
readers = new ArrayList<>(sources.length);
|
||||||
readService = reconstructor.createReadService();
|
readService = reconstructor.createReadService();
|
||||||
|
|
|
@ -0,0 +1,99 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.datanode.erasurecode;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.fs.StorageType;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Stores striped block info that can be used for block reconstruction.
|
||||||
|
*/
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
public class StripedReconstructionInfo {
|
||||||
|
|
||||||
|
private final ExtendedBlock blockGroup;
|
||||||
|
private final ErasureCodingPolicy ecPolicy;
|
||||||
|
|
||||||
|
// source info
|
||||||
|
private final byte[] liveIndices;
|
||||||
|
private final DatanodeInfo[] sources;
|
||||||
|
|
||||||
|
// target info
|
||||||
|
private final byte[] targetIndices;
|
||||||
|
private final DatanodeInfo[] targets;
|
||||||
|
private final StorageType[] targetStorageTypes;
|
||||||
|
|
||||||
|
public StripedReconstructionInfo(ExtendedBlock blockGroup,
|
||||||
|
ErasureCodingPolicy ecPolicy, byte[] liveIndices, DatanodeInfo[] sources,
|
||||||
|
byte[] targetIndices) {
|
||||||
|
this(blockGroup, ecPolicy, liveIndices, sources, targetIndices, null, null);
|
||||||
|
}
|
||||||
|
|
||||||
|
StripedReconstructionInfo(ExtendedBlock blockGroup,
|
||||||
|
ErasureCodingPolicy ecPolicy, byte[] liveIndices, DatanodeInfo[] sources,
|
||||||
|
DatanodeInfo[] targets, StorageType[] targetStorageTypes) {
|
||||||
|
this(blockGroup, ecPolicy, liveIndices, sources, null, targets,
|
||||||
|
targetStorageTypes);
|
||||||
|
}
|
||||||
|
|
||||||
|
private StripedReconstructionInfo(ExtendedBlock blockGroup,
|
||||||
|
ErasureCodingPolicy ecPolicy, byte[] liveIndices, DatanodeInfo[] sources,
|
||||||
|
byte[] targetIndices, DatanodeInfo[] targets,
|
||||||
|
StorageType[] targetStorageTypes) {
|
||||||
|
|
||||||
|
this.blockGroup = blockGroup;
|
||||||
|
this.ecPolicy = ecPolicy;
|
||||||
|
this.liveIndices = liveIndices;
|
||||||
|
this.sources = sources;
|
||||||
|
this.targetIndices = targetIndices;
|
||||||
|
this.targets = targets;
|
||||||
|
this.targetStorageTypes = targetStorageTypes;
|
||||||
|
}
|
||||||
|
|
||||||
|
ExtendedBlock getBlockGroup() {
|
||||||
|
return blockGroup;
|
||||||
|
}
|
||||||
|
|
||||||
|
ErasureCodingPolicy getEcPolicy() {
|
||||||
|
return ecPolicy;
|
||||||
|
}
|
||||||
|
|
||||||
|
byte[] getLiveIndices() {
|
||||||
|
return liveIndices;
|
||||||
|
}
|
||||||
|
|
||||||
|
DatanodeInfo[] getSources() {
|
||||||
|
return sources;
|
||||||
|
}
|
||||||
|
|
||||||
|
byte[] getTargetIndices() {
|
||||||
|
return targetIndices;
|
||||||
|
}
|
||||||
|
|
||||||
|
DatanodeInfo[] getTargets() {
|
||||||
|
return targets;
|
||||||
|
}
|
||||||
|
|
||||||
|
StorageType[] getTargetStorageTypes() {
|
||||||
|
return targetStorageTypes;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
||||||
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.BlockECReconstructionCommand.BlockECReconstructionInfo;
|
|
||||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||||
import org.apache.hadoop.io.erasurecode.CodecUtil;
|
import org.apache.hadoop.io.erasurecode.CodecUtil;
|
||||||
import org.apache.hadoop.io.erasurecode.ErasureCoderOptions;
|
import org.apache.hadoop.io.erasurecode.ErasureCoderOptions;
|
||||||
|
@ -39,6 +38,7 @@ import java.nio.ByteBuffer;
|
||||||
import java.util.BitSet;
|
import java.util.BitSet;
|
||||||
import java.util.concurrent.CompletionService;
|
import java.util.concurrent.CompletionService;
|
||||||
import java.util.concurrent.ExecutorCompletionService;
|
import java.util.concurrent.ExecutorCompletionService;
|
||||||
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* StripedReconstructor reconstruct one or more missed striped block in the
|
* StripedReconstructor reconstruct one or more missed striped block in the
|
||||||
|
@ -94,58 +94,50 @@ import java.util.concurrent.ExecutorCompletionService;
|
||||||
* reconstructed result received by targets?
|
* reconstructed result received by targets?
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private
|
||||||
class StripedReconstructor implements Runnable {
|
abstract class StripedReconstructor {
|
||||||
private static final Logger LOG = DataNode.LOG;
|
protected static final Logger LOG = DataNode.LOG;
|
||||||
|
|
||||||
private final ErasureCodingWorker worker;
|
|
||||||
private final DataNode datanode;
|
|
||||||
private final Configuration conf;
|
private final Configuration conf;
|
||||||
|
private final DataNode datanode;
|
||||||
private final ErasureCodingPolicy ecPolicy;
|
private final ErasureCodingPolicy ecPolicy;
|
||||||
|
|
||||||
private RawErasureDecoder decoder;
|
private RawErasureDecoder decoder;
|
||||||
|
|
||||||
private final ExtendedBlock blockGroup;
|
private final ExtendedBlock blockGroup;
|
||||||
private final BitSet liveBitSet;
|
|
||||||
|
|
||||||
// position in striped internal block
|
// position in striped internal block
|
||||||
private long positionInBlock;
|
private long positionInBlock;
|
||||||
|
|
||||||
private StripedReader stripedReader;
|
private StripedReader stripedReader;
|
||||||
|
private ThreadPoolExecutor stripedReadPool;
|
||||||
private StripedWriter stripedWriter;
|
|
||||||
|
|
||||||
private final CachingStrategy cachingStrategy;
|
private final CachingStrategy cachingStrategy;
|
||||||
|
private long maxTargetLength = 0L;
|
||||||
|
private final BitSet liveBitSet;
|
||||||
|
|
||||||
StripedReconstructor(ErasureCodingWorker worker,
|
StripedReconstructor(ErasureCodingWorker worker,
|
||||||
BlockECReconstructionInfo reconstructionInfo) {
|
StripedReconstructionInfo stripedReconInfo) {
|
||||||
this.worker = worker;
|
this.stripedReadPool = worker.getStripedReadPool();
|
||||||
this.datanode = worker.getDatanode();
|
this.datanode = worker.getDatanode();
|
||||||
this.conf = worker.getConf();
|
this.conf = worker.getConf();
|
||||||
|
this.ecPolicy = stripedReconInfo.getEcPolicy();
|
||||||
ecPolicy = reconstructionInfo.getErasureCodingPolicy();
|
liveBitSet = new BitSet(
|
||||||
|
ecPolicy.getNumDataUnits() + ecPolicy.getNumParityUnits());
|
||||||
blockGroup = reconstructionInfo.getExtendedBlock();
|
for (int i = 0; i < stripedReconInfo.getLiveIndices().length; i++) {
|
||||||
byte[] liveIndices = reconstructionInfo.getLiveBlockIndices();
|
liveBitSet.set(stripedReconInfo.getLiveIndices()[i]);
|
||||||
liveBitSet = new BitSet(ecPolicy.getNumDataUnits() +
|
|
||||||
ecPolicy.getNumParityUnits());
|
|
||||||
for (int i = 0; i < liveIndices.length; i++) {
|
|
||||||
liveBitSet.set(liveIndices[i]);
|
|
||||||
}
|
}
|
||||||
|
blockGroup = stripedReconInfo.getBlockGroup();
|
||||||
stripedReader = new StripedReader(this, datanode,
|
stripedReader = new StripedReader(this, datanode, conf, stripedReconInfo);
|
||||||
conf, reconstructionInfo);
|
|
||||||
stripedWriter = new StripedWriter(this, datanode,
|
|
||||||
conf, reconstructionInfo);
|
|
||||||
|
|
||||||
cachingStrategy = CachingStrategy.newDefaultStrategy();
|
cachingStrategy = CachingStrategy.newDefaultStrategy();
|
||||||
|
|
||||||
positionInBlock = 0L;
|
positionInBlock = 0L;
|
||||||
}
|
}
|
||||||
|
|
||||||
BitSet getLiveBitSet() {
|
/**
|
||||||
return liveBitSet;
|
* Reconstruct one or more missed striped block in the striped block group,
|
||||||
}
|
* the minimum number of live striped blocks should be no less than data
|
||||||
|
* block number.
|
||||||
|
*
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
abstract void reconstruct() throws IOException;
|
||||||
|
|
||||||
ByteBuffer allocateBuffer(int length) {
|
ByteBuffer allocateBuffer(int length) {
|
||||||
return ByteBuffer.allocate(length);
|
return ByteBuffer.allocate(length);
|
||||||
|
@ -160,61 +152,8 @@ class StripedReconstructor implements Runnable {
|
||||||
ecPolicy, i);
|
ecPolicy, i);
|
||||||
}
|
}
|
||||||
|
|
||||||
boolean hasValidTargets() {
|
|
||||||
return stripedWriter.hasValidTargets();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void run() {
|
|
||||||
datanode.incrementXmitsInProgress();
|
|
||||||
try {
|
|
||||||
stripedReader.init();
|
|
||||||
|
|
||||||
stripedWriter.init();
|
|
||||||
|
|
||||||
reconstructAndTransfer();
|
|
||||||
|
|
||||||
stripedWriter.endTargetBlocks();
|
|
||||||
|
|
||||||
// Currently we don't check the acks for packets, this is similar as
|
|
||||||
// block replication.
|
|
||||||
} catch (Throwable e) {
|
|
||||||
LOG.warn("Failed to reconstruct striped block: {}", blockGroup, e);
|
|
||||||
datanode.getMetrics().incrECFailedReconstructionTasks();
|
|
||||||
} finally {
|
|
||||||
datanode.decrementXmitsInProgress();
|
|
||||||
datanode.getMetrics().incrECReconstructionTasks();
|
|
||||||
stripedReader.close();
|
|
||||||
stripedWriter.close();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void reconstructAndTransfer() throws IOException {
|
|
||||||
while (positionInBlock < stripedWriter.getMaxTargetLength()) {
|
|
||||||
long remaining = stripedWriter.getMaxTargetLength() - positionInBlock;
|
|
||||||
final int toReconstructLen =
|
|
||||||
(int) Math.min(stripedReader.getBufferSize(), remaining);
|
|
||||||
// step1: read from minimum source DNs required for reconstruction.
|
|
||||||
// The returned success list is the source DNs we do real read from
|
|
||||||
stripedReader.readMinimumSources(toReconstructLen);
|
|
||||||
|
|
||||||
// step2: decode to reconstruct targets
|
|
||||||
reconstructTargets(toReconstructLen);
|
|
||||||
|
|
||||||
// step3: transfer data
|
|
||||||
if (stripedWriter.transferData2Targets() == 0) {
|
|
||||||
String error = "Transfer failed for all targets.";
|
|
||||||
throw new IOException(error);
|
|
||||||
}
|
|
||||||
|
|
||||||
positionInBlock += toReconstructLen;
|
|
||||||
|
|
||||||
clearBuffers();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Initialize decoder
|
// Initialize decoder
|
||||||
private void initDecoderIfNecessary() {
|
protected void initDecoderIfNecessary() {
|
||||||
if (decoder == null) {
|
if (decoder == null) {
|
||||||
ErasureCoderOptions coderOptions = new ErasureCoderOptions(
|
ErasureCoderOptions coderOptions = new ErasureCoderOptions(
|
||||||
ecPolicy.getNumDataUnits(), ecPolicy.getNumParityUnits());
|
ecPolicy.getNumDataUnits(), ecPolicy.getNumParityUnits());
|
||||||
|
@ -223,32 +162,10 @@ class StripedReconstructor implements Runnable {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private void reconstructTargets(int toReconstructLen) {
|
|
||||||
initDecoderIfNecessary();
|
|
||||||
|
|
||||||
ByteBuffer[] inputs = stripedReader.getInputBuffers(toReconstructLen);
|
|
||||||
|
|
||||||
int[] erasedIndices = stripedWriter.getRealTargetIndices();
|
|
||||||
ByteBuffer[] outputs = stripedWriter.getRealTargetBuffers(toReconstructLen);
|
|
||||||
|
|
||||||
decoder.decode(inputs, erasedIndices, outputs);
|
|
||||||
|
|
||||||
stripedWriter.updateRealTargetBuffers(toReconstructLen);
|
|
||||||
}
|
|
||||||
|
|
||||||
long getPositionInBlock() {
|
long getPositionInBlock() {
|
||||||
return positionInBlock;
|
return positionInBlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Clear all associated buffers.
|
|
||||||
*/
|
|
||||||
private void clearBuffers() {
|
|
||||||
stripedReader.clearBuffers();
|
|
||||||
|
|
||||||
stripedWriter.clearBuffers();
|
|
||||||
}
|
|
||||||
|
|
||||||
InetSocketAddress getSocketAddress4Transfer(DatanodeInfo dnInfo) {
|
InetSocketAddress getSocketAddress4Transfer(DatanodeInfo dnInfo) {
|
||||||
return NetUtils.createSocketAddr(dnInfo.getXferAddr(
|
return NetUtils.createSocketAddr(dnInfo.getXferAddr(
|
||||||
datanode.getDnConf().getConnectToDnViaHostname()));
|
datanode.getDnConf().getConnectToDnViaHostname()));
|
||||||
|
@ -258,7 +175,7 @@ class StripedReconstructor implements Runnable {
|
||||||
return stripedReader.getBufferSize();
|
return stripedReader.getBufferSize();
|
||||||
}
|
}
|
||||||
|
|
||||||
DataChecksum getChecksum() {
|
public DataChecksum getChecksum() {
|
||||||
return stripedReader.getChecksum();
|
return stripedReader.getChecksum();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -267,10 +184,42 @@ class StripedReconstructor implements Runnable {
|
||||||
}
|
}
|
||||||
|
|
||||||
CompletionService<Void> createReadService() {
|
CompletionService<Void> createReadService() {
|
||||||
return new ExecutorCompletionService<>(worker.getStripedReadPool());
|
return new ExecutorCompletionService<>(stripedReadPool);
|
||||||
}
|
}
|
||||||
|
|
||||||
ExtendedBlock getBlockGroup() {
|
ExtendedBlock getBlockGroup() {
|
||||||
return blockGroup;
|
return blockGroup;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
BitSet getLiveBitSet() {
|
||||||
|
return liveBitSet;
|
||||||
|
}
|
||||||
|
|
||||||
|
long getMaxTargetLength() {
|
||||||
|
return maxTargetLength;
|
||||||
|
}
|
||||||
|
|
||||||
|
void setMaxTargetLength(long maxTargetLength) {
|
||||||
|
this.maxTargetLength = maxTargetLength;
|
||||||
|
}
|
||||||
|
|
||||||
|
void updatePositionInBlock(long positionInBlockArg) {
|
||||||
|
this.positionInBlock += positionInBlockArg;
|
||||||
|
}
|
||||||
|
|
||||||
|
RawErasureDecoder getDecoder() {
|
||||||
|
return decoder;
|
||||||
|
}
|
||||||
|
|
||||||
|
StripedReader getStripedReader() {
|
||||||
|
return stripedReader;
|
||||||
|
}
|
||||||
|
|
||||||
|
Configuration getConf() {
|
||||||
|
return conf;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataNode getDatanode() {
|
||||||
|
return datanode;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -22,11 +22,9 @@ import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.StorageType;
|
import org.apache.hadoop.fs.StorageType;
|
||||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.datatransfer.PacketHeader;
|
import org.apache.hadoop.hdfs.protocol.datatransfer.PacketHeader;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
import org.apache.hadoop.hdfs.server.datanode.CachingStrategy;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.BlockECReconstructionCommand.BlockECReconstructionInfo;
|
|
||||||
import org.apache.hadoop.util.DataChecksum;
|
import org.apache.hadoop.util.DataChecksum;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
|
|
||||||
|
@ -57,7 +55,6 @@ class StripedWriter {
|
||||||
private final short[] targetIndices;
|
private final short[] targetIndices;
|
||||||
private boolean hasValidTargets;
|
private boolean hasValidTargets;
|
||||||
private final StorageType[] targetStorageTypes;
|
private final StorageType[] targetStorageTypes;
|
||||||
private long maxTargetLength;
|
|
||||||
|
|
||||||
private StripedBlockWriter[] writers;
|
private StripedBlockWriter[] writers;
|
||||||
|
|
||||||
|
@ -67,20 +64,19 @@ class StripedWriter {
|
||||||
private int bytesPerChecksum;
|
private int bytesPerChecksum;
|
||||||
private int checksumSize;
|
private int checksumSize;
|
||||||
|
|
||||||
StripedWriter(StripedReconstructor reconstructor,
|
StripedWriter(StripedReconstructor reconstructor, DataNode datanode,
|
||||||
DataNode datanode,
|
Configuration conf, StripedReconstructionInfo stripedReconInfo) {
|
||||||
Configuration conf,
|
|
||||||
BlockECReconstructionInfo reconstructionInfo) {
|
|
||||||
this.reconstructor = reconstructor;
|
this.reconstructor = reconstructor;
|
||||||
this.datanode = datanode;
|
this.datanode = datanode;
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
|
|
||||||
ErasureCodingPolicy ecPolicy = reconstructionInfo.getErasureCodingPolicy();
|
dataBlkNum = stripedReconInfo.getEcPolicy().getNumDataUnits();
|
||||||
dataBlkNum = ecPolicy.getNumDataUnits();
|
parityBlkNum = stripedReconInfo.getEcPolicy().getNumParityUnits();
|
||||||
parityBlkNum = ecPolicy.getNumParityUnits();
|
|
||||||
|
|
||||||
targets = reconstructionInfo.getTargetDnInfos();
|
this.targets = stripedReconInfo.getTargets();
|
||||||
targetStorageTypes = reconstructionInfo.getTargetStorageTypes();
|
assert targets != null;
|
||||||
|
this.targetStorageTypes = stripedReconInfo.getTargetStorageTypes();
|
||||||
|
assert targetStorageTypes != null;
|
||||||
|
|
||||||
writers = new StripedBlockWriter[targets.length];
|
writers = new StripedBlockWriter[targets.length];
|
||||||
|
|
||||||
|
@ -88,12 +84,12 @@ class StripedWriter {
|
||||||
Preconditions.checkArgument(targetIndices.length <= parityBlkNum,
|
Preconditions.checkArgument(targetIndices.length <= parityBlkNum,
|
||||||
"Too much missed striped blocks.");
|
"Too much missed striped blocks.");
|
||||||
initTargetIndices();
|
initTargetIndices();
|
||||||
|
long maxTargetLength = 0L;
|
||||||
maxTargetLength = 0L;
|
|
||||||
for (short targetIndex : targetIndices) {
|
for (short targetIndex : targetIndices) {
|
||||||
maxTargetLength = Math.max(maxTargetLength,
|
maxTargetLength = Math.max(maxTargetLength,
|
||||||
reconstructor.getBlockLen(targetIndex));
|
reconstructor.getBlockLen(targetIndex));
|
||||||
}
|
}
|
||||||
|
reconstructor.setMaxTargetLength(maxTargetLength);
|
||||||
|
|
||||||
// targetsStatus store whether some target is success, it will record
|
// targetsStatus store whether some target is success, it will record
|
||||||
// any failed target once, if some target failed (invalid DN or transfer
|
// any failed target once, if some target failed (invalid DN or transfer
|
||||||
|
@ -126,7 +122,6 @@ class StripedWriter {
|
||||||
BitSet bitset = reconstructor.getLiveBitSet();
|
BitSet bitset = reconstructor.getLiveBitSet();
|
||||||
|
|
||||||
int m = 0;
|
int m = 0;
|
||||||
int k = 0;
|
|
||||||
hasValidTargets = false;
|
hasValidTargets = false;
|
||||||
for (int i = 0; i < dataBlkNum + parityBlkNum; i++) {
|
for (int i = 0; i < dataBlkNum + parityBlkNum; i++) {
|
||||||
if (!bitset.get(i)) {
|
if (!bitset.get(i)) {
|
||||||
|
@ -257,10 +252,6 @@ class StripedWriter {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
long getMaxTargetLength() {
|
|
||||||
return maxTargetLength;
|
|
||||||
}
|
|
||||||
|
|
||||||
byte[] getChecksumBuf() {
|
byte[] getChecksumBuf() {
|
||||||
return checksumBuf;
|
return checksumBuf;
|
||||||
}
|
}
|
||||||
|
|
|
@ -36,6 +36,7 @@ import java.util.concurrent.TimeUnit;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.BlockLocation;
|
import org.apache.hadoop.fs.BlockLocation;
|
||||||
import org.apache.hadoop.fs.FSDataOutputStream;
|
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
|
import org.apache.hadoop.fs.FileChecksum;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hdfs.DFSClient;
|
import org.apache.hadoop.hdfs.DFSClient;
|
||||||
|
@ -276,6 +277,52 @@ public class TestDecommissionWithStriped {
|
||||||
cleanupFile(dfs, ecFile);
|
cleanupFile(dfs, ecFile);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tests to verify that the file checksum should be able to compute after the
|
||||||
|
* decommission operation.
|
||||||
|
*
|
||||||
|
* Below is the block indices list after the decommission. ' represents
|
||||||
|
* decommissioned node index.
|
||||||
|
*
|
||||||
|
* 0, 2, 3, 4, 5, 6, 7, 8, 1, 1'
|
||||||
|
*
|
||||||
|
* Here, this list contains duplicated blocks and does not maintaining any
|
||||||
|
* order.
|
||||||
|
*/
|
||||||
|
@Test(timeout = 120000)
|
||||||
|
public void testFileChecksumAfterDecommission() throws Exception {
|
||||||
|
LOG.info("Starting test testFileChecksumAfterDecommission");
|
||||||
|
|
||||||
|
final Path ecFile = new Path(ecDir, "testFileChecksumAfterDecommission");
|
||||||
|
int writeBytes = BLOCK_STRIPED_CELL_SIZE * NUM_DATA_BLOCKS;
|
||||||
|
writeStripedFile(dfs, ecFile, writeBytes);
|
||||||
|
Assert.assertEquals(0, bm.numOfUnderReplicatedBlocks());
|
||||||
|
FileChecksum fileChecksum1 = dfs.getFileChecksum(ecFile, writeBytes);
|
||||||
|
|
||||||
|
final List<DatanodeInfo> decommisionNodes = new ArrayList<DatanodeInfo>();
|
||||||
|
LocatedBlock lb = dfs.getClient().getLocatedBlocks(ecFile.toString(), 0)
|
||||||
|
.get(0);
|
||||||
|
DatanodeInfo[] dnLocs = lb.getLocations();
|
||||||
|
assertEquals(NUM_DATA_BLOCKS + NUM_PARITY_BLOCKS, dnLocs.length);
|
||||||
|
int decommNodeIndex = 1;
|
||||||
|
|
||||||
|
// add the node which will be decommissioning
|
||||||
|
decommisionNodes.add(dnLocs[decommNodeIndex]);
|
||||||
|
decommissionNode(0, decommisionNodes, AdminStates.DECOMMISSIONED);
|
||||||
|
assertEquals(decommisionNodes.size(), fsn.getNumDecomLiveDataNodes());
|
||||||
|
assertNull(checkFile(dfs, ecFile, 9, decommisionNodes, numDNs));
|
||||||
|
StripedFileTestUtil.checkData(dfs, ecFile, writeBytes, decommisionNodes,
|
||||||
|
null);
|
||||||
|
|
||||||
|
// verify checksum
|
||||||
|
FileChecksum fileChecksum2 = dfs.getFileChecksum(ecFile, writeBytes);
|
||||||
|
LOG.info("fileChecksum1:" + fileChecksum1);
|
||||||
|
LOG.info("fileChecksum2:" + fileChecksum2);
|
||||||
|
|
||||||
|
Assert.assertTrue("Checksum mismatches!",
|
||||||
|
fileChecksum1.equals(fileChecksum2));
|
||||||
|
}
|
||||||
|
|
||||||
private void testDecommission(int writeBytes, int storageCount,
|
private void testDecommission(int writeBytes, int storageCount,
|
||||||
int decomNodeCount, String filename) throws IOException, Exception {
|
int decomNodeCount, String filename) throws IOException, Exception {
|
||||||
Path ecFile = new Path(ecDir, filename);
|
Path ecFile = new Path(ecDir, filename);
|
||||||
|
|
|
@ -163,17 +163,40 @@ public class TestFileChecksum {
|
||||||
Assert.assertFalse(stripedFileChecksum1.equals(replicatedFileChecksum));
|
Assert.assertFalse(stripedFileChecksum1.equals(replicatedFileChecksum));
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
// TODO: allow datanode failure, HDFS-9833
|
|
||||||
@Test
|
@Test
|
||||||
public void testStripedAndReplicatedWithFailure() throws Exception {
|
public void testStripedFileChecksumWithMissedDataBlocks1() throws Exception {
|
||||||
FileChecksum stripedFileChecksum1 = getFileChecksum(stripedFile1,
|
FileChecksum stripedFileChecksum1 = getFileChecksum(stripedFile1, fileSize,
|
||||||
10, true);
|
false);
|
||||||
FileChecksum replicatedFileChecksum = getFileChecksum(replicatedFile,
|
FileChecksum stripedFileChecksumRecon = getFileChecksum(stripedFile1,
|
||||||
10, true);
|
fileSize, true);
|
||||||
|
|
||||||
Assert.assertFalse(stripedFileChecksum1.equals(replicatedFileChecksum));
|
LOG.info("stripedFileChecksum1:" + stripedFileChecksum1);
|
||||||
}*/
|
LOG.info("stripedFileChecksumRecon:" + stripedFileChecksumRecon);
|
||||||
|
|
||||||
|
Assert.assertTrue("Checksum mismatches!",
|
||||||
|
stripedFileChecksum1.equals(stripedFileChecksumRecon));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testStripedFileChecksumWithMissedDataBlocks2() throws Exception {
|
||||||
|
FileChecksum stripedFileChecksum1 = getFileChecksum(stripedFile1, -1,
|
||||||
|
false);
|
||||||
|
FileChecksum stripedFileChecksum2 = getFileChecksum(stripedFile2, -1,
|
||||||
|
false);
|
||||||
|
FileChecksum stripedFileChecksum2Recon = getFileChecksum(stripedFile2, -1,
|
||||||
|
true);
|
||||||
|
|
||||||
|
LOG.info("stripedFileChecksum1:" + stripedFileChecksum1);
|
||||||
|
LOG.info("stripedFileChecksum2:" + stripedFileChecksum1);
|
||||||
|
LOG.info("stripedFileChecksum2Recon:" + stripedFileChecksum2Recon);
|
||||||
|
|
||||||
|
Assert.assertTrue("Checksum mismatches!",
|
||||||
|
stripedFileChecksum1.equals(stripedFileChecksum2));
|
||||||
|
Assert.assertTrue("Checksum mismatches!",
|
||||||
|
stripedFileChecksum1.equals(stripedFileChecksum2Recon));
|
||||||
|
Assert.assertTrue("Checksum mismatches!",
|
||||||
|
stripedFileChecksum2.equals(stripedFileChecksum2Recon));
|
||||||
|
}
|
||||||
|
|
||||||
private FileChecksum getFileChecksum(String filePath, int range,
|
private FileChecksum getFileChecksum(String filePath, int range,
|
||||||
boolean killDn) throws Exception {
|
boolean killDn) throws Exception {
|
||||||
|
|
Loading…
Reference in New Issue