HDFS-15451. Do not discard non-initial block report for provided storage. (#2119). Contributed by Shanyu Zhao.
Signed-off-by: He Xiaoqiao <hexiaoqiao@apache.org>
This commit is contained in:
parent
0789ae5b78
commit
10c9df1d0a
|
@ -2745,6 +2745,7 @@ public class BlockManager implements BlockStatsMXBean {
|
||||||
storageInfo = node.updateStorage(storage);
|
storageInfo = node.updateStorage(storage);
|
||||||
}
|
}
|
||||||
if (namesystem.isInStartupSafeMode()
|
if (namesystem.isInStartupSafeMode()
|
||||||
|
&& !StorageType.PROVIDED.equals(storageInfo.getStorageType())
|
||||||
&& storageInfo.getBlockReportCount() > 0) {
|
&& storageInfo.getBlockReportCount() > 0) {
|
||||||
blockLog.info("BLOCK* processReport 0x{}: "
|
blockLog.info("BLOCK* processReport 0x{}: "
|
||||||
+ "discarded non-initial block report from {}"
|
+ "discarded non-initial block report from {}"
|
||||||
|
|
|
@ -49,9 +49,11 @@ import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor.BlockTargetPair;
|
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor.BlockTargetPair;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.FinalizedReplica;
|
import org.apache.hadoop.hdfs.server.datanode.FinalizedReplica;
|
||||||
|
import org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestProvidedImpl;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.InternalDataNodeTestUtils;
|
import org.apache.hadoop.hdfs.server.datanode.InternalDataNodeTestUtils;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.ReplicaBeingWritten;
|
import org.apache.hadoop.hdfs.server.datanode.ReplicaBeingWritten;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.CacheManager;
|
import org.apache.hadoop.hdfs.server.namenode.CacheManager;
|
||||||
|
@ -1051,6 +1053,57 @@ public class TestBlockManager {
|
||||||
(ds) >= 0);
|
(ds) >= 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSafeModeWithProvidedStorageBR() throws Exception {
|
||||||
|
DatanodeDescriptor node0 = spy(nodes.get(0));
|
||||||
|
DatanodeStorageInfo ds0 = node0.getStorageInfos()[0];
|
||||||
|
node0.setAlive(true);
|
||||||
|
DatanodeDescriptor node1 = spy(nodes.get(1));
|
||||||
|
DatanodeStorageInfo ds1 = node1.getStorageInfos()[0];
|
||||||
|
node1.setAlive(true);
|
||||||
|
|
||||||
|
String providedStorageID = DFSConfigKeys.DFS_PROVIDER_STORAGEUUID_DEFAULT;
|
||||||
|
DatanodeStorage providedStorage = new DatanodeStorage(
|
||||||
|
providedStorageID, DatanodeStorage.State.NORMAL, StorageType.PROVIDED);
|
||||||
|
|
||||||
|
// create block manager with provided storage enabled
|
||||||
|
Configuration conf = new HdfsConfiguration();
|
||||||
|
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_PROVIDED_ENABLED, true);
|
||||||
|
conf.setClass(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_CLASS,
|
||||||
|
TestProvidedImpl.TestFileRegionBlockAliasMap.class,
|
||||||
|
BlockAliasMap.class);
|
||||||
|
BlockManager bmPs = new BlockManager(fsn, false, conf);
|
||||||
|
bmPs.setBlockPoolId("BP-12344-10.1.1.2-12344");
|
||||||
|
|
||||||
|
// pretend to be in safemode
|
||||||
|
doReturn(true).when(fsn).isInStartupSafeMode();
|
||||||
|
|
||||||
|
// register new node
|
||||||
|
DatanodeRegistration nodeReg0 =
|
||||||
|
new DatanodeRegistration(node0, null, null, "");
|
||||||
|
bmPs.getDatanodeManager().registerDatanode(nodeReg0);
|
||||||
|
bmPs.getDatanodeManager().addDatanode(node0);
|
||||||
|
DatanodeRegistration nodeReg1 =
|
||||||
|
new DatanodeRegistration(node1, null, null, "");
|
||||||
|
bmPs.getDatanodeManager().registerDatanode(nodeReg1);
|
||||||
|
bmPs.getDatanodeManager().addDatanode(node1);
|
||||||
|
|
||||||
|
// process reports of provided storage and disk storage
|
||||||
|
bmPs.processReport(node0, providedStorage, BlockListAsLongs.EMPTY, null);
|
||||||
|
bmPs.processReport(node0, new DatanodeStorage(ds0.getStorageID()),
|
||||||
|
BlockListAsLongs.EMPTY, null);
|
||||||
|
bmPs.processReport(node1, providedStorage, BlockListAsLongs.EMPTY, null);
|
||||||
|
bmPs.processReport(node1, new DatanodeStorage(ds1.getStorageID()),
|
||||||
|
BlockListAsLongs.EMPTY, null);
|
||||||
|
|
||||||
|
// The provided stoage report should not affect disk storage report
|
||||||
|
DatanodeStorageInfo dsPs =
|
||||||
|
bmPs.getProvidedStorageMap().getProvidedStorageInfo();
|
||||||
|
assertEquals(2, dsPs.getBlockReportCount());
|
||||||
|
assertEquals(1, ds0.getBlockReportCount());
|
||||||
|
assertEquals(1, ds1.getBlockReportCount());
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testFullBR() throws Exception {
|
public void testFullBR() throws Exception {
|
||||||
doReturn(true).when(fsn).isRunning();
|
doReturn(true).when(fsn).isRunning();
|
||||||
|
|
Loading…
Reference in New Issue