HDFS-7261. storageMap is accessed without synchronization in DatanodeDescriptor#updateHeartbeatState() (Brahma Reddy Battula via Colin P. McCabe)
(cherry picked from commit 1feb9569f3
)
This commit is contained in:
parent
cc130a033a
commit
02ed22cd2d
|
@ -64,6 +64,10 @@ Release 2.8.0 - UNRELEASED
|
|||
HDFS-8002. Website refers to /trash directory. (Brahma Reddy Battula via
|
||||
aajisaka)
|
||||
|
||||
HDFS-7261. storageMap is accessed without synchronization in
|
||||
DatanodeDescriptor#updateHeartbeatState() (Brahma Reddy Battula via Colin
|
||||
P. McCabe)
|
||||
|
||||
Release 2.7.0 - UNRELEASED
|
||||
|
||||
INCOMPATIBLE CHANGES
|
||||
|
|
|
@ -449,8 +449,10 @@ public class DatanodeDescriptor extends DatanodeInfo {
|
|||
if (checkFailedStorages) {
|
||||
LOG.info("Number of failed storage changes from "
|
||||
+ this.volumeFailures + " to " + volFailures);
|
||||
failedStorageInfos = new HashSet<DatanodeStorageInfo>(
|
||||
storageMap.values());
|
||||
synchronized (storageMap) {
|
||||
failedStorageInfos =
|
||||
new HashSet<DatanodeStorageInfo>(storageMap.values());
|
||||
}
|
||||
}
|
||||
|
||||
setCacheCapacity(cacheCapacity);
|
||||
|
@ -482,8 +484,11 @@ public class DatanodeDescriptor extends DatanodeInfo {
|
|||
if (checkFailedStorages) {
|
||||
updateFailedStorage(failedStorageInfos);
|
||||
}
|
||||
|
||||
if (storageMap.size() != reports.length) {
|
||||
long storageMapSize;
|
||||
synchronized (storageMap) {
|
||||
storageMapSize = storageMap.size();
|
||||
}
|
||||
if (storageMapSize != reports.length) {
|
||||
pruneStorageMap(reports);
|
||||
}
|
||||
}
|
||||
|
@ -493,14 +498,14 @@ public class DatanodeDescriptor extends DatanodeInfo {
|
|||
* as long as they have associated block replicas.
|
||||
*/
|
||||
private void pruneStorageMap(final StorageReport[] reports) {
|
||||
if (LOG.isDebugEnabled()) {
|
||||
LOG.debug("Number of storages reported in heartbeat=" + reports.length +
|
||||
"; Number of storages in storageMap=" + storageMap.size());
|
||||
}
|
||||
|
||||
HashMap<String, DatanodeStorageInfo> excessStorages;
|
||||
|
||||
synchronized (storageMap) {
|
||||
if (LOG.isDebugEnabled()) {
|
||||
LOG.debug("Number of storages reported in heartbeat=" + reports.length
|
||||
+ "; Number of storages in storageMap=" + storageMap.size());
|
||||
}
|
||||
|
||||
HashMap<String, DatanodeStorageInfo> excessStorages;
|
||||
|
||||
// Init excessStorages with all known storages.
|
||||
excessStorages = new HashMap<String, DatanodeStorageInfo>(storageMap);
|
||||
|
||||
|
@ -517,8 +522,8 @@ public class DatanodeDescriptor extends DatanodeInfo {
|
|||
LOG.info("Removed storage " + storageInfo + " from DataNode" + this);
|
||||
} else if (LOG.isDebugEnabled()) {
|
||||
// This can occur until all block reports are received.
|
||||
LOG.debug("Deferring removal of stale storage " + storageInfo +
|
||||
" with " + storageInfo.numBlocks() + " blocks");
|
||||
LOG.debug("Deferring removal of stale storage " + storageInfo
|
||||
+ " with " + storageInfo.numBlocks() + " blocks");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue