From 0128a87eca1a3205631756003aeee241feb959ab Mon Sep 17 00:00:00 2001 From: Suresh Srinivas Date: Wed, 19 Jun 2013 04:04:01 +0000 Subject: [PATCH] HDFS-4917. Revert r1494434 previous merge that included extraneous change. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1494442 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 --- hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-dfs.sh | 2 +- .../hdfs/server/blockmanagement/UnderReplicatedBlocks.java | 6 +++--- 3 files changed, 4 insertions(+), 7 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index d6ffe9e2f18..4e274766d13 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -397,9 +397,6 @@ Release 2.1.0-beta - UNRELEASED HDFS-4906. HDFS Output streams should not accept writes after being closed. (atm) - HDFS-4917. Start-dfs.sh cannot pass the parameters correctly. - (Fengdong Yu via suresh) - BREAKDOWN OF HDFS-347 SUBTASKS AND RELATED JIRAS HDFS-4353. Encapsulate connections to peers in Peer and PeerServer classes. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-dfs.sh b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-dfs.sh index 8cbea16aa61..fb4edd069fe 100755 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-dfs.sh +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-dfs.sh @@ -47,7 +47,7 @@ if [ $# -ge 1 ]; then fi #Add other possible options -nameStartOpt="$nameStartOpt $@" +nameStartOpt="$nameStartOpts $@" #--------------------------------------------------------- # namenodes diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/UnderReplicatedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/UnderReplicatedBlocks.java index 26011b95d12..83a29d2f91b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/UnderReplicatedBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/UnderReplicatedBlocks.java @@ -190,7 +190,7 @@ class UnderReplicatedBlocks implements Iterable { assert curReplicas >= 0 : "Negative replicas!"; int priLevel = getPriority(block, curReplicas, decomissionedReplicas, expectedReplicas); - if(priorityQueues.get(priLevel).add(block)) { + if(priLevel != LEVEL && priorityQueues.get(priLevel).add(block)) { if(NameNode.blockStateChangeLog.isDebugEnabled()) { NameNode.blockStateChangeLog.debug( "BLOCK* NameSystem.UnderReplicationBlock.add:" @@ -293,10 +293,10 @@ class UnderReplicatedBlocks implements Iterable { " curPri " + curPri + " oldPri " + oldPri); } - if(oldPri != curPri) { + if(oldPri != LEVEL && oldPri != curPri) { remove(block, oldPri); } - if(priorityQueues.get(curPri).add(block)) { + if(curPri != LEVEL && priorityQueues.get(curPri).add(block)) { if(NameNode.blockStateChangeLog.isDebugEnabled()) { NameNode.blockStateChangeLog.debug( "BLOCK* NameSystem.UnderReplicationBlock.update:"