Merge trunk into HA branch.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-1623@1242871 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
commit
05d56e5e7e
|
@ -13,11 +13,6 @@ Trunk (unreleased changes)
|
|||
(suresh)
|
||||
|
||||
IMPROVEMENTS
|
||||
HADOOP-8048. Allow merging of Credentials (Daryn Sharp via tgraves)
|
||||
|
||||
HADOOP-8032. mvn site:stage-deploy should be able to use the scp protocol
|
||||
to stage documents (Ravi Prakash via tgraves)
|
||||
|
||||
HADOOP-8017. Configure hadoop-main pom to get rid of M2E plugin execution
|
||||
not covered (Eric Charles via bobby)
|
||||
|
||||
|
@ -160,6 +155,11 @@ Release 0.23.2 - UNRELEASED
|
|||
|
||||
IMPROVEMENTS
|
||||
|
||||
HADOOP-8048. Allow merging of Credentials (Daryn Sharp via tgraves)
|
||||
|
||||
HADOOP-8032. mvn site:stage-deploy should be able to use the scp protocol
|
||||
to stage documents (Ravi Prakash via tgraves)
|
||||
|
||||
HADOOP-7923. Automate the updating of version numbers in the doc system.
|
||||
(szetszwo)
|
||||
|
||||
|
|
|
@ -126,6 +126,9 @@ Trunk (unreleased changes)
|
|||
HDFS-2786. Fix host-based token incompatibilities in DFSUtil. (Kihwal Lee
|
||||
via jitendra)
|
||||
|
||||
HDFS-2486. Remove unnecessary priority level checks in
|
||||
UnderReplicatedBlocks. (Uma Maheswara Rao G via szetszwo)
|
||||
|
||||
OPTIMIZATIONS
|
||||
HDFS-2477. Optimize computing the diff between a block report and the
|
||||
namenode state. (Tomasz Nykiel via hairong)
|
||||
|
|
|
@ -189,7 +189,7 @@ class UnderReplicatedBlocks implements Iterable<Block> {
|
|||
assert curReplicas >= 0 : "Negative replicas!";
|
||||
int priLevel = getPriority(block, curReplicas, decomissionedReplicas,
|
||||
expectedReplicas);
|
||||
if(priLevel != LEVEL && priorityQueues.get(priLevel).add(block)) {
|
||||
if(priorityQueues.get(priLevel).add(block)) {
|
||||
if(NameNode.stateChangeLog.isDebugEnabled()) {
|
||||
NameNode.stateChangeLog.debug(
|
||||
"BLOCK* NameSystem.UnderReplicationBlock.add:"
|
||||
|
@ -292,10 +292,10 @@ class UnderReplicatedBlocks implements Iterable<Block> {
|
|||
" curPri " + curPri +
|
||||
" oldPri " + oldPri);
|
||||
}
|
||||
if(oldPri != LEVEL && oldPri != curPri) {
|
||||
if(oldPri != curPri) {
|
||||
remove(block, oldPri);
|
||||
}
|
||||
if(curPri != LEVEL && priorityQueues.get(curPri).add(block)) {
|
||||
if(priorityQueues.get(curPri).add(block)) {
|
||||
if(NameNode.stateChangeLog.isDebugEnabled()) {
|
||||
NameNode.stateChangeLog.debug(
|
||||
"BLOCK* NameSystem.UnderReplicationBlock.update:"
|
||||
|
|
|
@ -98,8 +98,11 @@ Release 0.23.2 - UNRELEASED
|
|||
OPTIMIZATIONS
|
||||
|
||||
BUG FIXES
|
||||
MAPREDUCE-3840. JobEndNotifier doesn't use the proxyToUse during connecting
|
||||
(Ravi Prakash via bobby)
|
||||
MAPREDUCE-3680. FifoScheduler web service rest API can print out invalid
|
||||
JSON. (B Anil Kumar via tgraves)
|
||||
|
||||
MAPREDUCE-3840. JobEndNotifier doesn't use the proxyToUse during connecting
|
||||
(Ravi Prakash via bobby)
|
||||
|
||||
Release 0.23.1 - 2012-02-08
|
||||
|
||||
|
|
|
@ -148,7 +148,12 @@ public class FifoScheduler implements ResourceScheduler {
|
|||
QueueInfo queueInfo = recordFactory.newRecordInstance(QueueInfo.class);
|
||||
queueInfo.setQueueName(DEFAULT_QUEUE.getQueueName());
|
||||
queueInfo.setCapacity(1.0f);
|
||||
queueInfo.setCurrentCapacity((float)usedResource.getMemory() / clusterResource.getMemory());
|
||||
if (clusterResource.getMemory() == 0) {
|
||||
queueInfo.setCurrentCapacity(0.0f);
|
||||
} else {
|
||||
queueInfo.setCurrentCapacity((float) usedResource.getMemory()
|
||||
/ clusterResource.getMemory());
|
||||
}
|
||||
queueInfo.setMaximumCapacity(1.0f);
|
||||
queueInfo.setChildQueues(new ArrayList<QueueInfo>());
|
||||
queueInfo.setQueueState(QueueState.RUNNING);
|
||||
|
|
|
@ -29,6 +29,7 @@ import org.apache.hadoop.net.NetworkTopology;
|
|||
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||
import org.apache.hadoop.yarn.api.records.Priority;
|
||||
import org.apache.hadoop.yarn.api.records.QueueInfo;
|
||||
import org.apache.hadoop.yarn.api.records.Resource;
|
||||
import org.apache.hadoop.yarn.event.AsyncDispatcher;
|
||||
import org.apache.hadoop.yarn.server.resourcemanager.Application;
|
||||
|
@ -74,6 +75,13 @@ public class TestFifoScheduler {
|
|||
.getRMContext());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFifoSchedulerCapacityWhenNoNMs() {
|
||||
FifoScheduler scheduler = new FifoScheduler();
|
||||
QueueInfo queueInfo = scheduler.getQueueInfo(null, false, false);
|
||||
Assert.assertEquals(0.0f, queueInfo.getCurrentCapacity());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testAppAttemptMetrics() throws Exception {
|
||||
AsyncDispatcher dispatcher = new InlineDispatcher();
|
||||
|
|
|
@ -539,7 +539,7 @@ public class TestRMWebServices extends JerseyTest {
|
|||
assertEquals("type doesn't match", "fifoScheduler", type);
|
||||
assertEquals("qstate doesn't match", QueueState.RUNNING.toString(), state);
|
||||
assertEquals("capacity doesn't match", 1.0, capacity, 0.0);
|
||||
assertEquals("usedCapacity doesn't match", Float.NaN, usedCapacity, 0.0);
|
||||
assertEquals("usedCapacity doesn't match", 0.0, usedCapacity, 0.0);
|
||||
assertEquals("minQueueMemoryCapacity doesn't match", 1024, minQueueCapacity);
|
||||
assertEquals("maxQueueMemoryCapacity doesn't match", 10240,
|
||||
maxQueueCapacity);
|
||||
|
|
Loading…
Reference in New Issue