svn merge -c 1458906 FIXES: MAPREDUCE-4972. Coverage fixing for org.apache.hadoop.mapreduce.jobhistory (Aleksey Gorshkov via bobby)

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1458915 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Robert Joseph Evans 2013-03-20 16:11:58 +00:00
parent aa874f3952
commit 05fabc2013
6 changed files with 465 additions and 41 deletions

View File

@ -612,6 +612,9 @@ Release 0.23.7 - UNRELEASED
MAPREDUCE-5027. Shuffle does not limit number of outstanding connections
(Robert Parker via jeagles)
MAPREDUCE-4972. Coverage fixing for org.apache.hadoop.mapreduce.jobhistory
(Aleksey Gorshkov via bobby)
OPTIMIZATIONS
MAPREDUCE-4946. Fix a performance problem for large jobs by reducing the

View File

@ -0,0 +1,397 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapreduce.jobhistory;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.DataInputStream;
import java.util.ArrayList;
import java.util.Arrays;
import static junit.framework.Assert.*;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.mapred.JobPriority;
import org.apache.hadoop.mapreduce.Counters;
import org.apache.hadoop.mapreduce.JobID;
import org.apache.hadoop.mapreduce.TaskAttemptID;
import org.apache.hadoop.mapreduce.TaskID;
import org.apache.hadoop.mapreduce.TaskType;
import org.junit.Test;
public class TestEvents {
/**
* test a getters of TaskAttemptFinishedEvent and TaskAttemptFinished
*
* @throws Exception
*/
@Test(timeout = 10000)
public void testTaskAttemptFinishedEvent() throws Exception {
JobID jid = new JobID("001", 1);
TaskID tid = new TaskID(jid, TaskType.REDUCE, 2);
TaskAttemptID taskAttemptId = new TaskAttemptID(tid, 3);
Counters counters = new Counters();
TaskAttemptFinishedEvent test = new TaskAttemptFinishedEvent(taskAttemptId,
TaskType.REDUCE, "TEST", 123L, "RAKNAME", "HOSTNAME", "STATUS",
counters);
assertEquals(test.getAttemptId().toString(), taskAttemptId.toString());
assertEquals(test.getCounters(), counters);
assertEquals(test.getFinishTime(), 123L);
assertEquals(test.getHostname(), "HOSTNAME");
assertEquals(test.getRackName(), "RAKNAME");
assertEquals(test.getState(), "STATUS");
assertEquals(test.getTaskId(), tid);
assertEquals(test.getTaskStatus(), "TEST");
assertEquals(test.getTaskType(), TaskType.REDUCE);
}
/**
* simple test JobPriorityChangeEvent and JobPriorityChange
*
* @throws Exception
*/
@Test(timeout = 10000)
public void testJobPriorityChange() throws Exception {
org.apache.hadoop.mapreduce.JobID jid = new JobID("001", 1);
JobPriorityChangeEvent test = new JobPriorityChangeEvent(jid,
JobPriority.LOW);
assertEquals(test.getJobId().toString(), jid.toString());
assertEquals(test.getPriority(), JobPriority.LOW);
}
/**
* simple test TaskUpdatedEvent and TaskUpdated
*
* @throws Exception
*/
@Test(timeout = 10000)
public void testTaskUpdated() throws Exception {
JobID jid = new JobID("001", 1);
TaskID tid = new TaskID(jid, TaskType.REDUCE, 2);
TaskUpdatedEvent test = new TaskUpdatedEvent(tid, 1234L);
assertEquals(test.getTaskId().toString(), tid.toString());
assertEquals(test.getFinishTime(), 1234L);
}
/*
* test EventReader EventReader should read the list of events and return
* instance of HistoryEvent Different HistoryEvent should have a different
* datum.
*/
@Test(timeout = 10000)
public void testEvents() throws Exception {
EventReader reader = new EventReader(new DataInputStream(
new ByteArrayInputStream(getEvents())));
HistoryEvent e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.JOB_PRIORITY_CHANGED));
assertEquals("ID", ((JobPriorityChange) e.getDatum()).jobid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.JOB_STATUS_CHANGED));
assertEquals("ID", ((JobStatusChanged) e.getDatum()).jobid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.TASK_UPDATED));
assertEquals("ID", ((TaskUpdated) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_KILLED));
assertEquals("task_1_2_r03_4",
((TaskAttemptUnsuccessfulCompletion) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.JOB_KILLED));
assertEquals("ID",
((JobUnsuccessfulCompletion) e.getDatum()).jobid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_STARTED));
assertEquals("task_1_2_r03_4",
((TaskAttemptStarted) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_FINISHED));
assertEquals("task_1_2_r03_4",
((TaskAttemptFinished) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_KILLED));
assertEquals("task_1_2_r03_4",
((TaskAttemptUnsuccessfulCompletion) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_KILLED));
assertEquals("task_1_2_r03_4",
((TaskAttemptUnsuccessfulCompletion) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_STARTED));
assertEquals("task_1_2_r03_4",
((TaskAttemptStarted) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_FINISHED));
assertEquals("task_1_2_r03_4",
((TaskAttemptFinished) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_KILLED));
assertEquals("task_1_2_r03_4",
((TaskAttemptUnsuccessfulCompletion) e.getDatum()).taskid.toString());
e = reader.getNextEvent();
assertTrue(e.getEventType().equals(EventType.REDUCE_ATTEMPT_KILLED));
assertEquals("task_1_2_r03_4",
((TaskAttemptUnsuccessfulCompletion) e.getDatum()).taskid.toString());
reader.close();
}
/*
* makes array of bytes with History events
*/
private byte[] getEvents() throws Exception {
ByteArrayOutputStream output = new ByteArrayOutputStream();
FSDataOutputStream fsOutput = new FSDataOutputStream(output,
new FileSystem.Statistics("scheme"));
EventWriter writer = new EventWriter(fsOutput);
writer.write(getJobPriorityChangedEvent());
writer.write(getJobStatusChangedEvent());
writer.write(getTaskUpdatedEvent());
writer.write(getReduceAttemptKilledEvent());
writer.write(getJobKilledEvent());
writer.write(getSetupAttemptStartedEvent());
writer.write(getTaskAttemptFinishedEvent());
writer.write(getSetupAttemptFieledEvent());
writer.write(getSetupAttemptKilledEvent());
writer.write(getCleanupAttemptStartedEvent());
writer.write(getCleanupAttemptFinishedEvent());
writer.write(getCleanupAttemptFiledEvent());
writer.write(getCleanupAttemptKilledEvent());
writer.flush();
writer.close();
return output.toByteArray();
}
private FakeEvent getCleanupAttemptKilledEvent() {
FakeEvent result = new FakeEvent(EventType.CLEANUP_ATTEMPT_KILLED);
result.setDatum(getTaskAttemptUnsuccessfulCompletion());
return result;
}
private FakeEvent getCleanupAttemptFiledEvent() {
FakeEvent result = new FakeEvent(EventType.CLEANUP_ATTEMPT_FAILED);
result.setDatum(getTaskAttemptUnsuccessfulCompletion());
return result;
}
private TaskAttemptUnsuccessfulCompletion getTaskAttemptUnsuccessfulCompletion() {
TaskAttemptUnsuccessfulCompletion datum = new TaskAttemptUnsuccessfulCompletion();
datum.attemptId = "attempt_1_2_r3_4_5";
datum.clockSplits = Arrays.asList(1, 2, 3);
datum.cpuUsages = Arrays.asList(100, 200, 300);
datum.error = "Error";
datum.finishTime = 2;
datum.hostname = "hostname";
datum.rackname = "rackname";
datum.physMemKbytes = Arrays.asList(1000, 2000, 3000);
datum.taskid = "task_1_2_r03_4";
datum.port = 1000;
datum.taskType = "REDUCE";
datum.status = "STATUS";
datum.counters = getCounters();
datum.vMemKbytes = Arrays.asList(1000, 2000, 3000);
return datum;
}
private JhCounters getCounters() {
JhCounters counters = new JhCounters();
counters.groups = new ArrayList<JhCounterGroup>(0);
counters.name = "name";
return counters;
}
private FakeEvent getCleanupAttemptFinishedEvent() {
FakeEvent result = new FakeEvent(EventType.CLEANUP_ATTEMPT_FINISHED);
TaskAttemptFinished datum = new TaskAttemptFinished();
datum.attemptId = "attempt_1_2_r3_4_5";
datum.counters = getCounters();
datum.finishTime = 2;
datum.hostname = "hostname";
datum.rackname = "rackName";
datum.state = "state";
datum.taskid = "task_1_2_r03_4";
datum.taskStatus = "taskStatus";
datum.taskType = "REDUCE";
result.setDatum(datum);
return result;
}
private FakeEvent getCleanupAttemptStartedEvent() {
FakeEvent result = new FakeEvent(EventType.CLEANUP_ATTEMPT_STARTED);
TaskAttemptStarted datum = new TaskAttemptStarted();
datum.attemptId = "attempt_1_2_r3_4_5";
datum.avataar = "avatar";
datum.containerId = "containerId";
datum.httpPort = 10000;
datum.locality = "locality";
datum.shufflePort = 10001;
datum.startTime = 1;
datum.taskid = "task_1_2_r03_4";
datum.taskType = "taskType";
datum.trackerName = "trackerName";
result.setDatum(datum);
return result;
}
private FakeEvent getSetupAttemptKilledEvent() {
FakeEvent result = new FakeEvent(EventType.SETUP_ATTEMPT_KILLED);
result.setDatum(getTaskAttemptUnsuccessfulCompletion());
return result;
}
private FakeEvent getSetupAttemptFieledEvent() {
FakeEvent result = new FakeEvent(EventType.SETUP_ATTEMPT_FAILED);
result.setDatum(getTaskAttemptUnsuccessfulCompletion());
return result;
}
private FakeEvent getTaskAttemptFinishedEvent() {
FakeEvent result = new FakeEvent(EventType.SETUP_ATTEMPT_FINISHED);
TaskAttemptFinished datum = new TaskAttemptFinished();
datum.attemptId = "attempt_1_2_r3_4_5";
datum.counters = getCounters();
datum.finishTime = 2;
datum.hostname = "hostname";
datum.rackname = "rackname";
datum.state = "state";
datum.taskid = "task_1_2_r03_4";
datum.taskStatus = "taskStatus";
datum.taskType = "REDUCE";
result.setDatum(datum);
return result;
}
private FakeEvent getSetupAttemptStartedEvent() {
FakeEvent result = new FakeEvent(EventType.SETUP_ATTEMPT_STARTED);
TaskAttemptStarted datum = new TaskAttemptStarted();
datum.attemptId = "ID";
datum.avataar = "avataar";
datum.containerId = "containerId";
datum.httpPort = 10000;
datum.locality = "locality";
datum.shufflePort = 10001;
datum.startTime = 1;
datum.taskid = "task_1_2_r03_4";
datum.taskType = "taskType";
datum.trackerName = "trackerName";
result.setDatum(datum);
return result;
}
private FakeEvent getJobKilledEvent() {
FakeEvent result = new FakeEvent(EventType.JOB_KILLED);
JobUnsuccessfulCompletion datum = new JobUnsuccessfulCompletion();
datum.finishedMaps = 1;
datum.finishedReduces = 2;
datum.finishTime = 3;
datum.jobid = "ID";
datum.jobStatus = "STATUS";
result.setDatum(datum);
return result;
}
private FakeEvent getReduceAttemptKilledEvent() {
FakeEvent result = new FakeEvent(EventType.REDUCE_ATTEMPT_KILLED);
result.setDatum(getTaskAttemptUnsuccessfulCompletion());
return result;
}
private FakeEvent getJobPriorityChangedEvent() {
FakeEvent result = new FakeEvent(EventType.JOB_PRIORITY_CHANGED);
JobPriorityChange datum = new JobPriorityChange();
datum.jobid = "ID";
datum.priority = "priority";
result.setDatum(datum);
return result;
}
private FakeEvent getJobStatusChangedEvent() {
FakeEvent result = new FakeEvent(EventType.JOB_STATUS_CHANGED);
JobStatusChanged datum = new JobStatusChanged();
datum.jobid = "ID";
datum.jobStatus = "newStatus";
result.setDatum(datum);
return result;
}
private FakeEvent getTaskUpdatedEvent() {
FakeEvent result = new FakeEvent(EventType.TASK_UPDATED);
TaskUpdated datum = new TaskUpdated();
datum.finishTime = 2;
datum.taskid = "ID";
result.setDatum(datum);
return result;
}
private class FakeEvent implements HistoryEvent {
private EventType eventType;
private Object datum;
public FakeEvent(EventType eventType) {
this.eventType = eventType;
}
@Override
public EventType getEventType() {
return eventType;
}
@Override
public Object getDatum() {
return datum;
}
@Override
public void setDatum(Object datum) {
this.datum = datum;
}
}
}

View File

@ -31,6 +31,7 @@ import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileContext;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapreduce.Counters;
import org.apache.hadoop.mapreduce.JobID;
import org.apache.hadoop.mapreduce.MRJobConfig;
import org.apache.hadoop.mapreduce.TaskID;
import org.apache.hadoop.mapreduce.TaskType;
@ -53,7 +54,7 @@ public class TestJobHistoryEventHandler {
private static final Log LOG = LogFactory
.getLog(TestJobHistoryEventHandler.class);
@Test
@Test (timeout=50000)
public void testFirstFlushOnCompletionEvent() throws Exception {
TestParams t = new TestParams();
Configuration conf = new Configuration();
@ -96,7 +97,7 @@ public class TestJobHistoryEventHandler {
}
}
@Test
@Test (timeout=50000)
public void testMaxUnflushedCompletionEvents() throws Exception {
TestParams t = new TestParams();
Configuration conf = new Configuration();
@ -141,7 +142,7 @@ public class TestJobHistoryEventHandler {
}
}
@Test
@Test (timeout=50000)
public void testUnflushedTimer() throws Exception {
TestParams t = new TestParams();
Configuration conf = new Configuration();
@ -182,7 +183,7 @@ public class TestJobHistoryEventHandler {
}
}
@Test
@Test (timeout=50000)
public void testBatchedFlushJobEndMultiplier() throws Exception {
TestParams t = new TestParams();
Configuration conf = new Configuration();
@ -279,12 +280,8 @@ public class TestJobHistoryEventHandler {
}
private JobHistoryEvent getEventToEnqueue(JobId jobId) {
JobHistoryEvent toReturn = Mockito.mock(JobHistoryEvent.class);
HistoryEvent he = Mockito.mock(HistoryEvent.class);
Mockito.when(he.getEventType()).thenReturn(EventType.JOB_STATUS_CHANGED);
Mockito.when(toReturn.getHistoryEvent()).thenReturn(he);
Mockito.when(toReturn.getJobID()).thenReturn(jobId);
return toReturn;
HistoryEvent toReturn = new JobStatusChangedEvent(new JobID(Integer.toString(jobId.getId()), jobId.getId()), "change status");
return new JobHistoryEvent(jobId, toReturn);
}
@Test
@ -344,8 +341,6 @@ public class TestJobHistoryEventHandler {
class JHEvenHandlerForTest extends JobHistoryEventHandler {
private EventWriter eventWriter;
volatile int handleEventCompleteCalls = 0;
volatile int handleEventStartedCalls = 0;
public JHEvenHandlerForTest(AppContext context, int startCount) {
super(context, startCount);
@ -375,13 +370,12 @@ class JHEvenHandlerForTest extends JobHistoryEventHandler {
* Class to help with testSigTermedFunctionality
*/
class JHEventHandlerForSigtermTest extends JobHistoryEventHandler {
private MetaInfo metaInfo;
public JHEventHandlerForSigtermTest(AppContext context, int startCount) {
super(context, startCount);
}
public void addToFileMap(JobId jobId) {
metaInfo = Mockito.mock(MetaInfo.class);
MetaInfo metaInfo = Mockito.mock(MetaInfo.class);
Mockito.when(metaInfo.isWriterActive()).thenReturn(true);
fileMap.put(jobId, metaInfo);
}

View File

@ -79,7 +79,7 @@ public class TestJobHistoryEntities {
}
/* Verify some expected values based on the history file */
@Test
@Test (timeout=10000)
public void testCompletedJob() throws Exception {
HistoryFileInfo info = mock(HistoryFileInfo.class);
when(info.getConfFile()).thenReturn(fullConfPath);
@ -104,7 +104,7 @@ public class TestJobHistoryEntities {
assertEquals(JobState.SUCCEEDED, jobReport.getJobState());
}
@Test
@Test (timeout=10000)
public void testCompletedTask() throws Exception {
HistoryFileInfo info = mock(HistoryFileInfo.class);
when(info.getConfFile()).thenReturn(fullConfPath);
@ -133,7 +133,7 @@ public class TestJobHistoryEntities {
assertEquals(rt1Id, rt1Report.getTaskId());
}
@Test
@Test (timeout=10000)
public void testCompletedTaskAttempt() throws Exception {
HistoryFileInfo info = mock(HistoryFileInfo.class);
when(info.getConfFile()).thenReturn(fullConfPath);

View File

@ -25,7 +25,6 @@ import junit.framework.Assert;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.mapreduce.MRJobConfig;
import org.apache.hadoop.mapreduce.TypeConverter;
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryEvent;
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler;
@ -67,8 +66,17 @@ public class TestJobHistoryEvents {
* completed maps
*/
HistoryContext context = new JobHistory();
// test start and stop states
((JobHistory)context).init(conf);
Job parsedJob = context.getJob(jobId);
((JobHistory)context).start();
Assert.assertTrue( context.getStartTime()>0);
Assert.assertEquals(((JobHistory)context).getServiceState(),Service.STATE.STARTED);
((JobHistory)context).stop();
Assert.assertEquals(((JobHistory)context).getServiceState(),Service.STATE.STOPPED);
Job parsedJob = context.getJob(jobId);
Assert.assertEquals("CompletedMaps not correct", 2,
parsedJob.getCompletedMaps());
Assert.assertEquals(System.getProperty("user.name"), parsedJob.getUserName());
@ -177,9 +185,8 @@ public class TestJobHistoryEvents {
@Override
protected EventHandler<JobHistoryEvent> createJobHistoryHandler(
AppContext context) {
JobHistoryEventHandler eventHandler = new JobHistoryEventHandler(
context, getStartCount());
return eventHandler;
return new JobHistoryEventHandler(
context, getStartCount());
}
}

View File

@ -18,7 +18,9 @@
package org.apache.hadoop.mapreduce.v2.hs;
import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.io.PrintStream;
import java.util.Arrays;
import java.util.HashMap;
import java.util.List;
@ -40,6 +42,7 @@ import org.apache.hadoop.mapreduce.TaskID;
import org.apache.hadoop.mapreduce.TypeConverter;
import org.apache.hadoop.mapreduce.jobhistory.EventReader;
import org.apache.hadoop.mapreduce.jobhistory.HistoryEvent;
import org.apache.hadoop.mapreduce.jobhistory.HistoryViewer;
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser;
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.AMInfo;
import org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo;
@ -60,7 +63,6 @@ import org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType;
import org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.HistoryFileInfo;
import org.apache.hadoop.mapreduce.v2.hs.TestJobHistoryEvents.MRAppWithHistory;
import org.apache.hadoop.mapreduce.v2.jobhistory.FileNameIndexUtils;
import org.apache.hadoop.mapreduce.v2.jobhistory.JHAdminConfig;
import org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils;
import org.apache.hadoop.mapreduce.v2.jobhistory.JobIndexInfo;
import org.apache.hadoop.net.DNSToSwitchMapping;
@ -78,6 +80,8 @@ public class TestJobHistoryParsing {
private static final String RACK_NAME = "/MyRackName";
private ByteArrayOutputStream outContent = new ByteArrayOutputStream();
public static class MyResolver implements DNSToSwitchMapping {
@Override
public List<String> resolve(List<String> names) {
@ -89,14 +93,14 @@ public class TestJobHistoryParsing {
}
}
@Test
@Test (timeout=50000)
public void testJobInfo() throws Exception {
JobInfo info = new JobInfo();
Assert.assertEquals("NORMAL", info.getPriority());
info.printAll();
}
@Test
@Test (timeout=50000)
public void testHistoryParsing() throws Exception {
LOG.info("STARTING testHistoryParsing()");
try {
@ -106,7 +110,7 @@ public class TestJobHistoryParsing {
}
}
@Test
@Test (timeout=50000)
public void testHistoryParsingWithParseErrors() throws Exception {
LOG.info("STARTING testHistoryParsingWithParseErrors()");
try {
@ -321,18 +325,37 @@ public class TestJobHistoryParsing {
}
}
}
// test output for HistoryViewer
PrintStream stdps=System.out;
try {
System.setOut(new PrintStream(outContent));
HistoryViewer viewer = new HistoryViewer(fc.makeQualified(
fileInfo.getHistoryFile()).toString(), conf, true);
viewer.print();
for (TaskInfo taskInfo : allTasks.values()) {
String test= (taskInfo.getTaskStatus()==null?"":taskInfo.getTaskStatus())+" "+taskInfo.getTaskType()+" task list for "+taskInfo.getTaskId().getJobID();
Assert.assertTrue(outContent.toString().indexOf(test)>0);
Assert.assertTrue(outContent.toString().indexOf(taskInfo.getTaskId().toString())>0);
}
} finally {
System.setOut(stdps);
}
}
// Computes finished maps similar to RecoveryService...
private long computeFinishedMaps(JobInfo jobInfo,
int numMaps, int numSuccessfulMaps) {
private long computeFinishedMaps(JobInfo jobInfo, int numMaps,
int numSuccessfulMaps) {
if (numMaps == numSuccessfulMaps) {
return jobInfo.getFinishedMaps();
}
long numFinishedMaps = 0;
Map<org.apache.hadoop.mapreduce.TaskID, TaskInfo> taskInfos =
jobInfo.getAllTasks();
Map<org.apache.hadoop.mapreduce.TaskID, TaskInfo> taskInfos = jobInfo
.getAllTasks();
for (TaskInfo taskInfo : taskInfos.values()) {
if (TaskState.SUCCEEDED.toString().equals(taskInfo.getTaskStatus())) {
++numFinishedMaps;
@ -341,7 +364,7 @@ public class TestJobHistoryParsing {
return numFinishedMaps;
}
@Test
@Test (timeout=50000)
public void testHistoryParsingForFailedAttempts() throws Exception {
LOG.info("STARTING testHistoryParsingForFailedAttempts");
try {
@ -468,7 +491,7 @@ public class TestJobHistoryParsing {
}
}
@Test
@Test (timeout=50000)
public void testScanningOldDirs() throws Exception {
LOG.info("STARTING testScanningOldDirs");
try {