MAPREDUCE-5960. JobSubmitter's check whether job.jar is local is incorrect with no authority in job jar path. Contributed by Gera Shegalov
This commit is contained in:
parent
80d7d183cd
commit
10f9f5101c
|
@ -430,6 +430,9 @@ public class FileContext {
|
|||
final Configuration aConf) throws UnsupportedFileSystemException {
|
||||
UserGroupInformation currentUser = null;
|
||||
AbstractFileSystem defaultAfs = null;
|
||||
if (defaultFsUri.getScheme() == null) {
|
||||
return getFileContext(aConf);
|
||||
}
|
||||
try {
|
||||
currentUser = UserGroupInformation.getCurrentUser();
|
||||
defaultAfs = getAbstractFileSystem(currentUser, defaultFsUri, aConf);
|
||||
|
|
|
@ -459,6 +459,9 @@ Release 2.6.0 - UNRELEASED
|
|||
MAPREDUCE-6048. Fixed TestJavaSerialization failure. (Varun Vasudev via
|
||||
jianhe)
|
||||
|
||||
MAPREDUCE-5960. JobSubmitter's check whether job.jar is local is incorrect
|
||||
with no authority in job jar path. (Gera Shegalov via jlowe)
|
||||
|
||||
Release 2.5.2 - UNRELEASED
|
||||
|
||||
INCOMPATIBLE CHANGES
|
||||
|
|
|
@ -657,9 +657,11 @@ public abstract class TaskAttemptImpl implements
|
|||
// //////////// Set up JobJar to be localized properly on the remote NM.
|
||||
String jobJar = conf.get(MRJobConfig.JAR);
|
||||
if (jobJar != null) {
|
||||
Path remoteJobJar = (new Path(jobJar)).makeQualified(remoteFS
|
||||
.getUri(), remoteFS.getWorkingDirectory());
|
||||
LocalResource rc = createLocalResource(remoteFS, remoteJobJar,
|
||||
final Path jobJarPath = new Path(jobJar);
|
||||
final FileSystem jobJarFs = FileSystem.get(jobJarPath.toUri(), conf);
|
||||
Path remoteJobJar = jobJarPath.makeQualified(jobJarFs.getUri(),
|
||||
jobJarFs.getWorkingDirectory());
|
||||
LocalResource rc = createLocalResource(jobJarFs, remoteJobJar,
|
||||
LocalResourceType.PATTERN, LocalResourceVisibility.APPLICATION);
|
||||
String pattern = conf.getPattern(JobContext.JAR_UNPACK_PATTERN,
|
||||
JobConf.UNPACK_JAR_PATTERN_DEFAULT).pattern();
|
||||
|
|
|
@ -250,11 +250,10 @@ class JobSubmitter {
|
|||
}
|
||||
Path jobJarPath = new Path(jobJar);
|
||||
URI jobJarURI = jobJarPath.toUri();
|
||||
// If the job jar is already in fs, we don't need to copy it from local fs
|
||||
if (jobJarURI.getScheme() == null || jobJarURI.getAuthority() == null
|
||||
|| !(jobJarURI.getScheme().equals(jtFs.getUri().getScheme())
|
||||
&& jobJarURI.getAuthority().equals(
|
||||
jtFs.getUri().getAuthority()))) {
|
||||
// If the job jar is already in a global fs,
|
||||
// we don't need to copy it from local fs
|
||||
if ( jobJarURI.getScheme() == null
|
||||
|| jobJarURI.getScheme().equals("file")) {
|
||||
copyJar(jobJarPath, JobSubmissionFiles.getJobJar(submitJobDir),
|
||||
replication);
|
||||
job.setJar(JobSubmissionFiles.getJobJar(submitJobDir).toString());
|
||||
|
|
|
@ -357,8 +357,9 @@ public class YARNRunner implements ClientProtocol {
|
|||
jobConfPath, LocalResourceType.FILE));
|
||||
if (jobConf.get(MRJobConfig.JAR) != null) {
|
||||
Path jobJarPath = new Path(jobConf.get(MRJobConfig.JAR));
|
||||
LocalResource rc = createApplicationResource(defaultFileContext,
|
||||
jobJarPath,
|
||||
LocalResource rc = createApplicationResource(
|
||||
FileContext.getFileContext(jobJarPath.toUri(), jobConf),
|
||||
jobJarPath,
|
||||
LocalResourceType.PATTERN);
|
||||
String pattern = conf.getPattern(JobContext.JAR_UNPACK_PATTERN,
|
||||
JobConf.UNPACK_JAR_PATTERN_DEFAULT).pattern();
|
||||
|
|
|
@ -40,6 +40,7 @@ import org.apache.commons.logging.LogFactory;
|
|||
import org.apache.hadoop.FailingMapper;
|
||||
import org.apache.hadoop.RandomTextWriterJob;
|
||||
import org.apache.hadoop.RandomTextWriterJob.RandomInputFormat;
|
||||
import org.apache.hadoop.fs.viewfs.ConfigUtil;
|
||||
import org.apache.hadoop.mapreduce.SleepJob;
|
||||
import org.apache.hadoop.mapreduce.SleepJob.SleepMapper;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
|
@ -89,6 +90,7 @@ import org.apache.hadoop.security.UserGroupInformation;
|
|||
import org.apache.hadoop.security.token.Token;
|
||||
import org.apache.hadoop.security.token.TokenIdentifier;
|
||||
import org.apache.hadoop.util.ApplicationClassLoader;
|
||||
import org.apache.hadoop.util.ClassUtil;
|
||||
import org.apache.hadoop.util.JarFinder;
|
||||
import org.apache.hadoop.util.Shell;
|
||||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||
|
@ -97,6 +99,7 @@ import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
|||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppState;
|
||||
import org.apache.hadoop.yarn.util.ConverterUtils;
|
||||
import org.apache.log4j.Level;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
|
@ -111,6 +114,9 @@ public class TestMRJobs {
|
|||
private static final String TEST_IO_SORT_MB = "11";
|
||||
private static final String TEST_GROUP_MAX = "200";
|
||||
|
||||
private static final int DEFAULT_REDUCES = 2;
|
||||
protected int numSleepReducers = DEFAULT_REDUCES;
|
||||
|
||||
protected static MiniMRYarnCluster mrCluster;
|
||||
protected static MiniDFSCluster dfsCluster;
|
||||
|
||||
|
@ -175,10 +181,23 @@ public class TestMRJobs {
|
|||
}
|
||||
}
|
||||
|
||||
@After
|
||||
public void resetInit() {
|
||||
numSleepReducers = DEFAULT_REDUCES;
|
||||
}
|
||||
|
||||
@Test (timeout = 300000)
|
||||
public void testSleepJob() throws IOException, InterruptedException,
|
||||
ClassNotFoundException {
|
||||
LOG.info("\n\n\nStarting testSleepJob().");
|
||||
public void testSleepJob() throws Exception {
|
||||
testSleepJobInternal(false);
|
||||
}
|
||||
|
||||
@Test (timeout = 300000)
|
||||
public void testSleepJobWithRemoteJar() throws Exception {
|
||||
testSleepJobInternal(true);
|
||||
}
|
||||
|
||||
private void testSleepJobInternal(boolean useRemoteJar) throws Exception {
|
||||
LOG.info("\n\n\nStarting testSleepJob: useRemoteJar=" + useRemoteJar);
|
||||
|
||||
if (!(new File(MiniMRYarnCluster.APPJAR)).exists()) {
|
||||
LOG.info("MRAppJar " + MiniMRYarnCluster.APPJAR
|
||||
|
@ -192,14 +211,20 @@ public class TestMRJobs {
|
|||
|
||||
SleepJob sleepJob = new SleepJob();
|
||||
sleepJob.setConf(sleepConf);
|
||||
|
||||
int numReduces = sleepConf.getInt("TestMRJobs.testSleepJob.reduces", 2); // or sleepConf.getConfig().getInt(MRJobConfig.NUM_REDUCES, 2);
|
||||
|
||||
// job with 3 maps (10s) and numReduces reduces (5s), 1 "record" each:
|
||||
Job job = sleepJob.createJob(3, numReduces, 10000, 1, 5000, 1);
|
||||
Job job = sleepJob.createJob(3, numSleepReducers, 10000, 1, 5000, 1);
|
||||
|
||||
job.addFileToClassPath(APP_JAR); // The AppMaster jar itself.
|
||||
job.setJarByClass(SleepJob.class);
|
||||
if (useRemoteJar) {
|
||||
final Path localJar = new Path(
|
||||
ClassUtil.findContainingJar(SleepJob.class));
|
||||
ConfigUtil.addLink(job.getConfiguration(), "/jobjars",
|
||||
localFs.makeQualified(localJar.getParent()).toUri());
|
||||
job.setJar("viewfs:///jobjars/" + localJar.getName());
|
||||
} else {
|
||||
job.setJarByClass(SleepJob.class);
|
||||
}
|
||||
job.setMaxMapAttempts(1); // speed up failures
|
||||
job.submit();
|
||||
String trackingUrl = job.getTrackingURL();
|
||||
|
@ -381,7 +406,7 @@ public class TestMRJobs {
|
|||
.getValue());
|
||||
Assert.assertEquals(3, counters.findCounter(JobCounter.TOTAL_LAUNCHED_MAPS)
|
||||
.getValue());
|
||||
Assert.assertEquals(2,
|
||||
Assert.assertEquals(numSleepReducers,
|
||||
counters.findCounter(JobCounter.TOTAL_LAUNCHED_REDUCES).getValue());
|
||||
Assert
|
||||
.assertTrue(counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS) != null
|
||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.hadoop.mapreduce.v2;
|
|||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.util.Arrays;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
|
@ -40,8 +39,7 @@ import org.junit.Test;
|
|||
public class TestUberAM extends TestMRJobs {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(TestUberAM.class);
|
||||
private int numSleepReducers;
|
||||
|
||||
|
||||
@BeforeClass
|
||||
public static void setup() throws IOException {
|
||||
TestMRJobs.setup();
|
||||
|
@ -54,21 +52,15 @@ public class TestUberAM extends TestMRJobs {
|
|||
@Override
|
||||
@Test
|
||||
public void testSleepJob()
|
||||
throws IOException, InterruptedException, ClassNotFoundException {
|
||||
throws Exception {
|
||||
numSleepReducers = 1;
|
||||
if (mrCluster != null) {
|
||||
mrCluster.getConfig().setInt("TestMRJobs.testSleepJob.reduces", numSleepReducers);
|
||||
}
|
||||
super.testSleepJob();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSleepJobWithMultipleReducers()
|
||||
throws IOException, InterruptedException, ClassNotFoundException {
|
||||
throws Exception {
|
||||
numSleepReducers = 3;
|
||||
if (mrCluster != null) {
|
||||
mrCluster.getConfig().setInt("TestMRJobs.testSleepJob.reduces", numSleepReducers);
|
||||
}
|
||||
super.testSleepJob();
|
||||
}
|
||||
|
||||
|
@ -76,20 +68,7 @@ public class TestUberAM extends TestMRJobs {
|
|||
protected void verifySleepJobCounters(Job job) throws InterruptedException,
|
||||
IOException {
|
||||
Counters counters = job.getCounters();
|
||||
|
||||
Assert.assertEquals(3, counters.findCounter(JobCounter.OTHER_LOCAL_MAPS)
|
||||
.getValue());
|
||||
Assert.assertEquals(3, counters.findCounter(JobCounter.TOTAL_LAUNCHED_MAPS)
|
||||
.getValue());
|
||||
Assert.assertEquals(numSleepReducers,
|
||||
counters.findCounter(JobCounter.TOTAL_LAUNCHED_REDUCES).getValue());
|
||||
Assert
|
||||
.assertTrue(counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS) != null
|
||||
&& counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS).getValue() != 0);
|
||||
Assert
|
||||
.assertTrue(counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS) != null
|
||||
&& counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS).getValue() != 0);
|
||||
|
||||
super.verifySleepJobCounters(job);
|
||||
Assert.assertEquals(3,
|
||||
counters.findCounter(JobCounter.NUM_UBER_SUBMAPS).getValue());
|
||||
Assert.assertEquals(numSleepReducers,
|
||||
|
@ -168,16 +147,7 @@ public class TestUberAM extends TestMRJobs {
|
|||
protected void verifyFailingMapperCounters(Job job)
|
||||
throws InterruptedException, IOException {
|
||||
Counters counters = job.getCounters();
|
||||
Assert.assertEquals(2, counters.findCounter(JobCounter.OTHER_LOCAL_MAPS)
|
||||
.getValue());
|
||||
Assert.assertEquals(2, counters.findCounter(JobCounter.TOTAL_LAUNCHED_MAPS)
|
||||
.getValue());
|
||||
Assert.assertEquals(2, counters.findCounter(JobCounter.NUM_FAILED_MAPS)
|
||||
.getValue());
|
||||
Assert
|
||||
.assertTrue(counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS) != null
|
||||
&& counters.findCounter(JobCounter.SLOTS_MILLIS_MAPS).getValue() != 0);
|
||||
|
||||
super.verifyFailingMapperCounters(job);
|
||||
Assert.assertEquals(2,
|
||||
counters.findCounter(JobCounter.TOTAL_LAUNCHED_UBERTASKS).getValue());
|
||||
Assert.assertEquals(2, counters.findCounter(JobCounter.NUM_UBER_SUBMAPS)
|
||||
|
|
Loading…
Reference in New Issue