MAPREDUCE-1700. User supplied dependencies may conflict with MapReduce system JARs.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1430929 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
eae2a30462
commit
0ba7078ef4
|
@ -247,6 +247,9 @@ Release 2.0.3-alpha - Unreleased
|
|||
MAPREDUCE-4278. Cannot run two local jobs in parallel from the same
|
||||
gateway. (Sandy Ryza via tomwhite)
|
||||
|
||||
MAPREDUCE-1700. User supplied dependencies may conflict with MapReduce
|
||||
system JARs. (tomwhite)
|
||||
|
||||
Release 2.0.2-alpha - 2012-09-07
|
||||
|
||||
INCOMPATIBLE CHANGES
|
||||
|
|
|
@ -47,6 +47,7 @@ import org.apache.hadoop.mapreduce.filecache.DistributedCache;
|
|||
import org.apache.hadoop.mapreduce.security.TokenCache;
|
||||
import org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier;
|
||||
import org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager;
|
||||
import org.apache.hadoop.mapreduce.v2.util.MRApps;
|
||||
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
|
||||
import org.apache.hadoop.metrics2.source.JvmMetrics;
|
||||
import org.apache.hadoop.net.NetUtils;
|
||||
|
@ -254,6 +255,9 @@ class YarnChild {
|
|||
final JobConf job = new JobConf(MRJobConfig.JOB_CONF_FILE);
|
||||
job.setCredentials(credentials);
|
||||
|
||||
// set job classloader if configured
|
||||
MRApps.setJobClassLoader(job);
|
||||
|
||||
String appAttemptIdEnv = System
|
||||
.getenv(MRJobConfig.APPLICATION_ATTEMPT_ID_ENV);
|
||||
LOG.debug("APPLICATION_ATTEMPT_ID: " + appAttemptIdEnv);
|
||||
|
|
|
@ -1223,6 +1223,8 @@ public class MRAppMaster extends CompositeService {
|
|||
// SIGTERM I have a chance to write out the job history. I'll be closing
|
||||
// the objects myself.
|
||||
conf.setBoolean("fs.automatic.close", false);
|
||||
// set job classloader if configured
|
||||
MRApps.setJobClassLoader(conf);
|
||||
initAndStartAppMaster(appMaster, conf, jobUserName);
|
||||
} catch (Throwable t) {
|
||||
LOG.fatal("Error starting MRAppMaster", t);
|
||||
|
|
|
@ -165,6 +165,7 @@ public abstract class TaskAttemptImpl implements
|
|||
private Token<JobTokenIdentifier> jobToken;
|
||||
private static AtomicBoolean initialClasspathFlag = new AtomicBoolean();
|
||||
private static String initialClasspath = null;
|
||||
private static String initialAppClasspath = null;
|
||||
private static Object commonContainerSpecLock = new Object();
|
||||
private static ContainerLaunchContext commonContainerSpec = null;
|
||||
private static final Object classpathLock = new Object();
|
||||
|
@ -599,6 +600,7 @@ public abstract class TaskAttemptImpl implements
|
|||
Map<String, String> env = new HashMap<String, String>();
|
||||
MRApps.setClasspath(env, conf);
|
||||
initialClasspath = env.get(Environment.CLASSPATH.name());
|
||||
initialAppClasspath = env.get(Environment.APP_CLASSPATH.name());
|
||||
initialClasspathFlag.set(true);
|
||||
return initialClasspath;
|
||||
}
|
||||
|
@ -697,6 +699,13 @@ public abstract class TaskAttemptImpl implements
|
|||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
getInitialClasspath(conf));
|
||||
|
||||
if (initialAppClasspath != null) {
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.APP_CLASSPATH.name(),
|
||||
initialAppClasspath);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
throw new YarnException(e);
|
||||
}
|
||||
|
|
|
@ -19,12 +19,18 @@
|
|||
package org.apache.hadoop.mapreduce.v2.util;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.MalformedURLException;
|
||||
import java.net.URI;
|
||||
import java.security.AccessController;
|
||||
import java.security.PrivilegedActionException;
|
||||
import java.security.PrivilegedExceptionAction;
|
||||
import java.util.Arrays;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||
import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
|
@ -50,6 +56,7 @@ import org.apache.hadoop.yarn.api.records.LocalResource;
|
|||
import org.apache.hadoop.yarn.api.records.LocalResourceType;
|
||||
import org.apache.hadoop.yarn.api.records.LocalResourceVisibility;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import org.apache.hadoop.yarn.util.ApplicationClassLoader;
|
||||
import org.apache.hadoop.yarn.util.Apps;
|
||||
import org.apache.hadoop.yarn.util.BuilderUtils;
|
||||
import org.apache.hadoop.yarn.util.ConverterUtils;
|
||||
|
@ -62,6 +69,8 @@ import com.google.common.base.Charsets;
|
|||
@Private
|
||||
@Unstable
|
||||
public class MRApps extends Apps {
|
||||
public static final Log LOG = LogFactory.getLog(MRApps.class);
|
||||
|
||||
public static String toString(JobId jid) {
|
||||
return jid.toString();
|
||||
}
|
||||
|
@ -157,38 +166,42 @@ public class MRApps extends Apps {
|
|||
boolean userClassesTakesPrecedence =
|
||||
conf.getBoolean(MRJobConfig.MAPREDUCE_JOB_USER_CLASSPATH_FIRST, false);
|
||||
|
||||
String classpathEnvVar =
|
||||
conf.getBoolean(MRJobConfig.MAPREDUCE_JOB_CLASSLOADER, false)
|
||||
? Environment.APP_CLASSPATH.name() : Environment.CLASSPATH.name();
|
||||
|
||||
Apps.addToEnvironment(environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
Environment.PWD.$());
|
||||
if (!userClassesTakesPrecedence) {
|
||||
MRApps.setMRFrameworkClasspath(environment, conf);
|
||||
}
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
MRJobConfig.JOB_JAR + Path.SEPARATOR + MRJobConfig.JOB_JAR);
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
MRJobConfig.JOB_JAR + Path.SEPARATOR + "classes" + Path.SEPARATOR);
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
MRJobConfig.JOB_JAR + Path.SEPARATOR + "lib" + Path.SEPARATOR + "*");
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
Environment.PWD.$() + Path.SEPARATOR + "*");
|
||||
// a * in the classpath will only find a .jar, so we need to filter out
|
||||
// all .jars and add everything else
|
||||
addToClasspathIfNotJar(DistributedCache.getFileClassPaths(conf),
|
||||
DistributedCache.getCacheFiles(conf),
|
||||
conf,
|
||||
environment);
|
||||
environment, classpathEnvVar);
|
||||
addToClasspathIfNotJar(DistributedCache.getArchiveClassPaths(conf),
|
||||
DistributedCache.getCacheArchives(conf),
|
||||
conf,
|
||||
environment);
|
||||
environment, classpathEnvVar);
|
||||
if (userClassesTakesPrecedence) {
|
||||
MRApps.setMRFrameworkClasspath(environment, conf);
|
||||
}
|
||||
|
@ -204,7 +217,8 @@ public class MRApps extends Apps {
|
|||
*/
|
||||
private static void addToClasspathIfNotJar(Path[] paths,
|
||||
URI[] withLinks, Configuration conf,
|
||||
Map<String, String> environment) throws IOException {
|
||||
Map<String, String> environment,
|
||||
String classpathEnvVar) throws IOException {
|
||||
if (paths != null) {
|
||||
HashMap<Path, String> linkLookup = new HashMap<Path, String>();
|
||||
if (withLinks != null) {
|
||||
|
@ -232,13 +246,64 @@ public class MRApps extends Apps {
|
|||
if(!name.toLowerCase().endsWith(".jar")) {
|
||||
Apps.addToEnvironment(
|
||||
environment,
|
||||
Environment.CLASSPATH.name(),
|
||||
classpathEnvVar,
|
||||
Environment.PWD.$() + Path.SEPARATOR + name);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Sets a {@link ApplicationClassLoader} on the given configuration and as
|
||||
* the context classloader, if
|
||||
* {@link MRJobConfig#MAPREDUCE_JOB_CLASSLOADER} is set to true, and
|
||||
* the APP_CLASSPATH environment variable is set.
|
||||
* @param conf
|
||||
* @throws IOException
|
||||
*/
|
||||
public static void setJobClassLoader(Configuration conf)
|
||||
throws IOException {
|
||||
if (conf.getBoolean(MRJobConfig.MAPREDUCE_JOB_CLASSLOADER, false)) {
|
||||
String appClasspath = System.getenv(Environment.APP_CLASSPATH.key());
|
||||
if (appClasspath == null) {
|
||||
LOG.warn("Not using job classloader since APP_CLASSPATH is not set.");
|
||||
} else {
|
||||
LOG.info("Using job classloader");
|
||||
if (LOG.isDebugEnabled()) {
|
||||
LOG.debug("APP_CLASSPATH=" + appClasspath);
|
||||
}
|
||||
String[] systemClasses = conf.getStrings(
|
||||
MRJobConfig.MAPREDUCE_JOB_CLASSLOADER_SYSTEM_CLASSES);
|
||||
ClassLoader jobClassLoader = createJobClassLoader(appClasspath,
|
||||
systemClasses);
|
||||
if (jobClassLoader != null) {
|
||||
conf.setClassLoader(jobClassLoader);
|
||||
Thread.currentThread().setContextClassLoader(jobClassLoader);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private static ClassLoader createJobClassLoader(final String appClasspath,
|
||||
final String[] systemClasses) throws IOException {
|
||||
try {
|
||||
return AccessController.doPrivileged(
|
||||
new PrivilegedExceptionAction<ClassLoader>() {
|
||||
@Override
|
||||
public ClassLoader run() throws MalformedURLException {
|
||||
return new ApplicationClassLoader(appClasspath,
|
||||
MRApps.class.getClassLoader(), Arrays.asList(systemClasses));
|
||||
}
|
||||
});
|
||||
} catch (PrivilegedActionException e) {
|
||||
Throwable t = e.getCause();
|
||||
if (t instanceof MalformedURLException) {
|
||||
throw (MalformedURLException) t;
|
||||
}
|
||||
throw new IOException(e);
|
||||
}
|
||||
}
|
||||
|
||||
private static final String STAGING_CONSTANT = ".staging";
|
||||
public static Path getStagingAreaDir(Configuration conf, String user) {
|
||||
return new Path(conf.get(MRJobConfig.MR_AM_STAGING_DIR,
|
||||
|
|
|
@ -235,6 +235,22 @@ public class TestMRApps {
|
|||
index, 0);
|
||||
}
|
||||
|
||||
@Test public void testSetClasspathWithJobClassloader() throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
conf.setBoolean(MRJobConfig.MAPREDUCE_JOB_CLASSLOADER, true);
|
||||
Map<String, String> env = new HashMap<String, String>();
|
||||
MRApps.setClasspath(env, conf);
|
||||
String cp = env.get("CLASSPATH");
|
||||
String appCp = env.get("APP_CLASSPATH");
|
||||
assertSame("MAPREDUCE_JOB_CLASSLOADER true, but job.jar is"
|
||||
+ " in the classpath!", cp.indexOf("jar:job"), -1);
|
||||
assertSame("MAPREDUCE_JOB_CLASSLOADER true, but PWD is"
|
||||
+ " in the classpath!", cp.indexOf("PWD"), -1);
|
||||
assertEquals("MAPREDUCE_JOB_CLASSLOADER true, but job.jar is not"
|
||||
+ " in the app classpath!",
|
||||
"$PWD:job.jar/job.jar:job.jar/classes/:job.jar/lib/*:$PWD/*", appCp);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSetupDistributedCacheEmpty() throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
|
|
|
@ -127,6 +127,10 @@ public interface MRJobConfig {
|
|||
|
||||
public static final String MAPREDUCE_JOB_USER_CLASSPATH_FIRST = "mapreduce.job.user.classpath.first";
|
||||
|
||||
public static final String MAPREDUCE_JOB_CLASSLOADER = "mapreduce.job.classloader";
|
||||
|
||||
public static final String MAPREDUCE_JOB_CLASSLOADER_SYSTEM_CLASSES = "mapreduce.job.classloader.system.classes";
|
||||
|
||||
public static final String IO_SORT_FACTOR = "mapreduce.task.io.sort.factor";
|
||||
|
||||
public static final String IO_SORT_MB = "mapreduce.task.io.sort.mb";
|
||||
|
|
|
@ -961,6 +961,23 @@
|
|||
<value>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*</value>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>mapreduce.job.classloader</name>
|
||||
<value>false</value>
|
||||
<description>Whether to use a separate (isolated) classloader for
|
||||
user classes in the task JVM.</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>mapreduce.job.classloader.system.classes</name>
|
||||
<value>java.,javax.,org.apache.commons.logging.,org.apache.log4j.,org.apache.hadoop.</value>
|
||||
<description>A comma-separated list of classes that should be loaded from the
|
||||
system classpath, not the user-supplied JARs, when mapreduce.job.classloader
|
||||
is enabled. Names ending in '.' (period) are treated as package names,
|
||||
and names starting with a '-' are treated as negative matches.
|
||||
</description>
|
||||
</property>
|
||||
|
||||
<!-- jobhistory properties -->
|
||||
|
||||
<property>
|
||||
|
|
Loading…
Reference in New Issue