Merging r1523402 through r1523803 from trunk to branch HDFS-2832
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-2832@1523804 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
commit
3e3a454bd2
|
@ -408,6 +408,9 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
HADOOP-9945. HAServiceState should have a state for stopped services.
|
HADOOP-9945. HAServiceState should have a state for stopped services.
|
||||||
(Karthik Kambatla via atm)
|
(Karthik Kambatla via atm)
|
||||||
|
|
||||||
|
HADOOP-9962. in order to avoid dependency divergence within Hadoop itself
|
||||||
|
lets enable DependencyConvergence. (rvs via tucu)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
@ -468,6 +471,11 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
HADOOP-9350. Hadoop not building against Java7 on OSX
|
HADOOP-9350. Hadoop not building against Java7 on OSX
|
||||||
(Robert Kanter via stevel)
|
(Robert Kanter via stevel)
|
||||||
|
|
||||||
|
HADOOP-9935. set junit dependency to test scope. (André Kelpe via cnauroth)
|
||||||
|
|
||||||
|
HADOOP-9961. versions of a few transitive dependencies diverged between hadoop
|
||||||
|
subprojects. (rvs via tucu)
|
||||||
|
|
||||||
Release 2.1.0-beta - 2013-08-22
|
Release 2.1.0-beta - 2013-08-22
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -250,7 +250,6 @@
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.commons</groupId>
|
<groupId>org.apache.commons</groupId>
|
||||||
<artifactId>commons-compress</artifactId>
|
<artifactId>commons-compress</artifactId>
|
||||||
<version>1.4</version>
|
|
||||||
</dependency>
|
</dependency>
|
||||||
</dependencies>
|
</dependencies>
|
||||||
|
|
||||||
|
|
|
@ -929,6 +929,8 @@ KVNO Timestamp Principal
|
||||||
*-------------------------+-------------------------+------------------------+
|
*-------------------------+-------------------------+------------------------+
|
||||||
| <<<banned.users>>> | hfds,yarn,mapred,bin | Banned users. |
|
| <<<banned.users>>> | hfds,yarn,mapred,bin | Banned users. |
|
||||||
*-------------------------+-------------------------+------------------------+
|
*-------------------------+-------------------------+------------------------+
|
||||||
|
| <<<allowed.system.users>>> | foo,bar | Allowed system users. |
|
||||||
|
*-------------------------+-------------------------+------------------------+
|
||||||
| <<<min.user.id>>> | 1000 | Prevent other super-users. |
|
| <<<min.user.id>>> | 1000 | Prevent other super-users. |
|
||||||
*-------------------------+-------------------------+------------------------+
|
*-------------------------+-------------------------+------------------------+
|
||||||
|
|
||||||
|
|
|
@ -86,13 +86,11 @@
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.netty</groupId>
|
<groupId>io.netty</groupId>
|
||||||
<artifactId>netty</artifactId>
|
<artifactId>netty</artifactId>
|
||||||
<version>3.6.2.Final</version>
|
|
||||||
<scope>compile</scope>
|
<scope>compile</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>com.google.guava</groupId>
|
<groupId>com.google.guava</groupId>
|
||||||
<artifactId>guava</artifactId>
|
<artifactId>guava</artifactId>
|
||||||
<version>11.0.2</version>
|
|
||||||
</dependency>
|
</dependency>
|
||||||
</dependencies>
|
</dependencies>
|
||||||
|
|
||||||
|
|
|
@ -49,7 +49,6 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.netty</groupId>
|
<groupId>io.netty</groupId>
|
||||||
<artifactId>netty</artifactId>
|
<artifactId>netty</artifactId>
|
||||||
<version>3.6.2.Final</version>
|
|
||||||
<scope>compile</scope>
|
<scope>compile</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
|
|
|
@ -36,7 +36,25 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||||
<hadoop.common.build.dir>${basedir}/../../../../../hadoop-common-project/hadoop-common/target</hadoop.common.build.dir>
|
<hadoop.common.build.dir>${basedir}/../../../../../hadoop-common-project/hadoop-common/target</hadoop.common.build.dir>
|
||||||
</properties>
|
</properties>
|
||||||
|
|
||||||
|
<dependencyManagement>
|
||||||
|
<dependencies>
|
||||||
|
<!-- This is a really old version of netty, that gets privatized
|
||||||
|
via shading and hence it is not managed via a parent pom -->
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.jboss.netty</groupId>
|
||||||
|
<artifactId>netty</artifactId>
|
||||||
|
<version>3.2.4.Final</version>
|
||||||
|
</dependency>
|
||||||
|
</dependencies>
|
||||||
|
</dependencyManagement>
|
||||||
|
|
||||||
<dependencies>
|
<dependencies>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.jboss.netty</groupId>
|
||||||
|
<artifactId>netty</artifactId>
|
||||||
|
<scope>compile</scope>
|
||||||
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>commons-logging</groupId>
|
<groupId>commons-logging</groupId>
|
||||||
<artifactId>commons-logging</artifactId>
|
<artifactId>commons-logging</artifactId>
|
||||||
|
|
|
@ -190,6 +190,8 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
but just before ClientService to avoid race conditions during RM restart.
|
but just before ClientService to avoid race conditions during RM restart.
|
||||||
(Jian He via vinodkv)
|
(Jian He via vinodkv)
|
||||||
|
|
||||||
|
MAPREDUCE-5379. Include token tracking ids in jobconf. (kkambatl via tucu)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
MAPREDUCE-5446. TestJobHistoryEvents and TestJobHistoryParsing have race
|
MAPREDUCE-5446. TestJobHistoryEvents and TestJobHistoryParsing have race
|
||||||
|
@ -265,6 +267,9 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
MAPREDUCE-5164. mapred job and queue commands omit HADOOP_CLIENT_OPTS
|
MAPREDUCE-5164. mapred job and queue commands omit HADOOP_CLIENT_OPTS
|
||||||
(Nemon Lou via devaraj)
|
(Nemon Lou via devaraj)
|
||||||
|
|
||||||
|
MAPREDUCE-5493. Cleanup in-memory & on-disk segments to prevent leak on
|
||||||
|
shuffle completion. (jlowe via acmurthy)
|
||||||
|
|
||||||
Release 2.1.0-beta - 2013-08-22
|
Release 2.1.0-beta - 2013-08-22
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -19,6 +19,7 @@
|
||||||
option java_package = "org.apache.hadoop.yarn.proto";
|
option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "MRClientProtocol";
|
option java_outer_classname = "MRClientProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
|
package hadoop.mapreduce;
|
||||||
|
|
||||||
import "Security.proto";
|
import "Security.proto";
|
||||||
import "mr_service_protos.proto";
|
import "mr_service_protos.proto";
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.proto";
|
||||||
option java_outer_classname = "MRProtos";
|
option java_outer_classname = "MRProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.mapreduce;
|
||||||
|
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
|
|
||||||
|
@ -29,7 +30,7 @@ enum TaskTypeProto {
|
||||||
}
|
}
|
||||||
|
|
||||||
message JobIdProto {
|
message JobIdProto {
|
||||||
optional ApplicationIdProto app_id = 1;
|
optional hadoop.yarn.ApplicationIdProto app_id = 1;
|
||||||
optional int32 id = 2;
|
optional int32 id = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -115,7 +116,7 @@ message TaskAttemptReportProto {
|
||||||
optional string node_manager_host = 12;
|
optional string node_manager_host = 12;
|
||||||
optional int32 node_manager_port = 13;
|
optional int32 node_manager_port = 13;
|
||||||
optional int32 node_manager_http_port = 14;
|
optional int32 node_manager_http_port = 14;
|
||||||
optional ContainerIdProto container_id = 15;
|
optional hadoop.yarn.ContainerIdProto container_id = 15;
|
||||||
}
|
}
|
||||||
|
|
||||||
enum JobStateProto {
|
enum JobStateProto {
|
||||||
|
@ -148,9 +149,9 @@ message JobReportProto {
|
||||||
}
|
}
|
||||||
|
|
||||||
message AMInfoProto {
|
message AMInfoProto {
|
||||||
optional ApplicationAttemptIdProto application_attempt_id = 1;
|
optional hadoop.yarn.ApplicationAttemptIdProto application_attempt_id = 1;
|
||||||
optional int64 start_time = 2;
|
optional int64 start_time = 2;
|
||||||
optional ContainerIdProto container_id = 3;
|
optional hadoop.yarn.ContainerIdProto container_id = 3;
|
||||||
optional string node_manager_host = 4;
|
optional string node_manager_host = 4;
|
||||||
optional int32 node_manager_port = 5;
|
optional int32 node_manager_port = 5;
|
||||||
optional int32 node_manager_http_port = 6;
|
optional int32 node_manager_http_port = 6;
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.proto";
|
||||||
option java_outer_classname = "MRServiceProtos";
|
option java_outer_classname = "MRServiceProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.mapreduce;
|
||||||
|
|
||||||
import "Security.proto";
|
import "Security.proto";
|
||||||
import "mr_protos.proto";
|
import "mr_protos.proto";
|
||||||
|
|
|
@ -24,6 +24,7 @@ import java.net.URI;
|
||||||
import java.net.URISyntaxException;
|
import java.net.URISyntaxException;
|
||||||
import java.net.UnknownHostException;
|
import java.net.UnknownHostException;
|
||||||
import java.security.NoSuchAlgorithmException;
|
import java.security.NoSuchAlgorithmException;
|
||||||
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
import java.util.Comparator;
|
import java.util.Comparator;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
@ -56,6 +57,7 @@ import org.apache.hadoop.security.Credentials;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
import org.apache.hadoop.security.authorize.AccessControlList;
|
import org.apache.hadoop.security.authorize.AccessControlList;
|
||||||
import org.apache.hadoop.security.token.Token;
|
import org.apache.hadoop.security.token.Token;
|
||||||
|
import org.apache.hadoop.security.token.TokenIdentifier;
|
||||||
import org.apache.hadoop.util.ReflectionUtils;
|
import org.apache.hadoop.util.ReflectionUtils;
|
||||||
import org.codehaus.jackson.JsonParseException;
|
import org.codehaus.jackson.JsonParseException;
|
||||||
import org.codehaus.jackson.map.JsonMappingException;
|
import org.codehaus.jackson.map.JsonMappingException;
|
||||||
|
@ -405,6 +407,19 @@ class JobSubmitter {
|
||||||
// different job.
|
// different job.
|
||||||
TokenCache.cleanUpTokenReferral(conf);
|
TokenCache.cleanUpTokenReferral(conf);
|
||||||
|
|
||||||
|
if (conf.getBoolean(
|
||||||
|
MRJobConfig.JOB_TOKEN_TRACKING_IDS_ENABLED,
|
||||||
|
MRJobConfig.DEFAULT_JOB_TOKEN_TRACKING_IDS_ENABLED)) {
|
||||||
|
// Add HDFS tracking ids
|
||||||
|
ArrayList<String> trackingIds = new ArrayList<String>();
|
||||||
|
for (Token<? extends TokenIdentifier> t :
|
||||||
|
job.getCredentials().getAllTokens()) {
|
||||||
|
trackingIds.add(t.decodeIdentifier().getTrackingId());
|
||||||
|
}
|
||||||
|
conf.setStrings(MRJobConfig.JOB_TOKEN_TRACKING_IDS,
|
||||||
|
trackingIds.toArray(new String[trackingIds.size()]));
|
||||||
|
}
|
||||||
|
|
||||||
// Write job file to submit dir
|
// Write job file to submit dir
|
||||||
writeConf(conf, submitJobFile);
|
writeConf(conf, submitJobFile);
|
||||||
|
|
||||||
|
|
|
@ -313,6 +313,13 @@ public interface MRJobConfig {
|
||||||
public static final String MAPREDUCE_JOB_CREDENTIALS_BINARY =
|
public static final String MAPREDUCE_JOB_CREDENTIALS_BINARY =
|
||||||
"mapreduce.job.credentials.binary";
|
"mapreduce.job.credentials.binary";
|
||||||
|
|
||||||
|
/* Configs for tracking ids of tokens used by a job */
|
||||||
|
public static final String JOB_TOKEN_TRACKING_IDS_ENABLED =
|
||||||
|
"mapreduce.job.token.tracking.ids.enabled";
|
||||||
|
public static final boolean DEFAULT_JOB_TOKEN_TRACKING_IDS_ENABLED = false;
|
||||||
|
public static final String JOB_TOKEN_TRACKING_IDS =
|
||||||
|
"mapreduce.job.token.tracking.ids";
|
||||||
|
|
||||||
public static final String JOB_SUBMITHOST =
|
public static final String JOB_SUBMITHOST =
|
||||||
"mapreduce.job.submithostname";
|
"mapreduce.job.submithostname";
|
||||||
public static final String JOB_SUBMITHOSTADDR =
|
public static final String JOB_SUBMITHOSTADDR =
|
||||||
|
|
|
@ -355,8 +355,11 @@ public class MergeManagerImpl<K, V> implements MergeManager<K, V> {
|
||||||
|
|
||||||
List<InMemoryMapOutput<K, V>> memory =
|
List<InMemoryMapOutput<K, V>> memory =
|
||||||
new ArrayList<InMemoryMapOutput<K, V>>(inMemoryMergedMapOutputs);
|
new ArrayList<InMemoryMapOutput<K, V>>(inMemoryMergedMapOutputs);
|
||||||
|
inMemoryMergedMapOutputs.clear();
|
||||||
memory.addAll(inMemoryMapOutputs);
|
memory.addAll(inMemoryMapOutputs);
|
||||||
|
inMemoryMapOutputs.clear();
|
||||||
List<CompressAwarePath> disk = new ArrayList<CompressAwarePath>(onDiskMapOutputs);
|
List<CompressAwarePath> disk = new ArrayList<CompressAwarePath>(onDiskMapOutputs);
|
||||||
|
onDiskMapOutputs.clear();
|
||||||
return finalMerge(jobConf, rfs, memory, disk);
|
return finalMerge(jobConf, rfs, memory, disk);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -748,6 +748,23 @@
|
||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>mapreduce.job.token.tracking.ids.enabled</name>
|
||||||
|
<value>false</value>
|
||||||
|
<description>Whether to write tracking ids of tokens to
|
||||||
|
job-conf. When true, the configuration property
|
||||||
|
"mapreduce.job.token.tracking.ids" is set to the token-tracking-ids of
|
||||||
|
the job</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>mapreduce.job.token.tracking.ids</name>
|
||||||
|
<value></value>
|
||||||
|
<description>When mapreduce.job.token.tracking.ids.enabled is
|
||||||
|
set to true, this is set by the framework to the
|
||||||
|
token-tracking-ids used by the job.</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>mapreduce.task.merge.progress.records</name>
|
<name>mapreduce.task.merge.progress.records</name>
|
||||||
<value>10000</value>
|
<value>10000</value>
|
||||||
|
|
|
@ -82,7 +82,7 @@ public class TestMerger {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testInMemoryMerger() throws IOException {
|
public void testInMemoryMerger() throws Throwable {
|
||||||
JobID jobId = new JobID("a", 0);
|
JobID jobId = new JobID("a", 0);
|
||||||
TaskAttemptID reduceId = new TaskAttemptID(
|
TaskAttemptID reduceId = new TaskAttemptID(
|
||||||
new TaskID(jobId, TaskType.REDUCE, 0), 0);
|
new TaskID(jobId, TaskType.REDUCE, 0), 0);
|
||||||
|
@ -132,6 +132,11 @@ public class TestMerger {
|
||||||
readOnDiskMapOutput(conf, fs, outPath, keys, values);
|
readOnDiskMapOutput(conf, fs, outPath, keys, values);
|
||||||
Assert.assertEquals(keys, Arrays.asList("apple", "banana", "carrot"));
|
Assert.assertEquals(keys, Arrays.asList("apple", "banana", "carrot"));
|
||||||
Assert.assertEquals(values, Arrays.asList("disgusting", "pretty good", "delicious"));
|
Assert.assertEquals(values, Arrays.asList("disgusting", "pretty good", "delicious"));
|
||||||
|
|
||||||
|
mergeManager.close();
|
||||||
|
Assert.assertEquals(0, mergeManager.inMemoryMapOutputs.size());
|
||||||
|
Assert.assertEquals(0, mergeManager.inMemoryMergedMapOutputs.size());
|
||||||
|
Assert.assertEquals(0, mergeManager.onDiskMapOutputs.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
private byte[] writeMapOutput(Configuration conf, Map<String, String> keysToValues)
|
private byte[] writeMapOutput(Configuration conf, Map<String, String> keysToValues)
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.hs.proto";
|
||||||
option java_outer_classname = "HSAdminRefreshProtocolProtos";
|
option java_outer_classname = "HSAdminRefreshProtocolProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.mapreduce;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* refresh admin acls request.
|
* refresh admin acls request.
|
||||||
|
|
|
@ -318,6 +318,11 @@
|
||||||
<artifactId>commons-math</artifactId>
|
<artifactId>commons-math</artifactId>
|
||||||
<version>2.1</version>
|
<version>2.1</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.commons</groupId>
|
||||||
|
<artifactId>commons-compress</artifactId>
|
||||||
|
<version>1.4.1</version>
|
||||||
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>xmlenc</groupId>
|
<groupId>xmlenc</groupId>
|
||||||
<artifactId>xmlenc</artifactId>
|
<artifactId>xmlenc</artifactId>
|
||||||
|
@ -360,6 +365,23 @@
|
||||||
<version>6.1.26</version>
|
<version>6.1.26</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.glassfish</groupId>
|
||||||
|
<artifactId>javax.servlet</artifactId>
|
||||||
|
<version>3.1</version>
|
||||||
|
</dependency>
|
||||||
|
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.codehaus.plexus</groupId>
|
||||||
|
<artifactId>plexus-utils</artifactId>
|
||||||
|
<version>2.0.5</version>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.codehaus.plexus</groupId>
|
||||||
|
<artifactId>plexus-component-annotations</artifactId>
|
||||||
|
<version>1.5.5</version>
|
||||||
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>asm</groupId>
|
<groupId>asm</groupId>
|
||||||
<artifactId>asm</artifactId>
|
<artifactId>asm</artifactId>
|
||||||
|
@ -420,7 +442,7 @@
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.netty</groupId>
|
<groupId>io.netty</groupId>
|
||||||
<artifactId>netty</artifactId>
|
<artifactId>netty</artifactId>
|
||||||
<version>3.5.11.Final</version>
|
<version>3.6.2.Final</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
|
@ -544,6 +566,7 @@
|
||||||
<groupId>junit</groupId>
|
<groupId>junit</groupId>
|
||||||
<artifactId>junit</artifactId>
|
<artifactId>junit</artifactId>
|
||||||
<version>4.10</version>
|
<version>4.10</version>
|
||||||
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>commons-lang</groupId>
|
<groupId>commons-lang</groupId>
|
||||||
|
@ -604,6 +627,7 @@
|
||||||
<groupId>org.mockito</groupId>
|
<groupId>org.mockito</groupId>
|
||||||
<artifactId>mockito-all</artifactId>
|
<artifactId>mockito-all</artifactId>
|
||||||
<version>1.8.5</version>
|
<version>1.8.5</version>
|
||||||
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.avro</groupId>
|
<groupId>org.apache.avro</groupId>
|
||||||
|
@ -664,6 +688,10 @@
|
||||||
<groupId>com.sun.jmx</groupId>
|
<groupId>com.sun.jmx</groupId>
|
||||||
<artifactId>jmxri</artifactId>
|
<artifactId>jmxri</artifactId>
|
||||||
</exclusion>
|
</exclusion>
|
||||||
|
<exclusion>
|
||||||
|
<groupId>org.jboss.netty</groupId>
|
||||||
|
<artifactId>netty</artifactId>
|
||||||
|
</exclusion>
|
||||||
</exclusions>
|
</exclusions>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
|
@ -672,6 +700,12 @@
|
||||||
<version>3.4.2</version>
|
<version>3.4.2</version>
|
||||||
<type>test-jar</type>
|
<type>test-jar</type>
|
||||||
<scope>test</scope>
|
<scope>test</scope>
|
||||||
|
<exclusions>
|
||||||
|
<exclusion>
|
||||||
|
<groupId>org.jboss.netty</groupId>
|
||||||
|
<artifactId>netty</artifactId>
|
||||||
|
</exclusion>
|
||||||
|
</exclusions>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.bookkeeper</groupId>
|
<groupId>org.apache.bookkeeper</groupId>
|
||||||
|
@ -706,7 +740,7 @@
|
||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-dependency-plugin</artifactId>
|
<artifactId>maven-dependency-plugin</artifactId>
|
||||||
<version>2.1</version>
|
<version>2.2</version>
|
||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.codehaus.mojo</groupId>
|
<groupId>org.codehaus.mojo</groupId>
|
||||||
|
@ -874,6 +908,26 @@
|
||||||
<includeReports>false</includeReports>
|
<includeReports>false</includeReports>
|
||||||
</configuration>
|
</configuration>
|
||||||
</plugin>
|
</plugin>
|
||||||
|
<plugin>
|
||||||
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
|
<artifactId>maven-enforcer-plugin</artifactId>
|
||||||
|
<executions>
|
||||||
|
<execution>
|
||||||
|
<id>depcheck</id>
|
||||||
|
<configuration>
|
||||||
|
<rules>
|
||||||
|
<DependencyConvergence>
|
||||||
|
<uniqueVersions>true</uniqueVersions>
|
||||||
|
</DependencyConvergence>
|
||||||
|
</rules>
|
||||||
|
</configuration>
|
||||||
|
<goals>
|
||||||
|
<goal>enforce</goal>
|
||||||
|
</goals>
|
||||||
|
<phase>verify</phase>
|
||||||
|
</execution>
|
||||||
|
</executions>
|
||||||
|
</plugin>
|
||||||
</plugins>
|
</plugins>
|
||||||
</build>
|
</build>
|
||||||
|
|
||||||
|
|
|
@ -29,6 +29,7 @@ Release 2.3.0 - UNRELEASED
|
||||||
YARN-905. Add state filters to nodes CLI (Wei Yan via Sandy Ryza)
|
YARN-905. Add state filters to nodes CLI (Wei Yan via Sandy Ryza)
|
||||||
YARN-1098. Separate out RM services into Always On and Active (Karthik
|
YARN-1098. Separate out RM services into Always On and Active (Karthik
|
||||||
Kambatla via bikas)
|
Kambatla via bikas)
|
||||||
|
YARN-1027. Implement RMHAProtocolService (Karthik Kambatla via bikas)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
|
@ -44,6 +45,10 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
YARN-707. Added user information also in the YARN ClientToken so that AMs
|
YARN-707. Added user information also in the YARN ClientToken so that AMs
|
||||||
can implement authorization based on incoming users. (Jason Lowe via vinodkv)
|
can implement authorization based on incoming users. (Jason Lowe via vinodkv)
|
||||||
|
|
||||||
|
YARN-1170. YARN & MapReduce proto definitions fixed to specify protobuf
|
||||||
|
package as hadoop.yarn and hadoop.mapreduce respectively. (Binglin Chang
|
||||||
|
via acmurthy)
|
||||||
|
|
||||||
NEW FEATURES
|
NEW FEATURES
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
|
@ -98,6 +103,9 @@ Release 2.1.1-beta - UNRELEASED
|
||||||
completions in addition to application events. (Alejandro Abdelnur via
|
completions in addition to application events. (Alejandro Abdelnur via
|
||||||
vinodkv)
|
vinodkv)
|
||||||
|
|
||||||
|
YARN-1137. Add support whitelist for system users to Yarn
|
||||||
|
container-executor.c. (rvs via tucu)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
|
@ -1,3 +1,4 @@
|
||||||
yarn.nodemanager.linux-container-executor.group=#configured value of yarn.nodemanager.linux-container-executor.group
|
yarn.nodemanager.linux-container-executor.group=#configured value of yarn.nodemanager.linux-container-executor.group
|
||||||
banned.users=#comma separated list of users who can not run applications
|
banned.users=#comma separated list of users who can not run applications
|
||||||
min.user.id=1000#Prevent other super-users
|
min.user.id=1000#Prevent other super-users
|
||||||
|
allowed.system.users=##comma separated list of system users who CAN run applications
|
||||||
|
|
|
@ -271,6 +271,11 @@ public class YarnConfiguration extends Configuration {
|
||||||
public static final String RECOVERY_ENABLED = RM_PREFIX + "recovery.enabled";
|
public static final String RECOVERY_ENABLED = RM_PREFIX + "recovery.enabled";
|
||||||
public static final boolean DEFAULT_RM_RECOVERY_ENABLED = false;
|
public static final boolean DEFAULT_RM_RECOVERY_ENABLED = false;
|
||||||
|
|
||||||
|
/** HA related configs */
|
||||||
|
public static final String RM_HA_PREFIX = RM_PREFIX + "ha.";
|
||||||
|
public static final String RM_HA_ENABLED = RM_HA_PREFIX + "enabled";
|
||||||
|
public static final boolean DEFAULT_RM_HA_ENABLED = false;
|
||||||
|
|
||||||
/** The class to use as the persistent store.*/
|
/** The class to use as the persistent store.*/
|
||||||
public static final String RM_STORE = RM_PREFIX + "store.class";
|
public static final String RM_STORE = RM_PREFIX + "store.class";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "ApplicationClientProtocol";
|
option java_outer_classname = "ApplicationClientProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "Security.proto";
|
import "Security.proto";
|
||||||
import "yarn_service_protos.proto";
|
import "yarn_service_protos.proto";
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "ApplicationMasterProtocol";
|
option java_outer_classname = "ApplicationMasterProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_service_protos.proto";
|
import "yarn_service_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "ContainerManagementProtocol";
|
option java_outer_classname = "ContainerManagementProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_service_protos.proto";
|
import "yarn_service_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "ResourceManagerAdministrationProtocol";
|
option java_outer_classname = "ResourceManagerAdministrationProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "server/yarn_server_resourcemanager_service_protos.proto";
|
import "server/yarn_server_resourcemanager_service_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnServerResourceManagerServiceProtos";
|
option java_outer_classname = "YarnServerResourceManagerServiceProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnProtos";
|
option java_outer_classname = "YarnProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "Security.proto";
|
import "Security.proto";
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnServiceProtos";
|
option java_outer_classname = "YarnServiceProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "Security.proto";
|
import "Security.proto";
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
|
|
|
@ -268,6 +268,14 @@
|
||||||
<!--value>hdfs://localhost:9000/rmstore</value-->
|
<!--value>hdfs://localhost:9000/rmstore</value-->
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<description>Enable RM high-availability. When enabled, the RM starts
|
||||||
|
in the Standby mode by default, and transitions to the Active mode when
|
||||||
|
prompted to.</description>
|
||||||
|
<name>yarn.resourcemanager.ha.enabled</name>
|
||||||
|
<value>false</value>
|
||||||
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<description>The maximum number of completed applications RM keeps. </description>
|
<description>The maximum number of completed applications RM keeps. </description>
|
||||||
<name>yarn.resourcemanager.max-completed-applications</name>
|
<name>yarn.resourcemanager.max-completed-applications</name>
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "ResourceTracker";
|
option java_outer_classname = "ResourceTracker";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_server_common_service_protos.proto";
|
import "yarn_server_common_service_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnServerCommonProtos";
|
option java_outer_classname = "YarnServerCommonProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnServerCommonServiceProtos";
|
option java_outer_classname = "YarnServerCommonServiceProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
import "yarn_server_common_protos.proto";
|
import "yarn_server_common_protos.proto";
|
||||||
|
|
|
@ -30,6 +30,7 @@
|
||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
#include <string.h>
|
#include <string.h>
|
||||||
|
#include <limits.h>
|
||||||
#include <sys/stat.h>
|
#include <sys/stat.h>
|
||||||
#include <sys/mount.h>
|
#include <sys/mount.h>
|
||||||
|
|
||||||
|
@ -492,6 +493,21 @@ static struct passwd* get_user_info(const char* user) {
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int is_whitelisted(const char *user) {
|
||||||
|
char **whitelist = get_values(ALLOWED_SYSTEM_USERS_KEY);
|
||||||
|
char **users = whitelist;
|
||||||
|
if (whitelist != NULL) {
|
||||||
|
for(; *users; ++users) {
|
||||||
|
if (strncmp(*users, user, LOGIN_NAME_MAX) == 0) {
|
||||||
|
free_values(whitelist);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
free_values(whitelist);
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Is the user a real user account?
|
* Is the user a real user account?
|
||||||
* Checks:
|
* Checks:
|
||||||
|
@ -526,9 +542,9 @@ struct passwd* check_user(const char *user) {
|
||||||
fflush(LOGFILE);
|
fflush(LOGFILE);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
if (user_info->pw_uid < min_uid) {
|
if (user_info->pw_uid < min_uid && !is_whitelisted(user)) {
|
||||||
fprintf(LOGFILE, "Requested user %s has id %d, which is below the "
|
fprintf(LOGFILE, "Requested user %s is not whitelisted and has id %d,"
|
||||||
"minimum allowed %d\n", user, user_info->pw_uid, min_uid);
|
"which is below the minimum allowed %d\n", user, user_info->pw_uid, min_uid);
|
||||||
fflush(LOGFILE);
|
fflush(LOGFILE);
|
||||||
free(user_info);
|
free(user_info);
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
|
@ -65,6 +65,7 @@ enum errorcodes {
|
||||||
#define CREDENTIALS_FILENAME "container_tokens"
|
#define CREDENTIALS_FILENAME "container_tokens"
|
||||||
#define MIN_USERID_KEY "min.user.id"
|
#define MIN_USERID_KEY "min.user.id"
|
||||||
#define BANNED_USERS_KEY "banned.users"
|
#define BANNED_USERS_KEY "banned.users"
|
||||||
|
#define ALLOWED_SYSTEM_USERS_KEY "allowed.system.users"
|
||||||
#define TMP_DIR "tmp"
|
#define TMP_DIR "tmp"
|
||||||
|
|
||||||
extern struct passwd *user_detail;
|
extern struct passwd *user_detail;
|
||||||
|
|
|
@ -99,6 +99,7 @@ int write_config_file(char *file_name) {
|
||||||
}
|
}
|
||||||
fprintf(file, "banned.users=bannedUser\n");
|
fprintf(file, "banned.users=bannedUser\n");
|
||||||
fprintf(file, "min.user.id=500\n");
|
fprintf(file, "min.user.id=500\n");
|
||||||
|
fprintf(file, "allowed.system.users=allowedUser,bin\n");
|
||||||
fclose(file);
|
fclose(file);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -195,6 +196,10 @@ void test_check_user() {
|
||||||
printf("FAIL: failed check for system user root\n");
|
printf("FAIL: failed check for system user root\n");
|
||||||
exit(1);
|
exit(1);
|
||||||
}
|
}
|
||||||
|
if (check_user("bin") == NULL) {
|
||||||
|
printf("FAIL: failed check for whitelisted system user bin\n");
|
||||||
|
exit(1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void test_resolve_config_path() {
|
void test_resolve_config_path() {
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "LocalizationProtocol";
|
option java_outer_classname = "LocalizationProtocol";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_server_nodemanager_service_protos.proto";
|
import "yarn_server_nodemanager_service_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto";
|
||||||
option java_outer_classname = "YarnServerNodemanagerServiceProtos";
|
option java_outer_classname = "YarnServerNodemanagerServiceProtos";
|
||||||
option java_generic_services = true;
|
option java_generic_services = true;
|
||||||
option java_generate_equals_and_hash = true;
|
option java_generate_equals_and_hash = true;
|
||||||
|
package hadoop.yarn;
|
||||||
|
|
||||||
import "yarn_protos.proto";
|
import "yarn_protos.proto";
|
||||||
|
|
||||||
|
|
|
@ -196,7 +196,7 @@ public class ClientRMService extends AbstractService implements
|
||||||
|
|
||||||
ApplicationId getNewApplicationId() {
|
ApplicationId getNewApplicationId() {
|
||||||
ApplicationId applicationId = org.apache.hadoop.yarn.server.utils.BuilderUtils
|
ApplicationId applicationId = org.apache.hadoop.yarn.server.utils.BuilderUtils
|
||||||
.newApplicationId(recordFactory, ResourceManager.clusterTimeStamp,
|
.newApplicationId(recordFactory, ResourceManager.getClusterTimeStamp(),
|
||||||
applicationCounter.incrementAndGet());
|
applicationCounter.incrementAndGet());
|
||||||
LOG.info("Allocated new applicationId: " + applicationId.getId());
|
LOG.info("Allocated new applicationId: " + applicationId.getId());
|
||||||
return applicationId;
|
return applicationId;
|
||||||
|
|
|
@ -0,0 +1,153 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.resourcemanager;
|
||||||
|
|
||||||
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.ha.HAServiceProtocol;
|
||||||
|
import org.apache.hadoop.ha.HAServiceStatus;
|
||||||
|
import org.apache.hadoop.ha.HealthCheckFailedException;
|
||||||
|
import org.apache.hadoop.service.AbstractService;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class RMHAProtocolService extends AbstractService implements
|
||||||
|
HAServiceProtocol {
|
||||||
|
private static final Log LOG = LogFactory.getLog(RMHAProtocolService.class);
|
||||||
|
|
||||||
|
private Configuration conf;
|
||||||
|
private ResourceManager rm;
|
||||||
|
@VisibleForTesting
|
||||||
|
protected HAServiceState haState = HAServiceState.INITIALIZING;
|
||||||
|
|
||||||
|
public RMHAProtocolService(ResourceManager resourceManager) {
|
||||||
|
super("RMHAProtocolService");
|
||||||
|
this.rm = resourceManager;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void serviceInit(Configuration conf) throws Exception {
|
||||||
|
this.conf = conf;
|
||||||
|
rm.createAndInitActiveServices();
|
||||||
|
super.serviceInit(this.conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void serviceStart() throws Exception {
|
||||||
|
boolean haEnabled = this.conf.getBoolean(YarnConfiguration.RM_HA_ENABLED,
|
||||||
|
YarnConfiguration.DEFAULT_RM_HA_ENABLED);
|
||||||
|
|
||||||
|
if (haEnabled) {
|
||||||
|
transitionToStandby(true);
|
||||||
|
} else {
|
||||||
|
transitionToActive();
|
||||||
|
}
|
||||||
|
|
||||||
|
super.serviceStart();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void serviceStop() throws Exception {
|
||||||
|
transitionToStandby(false);
|
||||||
|
haState = HAServiceState.STOPPING;
|
||||||
|
super.serviceStop();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void monitorHealth() throws HealthCheckFailedException {
|
||||||
|
if (haState == HAServiceState.ACTIVE && !rm.areActiveServicesRunning()) {
|
||||||
|
throw new HealthCheckFailedException(
|
||||||
|
"Active ResourceManager services are not running!");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private synchronized void transitionToActive() throws Exception {
|
||||||
|
if (haState == HAServiceState.ACTIVE) {
|
||||||
|
LOG.info("Already in active state");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.info("Transitioning to active");
|
||||||
|
rm.startActiveServices();
|
||||||
|
haState = HAServiceState.ACTIVE;
|
||||||
|
LOG.info("Transitioned to active");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void transitionToActive(StateChangeRequestInfo reqInfo) {
|
||||||
|
// TODO (YARN-1177): When automatic failover is enabled,
|
||||||
|
// check if transition should be allowed for this request
|
||||||
|
try {
|
||||||
|
transitionToActive();
|
||||||
|
} catch (Exception e) {
|
||||||
|
LOG.error("Error when transitioning to Active mode", e);
|
||||||
|
throw new YarnRuntimeException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private synchronized void transitionToStandby(boolean initialize)
|
||||||
|
throws Exception {
|
||||||
|
if (haState == HAServiceState.STANDBY) {
|
||||||
|
LOG.info("Already in standby state");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.info("Transitioning to standby");
|
||||||
|
if (haState == HAServiceState.ACTIVE) {
|
||||||
|
rm.stopActiveServices();
|
||||||
|
if (initialize) {
|
||||||
|
rm.createAndInitActiveServices();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
haState = HAServiceState.STANDBY;
|
||||||
|
LOG.info("Transitioned to standby");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized void transitionToStandby(StateChangeRequestInfo reqInfo) {
|
||||||
|
// TODO (YARN-1177): When automatic failover is enabled,
|
||||||
|
// check if transition should be allowed for this request
|
||||||
|
try {
|
||||||
|
transitionToStandby(true);
|
||||||
|
} catch (Exception e) {
|
||||||
|
LOG.error("Error when transitioning to Standby mode", e);
|
||||||
|
throw new YarnRuntimeException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public synchronized HAServiceStatus getServiceStatus() throws IOException {
|
||||||
|
HAServiceStatus ret = new HAServiceStatus(haState);
|
||||||
|
if (haState == HAServiceState.ACTIVE || haState == HAServiceState.STANDBY) {
|
||||||
|
ret.setReadyToBecomeActive();
|
||||||
|
} else {
|
||||||
|
ret.setNotReadyToBecomeActive("State is " + haState);
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
}
|
|
@ -105,7 +105,14 @@ public class ResourceManager extends CompositeService implements Recoverable {
|
||||||
public static final int SHUTDOWN_HOOK_PRIORITY = 30;
|
public static final int SHUTDOWN_HOOK_PRIORITY = 30;
|
||||||
|
|
||||||
private static final Log LOG = LogFactory.getLog(ResourceManager.class);
|
private static final Log LOG = LogFactory.getLog(ResourceManager.class);
|
||||||
public static final long clusterTimeStamp = System.currentTimeMillis();
|
private static long clusterTimeStamp = System.currentTimeMillis();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* "Always On" services. Services that need to run always irrespective of
|
||||||
|
* the HA state of the RM.
|
||||||
|
*/
|
||||||
|
@VisibleForTesting
|
||||||
|
protected RMHAProtocolService haService;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* "Active" services. Services that need to run only on the Active RM.
|
* "Active" services. Services that need to run only on the Active RM.
|
||||||
|
@ -156,13 +163,17 @@ public class ResourceManager extends CompositeService implements Recoverable {
|
||||||
return this.rmContext;
|
return this.rmContext;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static long getClusterTimeStamp() {
|
||||||
|
return clusterTimeStamp;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void serviceInit(Configuration conf) throws Exception {
|
protected void serviceInit(Configuration conf) throws Exception {
|
||||||
validateConfigs(conf);
|
validateConfigs(conf);
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
|
|
||||||
activeServices = new RMActiveServices();
|
haService = new RMHAProtocolService(this);
|
||||||
addService(activeServices);
|
addService(haService);
|
||||||
super.serviceInit(conf);
|
super.serviceInit(conf);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -470,6 +481,7 @@ public class ResourceManager extends CompositeService implements Recoverable {
|
||||||
LOG.error("Error closing store.", e);
|
LOG.error("Error closing store.", e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
super.serviceStop();
|
super.serviceStop();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -708,6 +720,43 @@ public class ResourceManager extends CompositeService implements Recoverable {
|
||||||
webApp = builder.start(new RMWebApp(this));
|
webApp = builder.start(new RMWebApp(this));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Helper method to create and init {@link #activeServices}. This creates an
|
||||||
|
* instance of {@link RMActiveServices} and initializes it.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
void createAndInitActiveServices() throws Exception {
|
||||||
|
activeServices = new RMActiveServices();
|
||||||
|
activeServices.init(conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Helper method to start {@link #activeServices}.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
void startActiveServices() throws Exception {
|
||||||
|
if (activeServices != null) {
|
||||||
|
clusterTimeStamp = System.currentTimeMillis();
|
||||||
|
activeServices.start();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Helper method to stop {@link #activeServices}.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
void stopActiveServices() throws Exception {
|
||||||
|
if (activeServices != null) {
|
||||||
|
activeServices.stop();
|
||||||
|
activeServices = null;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
protected boolean areActiveServicesRunning() {
|
||||||
|
return activeServices != null && activeServices.isInState(STATE.STARTED);
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void serviceStart() throws Exception {
|
protected void serviceStart() throws Exception {
|
||||||
try {
|
try {
|
||||||
|
@ -715,7 +764,6 @@ public class ResourceManager extends CompositeService implements Recoverable {
|
||||||
} catch(IOException ie) {
|
} catch(IOException ie) {
|
||||||
throw new YarnRuntimeException("Failed to login", ie);
|
throw new YarnRuntimeException("Failed to login", ie);
|
||||||
}
|
}
|
||||||
|
|
||||||
super.serviceStart();
|
super.serviceStart();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -229,7 +229,7 @@ public class ResourceTrackerService extends AbstractService implements
|
||||||
+ ", assigned nodeId " + nodeId;
|
+ ", assigned nodeId " + nodeId;
|
||||||
LOG.info(message);
|
LOG.info(message);
|
||||||
response.setNodeAction(NodeAction.NORMAL);
|
response.setNodeAction(NodeAction.NORMAL);
|
||||||
response.setRMIdentifier(ResourceManager.clusterTimeStamp);
|
response.setRMIdentifier(ResourceManager.getClusterTimeStamp());
|
||||||
return response;
|
return response;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -185,7 +185,7 @@ public class RMContainerTokenSecretManager extends
|
||||||
tokenIdentifier =
|
tokenIdentifier =
|
||||||
new ContainerTokenIdentifier(containerId, nodeId.toString(),
|
new ContainerTokenIdentifier(containerId, nodeId.toString(),
|
||||||
appSubmitter, capability, expiryTimeStamp, this.currentMasterKey
|
appSubmitter, capability, expiryTimeStamp, this.currentMasterKey
|
||||||
.getMasterKey().getKeyId(), ResourceManager.clusterTimeStamp);
|
.getMasterKey().getKeyId(), ResourceManager.getClusterTimeStamp());
|
||||||
password = this.createPassword(tokenIdentifier);
|
password = this.createPassword(tokenIdentifier);
|
||||||
|
|
||||||
} finally {
|
} finally {
|
||||||
|
|
|
@ -108,7 +108,7 @@ public class AppInfo {
|
||||||
this.diagnostics = "";
|
this.diagnostics = "";
|
||||||
}
|
}
|
||||||
this.finalStatus = app.getFinalApplicationStatus();
|
this.finalStatus = app.getFinalApplicationStatus();
|
||||||
this.clusterId = ResourceManager.clusterTimeStamp;
|
this.clusterId = ResourceManager.getClusterTimeStamp();
|
||||||
if (hasAccess) {
|
if (hasAccess) {
|
||||||
this.startedTime = app.getStartTime();
|
this.startedTime = app.getStartTime();
|
||||||
this.finishedTime = app.getFinishTime();
|
this.finishedTime = app.getFinishTime();
|
||||||
|
|
|
@ -44,7 +44,7 @@ public class ClusterInfo {
|
||||||
} // JAXB needs this
|
} // JAXB needs this
|
||||||
|
|
||||||
public ClusterInfo(ResourceManager rm) {
|
public ClusterInfo(ResourceManager rm) {
|
||||||
long ts = ResourceManager.clusterTimeStamp;
|
long ts = ResourceManager.getClusterTimeStamp();
|
||||||
|
|
||||||
this.id = ts;
|
this.id = ts;
|
||||||
this.state = rm.getServiceState();
|
this.state = rm.getServiceState();
|
||||||
|
|
|
@ -0,0 +1,151 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.resourcemanager;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.ha.HAServiceProtocol;
|
||||||
|
import org.apache.hadoop.ha.HAServiceProtocol.HAServiceState;
|
||||||
|
import org.apache.hadoop.ha.HAServiceProtocol.StateChangeRequestInfo;
|
||||||
|
import org.apache.hadoop.ha.HealthCheckFailedException;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
|
public class TestRMHA {
|
||||||
|
private Log LOG = LogFactory.getLog(TestRMHA.class);
|
||||||
|
private MockRM rm = null;
|
||||||
|
private static final String STATE_ERR =
|
||||||
|
"ResourceManager is in wrong HA state";
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setUp() throws Exception {
|
||||||
|
Configuration conf = new YarnConfiguration();
|
||||||
|
conf.setBoolean(YarnConfiguration.RM_HA_ENABLED, true);
|
||||||
|
rm = new MockRM(conf);
|
||||||
|
rm.init(conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void checkMonitorHealth() {
|
||||||
|
try {
|
||||||
|
rm.haService.monitorHealth();
|
||||||
|
} catch (HealthCheckFailedException e) {
|
||||||
|
fail("The RM is in bad health: it is Active, but the active services " +
|
||||||
|
"are not running");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void checkStandbyRMFunctionality() throws IOException {
|
||||||
|
assertEquals(STATE_ERR, HAServiceState.STANDBY,
|
||||||
|
rm.haService.getServiceStatus().getState());
|
||||||
|
assertFalse("Active RM services are started",
|
||||||
|
rm.areActiveServicesRunning());
|
||||||
|
assertTrue("RM is not ready to become active",
|
||||||
|
rm.haService.getServiceStatus().isReadyToBecomeActive());
|
||||||
|
}
|
||||||
|
|
||||||
|
private void checkActiveRMFunctionality() throws IOException {
|
||||||
|
assertEquals(STATE_ERR, HAServiceState.ACTIVE,
|
||||||
|
rm.haService.getServiceStatus().getState());
|
||||||
|
assertTrue("Active RM services aren't started",
|
||||||
|
rm.areActiveServicesRunning());
|
||||||
|
assertTrue("RM is not ready to become active",
|
||||||
|
rm.haService.getServiceStatus().isReadyToBecomeActive());
|
||||||
|
|
||||||
|
try {
|
||||||
|
rm.getNewAppId();
|
||||||
|
rm.registerNode("127.0.0.1:0", 2048);
|
||||||
|
rm.submitApp(1024);
|
||||||
|
} catch (Exception e) {
|
||||||
|
fail("Unable to perform Active RM functions");
|
||||||
|
LOG.error("ActiveRM check failed", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test to verify the following RM HA transitions to the following states.
|
||||||
|
* 1. Standby: Should be a no-op
|
||||||
|
* 2. Active: Active services should start
|
||||||
|
* 3. Active: Should be a no-op.
|
||||||
|
* While active, submit a couple of jobs
|
||||||
|
* 4. Standby: Active services should stop
|
||||||
|
* 5. Active: Active services should start
|
||||||
|
* 6. Stop the RM: All services should stop and RM should not be ready to
|
||||||
|
* become Active
|
||||||
|
*/
|
||||||
|
@Test (timeout = 30000)
|
||||||
|
public void testStartAndTransitions() throws IOException {
|
||||||
|
StateChangeRequestInfo requestInfo = new StateChangeRequestInfo(
|
||||||
|
HAServiceProtocol.RequestSource.REQUEST_BY_USER);
|
||||||
|
|
||||||
|
assertEquals(STATE_ERR, HAServiceState.INITIALIZING,
|
||||||
|
rm.haService.getServiceStatus().getState());
|
||||||
|
assertFalse("RM is ready to become active before being started",
|
||||||
|
rm.haService.getServiceStatus().isReadyToBecomeActive());
|
||||||
|
checkMonitorHealth();
|
||||||
|
|
||||||
|
rm.start();
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkStandbyRMFunctionality();
|
||||||
|
|
||||||
|
// 1. Transition to Standby - must be a no-op
|
||||||
|
rm.haService.transitionToStandby(requestInfo);
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkStandbyRMFunctionality();
|
||||||
|
|
||||||
|
// 2. Transition to active
|
||||||
|
rm.haService.transitionToActive(requestInfo);
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkActiveRMFunctionality();
|
||||||
|
|
||||||
|
// 3. Transition to active - no-op
|
||||||
|
rm.haService.transitionToActive(requestInfo);
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkActiveRMFunctionality();
|
||||||
|
|
||||||
|
// 4. Transition to standby
|
||||||
|
rm.haService.transitionToStandby(requestInfo);
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkStandbyRMFunctionality();
|
||||||
|
|
||||||
|
// 5. Transition to active to check Active->Standby->Active works
|
||||||
|
rm.haService.transitionToActive(requestInfo);
|
||||||
|
checkMonitorHealth();
|
||||||
|
checkActiveRMFunctionality();
|
||||||
|
|
||||||
|
// 6. Stop the RM. All services should stop and RM should not be ready to
|
||||||
|
// become active
|
||||||
|
rm.stop();
|
||||||
|
assertEquals(STATE_ERR, HAServiceState.STOPPING,
|
||||||
|
rm.haService.getServiceStatus().getState());
|
||||||
|
assertFalse("RM is ready to become active even after it is stopped",
|
||||||
|
rm.haService.getServiceStatus().isReadyToBecomeActive());
|
||||||
|
assertFalse("Active RM services are started",
|
||||||
|
rm.areActiveServicesRunning());
|
||||||
|
checkMonitorHealth();
|
||||||
|
}
|
||||||
|
}
|
|
@ -283,7 +283,7 @@ public class TestResourceTrackerService {
|
||||||
RegisterNodeManagerResponse response = nm.registerNode();
|
RegisterNodeManagerResponse response = nm.registerNode();
|
||||||
|
|
||||||
// Verify the RMIdentifier is correctly set in RegisterNodeManagerResponse
|
// Verify the RMIdentifier is correctly set in RegisterNodeManagerResponse
|
||||||
Assert.assertEquals(ResourceManager.clusterTimeStamp,
|
Assert.assertEquals(ResourceManager.getClusterTimeStamp(),
|
||||||
response.getRMIdentifier());
|
response.getRMIdentifier());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -88,7 +88,7 @@ public class TestApplicationMasterService {
|
||||||
ContainerTokenIdentifier tokenId =
|
ContainerTokenIdentifier tokenId =
|
||||||
BuilderUtils.newContainerTokenIdentifier(allocatedContainer
|
BuilderUtils.newContainerTokenIdentifier(allocatedContainer
|
||||||
.getContainerToken());
|
.getContainerToken());
|
||||||
Assert.assertEquals(MockRM.clusterTimeStamp, tokenId.getRMIdentifer());
|
Assert.assertEquals(MockRM.getClusterTimeStamp(), tokenId.getRMIdentifer());
|
||||||
rm.stop();
|
rm.stop();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -295,10 +295,10 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
String hadoopVersion, String resourceManagerVersionBuiltOn,
|
String hadoopVersion, String resourceManagerVersionBuiltOn,
|
||||||
String resourceManagerBuildVersion, String resourceManagerVersion) {
|
String resourceManagerBuildVersion, String resourceManagerVersion) {
|
||||||
|
|
||||||
assertEquals("clusterId doesn't match: ", ResourceManager.clusterTimeStamp,
|
assertEquals("clusterId doesn't match: ",
|
||||||
clusterid);
|
ResourceManager.getClusterTimeStamp(), clusterid);
|
||||||
assertEquals("startedOn doesn't match: ", ResourceManager.clusterTimeStamp,
|
assertEquals("startedOn doesn't match: ",
|
||||||
startedon);
|
ResourceManager.getClusterTimeStamp(), startedon);
|
||||||
assertTrue("stated doesn't match: " + state,
|
assertTrue("stated doesn't match: " + state,
|
||||||
state.matches(STATE.INITED.toString()));
|
state.matches(STATE.INITED.toString()));
|
||||||
|
|
||||||
|
|
|
@ -1181,8 +1181,8 @@ public class TestRMWebServicesApps extends JerseyTest {
|
||||||
trackingUI);
|
trackingUI);
|
||||||
WebServicesTestUtils.checkStringMatch("diagnostics", app.getDiagnostics()
|
WebServicesTestUtils.checkStringMatch("diagnostics", app.getDiagnostics()
|
||||||
.toString(), diagnostics);
|
.toString(), diagnostics);
|
||||||
assertEquals("clusterId doesn't match", ResourceManager.clusterTimeStamp,
|
assertEquals("clusterId doesn't match",
|
||||||
clusterId);
|
ResourceManager.getClusterTimeStamp(), clusterId);
|
||||||
assertEquals("startedTime doesn't match", app.getStartTime(), startedTime);
|
assertEquals("startedTime doesn't match", app.getStartTime(), startedTime);
|
||||||
assertEquals("finishedTime doesn't match", app.getFinishTime(),
|
assertEquals("finishedTime doesn't match", app.getFinishTime(),
|
||||||
finishedTime);
|
finishedTime);
|
||||||
|
|
9
pom.xml
9
pom.xml
|
@ -105,10 +105,15 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs
|
||||||
<build>
|
<build>
|
||||||
<pluginManagement>
|
<pluginManagement>
|
||||||
<plugins>
|
<plugins>
|
||||||
|
<plugin>
|
||||||
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
|
<artifactId>maven-dependency-plugin</artifactId>
|
||||||
|
<version>2.4</version>
|
||||||
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-enforcer-plugin</artifactId>
|
<artifactId>maven-enforcer-plugin</artifactId>
|
||||||
<version>1.0</version>
|
<version>1.3.1</version>
|
||||||
<configuration>
|
<configuration>
|
||||||
<rules>
|
<rules>
|
||||||
<requireMavenVersion>
|
<requireMavenVersion>
|
||||||
|
@ -224,7 +229,7 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs
|
||||||
<pluginExecutionFilter>
|
<pluginExecutionFilter>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-dependency-plugin</artifactId>
|
<artifactId>maven-dependency-plugin</artifactId>
|
||||||
<versionRange>[2.1,)</versionRange>
|
<versionRange>[2.4,)</versionRange>
|
||||||
<goals>
|
<goals>
|
||||||
<goal>copy-dependencies</goal>
|
<goal>copy-dependencies</goal>
|
||||||
<goal>build-classpath</goal>
|
<goal>build-classpath</goal>
|
||||||
|
|
Loading…
Reference in New Issue