From c4ea83c6420632d185cd13b1543fc2fef45cfd0b Mon Sep 17 00:00:00 2001 From: Chris Nauroth Date: Sun, 15 Sep 2013 04:36:48 +0000 Subject: [PATCH 1/8] =?UTF-8?q?HADOOP-9935.=20set=20junit=20dependency=20t?= =?UTF-8?q?o=20test=20scope.=20Contributed=20by=20Andr=C3=A9=20Kelpe.?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523403 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-common-project/hadoop-common/CHANGES.txt | 2 ++ hadoop-project/pom.xml | 2 ++ 2 files changed, 4 insertions(+) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 83c8c4701b2..5330b17768c 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -468,6 +468,8 @@ Release 2.1.1-beta - UNRELEASED HADOOP-9350. Hadoop not building against Java7 on OSX (Robert Kanter via stevel) + HADOOP-9935. set junit dependency to test scope. (André Kelpe via cnauroth) + Release 2.1.0-beta - 2013-08-22 INCOMPATIBLE CHANGES diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml index 316be600309..fd09dcfe225 100644 --- a/hadoop-project/pom.xml +++ b/hadoop-project/pom.xml @@ -544,6 +544,7 @@ junit junit 4.10 + test commons-lang @@ -604,6 +605,7 @@ org.mockito mockito-all 1.8.5 + test org.apache.avro From 89c7ee9245ab7841be205119ab8932f0e4a2505e Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Mon, 16 Sep 2013 11:01:28 +0000 Subject: [PATCH 2/8] YARN-1137. Add support whitelist for system users to Yarn container-executor.c. (rvs via tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523587 13f79535-47bb-0310-9956-ffa450edef68 --- .../src/site/apt/ClusterSetup.apt.vm | 2 ++ hadoop-yarn-project/CHANGES.txt | 3 +++ .../hadoop-yarn/conf/container-executor.cfg | 1 + .../impl/container-executor.c | 22 ++++++++++++++++--- .../impl/container-executor.h | 1 + .../test/test-container-executor.c | 5 +++++ 6 files changed, 31 insertions(+), 3 deletions(-) diff --git a/hadoop-common-project/hadoop-common/src/site/apt/ClusterSetup.apt.vm b/hadoop-common-project/hadoop-common/src/site/apt/ClusterSetup.apt.vm index 121754413d0..cae366abc77 100644 --- a/hadoop-common-project/hadoop-common/src/site/apt/ClusterSetup.apt.vm +++ b/hadoop-common-project/hadoop-common/src/site/apt/ClusterSetup.apt.vm @@ -929,6 +929,8 @@ KVNO Timestamp Principal *-------------------------+-------------------------+------------------------+ | <<>> | hfds,yarn,mapred,bin | Banned users. | *-------------------------+-------------------------+------------------------+ +| <<>> | foo,bar | Allowed system users. | +*-------------------------+-------------------------+------------------------+ | <<>> | 1000 | Prevent other super-users. | *-------------------------+-------------------------+------------------------+ diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index a02d7ef1c42..aeea96a6458 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -98,6 +98,9 @@ Release 2.1.1-beta - UNRELEASED completions in addition to application events. (Alejandro Abdelnur via vinodkv) + YARN-1137. Add support whitelist for system users to Yarn + container-executor.c. (rvs via tucu) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-yarn-project/hadoop-yarn/conf/container-executor.cfg b/hadoop-yarn-project/hadoop-yarn/conf/container-executor.cfg index fe1d6805296..d68cee8cc98 100644 --- a/hadoop-yarn-project/hadoop-yarn/conf/container-executor.cfg +++ b/hadoop-yarn-project/hadoop-yarn/conf/container-executor.cfg @@ -1,3 +1,4 @@ yarn.nodemanager.linux-container-executor.group=#configured value of yarn.nodemanager.linux-container-executor.group banned.users=#comma separated list of users who can not run applications min.user.id=1000#Prevent other super-users +allowed.system.users=##comma separated list of system users who CAN run applications diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c index af443770f85..307e0fafda7 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c @@ -30,6 +30,7 @@ #include #include #include +#include #include #include @@ -492,6 +493,21 @@ static struct passwd* get_user_info(const char* user) { return result; } +int is_whitelisted(const char *user) { + char **whitelist = get_values(ALLOWED_SYSTEM_USERS_KEY); + char **users = whitelist; + if (whitelist != NULL) { + for(; *users; ++users) { + if (strncmp(*users, user, LOGIN_NAME_MAX) == 0) { + free_values(whitelist); + return 1; + } + } + free_values(whitelist); + } + return 0; +} + /** * Is the user a real user account? * Checks: @@ -526,9 +542,9 @@ struct passwd* check_user(const char *user) { fflush(LOGFILE); return NULL; } - if (user_info->pw_uid < min_uid) { - fprintf(LOGFILE, "Requested user %s has id %d, which is below the " - "minimum allowed %d\n", user, user_info->pw_uid, min_uid); + if (user_info->pw_uid < min_uid && !is_whitelisted(user)) { + fprintf(LOGFILE, "Requested user %s is not whitelisted and has id %d," + "which is below the minimum allowed %d\n", user, user_info->pw_uid, min_uid); fflush(LOGFILE); free(user_info); return NULL; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.h b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.h index ec5a3742785..581ba049213 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.h +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.h @@ -65,6 +65,7 @@ enum errorcodes { #define CREDENTIALS_FILENAME "container_tokens" #define MIN_USERID_KEY "min.user.id" #define BANNED_USERS_KEY "banned.users" +#define ALLOWED_SYSTEM_USERS_KEY "allowed.system.users" #define TMP_DIR "tmp" extern struct passwd *user_detail; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c index 2563fa68eb3..b2d7d6f4997 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c @@ -99,6 +99,7 @@ int write_config_file(char *file_name) { } fprintf(file, "banned.users=bannedUser\n"); fprintf(file, "min.user.id=500\n"); + fprintf(file, "allowed.system.users=allowedUser,bin\n"); fclose(file); return 0; } @@ -195,6 +196,10 @@ void test_check_user() { printf("FAIL: failed check for system user root\n"); exit(1); } + if (check_user("bin") == NULL) { + printf("FAIL: failed check for whitelisted system user bin\n"); + exit(1); + } } void test_resolve_config_path() { From 27e7feb77734a3a6d8a6323d7a02aeff2362c4c8 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Mon, 16 Sep 2013 11:30:47 +0000 Subject: [PATCH 3/8] HADOOP-9961. versions of a few transitive dependencies diverged between hadoop subprojects. (rvs via tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523596 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 3 ++ hadoop-common-project/hadoop-common/pom.xml | 1 - hadoop-common-project/hadoop-nfs/pom.xml | 2 -- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 1 - .../hadoop-hdfs/src/contrib/bkjournal/pom.xml | 18 ++++++++++ hadoop-project/pom.xml | 36 +++++++++++++++++-- pom.xml | 9 +++-- 7 files changed, 62 insertions(+), 8 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 5330b17768c..17d2d6713c6 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -470,6 +470,9 @@ Release 2.1.1-beta - UNRELEASED HADOOP-9935. set junit dependency to test scope. (André Kelpe via cnauroth) + HADOOP-9961. versions of a few transitive dependencies diverged between hadoop + subprojects. (rvs via tucu) + Release 2.1.0-beta - 2013-08-22 INCOMPATIBLE CHANGES diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index 6f7fa0f0259..7b11971f458 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -250,7 +250,6 @@ org.apache.commons commons-compress - 1.4 diff --git a/hadoop-common-project/hadoop-nfs/pom.xml b/hadoop-common-project/hadoop-nfs/pom.xml index 468c84d9ebc..9e903b2fa0f 100644 --- a/hadoop-common-project/hadoop-nfs/pom.xml +++ b/hadoop-common-project/hadoop-nfs/pom.xml @@ -86,13 +86,11 @@ io.netty netty - 3.6.2.Final compile com.google.guava guava - 11.0.2 diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml index 9577fe180f7..7b0c31ecdb3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml @@ -49,7 +49,6 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd"> io.netty netty - 3.6.2.Final compile diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml b/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml index 537dee79c4d..253dba85e25 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml @@ -36,7 +36,25 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd"> ${basedir}/../../../../../hadoop-common-project/hadoop-common/target + + + + + org.jboss.netty + netty + 3.2.4.Final + + + + + + org.jboss.netty + netty + compile + + commons-logging commons-logging diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml index fd09dcfe225..0e1fab3c308 100644 --- a/hadoop-project/pom.xml +++ b/hadoop-project/pom.xml @@ -318,6 +318,11 @@ commons-math 2.1 + + org.apache.commons + commons-compress + 1.4.1 + xmlenc xmlenc @@ -360,6 +365,23 @@ 6.1.26 + + org.glassfish + javax.servlet + 3.1 + + + + org.codehaus.plexus + plexus-utils + 2.0.5 + + + org.codehaus.plexus + plexus-component-annotations + 1.5.5 + + asm asm @@ -420,7 +442,7 @@ io.netty netty - 3.5.11.Final + 3.6.2.Final @@ -666,6 +688,10 @@ com.sun.jmx jmxri + + org.jboss.netty + netty + @@ -674,6 +700,12 @@ 3.4.2 test-jar test + + + org.jboss.netty + netty + + org.apache.bookkeeper @@ -708,7 +740,7 @@ org.apache.maven.plugins maven-dependency-plugin - 2.1 + 2.2 org.codehaus.mojo diff --git a/pom.xml b/pom.xml index 31ccff8b8cc..dff990e24bc 100644 --- a/pom.xml +++ b/pom.xml @@ -105,10 +105,15 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs + + org.apache.maven.plugins + maven-dependency-plugin + 2.4 + org.apache.maven.plugins maven-enforcer-plugin - 1.0 + 1.3.1 @@ -224,7 +229,7 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs org.apache.maven.plugins maven-dependency-plugin - [2.1,) + [2.4,) copy-dependencies build-classpath From 9fe18b50ddc6070b5fc04347b0b4bc9e21ec4a30 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Mon, 16 Sep 2013 11:37:15 +0000 Subject: [PATCH 4/8] HADOOP-9962. in order to avoid dependency divergence within Hadoop itself lets enable DependencyConvergence. (rvs via tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523599 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 3 +++ hadoop-project/pom.xml | 20 +++++++++++++++++++ 2 files changed, 23 insertions(+) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 17d2d6713c6..10e71a48162 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -408,6 +408,9 @@ Release 2.1.1-beta - UNRELEASED HADOOP-9945. HAServiceState should have a state for stopped services. (Karthik Kambatla via atm) + HADOOP-9962. in order to avoid dependency divergence within Hadoop itself + lets enable DependencyConvergence. (rvs via tucu) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml index 0e1fab3c308..e33ea6d889e 100644 --- a/hadoop-project/pom.xml +++ b/hadoop-project/pom.xml @@ -908,6 +908,26 @@ false + + org.apache.maven.plugins + maven-enforcer-plugin + + + depcheck + + + + true + + + + + enforce + + verify + + + From e60b057c0a1da225c633fc1d7acc37043b482616 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Mon, 16 Sep 2013 12:17:17 +0000 Subject: [PATCH 5/8] MAPREDUCE-5379. Include token tracking ids in jobconf. (kkambatl via tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523605 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-mapreduce-project/CHANGES.txt | 2 ++ .../apache/hadoop/mapreduce/JobSubmitter.java | 15 +++++++++++++++ .../apache/hadoop/mapreduce/MRJobConfig.java | 7 +++++++ .../src/main/resources/mapred-default.xml | 17 +++++++++++++++++ 4 files changed, 41 insertions(+) diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 093273f0cb9..fc26e8d4fca 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -190,6 +190,8 @@ Release 2.1.1-beta - UNRELEASED but just before ClientService to avoid race conditions during RM restart. (Jian He via vinodkv) + MAPREDUCE-5379. Include token tracking ids in jobconf. (kkambatl via tucu) + OPTIMIZATIONS MAPREDUCE-5446. TestJobHistoryEvents and TestJobHistoryParsing have race diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmitter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmitter.java index cee7a33d7e6..d83a3dd7ab9 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmitter.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmitter.java @@ -24,6 +24,7 @@ import java.net.URISyntaxException; import java.net.UnknownHostException; import java.security.NoSuchAlgorithmException; +import java.util.ArrayList; import java.util.Arrays; import java.util.Comparator; import java.util.List; @@ -56,6 +57,7 @@ import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.authorize.AccessControlList; import org.apache.hadoop.security.token.Token; +import org.apache.hadoop.security.token.TokenIdentifier; import org.apache.hadoop.util.ReflectionUtils; import org.codehaus.jackson.JsonParseException; import org.codehaus.jackson.map.JsonMappingException; @@ -405,6 +407,19 @@ JobStatus submitJobInternal(Job job, Cluster cluster) // different job. TokenCache.cleanUpTokenReferral(conf); + if (conf.getBoolean( + MRJobConfig.JOB_TOKEN_TRACKING_IDS_ENABLED, + MRJobConfig.DEFAULT_JOB_TOKEN_TRACKING_IDS_ENABLED)) { + // Add HDFS tracking ids + ArrayList trackingIds = new ArrayList(); + for (Token t : + job.getCredentials().getAllTokens()) { + trackingIds.add(t.decodeIdentifier().getTrackingId()); + } + conf.setStrings(MRJobConfig.JOB_TOKEN_TRACKING_IDS, + trackingIds.toArray(new String[trackingIds.size()])); + } + // Write job file to submit dir writeConf(conf, submitJobFile); diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java index 0e8586000c9..85f6b96e291 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java @@ -313,6 +313,13 @@ public interface MRJobConfig { public static final String MAPREDUCE_JOB_CREDENTIALS_BINARY = "mapreduce.job.credentials.binary"; + /* Configs for tracking ids of tokens used by a job */ + public static final String JOB_TOKEN_TRACKING_IDS_ENABLED = + "mapreduce.job.token.tracking.ids.enabled"; + public static final boolean DEFAULT_JOB_TOKEN_TRACKING_IDS_ENABLED = false; + public static final String JOB_TOKEN_TRACKING_IDS = + "mapreduce.job.token.tracking.ids"; + public static final String JOB_SUBMITHOST = "mapreduce.job.submithostname"; public static final String JOB_SUBMITHOSTADDR = diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml index 25da3f75b34..cf28e4d288b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml @@ -748,6 +748,23 @@ + + mapreduce.job.token.tracking.ids.enabled + false + Whether to write tracking ids of tokens to + job-conf. When true, the configuration property + "mapreduce.job.token.tracking.ids" is set to the token-tracking-ids of + the job + + + + mapreduce.job.token.tracking.ids + + When mapreduce.job.token.tracking.ids.enabled is + set to true, this is set by the framework to the + token-tracking-ids used by the job. + + mapreduce.task.merge.progress.records 10000 From 60a1a70197093abf5da4ccfe9b755d40753c3fb3 Mon Sep 17 00:00:00 2001 From: Arun Murthy Date: Mon, 16 Sep 2013 14:04:04 +0000 Subject: [PATCH 6/8] YARN-1170. YARN & MapReduce proto definitions fixed to specify protobuf package as hadoop.yarn and hadoop.mapreduce respectively. Contibuted by Binglin Chang. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523652 13f79535-47bb-0310-9956-ffa450edef68 --- .../src/main/proto/MRClientProtocol.proto | 1 + .../src/main/proto/mr_protos.proto | 9 +++++---- .../src/main/proto/mr_service_protos.proto | 1 + .../src/main/proto/HSAdminRefreshProtocol.proto | 1 + hadoop-yarn-project/CHANGES.txt | 4 ++++ .../src/main/proto/applicationclient_protocol.proto | 1 + .../src/main/proto/applicationmaster_protocol.proto | 1 + .../src/main/proto/containermanagement_protocol.proto | 1 + .../server/resourcemanager_administration_protocol.proto | 1 + .../yarn_server_resourcemanager_service_protos.proto | 1 + .../hadoop-yarn-api/src/main/proto/yarn_protos.proto | 1 + .../src/main/proto/yarn_service_protos.proto | 1 + .../src/main/proto/ResourceTracker.proto | 1 + .../src/main/proto/yarn_server_common_protos.proto | 1 + .../main/proto/yarn_server_common_service_protos.proto | 1 + .../src/main/proto/LocalizationProtocol.proto | 1 + .../proto/yarn_server_nodemanager_service_protos.proto | 1 + 17 files changed, 24 insertions(+), 4 deletions(-) diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/MRClientProtocol.proto b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/MRClientProtocol.proto index 83a946f4cee..3f09719ba9b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/MRClientProtocol.proto +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/MRClientProtocol.proto @@ -19,6 +19,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "MRClientProtocol"; option java_generic_services = true; +package hadoop.mapreduce; import "Security.proto"; import "mr_service_protos.proto"; diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_protos.proto b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_protos.proto index c0a4e92c5b5..b74eef63578 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_protos.proto +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_protos.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.proto"; option java_outer_classname = "MRProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.mapreduce; import "yarn_protos.proto"; @@ -29,7 +30,7 @@ enum TaskTypeProto { } message JobIdProto { - optional ApplicationIdProto app_id = 1; + optional hadoop.yarn.ApplicationIdProto app_id = 1; optional int32 id = 2; } @@ -115,7 +116,7 @@ message TaskAttemptReportProto { optional string node_manager_host = 12; optional int32 node_manager_port = 13; optional int32 node_manager_http_port = 14; - optional ContainerIdProto container_id = 15; + optional hadoop.yarn.ContainerIdProto container_id = 15; } enum JobStateProto { @@ -148,9 +149,9 @@ message JobReportProto { } message AMInfoProto { - optional ApplicationAttemptIdProto application_attempt_id = 1; + optional hadoop.yarn.ApplicationAttemptIdProto application_attempt_id = 1; optional int64 start_time = 2; - optional ContainerIdProto container_id = 3; + optional hadoop.yarn.ContainerIdProto container_id = 3; optional string node_manager_host = 4; optional int32 node_manager_port = 5; optional int32 node_manager_http_port = 6; diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_service_protos.proto b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_service_protos.proto index ff965f30774..cb3c30c27a7 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_service_protos.proto +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/proto/mr_service_protos.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.proto"; option java_outer_classname = "MRServiceProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.mapreduce; import "Security.proto"; import "mr_protos.proto"; diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/main/proto/HSAdminRefreshProtocol.proto b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/main/proto/HSAdminRefreshProtocol.proto index e23f0a15221..1f95ee1c633 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/main/proto/HSAdminRefreshProtocol.proto +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/main/proto/HSAdminRefreshProtocol.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.mapreduce.v2.hs.proto"; option java_outer_classname = "HSAdminRefreshProtocolProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.mapreduce; /** * refresh admin acls request. diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index aeea96a6458..a2b65408170 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -44,6 +44,10 @@ Release 2.1.1-beta - UNRELEASED YARN-707. Added user information also in the YARN ClientToken so that AMs can implement authorization based on incoming users. (Jason Lowe via vinodkv) + YARN-1170. YARN & MapReduce proto definitions fixed to specify protobuf + package as hadoop.yarn and hadoop.mapreduce respectively. (Binglin Chang + via acmurthy) + NEW FEATURES IMPROVEMENTS diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationclient_protocol.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationclient_protocol.proto index 3f76849de93..af18c879521 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationclient_protocol.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationclient_protocol.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "ApplicationClientProtocol"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "Security.proto"; import "yarn_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationmaster_protocol.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationmaster_protocol.proto index f43209e48b9..d0e367291b5 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationmaster_protocol.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/applicationmaster_protocol.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "ApplicationMasterProtocol"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/containermanagement_protocol.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/containermanagement_protocol.proto index 98f438aa993..7b1647b5a1d 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/containermanagement_protocol.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/containermanagement_protocol.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "ContainerManagementProtocol"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/resourcemanager_administration_protocol.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/resourcemanager_administration_protocol.proto index 53dda07b8a6..c59c6b61f1e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/resourcemanager_administration_protocol.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/resourcemanager_administration_protocol.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "ResourceManagerAdministrationProtocol"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "server/yarn_server_resourcemanager_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/yarn_server_resourcemanager_service_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/yarn_server_resourcemanager_service_protos.proto index 4cad0abdc33..eeb14791fed 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/yarn_server_resourcemanager_service_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/server/yarn_server_resourcemanager_service_protos.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnServerResourceManagerServiceProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_protos.proto index a27cbb7cd51..f49b1653b9c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_protos.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "Security.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_service_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_service_protos.proto index 36e1d458efd..391019ade79 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_service_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_service_protos.proto @@ -26,6 +26,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnServiceProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "Security.proto"; import "yarn_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/ResourceTracker.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/ResourceTracker.proto index 6cb2f378724..1f91b6366a5 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/ResourceTracker.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/ResourceTracker.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "ResourceTracker"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_server_common_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_protos.proto index 8a9a534ae1a..4f5d16895be 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_protos.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnServerCommonProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_service_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_service_protos.proto index bc80971707b..f6d2526fb3e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_service_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/proto/yarn_server_common_service_protos.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnServerCommonServiceProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_protos.proto"; import "yarn_server_common_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/LocalizationProtocol.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/LocalizationProtocol.proto index 8fe4e988035..0f50dc3fd72 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/LocalizationProtocol.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/LocalizationProtocol.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "LocalizationProtocol"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_server_nodemanager_service_protos.proto"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/yarn_server_nodemanager_service_protos.proto b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/yarn_server_nodemanager_service_protos.proto index 166ea5b3702..6fde7cce867 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/yarn_server_nodemanager_service_protos.proto +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/proto/yarn_server_nodemanager_service_protos.proto @@ -20,6 +20,7 @@ option java_package = "org.apache.hadoop.yarn.proto"; option java_outer_classname = "YarnServerNodemanagerServiceProtos"; option java_generic_services = true; option java_generate_equals_and_hash = true; +package hadoop.yarn; import "yarn_protos.proto"; From 59587d9fad4319d10f9306a796fcd84d6b2f5adb Mon Sep 17 00:00:00 2001 From: Arun Murthy Date: Mon, 16 Sep 2013 14:18:42 +0000 Subject: [PATCH 7/8] MAPREDUCE-5493. Cleanup in-memory & on-disk segments to prevent leak on shuffle completion. Contributed by Jason Lowe. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523660 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-mapreduce-project/CHANGES.txt | 3 +++ .../hadoop/mapreduce/task/reduce/MergeManagerImpl.java | 3 +++ .../apache/hadoop/mapreduce/task/reduce/TestMerger.java | 7 ++++++- 3 files changed, 12 insertions(+), 1 deletion(-) diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index fc26e8d4fca..58cf3c6ff2b 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -267,6 +267,9 @@ Release 2.1.1-beta - UNRELEASED MAPREDUCE-5164. mapred job and queue commands omit HADOOP_CLIENT_OPTS (Nemon Lou via devaraj) + MAPREDUCE-5493. Cleanup in-memory & on-disk segments to prevent leak on + shuffle completion. (jlowe via acmurthy) + Release 2.1.0-beta - 2013-08-22 INCOMPATIBLE CHANGES diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/task/reduce/MergeManagerImpl.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/task/reduce/MergeManagerImpl.java index e52f4bf6593..ca3bed93998 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/task/reduce/MergeManagerImpl.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/task/reduce/MergeManagerImpl.java @@ -355,8 +355,11 @@ public RawKeyValueIterator close() throws Throwable { List> memory = new ArrayList>(inMemoryMergedMapOutputs); + inMemoryMergedMapOutputs.clear(); memory.addAll(inMemoryMapOutputs); + inMemoryMapOutputs.clear(); List disk = new ArrayList(onDiskMapOutputs); + onDiskMapOutputs.clear(); return finalMerge(jobConf, rfs, memory, disk); } diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestMerger.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestMerger.java index 41a1848c35f..f1d54359b82 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestMerger.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestMerger.java @@ -82,7 +82,7 @@ public void cleanup() throws IOException { } @Test - public void testInMemoryMerger() throws IOException { + public void testInMemoryMerger() throws Throwable { JobID jobId = new JobID("a", 0); TaskAttemptID reduceId = new TaskAttemptID( new TaskID(jobId, TaskType.REDUCE, 0), 0); @@ -132,6 +132,11 @@ public void testInMemoryMerger() throws IOException { readOnDiskMapOutput(conf, fs, outPath, keys, values); Assert.assertEquals(keys, Arrays.asList("apple", "banana", "carrot")); Assert.assertEquals(values, Arrays.asList("disgusting", "pretty good", "delicious")); + + mergeManager.close(); + Assert.assertEquals(0, mergeManager.inMemoryMapOutputs.size()); + Assert.assertEquals(0, mergeManager.inMemoryMergedMapOutputs.size()); + Assert.assertEquals(0, mergeManager.onDiskMapOutputs.size()); } private byte[] writeMapOutput(Configuration conf, Map keysToValues) From ead3dcc3dea0d5b6d3248c1e1a6300f0714a8456 Mon Sep 17 00:00:00 2001 From: Bikas Saha Date: Mon, 16 Sep 2013 18:33:59 +0000 Subject: [PATCH 8/8] YARN-1027. Implement RMHAProtocolService (Karthik Kambatla via bikas) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1523750 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 1 + .../hadoop/yarn/conf/YarnConfiguration.java | 5 + .../src/main/resources/yarn-default.xml | 8 + .../resourcemanager/ClientRMService.java | 2 +- .../resourcemanager/RMHAProtocolService.java | 153 ++++++++++++++++++ .../resourcemanager/ResourceManager.java | 56 ++++++- .../ResourceTrackerService.java | 2 +- .../RMContainerTokenSecretManager.java | 2 +- .../resourcemanager/webapp/dao/AppInfo.java | 2 +- .../webapp/dao/ClusterInfo.java | 2 +- .../yarn/server/resourcemanager/TestRMHA.java | 151 +++++++++++++++++ .../TestResourceTrackerService.java | 2 +- .../TestApplicationMasterService.java | 2 +- .../webapp/TestRMWebServices.java | 8 +- .../webapp/TestRMWebServicesApps.java | 4 +- 15 files changed, 383 insertions(+), 17 deletions(-) create mode 100644 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMHAProtocolService.java create mode 100644 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMHA.java diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index a2b65408170..17e8a4ef754 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -29,6 +29,7 @@ Release 2.3.0 - UNRELEASED YARN-905. Add state filters to nodes CLI (Wei Yan via Sandy Ryza) YARN-1098. Separate out RM services into Always On and Active (Karthik Kambatla via bikas) + YARN-1027. Implement RMHAProtocolService (Karthik Kambatla via bikas) OPTIMIZATIONS diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java index febf095d227..8142b78d58e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java @@ -270,6 +270,11 @@ public class YarnConfiguration extends Configuration { public static final String RECOVERY_ENABLED = RM_PREFIX + "recovery.enabled"; public static final boolean DEFAULT_RM_RECOVERY_ENABLED = false; + + /** HA related configs */ + public static final String RM_HA_PREFIX = RM_PREFIX + "ha."; + public static final String RM_HA_ENABLED = RM_HA_PREFIX + "enabled"; + public static final boolean DEFAULT_RM_HA_ENABLED = false; /** The class to use as the persistent store.*/ public static final String RM_STORE = RM_PREFIX + "store.class"; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml index ab8d50aab10..c0f3a1ec417 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml @@ -268,6 +268,14 @@ + + Enable RM high-availability. When enabled, the RM starts + in the Standby mode by default, and transitions to the Active mode when + prompted to. + yarn.resourcemanager.ha.enabled + false + + The maximum number of completed applications RM keeps. yarn.resourcemanager.max-completed-applications diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java index cc31643296f..ec730f0485e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java @@ -196,7 +196,7 @@ private boolean checkAccess(UserGroupInformation callerUGI, String owner, ApplicationId getNewApplicationId() { ApplicationId applicationId = org.apache.hadoop.yarn.server.utils.BuilderUtils - .newApplicationId(recordFactory, ResourceManager.clusterTimeStamp, + .newApplicationId(recordFactory, ResourceManager.getClusterTimeStamp(), applicationCounter.incrementAndGet()); LOG.info("Allocated new applicationId: " + applicationId.getId()); return applicationId; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMHAProtocolService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMHAProtocolService.java new file mode 100644 index 00000000000..b9aca3cbe43 --- /dev/null +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMHAProtocolService.java @@ -0,0 +1,153 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.yarn.server.resourcemanager; + +import com.google.common.annotations.VisibleForTesting; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.ha.HAServiceProtocol; +import org.apache.hadoop.ha.HAServiceStatus; +import org.apache.hadoop.ha.HealthCheckFailedException; +import org.apache.hadoop.service.AbstractService; +import org.apache.hadoop.yarn.conf.YarnConfiguration; +import org.apache.hadoop.yarn.exceptions.YarnRuntimeException; + +import java.io.IOException; + +@InterfaceAudience.Private +@InterfaceStability.Unstable +public class RMHAProtocolService extends AbstractService implements + HAServiceProtocol { + private static final Log LOG = LogFactory.getLog(RMHAProtocolService.class); + + private Configuration conf; + private ResourceManager rm; + @VisibleForTesting + protected HAServiceState haState = HAServiceState.INITIALIZING; + + public RMHAProtocolService(ResourceManager resourceManager) { + super("RMHAProtocolService"); + this.rm = resourceManager; + } + + @Override + public synchronized void serviceInit(Configuration conf) throws Exception { + this.conf = conf; + rm.createAndInitActiveServices(); + super.serviceInit(this.conf); + } + + @Override + public synchronized void serviceStart() throws Exception { + boolean haEnabled = this.conf.getBoolean(YarnConfiguration.RM_HA_ENABLED, + YarnConfiguration.DEFAULT_RM_HA_ENABLED); + + if (haEnabled) { + transitionToStandby(true); + } else { + transitionToActive(); + } + + super.serviceStart(); + } + + @Override + public synchronized void serviceStop() throws Exception { + transitionToStandby(false); + haState = HAServiceState.STOPPING; + super.serviceStop(); + } + + @Override + public synchronized void monitorHealth() throws HealthCheckFailedException { + if (haState == HAServiceState.ACTIVE && !rm.areActiveServicesRunning()) { + throw new HealthCheckFailedException( + "Active ResourceManager services are not running!"); + } + } + + private synchronized void transitionToActive() throws Exception { + if (haState == HAServiceState.ACTIVE) { + LOG.info("Already in active state"); + return; + } + + LOG.info("Transitioning to active"); + rm.startActiveServices(); + haState = HAServiceState.ACTIVE; + LOG.info("Transitioned to active"); + } + + @Override + public synchronized void transitionToActive(StateChangeRequestInfo reqInfo) { + // TODO (YARN-1177): When automatic failover is enabled, + // check if transition should be allowed for this request + try { + transitionToActive(); + } catch (Exception e) { + LOG.error("Error when transitioning to Active mode", e); + throw new YarnRuntimeException(e); + } + } + + private synchronized void transitionToStandby(boolean initialize) + throws Exception { + if (haState == HAServiceState.STANDBY) { + LOG.info("Already in standby state"); + return; + } + + LOG.info("Transitioning to standby"); + if (haState == HAServiceState.ACTIVE) { + rm.stopActiveServices(); + if (initialize) { + rm.createAndInitActiveServices(); + } + } + haState = HAServiceState.STANDBY; + LOG.info("Transitioned to standby"); + } + + @Override + public synchronized void transitionToStandby(StateChangeRequestInfo reqInfo) { + // TODO (YARN-1177): When automatic failover is enabled, + // check if transition should be allowed for this request + try { + transitionToStandby(true); + } catch (Exception e) { + LOG.error("Error when transitioning to Standby mode", e); + throw new YarnRuntimeException(e); + } + } + + @Override + public synchronized HAServiceStatus getServiceStatus() throws IOException { + HAServiceStatus ret = new HAServiceStatus(haState); + if (haState == HAServiceState.ACTIVE || haState == HAServiceState.STANDBY) { + ret.setReadyToBecomeActive(); + } else { + ret.setNotReadyToBecomeActive("State is " + haState); + } + return ret; + } +} diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java index 8c0b195f707..ff57fc57540 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java @@ -105,7 +105,14 @@ public class ResourceManager extends CompositeService implements Recoverable { public static final int SHUTDOWN_HOOK_PRIORITY = 30; private static final Log LOG = LogFactory.getLog(ResourceManager.class); - public static final long clusterTimeStamp = System.currentTimeMillis(); + private static long clusterTimeStamp = System.currentTimeMillis(); + + /** + * "Always On" services. Services that need to run always irrespective of + * the HA state of the RM. + */ + @VisibleForTesting + protected RMHAProtocolService haService; /** * "Active" services. Services that need to run only on the Active RM. @@ -155,14 +162,18 @@ public ResourceManager() { public RMContext getRMContext() { return this.rmContext; } + + public static long getClusterTimeStamp() { + return clusterTimeStamp; + } @Override protected void serviceInit(Configuration conf) throws Exception { validateConfigs(conf); this.conf = conf; - activeServices = new RMActiveServices(); - addService(activeServices); + haService = new RMHAProtocolService(this); + addService(haService); super.serviceInit(conf); } @@ -470,6 +481,7 @@ protected void serviceStop() throws Exception { LOG.error("Error closing store.", e); } } + super.serviceStop(); } } @@ -708,6 +720,43 @@ protected void startWepApp() { webApp = builder.start(new RMWebApp(this)); } + /** + * Helper method to create and init {@link #activeServices}. This creates an + * instance of {@link RMActiveServices} and initializes it. + * @throws Exception + */ + void createAndInitActiveServices() throws Exception { + activeServices = new RMActiveServices(); + activeServices.init(conf); + } + + /** + * Helper method to start {@link #activeServices}. + * @throws Exception + */ + void startActiveServices() throws Exception { + if (activeServices != null) { + clusterTimeStamp = System.currentTimeMillis(); + activeServices.start(); + } + } + + /** + * Helper method to stop {@link #activeServices}. + * @throws Exception + */ + void stopActiveServices() throws Exception { + if (activeServices != null) { + activeServices.stop(); + activeServices = null; + } + } + + @VisibleForTesting + protected boolean areActiveServicesRunning() { + return activeServices != null && activeServices.isInState(STATE.STARTED); + } + @Override protected void serviceStart() throws Exception { try { @@ -715,7 +764,6 @@ protected void serviceStart() throws Exception { } catch(IOException ie) { throw new YarnRuntimeException("Failed to login", ie); } - super.serviceStart(); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java index aa8f120a9cc..ec2d4350427 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java @@ -229,7 +229,7 @@ public RegisterNodeManagerResponse registerNodeManager( + ", assigned nodeId " + nodeId; LOG.info(message); response.setNodeAction(NodeAction.NORMAL); - response.setRMIdentifier(ResourceManager.clusterTimeStamp); + response.setRMIdentifier(ResourceManager.getClusterTimeStamp()); return response; } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/RMContainerTokenSecretManager.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/RMContainerTokenSecretManager.java index f2f4925a383..c62f2ee369c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/RMContainerTokenSecretManager.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/RMContainerTokenSecretManager.java @@ -185,7 +185,7 @@ public void run() { tokenIdentifier = new ContainerTokenIdentifier(containerId, nodeId.toString(), appSubmitter, capability, expiryTimeStamp, this.currentMasterKey - .getMasterKey().getKeyId(), ResourceManager.clusterTimeStamp); + .getMasterKey().getKeyId(), ResourceManager.getClusterTimeStamp()); password = this.createPassword(tokenIdentifier); } finally { diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/AppInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/AppInfo.java index fa7b4f784f2..7cb126f7ea8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/AppInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/AppInfo.java @@ -108,7 +108,7 @@ public AppInfo(RMApp app, Boolean hasAccess) { this.diagnostics = ""; } this.finalStatus = app.getFinalApplicationStatus(); - this.clusterId = ResourceManager.clusterTimeStamp; + this.clusterId = ResourceManager.getClusterTimeStamp(); if (hasAccess) { this.startedTime = app.getStartTime(); this.finishedTime = app.getFinishTime(); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/ClusterInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/ClusterInfo.java index 96cda2b8cb0..2da79605e21 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/ClusterInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/ClusterInfo.java @@ -44,7 +44,7 @@ public ClusterInfo() { } // JAXB needs this public ClusterInfo(ResourceManager rm) { - long ts = ResourceManager.clusterTimeStamp; + long ts = ResourceManager.getClusterTimeStamp(); this.id = ts; this.state = rm.getServiceState(); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMHA.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMHA.java new file mode 100644 index 00000000000..7415791f094 --- /dev/null +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMHA.java @@ -0,0 +1,151 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.yarn.server.resourcemanager; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.ha.HAServiceProtocol; +import org.apache.hadoop.ha.HAServiceProtocol.HAServiceState; +import org.apache.hadoop.ha.HAServiceProtocol.StateChangeRequestInfo; +import org.apache.hadoop.ha.HealthCheckFailedException; +import org.apache.hadoop.yarn.conf.YarnConfiguration; +import org.junit.Before; +import org.junit.Test; + +import java.io.IOException; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertFalse; +import static org.junit.Assert.assertTrue; +import static org.junit.Assert.fail; + +public class TestRMHA { + private Log LOG = LogFactory.getLog(TestRMHA.class); + private MockRM rm = null; + private static final String STATE_ERR = + "ResourceManager is in wrong HA state"; + + @Before + public void setUp() throws Exception { + Configuration conf = new YarnConfiguration(); + conf.setBoolean(YarnConfiguration.RM_HA_ENABLED, true); + rm = new MockRM(conf); + rm.init(conf); + } + + private void checkMonitorHealth() { + try { + rm.haService.monitorHealth(); + } catch (HealthCheckFailedException e) { + fail("The RM is in bad health: it is Active, but the active services " + + "are not running"); + } + } + + private void checkStandbyRMFunctionality() throws IOException { + assertEquals(STATE_ERR, HAServiceState.STANDBY, + rm.haService.getServiceStatus().getState()); + assertFalse("Active RM services are started", + rm.areActiveServicesRunning()); + assertTrue("RM is not ready to become active", + rm.haService.getServiceStatus().isReadyToBecomeActive()); + } + + private void checkActiveRMFunctionality() throws IOException { + assertEquals(STATE_ERR, HAServiceState.ACTIVE, + rm.haService.getServiceStatus().getState()); + assertTrue("Active RM services aren't started", + rm.areActiveServicesRunning()); + assertTrue("RM is not ready to become active", + rm.haService.getServiceStatus().isReadyToBecomeActive()); + + try { + rm.getNewAppId(); + rm.registerNode("127.0.0.1:0", 2048); + rm.submitApp(1024); + } catch (Exception e) { + fail("Unable to perform Active RM functions"); + LOG.error("ActiveRM check failed", e); + } + } + + /** + * Test to verify the following RM HA transitions to the following states. + * 1. Standby: Should be a no-op + * 2. Active: Active services should start + * 3. Active: Should be a no-op. + * While active, submit a couple of jobs + * 4. Standby: Active services should stop + * 5. Active: Active services should start + * 6. Stop the RM: All services should stop and RM should not be ready to + * become Active + */ + @Test (timeout = 30000) + public void testStartAndTransitions() throws IOException { + StateChangeRequestInfo requestInfo = new StateChangeRequestInfo( + HAServiceProtocol.RequestSource.REQUEST_BY_USER); + + assertEquals(STATE_ERR, HAServiceState.INITIALIZING, + rm.haService.getServiceStatus().getState()); + assertFalse("RM is ready to become active before being started", + rm.haService.getServiceStatus().isReadyToBecomeActive()); + checkMonitorHealth(); + + rm.start(); + checkMonitorHealth(); + checkStandbyRMFunctionality(); + + // 1. Transition to Standby - must be a no-op + rm.haService.transitionToStandby(requestInfo); + checkMonitorHealth(); + checkStandbyRMFunctionality(); + + // 2. Transition to active + rm.haService.transitionToActive(requestInfo); + checkMonitorHealth(); + checkActiveRMFunctionality(); + + // 3. Transition to active - no-op + rm.haService.transitionToActive(requestInfo); + checkMonitorHealth(); + checkActiveRMFunctionality(); + + // 4. Transition to standby + rm.haService.transitionToStandby(requestInfo); + checkMonitorHealth(); + checkStandbyRMFunctionality(); + + // 5. Transition to active to check Active->Standby->Active works + rm.haService.transitionToActive(requestInfo); + checkMonitorHealth(); + checkActiveRMFunctionality(); + + // 6. Stop the RM. All services should stop and RM should not be ready to + // become active + rm.stop(); + assertEquals(STATE_ERR, HAServiceState.STOPPING, + rm.haService.getServiceStatus().getState()); + assertFalse("RM is ready to become active even after it is stopped", + rm.haService.getServiceStatus().isReadyToBecomeActive()); + assertFalse("Active RM services are started", + rm.areActiveServicesRunning()); + checkMonitorHealth(); + } +} diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java index 83a6af8f40a..6026ef9ec52 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java @@ -283,7 +283,7 @@ public void testSetRMIdentifierInRegistration() throws Exception { RegisterNodeManagerResponse response = nm.registerNode(); // Verify the RMIdentifier is correctly set in RegisterNodeManagerResponse - Assert.assertEquals(ResourceManager.clusterTimeStamp, + Assert.assertEquals(ResourceManager.getClusterTimeStamp(), response.getRMIdentifier()); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java index 31e100c34da..021c1978673 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java @@ -88,7 +88,7 @@ public void testRMIdentifierOnContainerAllocation() throws Exception { ContainerTokenIdentifier tokenId = BuilderUtils.newContainerTokenIdentifier(allocatedContainer .getContainerToken()); - Assert.assertEquals(MockRM.clusterTimeStamp, tokenId.getRMIdentifer()); + Assert.assertEquals(MockRM.getClusterTimeStamp(), tokenId.getRMIdentifer()); rm.stop(); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServices.java index 89bec21cc24..2bae59daf20 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServices.java @@ -295,10 +295,10 @@ public void verifyClusterGeneric(long clusterid, long startedon, String hadoopVersion, String resourceManagerVersionBuiltOn, String resourceManagerBuildVersion, String resourceManagerVersion) { - assertEquals("clusterId doesn't match: ", ResourceManager.clusterTimeStamp, - clusterid); - assertEquals("startedOn doesn't match: ", ResourceManager.clusterTimeStamp, - startedon); + assertEquals("clusterId doesn't match: ", + ResourceManager.getClusterTimeStamp(), clusterid); + assertEquals("startedOn doesn't match: ", + ResourceManager.getClusterTimeStamp(), startedon); assertTrue("stated doesn't match: " + state, state.matches(STATE.INITED.toString())); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java index 5e7145283ce..77ccbaddf57 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java @@ -1181,8 +1181,8 @@ public void verifyAppInfoGeneric(RMApp app, String id, String user, trackingUI); WebServicesTestUtils.checkStringMatch("diagnostics", app.getDiagnostics() .toString(), diagnostics); - assertEquals("clusterId doesn't match", ResourceManager.clusterTimeStamp, - clusterId); + assertEquals("clusterId doesn't match", + ResourceManager.getClusterTimeStamp(), clusterId); assertEquals("startedTime doesn't match", app.getStartTime(), startedTime); assertEquals("finishedTime doesn't match", app.getFinishTime(), finishedTime);