Merge -c 1370666 from trunk to branch-2 to fix YARN-1. Promote YARN to be a sub-project of Apache Hadoop.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1370667 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
208da340c3
commit
92a99de731
|
@ -78,15 +78,15 @@ if $PATCH -p0 -E --dry-run < $PATCH_FILE 2>&1 > $TMP; then
|
||||||
if [[ -d hadoop-common-project ]]; then
|
if [[ -d hadoop-common-project ]]; then
|
||||||
echo Looks like this is being run at project root
|
echo Looks like this is being run at project root
|
||||||
|
|
||||||
# if all of the lines start with hadoop-common/, hadoop-hdfs/, or hadoop-mapreduce/, this is
|
# if all of the lines start with hadoop-common/, hadoop-hdfs/, hadoop-yarn/ or hadoop-mapreduce/, this is
|
||||||
# relative to the hadoop root instead of the subproject root, so we need
|
# relative to the hadoop root instead of the subproject root, so we need
|
||||||
# to chop off another layer
|
# to chop off another layer
|
||||||
elif [[ "$PREFIX_DIRS_AND_FILES" =~ ^(hadoop-common-project|hadoop-hdfs-project|hadoop-mapreduce-project)$ ]]; then
|
elif [[ "$PREFIX_DIRS_AND_FILES" =~ ^(hadoop-common-project|hadoop-hdfs-project|hadoop-yarn-project|hadoop-mapreduce-project)$ ]]; then
|
||||||
|
|
||||||
echo Looks like this is relative to project root. Increasing PLEVEL
|
echo Looks like this is relative to project root. Increasing PLEVEL
|
||||||
PLEVEL=$[$PLEVEL + 1]
|
PLEVEL=$[$PLEVEL + 1]
|
||||||
|
|
||||||
elif ! echo "$PREFIX_DIRS_AND_FILES" | grep -vxq 'hadoop-common-project\|hadoop-hdfs-project\|hadoop-mapreduce-project' ; then
|
elif ! echo "$PREFIX_DIRS_AND_FILES" | grep -vxq 'hadoop-common-project\|hadoop-hdfs-project\|hadoop-yarn-project\|hadoop-mapreduce-project' ; then
|
||||||
echo Looks like this is a cross-subproject patch. Try applying from the project root
|
echo Looks like this is a cross-subproject patch. Try applying from the project root
|
||||||
cleanup 1
|
cleanup 1
|
||||||
fi
|
fi
|
||||||
|
|
|
@ -23,19 +23,6 @@
|
||||||
</formats>
|
</formats>
|
||||||
<includeBaseDirectory>false</includeBaseDirectory>
|
<includeBaseDirectory>false</includeBaseDirectory>
|
||||||
<fileSets>
|
<fileSets>
|
||||||
<fileSet>
|
|
||||||
<directory>hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/target/native/target/usr/local/bin</directory>
|
|
||||||
<outputDirectory>bin</outputDirectory>
|
|
||||||
<fileMode>0755</fileMode>
|
|
||||||
</fileSet>
|
|
||||||
<fileSet>
|
|
||||||
<directory>hadoop-yarn/bin</directory>
|
|
||||||
<outputDirectory>bin</outputDirectory>
|
|
||||||
<includes>
|
|
||||||
<include>yarn</include>
|
|
||||||
</includes>
|
|
||||||
<fileMode>0755</fileMode>
|
|
||||||
</fileSet>
|
|
||||||
<fileSet>
|
<fileSet>
|
||||||
<directory>bin</directory>
|
<directory>bin</directory>
|
||||||
<outputDirectory>bin</outputDirectory>
|
<outputDirectory>bin</outputDirectory>
|
||||||
|
@ -52,25 +39,6 @@
|
||||||
</includes>
|
</includes>
|
||||||
<fileMode>0755</fileMode>
|
<fileMode>0755</fileMode>
|
||||||
</fileSet>
|
</fileSet>
|
||||||
<fileSet>
|
|
||||||
<directory>hadoop-yarn/bin</directory>
|
|
||||||
<outputDirectory>libexec</outputDirectory>
|
|
||||||
<includes>
|
|
||||||
<include>yarn-config.sh</include>
|
|
||||||
</includes>
|
|
||||||
<fileMode>0755</fileMode>
|
|
||||||
</fileSet>
|
|
||||||
<fileSet>
|
|
||||||
<directory>hadoop-yarn/bin</directory>
|
|
||||||
<outputDirectory>sbin</outputDirectory>
|
|
||||||
<includes>
|
|
||||||
<include>yarn-daemon.sh</include>
|
|
||||||
<include>yarn-daemons.sh</include>
|
|
||||||
<include>start-yarn.sh</include>
|
|
||||||
<include>stop-yarn.sh</include>
|
|
||||||
</includes>
|
|
||||||
<fileMode>0755</fileMode>
|
|
||||||
</fileSet>
|
|
||||||
<fileSet>
|
<fileSet>
|
||||||
<directory>bin</directory>
|
<directory>bin</directory>
|
||||||
<outputDirectory>sbin</outputDirectory>
|
<outputDirectory>sbin</outputDirectory>
|
||||||
|
@ -80,7 +48,7 @@
|
||||||
<fileMode>0755</fileMode>
|
<fileMode>0755</fileMode>
|
||||||
</fileSet>
|
</fileSet>
|
||||||
<fileSet>
|
<fileSet>
|
||||||
<directory>hadoop-yarn/conf</directory>
|
<directory>conf</directory>
|
||||||
<outputDirectory>etc/hadoop</outputDirectory>
|
<outputDirectory>etc/hadoop</outputDirectory>
|
||||||
<includes>
|
<includes>
|
||||||
<include>**/*</include>
|
<include>**/*</include>
|
||||||
|
|
|
@ -0,0 +1,145 @@
|
||||||
|
<!--
|
||||||
|
Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
contributor license agreements. See the NOTICE file distributed with
|
||||||
|
this work for additional information regarding copyright ownership.
|
||||||
|
The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
(the "License"); you may not use this file except in compliance with
|
||||||
|
the License. You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
-->
|
||||||
|
<assembly xmlns="http://maven.apache.org/plugins/maven-assembly-plugin/assembly/1.1.0"
|
||||||
|
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||||
|
xsi:schemaLocation="http://maven.apache.org/plugins/maven-assembly-plugin/assembly/1.1.0 http://maven.apache.org/xsd/assembly-1.1.0.xsd">
|
||||||
|
<id>hadoop-yarn-dist</id>
|
||||||
|
<formats>
|
||||||
|
<format>dir</format>
|
||||||
|
</formats>
|
||||||
|
<includeBaseDirectory>false</includeBaseDirectory>
|
||||||
|
<fileSets>
|
||||||
|
<fileSet>
|
||||||
|
<directory>hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/target/native/target/usr/local/bin</directory>
|
||||||
|
<outputDirectory>bin</outputDirectory>
|
||||||
|
<fileMode>0755</fileMode>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>hadoop-yarn/bin</directory>
|
||||||
|
<outputDirectory>bin</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>yarn</include>
|
||||||
|
</includes>
|
||||||
|
<fileMode>0755</fileMode>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>hadoop-yarn/bin</directory>
|
||||||
|
<outputDirectory>libexec</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>yarn-config.sh</include>
|
||||||
|
</includes>
|
||||||
|
<fileMode>0755</fileMode>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>hadoop-yarn/bin</directory>
|
||||||
|
<outputDirectory>sbin</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>yarn-daemon.sh</include>
|
||||||
|
<include>yarn-daemons.sh</include>
|
||||||
|
<include>start-yarn.sh</include>
|
||||||
|
<include>stop-yarn.sh</include>
|
||||||
|
</includes>
|
||||||
|
<fileMode>0755</fileMode>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>hadoop-yarn/conf</directory>
|
||||||
|
<outputDirectory>etc/hadoop</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>**/*</include>
|
||||||
|
</includes>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${basedir}</directory>
|
||||||
|
<outputDirectory>/share/doc/hadoop/${hadoop.component}</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>*.txt</include>
|
||||||
|
</includes>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${project.build.directory}/webapps</directory>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/webapps</outputDirectory>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${basedir}/src/main/conf</directory>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/templates</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>*-site.xml</include>
|
||||||
|
</includes>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${basedir}/src/main/packages/templates/conf</directory>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/templates/conf</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>*</include>
|
||||||
|
</includes>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${basedir}/dev-support/jdiff</directory>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/jdiff</outputDirectory>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${project.build.directory}/site/jdiff/xml</directory>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/jdiff</outputDirectory>
|
||||||
|
</fileSet>
|
||||||
|
<fileSet>
|
||||||
|
<directory>${project.build.directory}/site</directory>
|
||||||
|
<outputDirectory>/share/doc/hadoop/${hadoop.component}</outputDirectory>
|
||||||
|
</fileSet>
|
||||||
|
</fileSets>
|
||||||
|
<moduleSets>
|
||||||
|
<moduleSet>
|
||||||
|
<binaries>
|
||||||
|
<outputDirectory>share/hadoop/${hadoop.component}</outputDirectory>
|
||||||
|
<includeDependencies>false</includeDependencies>
|
||||||
|
<unpack>false</unpack>
|
||||||
|
</binaries>
|
||||||
|
</moduleSet>
|
||||||
|
<moduleSet>
|
||||||
|
<includes>
|
||||||
|
<include>org.apache.hadoop:hadoop-yarn-server-tests</include>
|
||||||
|
</includes>
|
||||||
|
<binaries>
|
||||||
|
<attachmentClassifier>tests</attachmentClassifier>
|
||||||
|
<outputDirectory>share/hadoop/${hadoop.component}</outputDirectory>
|
||||||
|
<includeDependencies>false</includeDependencies>
|
||||||
|
<unpack>false</unpack>
|
||||||
|
</binaries>
|
||||||
|
</moduleSet>
|
||||||
|
</moduleSets>
|
||||||
|
<dependencySets>
|
||||||
|
<dependencySet>
|
||||||
|
<useProjectArtifact>false</useProjectArtifact>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/lib</outputDirectory>
|
||||||
|
<!-- Exclude hadoop artifacts. They will be found via HADOOP* env -->
|
||||||
|
<excludes>
|
||||||
|
<exclude>org.apache.hadoop:hadoop-common</exclude>
|
||||||
|
<exclude>org.apache.hadoop:hadoop-hdfs</exclude>
|
||||||
|
<!-- use slf4j from common to avoid multiple binding warnings -->
|
||||||
|
<exclude>org.slf4j:slf4j-api</exclude>
|
||||||
|
<exclude>org.slf4j:slf4j-log4j12</exclude>
|
||||||
|
<exclude>org.hsqldb:hsqldb</exclude>
|
||||||
|
</excludes>
|
||||||
|
</dependencySet>
|
||||||
|
<dependencySet>
|
||||||
|
<useProjectArtifact>false</useProjectArtifact>
|
||||||
|
<outputDirectory>/share/hadoop/${hadoop.component}/lib-examples</outputDirectory>
|
||||||
|
<includes>
|
||||||
|
<include>org.hsqldb:hsqldb</include>
|
||||||
|
</includes>
|
||||||
|
</dependencySet>
|
||||||
|
</dependencySets>
|
||||||
|
</assembly>
|
|
@ -42,8 +42,8 @@ HADOOP_COMMON_LIB_JARS_DIR=${HADOOP_COMMON_LIB_JARS_DIR:-"share/hadoop/common/li
|
||||||
HADOOP_COMMON_LIB_NATIVE_DIR=${HADOOP_COMMON_LIB_NATIVE_DIR:-"lib/native"}
|
HADOOP_COMMON_LIB_NATIVE_DIR=${HADOOP_COMMON_LIB_NATIVE_DIR:-"lib/native"}
|
||||||
HDFS_DIR=${HDFS_DIR:-"share/hadoop/hdfs"}
|
HDFS_DIR=${HDFS_DIR:-"share/hadoop/hdfs"}
|
||||||
HDFS_LIB_JARS_DIR=${HDFS_LIB_JARS_DIR:-"share/hadoop/hdfs/lib"}
|
HDFS_LIB_JARS_DIR=${HDFS_LIB_JARS_DIR:-"share/hadoop/hdfs/lib"}
|
||||||
YARN_DIR=${YARN_DIR:-"share/hadoop/mapreduce"}
|
YARN_DIR=${YARN_DIR:-"share/hadoop/yarn"}
|
||||||
YARN_LIB_JARS_DIR=${YARN_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
|
YARN_LIB_JARS_DIR=${YARN_LIB_JARS_DIR:-"share/hadoop/yarn/lib"}
|
||||||
MAPRED_DIR=${MAPRED_DIR:-"share/hadoop/mapreduce"}
|
MAPRED_DIR=${MAPRED_DIR:-"share/hadoop/mapreduce"}
|
||||||
MAPRED_LIB_JARS_DIR=${MAPRED_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
|
MAPRED_LIB_JARS_DIR=${MAPRED_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
|
||||||
|
|
||||||
|
|
|
@ -120,6 +120,7 @@
|
||||||
run cp -r $ROOT/hadoop-common-project/hadoop-common/target/hadoop-common-${project.version}/* .
|
run cp -r $ROOT/hadoop-common-project/hadoop-common/target/hadoop-common-${project.version}/* .
|
||||||
run cp -r $ROOT/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-hdfs-${project.version}/* .
|
run cp -r $ROOT/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-hdfs-${project.version}/* .
|
||||||
run cp -r $ROOT/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-${project.version}/* .
|
run cp -r $ROOT/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-${project.version}/* .
|
||||||
|
run cp -r $ROOT/hadoop-yarn-project/target/hadoop-yarn-project-${project.version}/* .
|
||||||
run cp -r $ROOT/hadoop-mapreduce-project/target/hadoop-mapreduce-${project.version}/* .
|
run cp -r $ROOT/hadoop-mapreduce-project/target/hadoop-mapreduce-${project.version}/* .
|
||||||
run cp -r $ROOT/hadoop-tools/hadoop-tools-dist/target/hadoop-tools-dist-${project.version}/* .
|
run cp -r $ROOT/hadoop-tools/hadoop-tools-dist/target/hadoop-tools-dist-${project.version}/* .
|
||||||
echo
|
echo
|
||||||
|
|
|
@ -77,7 +77,7 @@
|
||||||
<configuration>
|
<configuration>
|
||||||
<executable>protoc</executable>
|
<executable>protoc</executable>
|
||||||
<arguments>
|
<arguments>
|
||||||
<argument>-I../../hadoop-yarn/hadoop-yarn-api/src/main/proto/</argument>
|
<argument>-I../../../hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/</argument>
|
||||||
<argument>-Isrc/main/proto/</argument>
|
<argument>-Isrc/main/proto/</argument>
|
||||||
<argument>--java_out=target/generated-sources/proto</argument>
|
<argument>--java_out=target/generated-sources/proto</argument>
|
||||||
<argument>src/main/proto/mr_protos.proto</argument>
|
<argument>src/main/proto/mr_protos.proto</argument>
|
||||||
|
|
|
@ -1,727 +0,0 @@
|
||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
#include "configuration.h"
|
|
||||||
#include "container-executor.h"
|
|
||||||
|
|
||||||
#include <errno.h>
|
|
||||||
#include <fcntl.h>
|
|
||||||
#include <unistd.h>
|
|
||||||
#include <signal.h>
|
|
||||||
#include <stdio.h>
|
|
||||||
#include <stdlib.h>
|
|
||||||
#include <string.h>
|
|
||||||
#include <sys/stat.h>
|
|
||||||
#include <sys/wait.h>
|
|
||||||
|
|
||||||
#define TEST_ROOT "/tmp/test-task-controller"
|
|
||||||
#define DONT_TOUCH_FILE "dont-touch-me"
|
|
||||||
|
|
||||||
static char* username = NULL;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Run the command using the effective user id.
|
|
||||||
* It can't use system, since bash seems to copy the real user id into the
|
|
||||||
* effective id.
|
|
||||||
*/
|
|
||||||
void run(const char *cmd) {
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: failed to fork - %s\n", strerror(errno));
|
|
||||||
} else if (child == 0) {
|
|
||||||
char *cmd_copy = strdup(cmd);
|
|
||||||
char *ptr;
|
|
||||||
int words = 1;
|
|
||||||
for(ptr = strchr(cmd_copy, ' '); ptr; ptr = strchr(ptr+1, ' ')) {
|
|
||||||
words += 1;
|
|
||||||
}
|
|
||||||
char **argv = malloc(sizeof(char *) * (words + 1));
|
|
||||||
ptr = strtok(cmd_copy, " ");
|
|
||||||
int i = 0;
|
|
||||||
argv[i++] = ptr;
|
|
||||||
while (ptr != NULL) {
|
|
||||||
ptr = strtok(NULL, " ");
|
|
||||||
argv[i++] = ptr;
|
|
||||||
}
|
|
||||||
if (execvp(argv[0], argv) != 0) {
|
|
||||||
printf("FAIL: exec failed in child %s - %s\n", cmd, strerror(errno));
|
|
||||||
exit(42);
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) <= 0) {
|
|
||||||
printf("FAIL: failed waiting for child process %s pid %d - %s\n",
|
|
||||||
cmd, child, strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (!WIFEXITED(status)) {
|
|
||||||
printf("FAIL: process %s pid %d did not exit\n", cmd, child);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (WEXITSTATUS(status) != 0) {
|
|
||||||
printf("FAIL: process %s pid %d exited with error status %d\n", cmd,
|
|
||||||
child, WEXITSTATUS(status));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
int write_config_file(char *file_name) {
|
|
||||||
FILE *file;
|
|
||||||
file = fopen(file_name, "w");
|
|
||||||
if (file == NULL) {
|
|
||||||
printf("Failed to open %s.\n", file_name);
|
|
||||||
return EXIT_FAILURE;
|
|
||||||
}
|
|
||||||
fprintf(file, "mapred.local.dir=" TEST_ROOT "/local-1");
|
|
||||||
int i;
|
|
||||||
for(i=2; i < 5; ++i) {
|
|
||||||
fprintf(file, "," TEST_ROOT "/local-%d", i);
|
|
||||||
}
|
|
||||||
fprintf(file, "\n");
|
|
||||||
fprintf(file, "mapreduce.cluster.local.dir=" TEST_ROOT "/local-1");
|
|
||||||
for(i=2; i < 5; ++i) {
|
|
||||||
fprintf(file, "," TEST_ROOT "/local-%d", i);
|
|
||||||
}
|
|
||||||
fprintf(file, "\n");
|
|
||||||
|
|
||||||
fprintf(file, "hadoop.log.dir=" TEST_ROOT "/logs\n");
|
|
||||||
fclose(file);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
void create_tt_roots() {
|
|
||||||
char** tt_roots = get_values("mapred.local.dir");
|
|
||||||
char** tt_root;
|
|
||||||
for(tt_root=tt_roots; *tt_root != NULL; ++tt_root) {
|
|
||||||
if (mkdir(*tt_root, 0755) != 0) {
|
|
||||||
printf("FAIL: Can't create directory %s - %s\n", *tt_root,
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "%s/usercache", *tt_root);
|
|
||||||
if (mkdir(buffer, 0755) != 0) {
|
|
||||||
printf("FAIL: Can't create directory %s - %s\n", buffer,
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
free_values(tt_roots);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_get_user_directory() {
|
|
||||||
char *user_dir = get_user_directory("/tmp", "user");
|
|
||||||
char *expected = "/tmp/usercache/user";
|
|
||||||
if (strcmp(user_dir, expected) != 0) {
|
|
||||||
printf("test_get_user_directory expected %s got %s\n", expected, user_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(user_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_get_job_directory() {
|
|
||||||
char *expected = "/tmp/usercache/user/appcache/job_200906101234_0001";
|
|
||||||
char *job_dir = (char *) get_job_directory("/tmp", "user",
|
|
||||||
"job_200906101234_0001");
|
|
||||||
if (strcmp(job_dir, expected) != 0) {
|
|
||||||
printf("test_get_job_directory expected %s got %s\n", expected, job_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_get_attempt_directory() {
|
|
||||||
char *attempt_dir = get_attempt_work_directory("/tmp", "owen", "job_1",
|
|
||||||
"attempt_1");
|
|
||||||
char *expected = "/tmp/usercache/owen/appcache/job_1/attempt_1";
|
|
||||||
if (strcmp(attempt_dir, expected) != 0) {
|
|
||||||
printf("Fail get_attempt_work_directory got %s expected %s\n",
|
|
||||||
attempt_dir, expected);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(attempt_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_get_task_launcher_file() {
|
|
||||||
char *expected_file = ("/tmp/usercache/user/appcache/job_200906101234_0001"
|
|
||||||
"/task.sh");
|
|
||||||
char *job_dir = get_job_directory("/tmp", "user",
|
|
||||||
"job_200906101234_0001");
|
|
||||||
char *task_file = get_task_launcher_file(job_dir);
|
|
||||||
if (strcmp(task_file, expected_file) != 0) {
|
|
||||||
printf("failure to match expected task file %s vs %s\n", task_file,
|
|
||||||
expected_file);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
free(task_file);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_get_job_log_dir() {
|
|
||||||
char *expected = TEST_ROOT "/logs/userlogs/job_200906101234_0001";
|
|
||||||
char *logdir = get_job_log_directory(TEST_ROOT "/logs/userlogs","job_200906101234_0001");
|
|
||||||
if (strcmp(logdir, expected) != 0) {
|
|
||||||
printf("Fail get_job_log_dir got %s expected %s\n", logdir, expected);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(logdir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_check_user() {
|
|
||||||
printf("\nTesting test_check_user\n");
|
|
||||||
struct passwd *user = check_user(username);
|
|
||||||
if (user == NULL) {
|
|
||||||
printf("FAIL: failed check for user %s\n", username);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(user);
|
|
||||||
if (check_user("lp") != NULL) {
|
|
||||||
printf("FAIL: failed check for system user lp\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (check_user("root") != NULL) {
|
|
||||||
printf("FAIL: failed check for system user root\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (check_user("mapred") != NULL) {
|
|
||||||
printf("FAIL: failed check for hadoop user mapred\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_check_configuration_permissions() {
|
|
||||||
printf("\nTesting check_configuration_permissions\n");
|
|
||||||
if (check_configuration_permissions("/etc/passwd") != 0) {
|
|
||||||
printf("FAIL: failed permission check on /etc/passwd\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (check_configuration_permissions(TEST_ROOT) == 0) {
|
|
||||||
printf("FAIL: failed permission check on %s\n", TEST_ROOT);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_delete_task() {
|
|
||||||
if (initialize_user(username)) {
|
|
||||||
printf("FAIL: failed to initialize user %s\n", username);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char* job_dir = get_job_directory(TEST_ROOT "/local-2", username, "job_1");
|
|
||||||
char* dont_touch = get_job_directory(TEST_ROOT "/local-2", username,
|
|
||||||
DONT_TOUCH_FILE);
|
|
||||||
char* task_dir = get_attempt_work_directory(TEST_ROOT "/local-2",
|
|
||||||
username, "job_1", "task_1");
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "mkdir -p %s/who/let/the/dogs/out/who/who", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
sprintf(buffer, "touch %s", dont_touch);
|
|
||||||
run(buffer);
|
|
||||||
|
|
||||||
// soft link to the canary file from the task directory
|
|
||||||
sprintf(buffer, "ln -s %s %s/who/softlink", dont_touch, task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// hard link to the canary file from the task directory
|
|
||||||
sprintf(buffer, "ln %s %s/who/hardlink", dont_touch, task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a dot file in the task directory
|
|
||||||
sprintf(buffer, "touch %s/who/let/.dotfile", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a no permission file
|
|
||||||
sprintf(buffer, "touch %s/who/let/protect", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
sprintf(buffer, "chmod 000 %s/who/let/protect", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a no permission directory
|
|
||||||
sprintf(buffer, "chmod 000 %s/who/let", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
|
|
||||||
// delete task directory
|
|
||||||
char * dirs[] = {job_dir, 0};
|
|
||||||
int ret = delete_as_user(username, "task_1" , dirs);
|
|
||||||
if (ret != 0) {
|
|
||||||
printf("FAIL: return code from delete_as_user is %d\n", ret);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
// check to make sure the task directory is gone
|
|
||||||
if (access(task_dir, R_OK) == 0) {
|
|
||||||
printf("FAIL: failed to delete the directory - %s\n", task_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
// check to make sure the job directory is not gone
|
|
||||||
if (access(job_dir, R_OK) != 0) {
|
|
||||||
printf("FAIL: accidently deleted the directory - %s\n", job_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
// but that the canary is not gone
|
|
||||||
if (access(dont_touch, R_OK) != 0) {
|
|
||||||
printf("FAIL: accidently deleted file %s\n", dont_touch);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
sprintf(buffer, "chmod -R 700 %s", job_dir);
|
|
||||||
run(buffer);
|
|
||||||
sprintf(buffer, "rm -fr %s", job_dir);
|
|
||||||
run(buffer);
|
|
||||||
free(job_dir);
|
|
||||||
free(task_dir);
|
|
||||||
free(dont_touch);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_delete_job() {
|
|
||||||
char* job_dir = get_job_directory(TEST_ROOT "/local-2", username, "job_2");
|
|
||||||
char* dont_touch = get_job_directory(TEST_ROOT "/local-2", username,
|
|
||||||
DONT_TOUCH_FILE);
|
|
||||||
char* task_dir = get_attempt_work_directory(TEST_ROOT "/local-2",
|
|
||||||
username, "job_2", "task_1");
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "mkdir -p %s/who/let/the/dogs/out/who/who", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
sprintf(buffer, "touch %s", dont_touch);
|
|
||||||
run(buffer);
|
|
||||||
|
|
||||||
// soft link to the canary file from the task directory
|
|
||||||
sprintf(buffer, "ln -s %s %s/who/softlink", dont_touch, task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// hard link to the canary file from the task directory
|
|
||||||
sprintf(buffer, "ln %s %s/who/hardlink", dont_touch, task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a dot file in the task directory
|
|
||||||
sprintf(buffer, "touch %s/who/let/.dotfile", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a no permission file
|
|
||||||
sprintf(buffer, "touch %s/who/let/protect", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
sprintf(buffer, "chmod 000 %s/who/let/protect", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
// create a no permission directory
|
|
||||||
sprintf(buffer, "chmod 000 %s/who/let", task_dir);
|
|
||||||
run(buffer);
|
|
||||||
|
|
||||||
// delete task directory
|
|
||||||
int ret = delete_as_user(username, job_dir, NULL);
|
|
||||||
if (ret != 0) {
|
|
||||||
printf("FAIL: return code from delete_as_user is %d\n", ret);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
// check to make sure the task directory is gone
|
|
||||||
if (access(task_dir, R_OK) == 0) {
|
|
||||||
printf("FAIL: failed to delete the directory - %s\n", task_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
// check to make sure the job directory is gone
|
|
||||||
if (access(job_dir, R_OK) == 0) {
|
|
||||||
printf("FAIL: didn't delete the directory - %s\n", job_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
// but that the canary is not gone
|
|
||||||
if (access(dont_touch, R_OK) != 0) {
|
|
||||||
printf("FAIL: accidently deleted file %s\n", dont_touch);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
free(task_dir);
|
|
||||||
free(dont_touch);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
void test_delete_user() {
|
|
||||||
printf("\nTesting delete_user\n");
|
|
||||||
char* job_dir = get_job_directory(TEST_ROOT "/local-1", username, "job_3");
|
|
||||||
if (mkdirs(job_dir, 0700) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "%s/local-1/usercache/%s", TEST_ROOT, username);
|
|
||||||
if (access(buffer, R_OK) != 0) {
|
|
||||||
printf("FAIL: directory missing before test\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (delete_as_user(username, buffer, NULL) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (access(buffer, R_OK) == 0) {
|
|
||||||
printf("FAIL: directory not deleted\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (access(TEST_ROOT "/local-1", R_OK) != 0) {
|
|
||||||
printf("FAIL: local-1 directory does not exist\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void run_test_in_child(const char* test_name, void (*func)()) {
|
|
||||||
printf("\nRunning test %s in child process\n", test_name);
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: fork failed\n");
|
|
||||||
exit(1);
|
|
||||||
} else if (child == 0) {
|
|
||||||
func();
|
|
||||||
exit(0);
|
|
||||||
} else {
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) == -1) {
|
|
||||||
printf("FAIL: waitpid %d failed - %s\n", child, strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (!WIFEXITED(status)) {
|
|
||||||
printf("FAIL: child %d didn't exit - %d\n", child, status);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (WEXITSTATUS(status) != 0) {
|
|
||||||
printf("FAIL: child %d exited with bad status %d\n",
|
|
||||||
child, WEXITSTATUS(status));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_signal_task() {
|
|
||||||
printf("\nTesting signal_task\n");
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: fork failed\n");
|
|
||||||
exit(1);
|
|
||||||
} else if (child == 0) {
|
|
||||||
if (change_user(user_detail->pw_uid, user_detail->pw_gid) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
sleep(3600);
|
|
||||||
exit(0);
|
|
||||||
} else {
|
|
||||||
printf("Child task launched as %d\n", child);
|
|
||||||
if (signal_user_task(username, child, SIGQUIT) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) == -1) {
|
|
||||||
printf("FAIL: waitpid failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (!WIFSIGNALED(status)) {
|
|
||||||
printf("FAIL: child wasn't signalled - %d\n", status);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (WTERMSIG(status) != SIGQUIT) {
|
|
||||||
printf("FAIL: child was killed with %d instead of %d\n",
|
|
||||||
WTERMSIG(status), SIGQUIT);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_signal_task_group() {
|
|
||||||
printf("\nTesting group signal_task\n");
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: fork failed\n");
|
|
||||||
exit(1);
|
|
||||||
} else if (child == 0) {
|
|
||||||
setpgrp();
|
|
||||||
if (change_user(user_detail->pw_uid, user_detail->pw_gid) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
sleep(3600);
|
|
||||||
exit(0);
|
|
||||||
}
|
|
||||||
printf("Child task launched as %d\n", child);
|
|
||||||
if (signal_user_task(username, child, SIGKILL) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) == -1) {
|
|
||||||
printf("FAIL: waitpid failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (!WIFSIGNALED(status)) {
|
|
||||||
printf("FAIL: child wasn't signalled - %d\n", status);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (WTERMSIG(status) != SIGKILL) {
|
|
||||||
printf("FAIL: child was killed with %d instead of %d\n",
|
|
||||||
WTERMSIG(status), SIGKILL);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_init_job() {
|
|
||||||
printf("\nTesting init job\n");
|
|
||||||
if (seteuid(0) != 0) {
|
|
||||||
printf("FAIL: seteuid to root failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
FILE* creds = fopen(TEST_ROOT "/creds.txt", "w");
|
|
||||||
if (creds == NULL) {
|
|
||||||
printf("FAIL: failed to create credentials file - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fprintf(creds, "secret key\n") < 0) {
|
|
||||||
printf("FAIL: fprintf failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fclose(creds) != 0) {
|
|
||||||
printf("FAIL: fclose failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
FILE* job_xml = fopen(TEST_ROOT "/job.xml", "w");
|
|
||||||
if (job_xml == NULL) {
|
|
||||||
printf("FAIL: failed to create job file - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fprintf(job_xml, "<jobconf/>\n") < 0) {
|
|
||||||
printf("FAIL: fprintf failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fclose(job_xml) != 0) {
|
|
||||||
printf("FAIL: fclose failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (seteuid(user_detail->pw_uid) != 0) {
|
|
||||||
printf("FAIL: failed to seteuid back to user - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: failed to fork process for init_job - %s\n",
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
} else if (child == 0) {
|
|
||||||
char *final_pgm[] = {"touch", "my-touch-file", 0};
|
|
||||||
if (initialize_job(username, "job_4", TEST_ROOT "/creds.txt", final_pgm) != 0) {
|
|
||||||
printf("FAIL: failed in child\n");
|
|
||||||
exit(42);
|
|
||||||
}
|
|
||||||
// should never return
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) <= 0) {
|
|
||||||
printf("FAIL: failed waiting for process %d - %s\n", child,
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (access(TEST_ROOT "/logs/userlogs/job_4", R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create job log directory\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char* job_dir = get_job_directory(TEST_ROOT "/local-1", username, "job_4");
|
|
||||||
if (access(job_dir, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create job directory %s\n", job_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "%s/jobToken", job_dir);
|
|
||||||
if (access(buffer, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create credentials %s\n", buffer);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
sprintf(buffer, "%s/my-touch-file", job_dir);
|
|
||||||
if (access(buffer, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create touch file %s\n", buffer);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
job_dir = get_job_log_directory("logs","job_4");
|
|
||||||
if (access(job_dir, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create job log directory %s\n", job_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(job_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
void test_run_task() {
|
|
||||||
printf("\nTesting run task\n");
|
|
||||||
if (seteuid(0) != 0) {
|
|
||||||
printf("FAIL: seteuid to root failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
FILE* creds = fopen(TEST_ROOT "/creds.txt", "w");
|
|
||||||
if (creds == NULL) {
|
|
||||||
printf("FAIL: failed to create credentials file - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fprintf(creds, "secret key\n") < 0) {
|
|
||||||
printf("FAIL: fprintf failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fclose(creds) != 0) {
|
|
||||||
printf("FAIL: fclose failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
const char* script_name = TEST_ROOT "/task-script";
|
|
||||||
FILE* script = fopen(script_name, "w");
|
|
||||||
if (script == NULL) {
|
|
||||||
printf("FAIL: failed to create script file - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (seteuid(user_detail->pw_uid) != 0) {
|
|
||||||
printf("FAIL: failed to seteuid back to user - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fprintf(script, "#!/bin/bash\n"
|
|
||||||
"touch foobar\n"
|
|
||||||
"exit 0") < 0) {
|
|
||||||
printf("FAIL: fprintf failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (fclose(script) != 0) {
|
|
||||||
printf("FAIL: fclose failed - %s\n", strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
fflush(stdout);
|
|
||||||
fflush(stderr);
|
|
||||||
char* task_dir = get_attempt_work_directory(TEST_ROOT "/local-1",
|
|
||||||
username, "job_4", "task_1");
|
|
||||||
pid_t child = fork();
|
|
||||||
if (child == -1) {
|
|
||||||
printf("FAIL: failed to fork process for init_job - %s\n",
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
} else if (child == 0) {
|
|
||||||
if (run_task_as_user(username, "job_4", "task_1",
|
|
||||||
task_dir, script_name, TEST_ROOT "creds.txt") != 0) {
|
|
||||||
printf("FAIL: failed in child\n");
|
|
||||||
exit(42);
|
|
||||||
}
|
|
||||||
// should never return
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
int status = 0;
|
|
||||||
if (waitpid(child, &status, 0) <= 0) {
|
|
||||||
printf("FAIL: failed waiting for process %d - %s\n", child,
|
|
||||||
strerror(errno));
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (access(TEST_ROOT "/logs/userlogs/job_4/task_1", R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create task log directory\n");
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
if (access(task_dir, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create task directory %s\n", task_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
char buffer[100000];
|
|
||||||
sprintf(buffer, "%s/foobar", task_dir);
|
|
||||||
if (access(buffer, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create touch file %s\n", buffer);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(task_dir);
|
|
||||||
task_dir = get_job_log_directory("logs", "job_4/task_1");
|
|
||||||
if (access(task_dir, R_OK) != 0) {
|
|
||||||
printf("FAIL: failed to create job log directory %s\n", task_dir);
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
free(task_dir);
|
|
||||||
}
|
|
||||||
|
|
||||||
int main(int argc, char **argv) {
|
|
||||||
LOGFILE = stdout;
|
|
||||||
int my_username = 0;
|
|
||||||
|
|
||||||
// clean up any junk from previous run
|
|
||||||
system("chmod -R u=rwx " TEST_ROOT "; rm -fr " TEST_ROOT);
|
|
||||||
|
|
||||||
if (mkdirs(TEST_ROOT "/logs/userlogs", 0755) != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (write_config_file(TEST_ROOT "/test.cfg") != 0) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
read_config(TEST_ROOT "/test.cfg");
|
|
||||||
|
|
||||||
create_tt_roots();
|
|
||||||
|
|
||||||
if (getuid() == 0 && argc == 2) {
|
|
||||||
username = argv[1];
|
|
||||||
} else {
|
|
||||||
username = strdup(getpwuid(getuid())->pw_name);
|
|
||||||
my_username = 1;
|
|
||||||
}
|
|
||||||
set_tasktracker_uid(geteuid(), getegid());
|
|
||||||
|
|
||||||
if (set_user(username)) {
|
|
||||||
exit(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
printf("\nStarting tests\n");
|
|
||||||
|
|
||||||
printf("\nTesting get_user_directory()\n");
|
|
||||||
test_get_user_directory();
|
|
||||||
|
|
||||||
printf("\nTesting get_job_directory()\n");
|
|
||||||
test_get_job_directory();
|
|
||||||
|
|
||||||
printf("\nTesting get_attempt_directory()\n");
|
|
||||||
test_get_attempt_directory();
|
|
||||||
|
|
||||||
printf("\nTesting get_task_launcher_file()\n");
|
|
||||||
test_get_task_launcher_file();
|
|
||||||
|
|
||||||
printf("\nTesting get_job_log_dir()\n");
|
|
||||||
test_get_job_log_dir();
|
|
||||||
|
|
||||||
test_check_configuration_permissions();
|
|
||||||
|
|
||||||
printf("\nTesting delete_task()\n");
|
|
||||||
test_delete_task();
|
|
||||||
|
|
||||||
printf("\nTesting delete_job()\n");
|
|
||||||
test_delete_job();
|
|
||||||
|
|
||||||
test_delete_user();
|
|
||||||
|
|
||||||
test_check_user();
|
|
||||||
|
|
||||||
// the tests that change user need to be run in a subshell, so that
|
|
||||||
// when they change user they don't give up our privs
|
|
||||||
run_test_in_child("test_signal_task", test_signal_task);
|
|
||||||
run_test_in_child("test_signal_task_group", test_signal_task_group);
|
|
||||||
|
|
||||||
// init job and run task can't be run if you aren't testing as root
|
|
||||||
if (getuid() == 0) {
|
|
||||||
// these tests do internal forks so that the change_owner and execs
|
|
||||||
// don't mess up our process.
|
|
||||||
test_init_job();
|
|
||||||
test_run_task();
|
|
||||||
}
|
|
||||||
|
|
||||||
seteuid(0);
|
|
||||||
run("rm -fr " TEST_ROOT);
|
|
||||||
printf("\nFinished tests\n");
|
|
||||||
|
|
||||||
if (my_username) {
|
|
||||||
free(username);
|
|
||||||
}
|
|
||||||
free_configurations();
|
|
||||||
return 0;
|
|
||||||
}
|
|
|
@ -39,9 +39,8 @@
|
||||||
</properties>
|
</properties>
|
||||||
|
|
||||||
<modules>
|
<modules>
|
||||||
<module>hadoop-yarn</module>
|
<module>hadoop-mapreduce-client</module>
|
||||||
<module>hadoop-mapreduce-client</module>
|
<module>hadoop-mapreduce-examples</module>
|
||||||
<module>hadoop-mapreduce-examples</module>
|
|
||||||
</modules>
|
</modules>
|
||||||
|
|
||||||
<dependencies>
|
<dependencies>
|
||||||
|
|
|
@ -0,0 +1,34 @@
|
||||||
|
Hadoop MapReduce Change Log
|
||||||
|
|
||||||
|
Trunk (unreleased changes)
|
||||||
|
|
||||||
|
INCOMPATIBLE CHANGES
|
||||||
|
|
||||||
|
NEW FEATURES
|
||||||
|
|
||||||
|
IMPROVEMENTS
|
||||||
|
|
||||||
|
BUG FIXES
|
||||||
|
|
||||||
|
Branch-2 ( Unreleased changes )
|
||||||
|
|
||||||
|
INCOMPATIBLE CHANGES
|
||||||
|
|
||||||
|
NEW FEATURES
|
||||||
|
|
||||||
|
IMPROVEMENTS
|
||||||
|
|
||||||
|
BUG FIXES
|
||||||
|
|
||||||
|
Release 2.1.0-alpha - Unreleased
|
||||||
|
|
||||||
|
INCOMPATIBLE CHANGES
|
||||||
|
|
||||||
|
NEW FEATURES
|
||||||
|
|
||||||
|
YARN-1. Promote YARN to be a sub-project of Apache Hadoop. (acmurthy)
|
||||||
|
|
||||||
|
IMPROVEMENTS
|
||||||
|
|
||||||
|
BUG FIXES
|
||||||
|
|
|
@ -0,0 +1,244 @@
|
||||||
|
|
||||||
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
|
|
||||||
|
1. Definitions.
|
||||||
|
|
||||||
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
|
the copyright owner that is granting the License.
|
||||||
|
|
||||||
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
|
other entities that control, are controlled by, or are under common
|
||||||
|
control with that entity. For the purposes of this definition,
|
||||||
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
|
exercising permissions granted by this License.
|
||||||
|
|
||||||
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright [yyyy] [name of copyright owner]
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
|
||||||
|
|
||||||
|
APACHE HADOOP SUBCOMPONENTS:
|
||||||
|
|
||||||
|
The Apache Hadoop project contains subcomponents with separate copyright
|
||||||
|
notices and license terms. Your use of the source code for the these
|
||||||
|
subcomponents is subject to the terms and conditions of the following
|
||||||
|
licenses.
|
||||||
|
|
||||||
|
For the org.apache.hadoop.util.bloom.* classes:
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* Copyright (c) 2005, European Commission project OneLab under contract
|
||||||
|
* 034819 (http://www.one-lab.org)
|
||||||
|
* All rights reserved.
|
||||||
|
* Redistribution and use in source and binary forms, with or
|
||||||
|
* without modification, are permitted provided that the following
|
||||||
|
* conditions are met:
|
||||||
|
* - Redistributions of source code must retain the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer.
|
||||||
|
* - Redistributions in binary form must reproduce the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer in
|
||||||
|
* the documentation and/or other materials provided with the distribution.
|
||||||
|
* - Neither the name of the University Catholique de Louvain - UCL
|
||||||
|
* nor the names of its contributors may be used to endorse or
|
||||||
|
* promote products derived from this software without specific prior
|
||||||
|
* written permission.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
|
||||||
|
* FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
|
||||||
|
* COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||||
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
|
||||||
|
* BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
||||||
|
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
||||||
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
|
||||||
|
* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||||||
|
* POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
|
@ -0,0 +1,2 @@
|
||||||
|
This product includes software developed by The Apache Software
|
||||||
|
Foundation (http://www.apache.org/).
|
|
@ -19,12 +19,12 @@
|
||||||
<parent>
|
<parent>
|
||||||
<artifactId>hadoop-yarn</artifactId>
|
<artifactId>hadoop-yarn</artifactId>
|
||||||
<groupId>org.apache.hadoop</groupId>
|
<groupId>org.apache.hadoop</groupId>
|
||||||
<version>2.0.1-SNAPSHOT</version>
|
<version>3.0.0-SNAPSHOT</version>
|
||||||
</parent>
|
</parent>
|
||||||
<modelVersion>4.0.0</modelVersion>
|
<modelVersion>4.0.0</modelVersion>
|
||||||
<groupId>org.apache.hadoop</groupId>
|
<groupId>org.apache.hadoop</groupId>
|
||||||
<artifactId>hadoop-yarn-api</artifactId>
|
<artifactId>hadoop-yarn-api</artifactId>
|
||||||
<version>2.0.1-SNAPSHOT</version>
|
<version>3.0.0-SNAPSHOT</version>
|
||||||
<name>hadoop-yarn-api</name>
|
<name>hadoop-yarn-api</name>
|
||||||
|
|
||||||
<properties>
|
<properties>
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue