MAPREDUCE-2948. Hadoop streaming test failure, post MR-2767 (mahadev) - merging r1166402 from trunk
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1166405 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
210eb9ee47
commit
f049f33e5a
|
@ -1213,6 +1213,8 @@ Release 0.23.0 - Unreleased
|
||||||
MAPREDUCE-2938. Log application submission failure in CapacityScheduler.
|
MAPREDUCE-2938. Log application submission failure in CapacityScheduler.
|
||||||
(acmurthy)
|
(acmurthy)
|
||||||
|
|
||||||
|
MAPREDUCE-2948. Hadoop streaming test failure, post MR-2767 (mahadev)
|
||||||
|
|
||||||
Release 0.22.0 - Unreleased
|
Release 0.22.0 - Unreleased
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -1,178 +0,0 @@
|
||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.apache.hadoop.streaming;
|
|
||||||
|
|
||||||
import java.io.DataOutputStream;
|
|
||||||
import java.io.File;
|
|
||||||
import java.io.FileOutputStream;
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.security.PrivilegedExceptionAction;
|
|
||||||
|
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
|
||||||
import org.apache.hadoop.fs.Path;
|
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
|
||||||
import org.apache.hadoop.mapred.ClusterWithLinuxTaskController;
|
|
||||||
import org.apache.hadoop.mapred.JobConf;
|
|
||||||
import org.apache.hadoop.security.Groups;
|
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
|
||||||
import org.apache.hadoop.util.Shell;
|
|
||||||
import org.apache.hadoop.util.ToolRunner;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Test Streaming with LinuxTaskController running the jobs as a user different
|
|
||||||
* from the user running the cluster. See {@link ClusterWithLinuxTaskController}
|
|
||||||
*/
|
|
||||||
public class TestStreamingAsDifferentUser extends
|
|
||||||
ClusterWithLinuxTaskController {
|
|
||||||
|
|
||||||
private Path inputPath = new Path("input");
|
|
||||||
private Path outputPath = new Path("output");
|
|
||||||
private String input = "roses.are.red\nviolets.are.blue\nbunnies.are.pink\n";
|
|
||||||
private String map =
|
|
||||||
UtilTest.makeJavaCommand(TrApp.class, new String[] { ".", "\\n" });
|
|
||||||
private String reduce =
|
|
||||||
UtilTest.makeJavaCommand(UniqApp.class, new String[] { "R" });
|
|
||||||
|
|
||||||
public void testStreaming()
|
|
||||||
throws Exception {
|
|
||||||
if (!shouldRun()) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
startCluster();
|
|
||||||
final JobConf myConf = getClusterConf();
|
|
||||||
jobOwner.doAs(new PrivilegedExceptionAction<Void>() {
|
|
||||||
public Void run() throws IOException{
|
|
||||||
|
|
||||||
FileSystem inFs = inputPath.getFileSystem(myConf);
|
|
||||||
FileSystem outFs = outputPath.getFileSystem(myConf);
|
|
||||||
outFs.delete(outputPath, true);
|
|
||||||
if (!inFs.mkdirs(inputPath)) {
|
|
||||||
throw new IOException("Mkdirs failed to create " + inFs.toString());
|
|
||||||
}
|
|
||||||
DataOutputStream file = inFs.create(new Path(inputPath, "part-0"));
|
|
||||||
file.writeBytes(input);
|
|
||||||
file.close();
|
|
||||||
final String[] args =
|
|
||||||
new String[] { "-input", inputPath.makeQualified(inFs).toString(),
|
|
||||||
"-output", outputPath.makeQualified(outFs).toString(), "-mapper",
|
|
||||||
map, "-reducer", reduce, "-jobconf",
|
|
||||||
"mapreduce.task.files.preserve.failedtasks=true", "-jobconf",
|
|
||||||
"stream.tmpdir=" + System.getProperty("test.build.data", "/tmp") };
|
|
||||||
|
|
||||||
StreamJob streamJob = new StreamJob(args, true);
|
|
||||||
streamJob.setConf(myConf);
|
|
||||||
assertTrue("Job has not succeeded", streamJob.go() == 0);
|
|
||||||
assertOwnerShip(outputPath);
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Verify if the permissions of distcache dir contents are valid once the job
|
|
||||||
* is finished
|
|
||||||
*/
|
|
||||||
public void testStreamingWithDistCache()
|
|
||||||
throws Exception {
|
|
||||||
if (!shouldRun()) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
startCluster();
|
|
||||||
final String[] localDirs = mrCluster.getTaskTrackerLocalDirs(0);
|
|
||||||
final JobConf myConf = getClusterConf();
|
|
||||||
|
|
||||||
// create file that will go into public distributed cache
|
|
||||||
File publicFile = new File(System.getProperty(
|
|
||||||
"test.build.data", "/tmp"), "publicFile");
|
|
||||||
FileOutputStream fstream = new FileOutputStream(publicFile);
|
|
||||||
fstream.write("public file contents".getBytes());
|
|
||||||
fstream.close();
|
|
||||||
|
|
||||||
// put the file(that should go into public dist cache) in dfs and set
|
|
||||||
// read and exe permissions for others
|
|
||||||
FileSystem dfs = dfsCluster.getFileSystem();
|
|
||||||
dfs.setPermission(new Path(dfs.getDefaultUri(myConf).toString() + "/tmp"),
|
|
||||||
new FsPermission((short)0755));
|
|
||||||
final String publicCacheFile = dfs.getDefaultUri(myConf).toString()
|
|
||||||
+ "/tmp/publicFile";
|
|
||||||
dfs.copyFromLocalFile(new Path(publicFile.getAbsolutePath()),
|
|
||||||
new Path(publicCacheFile));
|
|
||||||
dfs.setPermission(new Path(publicCacheFile), new FsPermission((short)0755));
|
|
||||||
final String taskTrackerUser
|
|
||||||
= UserGroupInformation.getCurrentUser().getShortUserName();
|
|
||||||
|
|
||||||
jobOwner.doAs(new PrivilegedExceptionAction<Void>() {
|
|
||||||
public Void run() throws Exception{
|
|
||||||
|
|
||||||
FileSystem inFs = inputPath.getFileSystem(myConf);
|
|
||||||
FileSystem outFs = outputPath.getFileSystem(myConf);
|
|
||||||
outFs.delete(outputPath, true);
|
|
||||||
if (!inFs.mkdirs(inputPath)) {
|
|
||||||
throw new IOException("Mkdirs failed to create " + inFs.toString());
|
|
||||||
}
|
|
||||||
|
|
||||||
// create input file
|
|
||||||
DataOutputStream file = inFs.create(new Path(inputPath, "part-0"));
|
|
||||||
file.writeBytes(input);
|
|
||||||
file.close();
|
|
||||||
|
|
||||||
// Create file that will be passed using -files option.
|
|
||||||
// This is private dist cache file
|
|
||||||
File privateFile = new File(System.getProperty(
|
|
||||||
"test.build.data", "/tmp"), "test.sh");
|
|
||||||
privateFile.createNewFile();
|
|
||||||
|
|
||||||
String[] args =
|
|
||||||
new String[] {
|
|
||||||
"-files", privateFile.toString() + "," + publicCacheFile,
|
|
||||||
"-Dmapreduce.task.files.preserve.failedtasks=true",
|
|
||||||
"-Dstream.tmpdir=" + System.getProperty("test.build.data", "/tmp"),
|
|
||||||
"-input", inputPath.makeQualified(inFs).toString(),
|
|
||||||
"-output", outputPath.makeQualified(outFs).toString(),
|
|
||||||
"-mapper", "pwd",
|
|
||||||
"-reducer", StreamJob.REDUCE_NONE
|
|
||||||
};
|
|
||||||
StreamJob streamJob = new StreamJob();
|
|
||||||
streamJob.setConf(myConf);
|
|
||||||
|
|
||||||
assertTrue("Job failed", ToolRunner.run(streamJob, args)==0);
|
|
||||||
|
|
||||||
// validate private cache files' permissions
|
|
||||||
checkPermissionsOnPrivateDistCache(localDirs,
|
|
||||||
jobOwner.getShortUserName(), taskTrackerUser,
|
|
||||||
taskTrackerSpecialGroup);
|
|
||||||
|
|
||||||
// check the file is present even after the job is over.
|
|
||||||
// work directory symlink cleanup should not have removed the target
|
|
||||||
// files.
|
|
||||||
checkPresenceOfPrivateDistCacheFiles(localDirs,
|
|
||||||
jobOwner.getShortUserName(), new String[] {"test.sh"});
|
|
||||||
|
|
||||||
// validate private cache files' permissions
|
|
||||||
checkPermissionsOnPublicDistCache(FileSystem.getLocal(myConf),
|
|
||||||
localDirs, taskTrackerUser, taskTrackerPrimaryGroup);
|
|
||||||
|
|
||||||
checkPresenceOfPublicDistCacheFiles(localDirs,
|
|
||||||
new String[] {"publicFile"});
|
|
||||||
assertOwnerShip(outputPath);
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
Loading…
Reference in New Issue