YARN-3443. Create a 'ResourceHandler' subsystem to ease addition of support for new resource types on the NM. Contributed by Sidharta Seethana.
This commit is contained in:
parent
a77d628339
commit
838b06ac87
|
@ -69,6 +69,9 @@ Release 2.8.0 - UNRELEASED
|
||||||
YARN-3347. Improve YARN log command to get AMContainer logs as well as
|
YARN-3347. Improve YARN log command to get AMContainer logs as well as
|
||||||
running containers logs. (Xuan Gong via junping_du)
|
running containers logs. (Xuan Gong via junping_du)
|
||||||
|
|
||||||
|
YARN-3443. Create a 'ResourceHandler' subsystem to ease addition of support
|
||||||
|
for new resource types on the NM. (Sidharta Seethana via junping_du)
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
|
|
||||||
YARN-1880. Cleanup TestApplicationClientProtocolOnHA
|
YARN-1880. Cleanup TestApplicationClientProtocolOnHA
|
||||||
|
|
|
@ -0,0 +1,119 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Collections;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Represents operations that require higher system privileges - e.g
|
||||||
|
* creating cgroups, launching containers as specified users, 'tc' commands etc
|
||||||
|
* that are completed using the container-executor binary
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class PrivilegedOperation {
|
||||||
|
|
||||||
|
public enum OperationType {
|
||||||
|
CHECK_SETUP("--checksetup"),
|
||||||
|
MOUNT_CGROUPS("--mount-cgroups"),
|
||||||
|
INITIALIZE_CONTAINER(""), //no CLI switch supported yet
|
||||||
|
LAUNCH_CONTAINER(""), //no CLI switch supported yet
|
||||||
|
SIGNAL_CONTAINER(""), //no CLI switch supported yet
|
||||||
|
DELETE_AS_USER(""), //no CLI switch supported yet
|
||||||
|
TC_MODIFY_STATE("--tc-modify-state"),
|
||||||
|
TC_READ_STATE("--tc-read-state"),
|
||||||
|
TC_READ_STATS("--tc-read-stats"),
|
||||||
|
ADD_PID_TO_CGROUP(""); //no CLI switch supported yet.
|
||||||
|
|
||||||
|
private final String option;
|
||||||
|
|
||||||
|
OperationType(String option) {
|
||||||
|
this.option = option;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getOption() {
|
||||||
|
return option;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public static final String CGROUP_ARG_PREFIX = "cgroups=";
|
||||||
|
|
||||||
|
private final OperationType opType;
|
||||||
|
private final List<String> args;
|
||||||
|
|
||||||
|
public PrivilegedOperation(OperationType opType, String arg) {
|
||||||
|
this.opType = opType;
|
||||||
|
this.args = new ArrayList<String>();
|
||||||
|
|
||||||
|
if (arg != null) {
|
||||||
|
this.args.add(arg);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public PrivilegedOperation(OperationType opType, List<String> args) {
|
||||||
|
this.opType = opType;
|
||||||
|
this.args = new ArrayList<String>();
|
||||||
|
|
||||||
|
if (args != null) {
|
||||||
|
this.args.addAll(args);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void appendArgs(String... args) {
|
||||||
|
for (String arg : args) {
|
||||||
|
this.args.add(arg);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void appendArgs(List<String> args) {
|
||||||
|
this.args.addAll(args);
|
||||||
|
}
|
||||||
|
|
||||||
|
public OperationType getOperationType() {
|
||||||
|
return opType;
|
||||||
|
}
|
||||||
|
|
||||||
|
public List<String> getArguments() {
|
||||||
|
return Collections.unmodifiableList(this.args);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean equals(Object other) {
|
||||||
|
if (other == null || !(other instanceof PrivilegedOperation)) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
PrivilegedOperation otherOp = (PrivilegedOperation) other;
|
||||||
|
|
||||||
|
return otherOp.opType.equals(opType) && otherOp.args.equals(args);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int hashCode() {
|
||||||
|
return opType.hashCode() + 97 * args.hashCode();
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,43 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged;
|
||||||
|
|
||||||
|
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||||
|
|
||||||
|
public class PrivilegedOperationException extends YarnException {
|
||||||
|
private static final long serialVersionUID = 1L;
|
||||||
|
|
||||||
|
public PrivilegedOperationException() {
|
||||||
|
super();
|
||||||
|
}
|
||||||
|
|
||||||
|
public PrivilegedOperationException(String message) {
|
||||||
|
super(message);
|
||||||
|
}
|
||||||
|
|
||||||
|
public PrivilegedOperationException(Throwable cause) {
|
||||||
|
super(cause);
|
||||||
|
}
|
||||||
|
|
||||||
|
public PrivilegedOperationException(String message, Throwable cause) {
|
||||||
|
super(message, cause);
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,255 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged;
|
||||||
|
|
||||||
|
import org.apache.commons.lang.StringUtils;
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.util.Shell.ShellCommandExecutor;
|
||||||
|
import org.apache.hadoop.util.Shell.ExitCodeException;
|
||||||
|
import org.apache.hadoop.yarn.api.ApplicationConstants;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* provides mechanisms to execute PrivilegedContainerOperations *
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class PrivilegedOperationExecutor {
|
||||||
|
private static final Log LOG = LogFactory.getLog(PrivilegedOperationExecutor
|
||||||
|
.class);
|
||||||
|
private volatile static PrivilegedOperationExecutor instance;
|
||||||
|
|
||||||
|
private String containerExecutorExe;
|
||||||
|
|
||||||
|
public static String getContainerExecutorExecutablePath(Configuration conf) {
|
||||||
|
String yarnHomeEnvVar =
|
||||||
|
System.getenv(ApplicationConstants.Environment.HADOOP_YARN_HOME.key());
|
||||||
|
File hadoopBin = new File(yarnHomeEnvVar, "bin");
|
||||||
|
String defaultPath =
|
||||||
|
new File(hadoopBin, "container-executor").getAbsolutePath();
|
||||||
|
return null == conf
|
||||||
|
? defaultPath
|
||||||
|
: conf.get(YarnConfiguration.NM_LINUX_CONTAINER_EXECUTOR_PATH,
|
||||||
|
defaultPath);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void init(Configuration conf) {
|
||||||
|
containerExecutorExe = getContainerExecutorExecutablePath(conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
private PrivilegedOperationExecutor(Configuration conf) {
|
||||||
|
init(conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
public static PrivilegedOperationExecutor getInstance(Configuration conf) {
|
||||||
|
if (instance == null) {
|
||||||
|
synchronized (PrivilegedOperationExecutor.class) {
|
||||||
|
if (instance == null) {
|
||||||
|
instance = new PrivilegedOperationExecutor(conf);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return instance;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param prefixCommands in some cases ( e.g priorities using nice ),
|
||||||
|
* prefix commands are necessary
|
||||||
|
* @param operation the type and arguments for the operation to be
|
||||||
|
* executed
|
||||||
|
* @return execution string array for priviledged operation
|
||||||
|
*/
|
||||||
|
|
||||||
|
public String[] getPrivilegedOperationExecutionCommand(List<String>
|
||||||
|
prefixCommands,
|
||||||
|
PrivilegedOperation operation) {
|
||||||
|
List<String> fullCommand = new ArrayList<String>();
|
||||||
|
|
||||||
|
if (prefixCommands != null && !prefixCommands.isEmpty()) {
|
||||||
|
fullCommand.addAll(prefixCommands);
|
||||||
|
}
|
||||||
|
|
||||||
|
fullCommand.add(containerExecutorExe);
|
||||||
|
fullCommand.add(operation.getOperationType().getOption());
|
||||||
|
fullCommand.addAll(operation.getArguments());
|
||||||
|
|
||||||
|
String[] fullCommandArray =
|
||||||
|
fullCommand.toArray(new String[fullCommand.size()]);
|
||||||
|
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Privileged Execution Command Array: " +
|
||||||
|
Arrays.toString(fullCommandArray));
|
||||||
|
}
|
||||||
|
|
||||||
|
return fullCommandArray;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Executes a privileged operation. It is up to the callers to ensure that
|
||||||
|
* each privileged operation's parameters are constructed correctly. The
|
||||||
|
* parameters are passed verbatim to the container-executor binary.
|
||||||
|
*
|
||||||
|
* @param prefixCommands in some cases ( e.g priorities using nice ),
|
||||||
|
* prefix commands are necessary
|
||||||
|
* @param operation the type and arguments for the operation to be executed
|
||||||
|
* @param workingDir (optional) working directory for execution
|
||||||
|
* @param env (optional) env of the command will include specified vars
|
||||||
|
* @param grabOutput return (possibly large) shell command output
|
||||||
|
* @return stdout contents from shell executor - useful for some privileged
|
||||||
|
* operations - e.g --tc_read
|
||||||
|
* @throws org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException
|
||||||
|
*/
|
||||||
|
public String executePrivilegedOperation(List<String> prefixCommands,
|
||||||
|
PrivilegedOperation operation, File workingDir,
|
||||||
|
Map<String, String> env, boolean grabOutput)
|
||||||
|
throws PrivilegedOperationException {
|
||||||
|
String[] fullCommandArray = getPrivilegedOperationExecutionCommand
|
||||||
|
(prefixCommands, operation);
|
||||||
|
ShellCommandExecutor exec = new ShellCommandExecutor(fullCommandArray,
|
||||||
|
workingDir, env);
|
||||||
|
|
||||||
|
try {
|
||||||
|
exec.execute();
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Privileged Execution Operation Output:");
|
||||||
|
LOG.debug(exec.getOutput());
|
||||||
|
}
|
||||||
|
} catch (ExitCodeException e) {
|
||||||
|
String logLine = new StringBuffer("Shell execution returned exit code: ")
|
||||||
|
.append(exec.getExitCode())
|
||||||
|
.append(". Privileged Execution Operation Output: ")
|
||||||
|
.append(System.lineSeparator()).append(exec.getOutput()).toString();
|
||||||
|
|
||||||
|
LOG.warn(logLine);
|
||||||
|
throw new PrivilegedOperationException(e);
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.warn("IOException executing command: ", e);
|
||||||
|
throw new PrivilegedOperationException(e);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (grabOutput) {
|
||||||
|
return exec.getOutput();
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Executes a privileged operation. It is up to the callers to ensure that
|
||||||
|
* each privileged operation's parameters are constructed correctly. The
|
||||||
|
* parameters are passed verbatim to the container-executor binary.
|
||||||
|
*
|
||||||
|
* @param operation the type and arguments for the operation to be executed
|
||||||
|
* @param grabOutput return (possibly large) shell command output
|
||||||
|
* @return stdout contents from shell executor - useful for some privileged
|
||||||
|
* operations - e.g --tc_read
|
||||||
|
* @throws org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException
|
||||||
|
*/
|
||||||
|
public String executePrivilegedOperation(PrivilegedOperation operation,
|
||||||
|
boolean grabOutput) throws PrivilegedOperationException {
|
||||||
|
return executePrivilegedOperation(null, operation, null, null, grabOutput);
|
||||||
|
}
|
||||||
|
|
||||||
|
//Utility functions for squashing together operations in supported ways
|
||||||
|
//At some point, we need to create a generalized mechanism that uses a set
|
||||||
|
//of squashing 'rules' to squash an set of PrivilegedOperations of varying
|
||||||
|
//types - e.g Launch Container + Add Pid to CGroup(s) + TC rules
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Squash operations for cgroups - e.g mount, add pid to cgroup etc .,
|
||||||
|
* For now, we only implement squashing for 'add pid to cgroup' since this
|
||||||
|
* is the only optimization relevant to launching containers
|
||||||
|
*
|
||||||
|
* @return single squashed cgroup operation. Null on failure.
|
||||||
|
*/
|
||||||
|
|
||||||
|
public static PrivilegedOperation squashCGroupOperations
|
||||||
|
(List<PrivilegedOperation> ops) throws PrivilegedOperationException {
|
||||||
|
if (ops.size() == 0) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
StringBuffer finalOpArg = new StringBuffer(PrivilegedOperation
|
||||||
|
.CGROUP_ARG_PREFIX);
|
||||||
|
boolean noneArgsOnly = true;
|
||||||
|
|
||||||
|
for (PrivilegedOperation op : ops) {
|
||||||
|
if (!op.getOperationType()
|
||||||
|
.equals(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP)) {
|
||||||
|
LOG.warn("Unsupported operation type: " + op.getOperationType());
|
||||||
|
throw new PrivilegedOperationException("Unsupported operation type:"
|
||||||
|
+ op.getOperationType());
|
||||||
|
}
|
||||||
|
|
||||||
|
List<String> args = op.getArguments();
|
||||||
|
if (args.size() != 1) {
|
||||||
|
LOG.warn("Invalid number of args: " + args.size());
|
||||||
|
throw new PrivilegedOperationException("Invalid number of args: "
|
||||||
|
+ args.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
String arg = args.get(0);
|
||||||
|
String tasksFile = StringUtils.substringAfter(arg,
|
||||||
|
PrivilegedOperation.CGROUP_ARG_PREFIX);
|
||||||
|
if (tasksFile == null || tasksFile.isEmpty()) {
|
||||||
|
LOG.warn("Invalid argument: " + arg);
|
||||||
|
throw new PrivilegedOperationException("Invalid argument: " + arg);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (tasksFile.equals("none")) {
|
||||||
|
//Don't append to finalOpArg
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (noneArgsOnly == false) {
|
||||||
|
//We have already appended at least one tasks file.
|
||||||
|
finalOpArg.append(",");
|
||||||
|
finalOpArg.append(tasksFile);
|
||||||
|
} else {
|
||||||
|
finalOpArg.append(tasksFile);
|
||||||
|
noneArgsOnly = false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (noneArgsOnly) {
|
||||||
|
finalOpArg.append("none"); //there were no tasks file to append
|
||||||
|
}
|
||||||
|
|
||||||
|
PrivilegedOperation finalOp = new PrivilegedOperation(
|
||||||
|
PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP, finalOpArg
|
||||||
|
.toString());
|
||||||
|
|
||||||
|
return finalOp;
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,132 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Provides CGroups functionality. Implementations are expected to be
|
||||||
|
* thread-safe
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public interface CGroupsHandler {
|
||||||
|
public enum CGroupController {
|
||||||
|
CPU("cpu"),
|
||||||
|
NET_CLS("net_cls");
|
||||||
|
|
||||||
|
private final String name;
|
||||||
|
|
||||||
|
CGroupController(String name) {
|
||||||
|
this.name = name;
|
||||||
|
}
|
||||||
|
|
||||||
|
String getName() {
|
||||||
|
return name;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public static final String CGROUP_FILE_TASKS = "tasks";
|
||||||
|
public static final String CGROUP_PARAM_CLASSID = "classid";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Mounts a cgroup controller
|
||||||
|
* @param controller - the controller being mounted
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
public void mountCGroupController(CGroupController controller)
|
||||||
|
throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a cgroup for a given controller
|
||||||
|
* @param controller - controller type for which the cgroup is being created
|
||||||
|
* @param cGroupId - id of the cgroup being created
|
||||||
|
* @return full path to created cgroup
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
public String createCGroup(CGroupController controller, String cGroupId)
|
||||||
|
throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Deletes the specified cgroup
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup being deleted
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
public void deleteCGroup(CGroupController controller, String cGroupId) throws
|
||||||
|
ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets the full path for the cgroup, given a controller and a cgroup id
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup
|
||||||
|
* @return full path for the cgroup
|
||||||
|
*/
|
||||||
|
public String getPathForCGroup(CGroupController controller, String
|
||||||
|
cGroupId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets the full path for the cgroup's tasks file, given a controller and a
|
||||||
|
* cgroup id
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup
|
||||||
|
* @return full path for the cgroup's tasks file
|
||||||
|
*/
|
||||||
|
public String getPathForCGroupTasks(CGroupController controller, String
|
||||||
|
cGroupId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets the full path for a cgroup parameter, given a controller,
|
||||||
|
* cgroup id and parameter name
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup
|
||||||
|
* @param param - cgroup parameter ( e.g classid )
|
||||||
|
* @return full path for the cgroup parameter
|
||||||
|
*/
|
||||||
|
public String getPathForCGroupParam(CGroupController controller, String
|
||||||
|
cGroupId, String param);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* updates a cgroup parameter, given a controller, cgroup id, parameter name
|
||||||
|
* and a parameter value
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup
|
||||||
|
* @param param - cgroup parameter ( e.g classid )
|
||||||
|
* @param value - value to be written to the parameter file
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
public void updateCGroupParam(CGroupController controller, String cGroupId,
|
||||||
|
String param, String value) throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* reads a cgroup parameter value, given a controller, cgroup id, parameter
|
||||||
|
* name
|
||||||
|
* @param controller - controller type for the cgroup
|
||||||
|
* @param cGroupId - id of the cgroup
|
||||||
|
* @param param - cgroup parameter ( e.g classid )
|
||||||
|
* @return parameter value as read from the parameter file
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
public String getCGroupParam(CGroupController controller, String cGroupId,
|
||||||
|
String param) throws ResourceHandlerException;
|
||||||
|
}
|
|
@ -0,0 +1,436 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
|
import org.apache.hadoop.io.IOUtils;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperation;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationExecutor;
|
||||||
|
import org.apache.hadoop.yarn.util.Clock;
|
||||||
|
import org.apache.hadoop.yarn.util.SystemClock;
|
||||||
|
|
||||||
|
import java.io.*;
|
||||||
|
import java.nio.file.Files;
|
||||||
|
import java.nio.file.Paths;
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.concurrent.locks.ReadWriteLock;
|
||||||
|
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
||||||
|
import java.util.regex.Matcher;
|
||||||
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Support for interacting with various CGroup subsystems. Thread-safe.
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
class CGroupsHandlerImpl implements CGroupsHandler {
|
||||||
|
|
||||||
|
private static final Log LOG = LogFactory.getLog(CGroupsHandlerImpl.class);
|
||||||
|
private static final String MTAB_FILE = "/proc/mounts";
|
||||||
|
private static final String CGROUPS_FSTYPE = "cgroup";
|
||||||
|
|
||||||
|
private final String cGroupPrefix;
|
||||||
|
private final boolean enableCGroupMount;
|
||||||
|
private final String cGroupMountPath;
|
||||||
|
private final long deleteCGroupTimeout;
|
||||||
|
private final long deleteCGroupDelay;
|
||||||
|
private final Map<CGroupController, String> controllerPaths;
|
||||||
|
private final ReadWriteLock rwLock;
|
||||||
|
private final PrivilegedOperationExecutor privilegedOperationExecutor;
|
||||||
|
private final Clock clock;
|
||||||
|
|
||||||
|
public CGroupsHandlerImpl(Configuration conf, PrivilegedOperationExecutor
|
||||||
|
privilegedOperationExecutor) throws ResourceHandlerException {
|
||||||
|
this.cGroupPrefix = conf.get(YarnConfiguration.
|
||||||
|
NM_LINUX_CONTAINER_CGROUPS_HIERARCHY, "/hadoop-yarn")
|
||||||
|
.replaceAll("^/", "").replaceAll("$/", "");
|
||||||
|
this.enableCGroupMount = conf.getBoolean(YarnConfiguration.
|
||||||
|
NM_LINUX_CONTAINER_CGROUPS_MOUNT, false);
|
||||||
|
this.cGroupMountPath = conf.get(YarnConfiguration.
|
||||||
|
NM_LINUX_CONTAINER_CGROUPS_MOUNT_PATH, null);
|
||||||
|
this.deleteCGroupTimeout = conf.getLong(
|
||||||
|
YarnConfiguration.NM_LINUX_CONTAINER_CGROUPS_DELETE_TIMEOUT,
|
||||||
|
YarnConfiguration.DEFAULT_NM_LINUX_CONTAINER_CGROUPS_DELETE_TIMEOUT);
|
||||||
|
this.deleteCGroupDelay =
|
||||||
|
conf.getLong(YarnConfiguration.NM_LINUX_CONTAINER_CGROUPS_DELETE_DELAY,
|
||||||
|
YarnConfiguration.DEFAULT_NM_LINUX_CONTAINER_CGROUPS_DELETE_DELAY);
|
||||||
|
this.controllerPaths = new HashMap<>();
|
||||||
|
this.rwLock = new ReentrantReadWriteLock();
|
||||||
|
this.privilegedOperationExecutor = privilegedOperationExecutor;
|
||||||
|
this.clock = new SystemClock();
|
||||||
|
|
||||||
|
init();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void init() throws ResourceHandlerException {
|
||||||
|
initializeControllerPaths();
|
||||||
|
}
|
||||||
|
|
||||||
|
private String getControllerPath(CGroupController controller) {
|
||||||
|
try {
|
||||||
|
rwLock.readLock().lock();
|
||||||
|
return controllerPaths.get(controller);
|
||||||
|
} finally {
|
||||||
|
rwLock.readLock().unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void initializeControllerPaths() throws ResourceHandlerException {
|
||||||
|
if (enableCGroupMount) {
|
||||||
|
//nothing to do here - we support 'deferred' mounting of specific
|
||||||
|
//controllers - we'll populate the path for a given controller when an
|
||||||
|
//explicit mountCGroupController request is issued.
|
||||||
|
LOG.info("CGroup controller mounting enabled.");
|
||||||
|
} else {
|
||||||
|
//cluster admins are expected to have mounted controllers in specific
|
||||||
|
//locations - we'll attempt to figure out mount points
|
||||||
|
initializeControllerPathsFromMtab();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void initializeControllerPathsFromMtab()
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
try {
|
||||||
|
Map<String, List<String>> parsedMtab = parseMtab();
|
||||||
|
|
||||||
|
//we want to do a bulk update without the paths changing concurrently
|
||||||
|
rwLock.writeLock().lock();
|
||||||
|
|
||||||
|
for (CGroupController controller : CGroupController.values()) {
|
||||||
|
String name = controller.getName();
|
||||||
|
String controllerPath = findControllerInMtab(name, parsedMtab);
|
||||||
|
|
||||||
|
if (controllerPath != null) {
|
||||||
|
File f = new File(controllerPath + "/" + this.cGroupPrefix);
|
||||||
|
|
||||||
|
if (FileUtil.canWrite(f)) {
|
||||||
|
controllerPaths.put(controller, controllerPath);
|
||||||
|
} else {
|
||||||
|
String error =
|
||||||
|
new StringBuffer("Mount point Based on mtab file: ")
|
||||||
|
.append(MTAB_FILE).append(
|
||||||
|
". Controller mount point not writable for: ")
|
||||||
|
.append(name).toString();
|
||||||
|
|
||||||
|
LOG.error(error);
|
||||||
|
throw new ResourceHandlerException(error);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
|
||||||
|
LOG.warn("Controller not mounted but automount disabled: " + name);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.warn("Failed to initialize controller paths! Exception: " + e);
|
||||||
|
throw new ResourceHandlerException(
|
||||||
|
"Failed to initialize controller paths!");
|
||||||
|
} finally {
|
||||||
|
rwLock.writeLock().unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* We are looking for entries of the form:
|
||||||
|
* none /cgroup/path/mem cgroup rw,memory 0 0
|
||||||
|
*
|
||||||
|
* Use a simple pattern that splits on the five spaces, and
|
||||||
|
* grabs the 2, 3, and 4th fields.
|
||||||
|
*/
|
||||||
|
|
||||||
|
private static final Pattern MTAB_FILE_FORMAT = Pattern.compile(
|
||||||
|
"^[^\\s]+\\s([^\\s]+)\\s([^\\s]+)\\s([^\\s]+)\\s[^\\s]+\\s[^\\s]+$");
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Returns a map: path -> mount options
|
||||||
|
* for mounts with type "cgroup". Cgroup controllers will
|
||||||
|
* appear in the list of options for a path.
|
||||||
|
*/
|
||||||
|
private Map<String, List<String>> parseMtab() throws IOException {
|
||||||
|
Map<String, List<String>> ret = new HashMap<String, List<String>>();
|
||||||
|
BufferedReader in = null;
|
||||||
|
|
||||||
|
try {
|
||||||
|
FileInputStream fis = new FileInputStream(new File(getMtabFileName()));
|
||||||
|
in = new BufferedReader(new InputStreamReader(fis, "UTF-8"));
|
||||||
|
|
||||||
|
for (String str = in.readLine(); str != null;
|
||||||
|
str = in.readLine()) {
|
||||||
|
Matcher m = MTAB_FILE_FORMAT.matcher(str);
|
||||||
|
boolean mat = m.find();
|
||||||
|
if (mat) {
|
||||||
|
String path = m.group(1);
|
||||||
|
String type = m.group(2);
|
||||||
|
String options = m.group(3);
|
||||||
|
|
||||||
|
if (type.equals(CGROUPS_FSTYPE)) {
|
||||||
|
List<String> value = Arrays.asList(options.split(","));
|
||||||
|
ret.put(path, value);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new IOException("Error while reading " + getMtabFileName(), e);
|
||||||
|
} finally {
|
||||||
|
IOUtils.cleanup(LOG, in);
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
private String findControllerInMtab(String controller,
|
||||||
|
Map<String, List<String>> entries) {
|
||||||
|
for (Map.Entry<String, List<String>> e : entries.entrySet()) {
|
||||||
|
if (e.getValue().contains(controller))
|
||||||
|
return e.getKey();
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
String getMtabFileName() {
|
||||||
|
return MTAB_FILE;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void mountCGroupController(CGroupController controller)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
if (!enableCGroupMount) {
|
||||||
|
LOG.warn("CGroup mounting is disabled - ignoring mount request for: " +
|
||||||
|
controller.getName());
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
String path = getControllerPath(controller);
|
||||||
|
|
||||||
|
if (path == null) {
|
||||||
|
try {
|
||||||
|
//lock out other readers/writers till we are done
|
||||||
|
rwLock.writeLock().lock();
|
||||||
|
|
||||||
|
String hierarchy = cGroupPrefix;
|
||||||
|
StringBuffer controllerPath = new StringBuffer()
|
||||||
|
.append(cGroupMountPath).append('/').append(controller.getName());
|
||||||
|
StringBuffer cGroupKV = new StringBuffer()
|
||||||
|
.append(controller.getName()).append('=').append(controllerPath);
|
||||||
|
PrivilegedOperation.OperationType opType = PrivilegedOperation
|
||||||
|
.OperationType.MOUNT_CGROUPS;
|
||||||
|
PrivilegedOperation op = new PrivilegedOperation(opType, (String) null);
|
||||||
|
|
||||||
|
op.appendArgs(hierarchy, cGroupKV.toString());
|
||||||
|
LOG.info("Mounting controller " + controller.getName() + " at " +
|
||||||
|
controllerPath);
|
||||||
|
privilegedOperationExecutor.executePrivilegedOperation(op, false);
|
||||||
|
|
||||||
|
//if privileged operation succeeds, update controller paths
|
||||||
|
controllerPaths.put(controller, controllerPath.toString());
|
||||||
|
|
||||||
|
return;
|
||||||
|
} catch (PrivilegedOperationException e) {
|
||||||
|
LOG.error("Failed to mount controller: " + controller.getName());
|
||||||
|
throw new ResourceHandlerException("Failed to mount controller: "
|
||||||
|
+ controller.getName());
|
||||||
|
} finally {
|
||||||
|
rwLock.writeLock().unlock();
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
LOG.info("CGroup controller already mounted at: " + path);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getPathForCGroup(CGroupController controller, String cGroupId) {
|
||||||
|
return new StringBuffer(getControllerPath(controller))
|
||||||
|
.append('/').append(cGroupPrefix).append("/")
|
||||||
|
.append(cGroupId).toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getPathForCGroupTasks(CGroupController controller,
|
||||||
|
String cGroupId) {
|
||||||
|
return new StringBuffer(getPathForCGroup(controller, cGroupId))
|
||||||
|
.append('/').append(CGROUP_FILE_TASKS).toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getPathForCGroupParam(CGroupController controller,
|
||||||
|
String cGroupId, String param) {
|
||||||
|
return new StringBuffer(getPathForCGroup(controller, cGroupId))
|
||||||
|
.append('/').append(controller.getName()).append('.')
|
||||||
|
.append(param).toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String createCGroup(CGroupController controller, String cGroupId)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
String path = getPathForCGroup(controller, cGroupId);
|
||||||
|
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("createCgroup: " + path);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!new File(path).mkdir()) {
|
||||||
|
throw new ResourceHandlerException("Failed to create cgroup at " + path);
|
||||||
|
}
|
||||||
|
|
||||||
|
return path;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Utility routine to print first line from cgroup tasks file
|
||||||
|
*/
|
||||||
|
private void logLineFromTasksFile(File cgf) {
|
||||||
|
String str;
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
try (BufferedReader inl =
|
||||||
|
new BufferedReader(new InputStreamReader(new FileInputStream(cgf
|
||||||
|
+ "/tasks"), "UTF-8"))) {
|
||||||
|
if ((str = inl.readLine()) != null) {
|
||||||
|
LOG.debug("First line in cgroup tasks file: " + cgf + " " + str);
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.warn("Failed to read cgroup tasks file. ", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* If tasks file is empty, delete the cgroup.
|
||||||
|
*
|
||||||
|
* @param cgf object referring to the cgroup to be deleted
|
||||||
|
* @return Boolean indicating whether cgroup was deleted
|
||||||
|
*/
|
||||||
|
boolean checkAndDeleteCgroup(File cgf) throws InterruptedException {
|
||||||
|
boolean deleted = false;
|
||||||
|
// FileInputStream in = null;
|
||||||
|
try (FileInputStream in = new FileInputStream(cgf + "/tasks")) {
|
||||||
|
if (in.read() == -1) {
|
||||||
|
/*
|
||||||
|
* "tasks" file is empty, sleep a bit more and then try to delete the
|
||||||
|
* cgroup. Some versions of linux will occasionally panic due to a race
|
||||||
|
* condition in this area, hence the paranoia.
|
||||||
|
*/
|
||||||
|
Thread.sleep(deleteCGroupDelay);
|
||||||
|
deleted = cgf.delete();
|
||||||
|
if (!deleted) {
|
||||||
|
LOG.warn("Failed attempt to delete cgroup: " + cgf);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
logLineFromTasksFile(cgf);
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.warn("Failed to read cgroup tasks file. ", e);
|
||||||
|
}
|
||||||
|
return deleted;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void deleteCGroup(CGroupController controller, String cGroupId)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
boolean deleted = false;
|
||||||
|
String cGroupPath = getPathForCGroup(controller, cGroupId);
|
||||||
|
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("deleteCGroup: " + cGroupPath);
|
||||||
|
}
|
||||||
|
|
||||||
|
long start = clock.getTime();
|
||||||
|
|
||||||
|
do {
|
||||||
|
try {
|
||||||
|
deleted = checkAndDeleteCgroup(new File(cGroupPath));
|
||||||
|
if (!deleted) {
|
||||||
|
Thread.sleep(deleteCGroupDelay);
|
||||||
|
}
|
||||||
|
} catch (InterruptedException ex) {
|
||||||
|
// NOP
|
||||||
|
}
|
||||||
|
} while (!deleted && (clock.getTime() - start) < deleteCGroupTimeout);
|
||||||
|
|
||||||
|
if (!deleted) {
|
||||||
|
LOG.warn("Unable to delete " + cGroupPath +
|
||||||
|
", tried to delete for " + deleteCGroupTimeout + "ms");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void updateCGroupParam(CGroupController controller, String cGroupId,
|
||||||
|
String param, String value) throws ResourceHandlerException {
|
||||||
|
String cGroupParamPath = getPathForCGroupParam(controller, cGroupId, param);
|
||||||
|
PrintWriter pw = null;
|
||||||
|
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug(
|
||||||
|
"updateCGroupParam for path: " + cGroupParamPath + " with value " +
|
||||||
|
value);
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
File file = new File(cGroupParamPath);
|
||||||
|
Writer w = new OutputStreamWriter(new FileOutputStream(file), "UTF-8");
|
||||||
|
pw = new PrintWriter(w);
|
||||||
|
pw.write(value);
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new ResourceHandlerException(new StringBuffer("Unable to write to ")
|
||||||
|
.append(cGroupParamPath).append(" with value: ").append(value)
|
||||||
|
.toString(), e);
|
||||||
|
} finally {
|
||||||
|
if (pw != null) {
|
||||||
|
boolean hasError = pw.checkError();
|
||||||
|
pw.close();
|
||||||
|
if (hasError) {
|
||||||
|
throw new ResourceHandlerException(
|
||||||
|
new StringBuffer("Unable to write to ")
|
||||||
|
.append(cGroupParamPath).append(" with value: ").append(value)
|
||||||
|
.toString());
|
||||||
|
}
|
||||||
|
if (pw.checkError()) {
|
||||||
|
throw new ResourceHandlerException("Error while closing cgroup file" +
|
||||||
|
" " + cGroupParamPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getCGroupParam(CGroupController controller, String cGroupId,
|
||||||
|
String param) throws ResourceHandlerException {
|
||||||
|
String cGroupParamPath = getPathForCGroupParam(controller, cGroupId, param);
|
||||||
|
|
||||||
|
try {
|
||||||
|
byte[] contents = Files.readAllBytes(Paths.get(cGroupParamPath));
|
||||||
|
return new String(contents, "UTF-8").trim();
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new ResourceHandlerException(
|
||||||
|
"Unable to read from " + cGroupParamPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,91 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperation;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Handler interface for resource subsystems' isolation and enforcement. e.g cpu, memory, network, disks etc
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public interface ResourceHandler {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Bootstrap resource susbsystem.
|
||||||
|
*
|
||||||
|
* @return (possibly empty) list of operations that require elevated
|
||||||
|
* privileges
|
||||||
|
*/
|
||||||
|
List<PrivilegedOperation> bootstrap(Configuration configuration)
|
||||||
|
throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Prepare a resource environment for container launch
|
||||||
|
*
|
||||||
|
* @param container Container being launched
|
||||||
|
* @return (possibly empty) list of operations that require elevated
|
||||||
|
* privileges e.g a) create a custom cgroup b) add pid for container to tasks
|
||||||
|
* file for a cgroup.
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
List<PrivilegedOperation> preStart(Container container)
|
||||||
|
throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Require state for container that was already launched
|
||||||
|
*
|
||||||
|
* @param containerId if of the container being reacquired.
|
||||||
|
* @return (possibly empty) list of operations that require elevated
|
||||||
|
* privileges
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
|
||||||
|
List<PrivilegedOperation> reacquireContainer(ContainerId containerId)
|
||||||
|
throws ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Perform any tasks necessary after container completion
|
||||||
|
* @param containerId of the container that was completed.
|
||||||
|
* @return (possibly empty) list of operations that require elevated
|
||||||
|
* privileges
|
||||||
|
* @throws ResourceHandlerException
|
||||||
|
*/
|
||||||
|
List<PrivilegedOperation> postComplete(ContainerId containerId) throws
|
||||||
|
ResourceHandlerException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Teardown environment for resource subsystem if requested. This method
|
||||||
|
* needs to be used with care since it could impact running containers.
|
||||||
|
*
|
||||||
|
* @return (possibly empty) list of operations that require elevated
|
||||||
|
* privileges
|
||||||
|
*/
|
||||||
|
List<PrivilegedOperation> teardown() throws ResourceHandlerException;
|
||||||
|
}
|
|
@ -0,0 +1,142 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperation;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Collections;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A helper class to delegate funcationality to a 'chain' of
|
||||||
|
* ResourceHandler(s)
|
||||||
|
*/
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class ResourceHandlerChain implements ResourceHandler {
|
||||||
|
private final List<ResourceHandler> resourceHandlers;
|
||||||
|
|
||||||
|
public ResourceHandlerChain(List<ResourceHandler> resourceHandlers) {
|
||||||
|
this.resourceHandlers = resourceHandlers;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<PrivilegedOperation> bootstrap(Configuration configuration)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
|
||||||
|
List<PrivilegedOperation> allOperations = new
|
||||||
|
ArrayList<PrivilegedOperation>();
|
||||||
|
|
||||||
|
for (ResourceHandler resourceHandler : resourceHandlers) {
|
||||||
|
List<PrivilegedOperation> handlerOperations =
|
||||||
|
resourceHandler.bootstrap(configuration);
|
||||||
|
if (handlerOperations != null) {
|
||||||
|
allOperations.addAll(handlerOperations);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return allOperations;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<PrivilegedOperation> preStart(Container container)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
List<PrivilegedOperation> allOperations = new
|
||||||
|
ArrayList<PrivilegedOperation>();
|
||||||
|
|
||||||
|
for (ResourceHandler resourceHandler : resourceHandlers) {
|
||||||
|
List<PrivilegedOperation> handlerOperations =
|
||||||
|
resourceHandler.preStart(container);
|
||||||
|
|
||||||
|
if (handlerOperations != null) {
|
||||||
|
allOperations.addAll(handlerOperations);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return allOperations;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<PrivilegedOperation> reacquireContainer(ContainerId containerId)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
List<PrivilegedOperation> allOperations = new
|
||||||
|
ArrayList<PrivilegedOperation>();
|
||||||
|
|
||||||
|
for (ResourceHandler resourceHandler : resourceHandlers) {
|
||||||
|
List<PrivilegedOperation> handlerOperations =
|
||||||
|
resourceHandler.reacquireContainer(containerId);
|
||||||
|
|
||||||
|
if (handlerOperations != null) {
|
||||||
|
allOperations.addAll(handlerOperations);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return allOperations;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<PrivilegedOperation> postComplete(ContainerId containerId)
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
List<PrivilegedOperation> allOperations = new
|
||||||
|
ArrayList<PrivilegedOperation>();
|
||||||
|
|
||||||
|
for (ResourceHandler resourceHandler : resourceHandlers) {
|
||||||
|
List<PrivilegedOperation> handlerOperations =
|
||||||
|
resourceHandler.postComplete(containerId);
|
||||||
|
|
||||||
|
if (handlerOperations != null) {
|
||||||
|
allOperations.addAll(handlerOperations);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return allOperations;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<PrivilegedOperation> teardown()
|
||||||
|
throws ResourceHandlerException {
|
||||||
|
List<PrivilegedOperation> allOperations = new
|
||||||
|
ArrayList<PrivilegedOperation>();
|
||||||
|
|
||||||
|
for (ResourceHandler resourceHandler : resourceHandlers) {
|
||||||
|
List<PrivilegedOperation> handlerOperations =
|
||||||
|
resourceHandler.teardown();
|
||||||
|
|
||||||
|
if (handlerOperations != null) {
|
||||||
|
allOperations.addAll(handlerOperations);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return allOperations;
|
||||||
|
}
|
||||||
|
|
||||||
|
List<ResourceHandler> getResourceHandlerList() {
|
||||||
|
return Collections.unmodifiableList(resourceHandlers);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
|
@ -0,0 +1,47 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||||
|
|
||||||
|
@InterfaceAudience.Private
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public class ResourceHandlerException extends YarnException {
|
||||||
|
private static final long serialVersionUID = 1L;
|
||||||
|
|
||||||
|
public ResourceHandlerException() {
|
||||||
|
super();
|
||||||
|
}
|
||||||
|
|
||||||
|
public ResourceHandlerException(String message) {
|
||||||
|
super(message);
|
||||||
|
}
|
||||||
|
|
||||||
|
public ResourceHandlerException(Throwable cause) {
|
||||||
|
super(cause);
|
||||||
|
}
|
||||||
|
|
||||||
|
public ResourceHandlerException(String message, Throwable cause) {
|
||||||
|
super(message, cause);
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,233 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
public class TestPrivilegedOperationExecutor {
|
||||||
|
private static final Log LOG = LogFactory
|
||||||
|
.getLog(TestPrivilegedOperationExecutor.class);
|
||||||
|
private String localDataDir;
|
||||||
|
private String customExecutorPath;
|
||||||
|
private Configuration nullConf = null;
|
||||||
|
private Configuration emptyConf;
|
||||||
|
private Configuration confWithExecutorPath;
|
||||||
|
|
||||||
|
private String cGroupTasksNone;
|
||||||
|
private String cGroupTasksInvalid;
|
||||||
|
private String cGroupTasks1;
|
||||||
|
private String cGroupTasks2;
|
||||||
|
private String cGroupTasks3;
|
||||||
|
private PrivilegedOperation opDisallowed;
|
||||||
|
private PrivilegedOperation opTasksNone;
|
||||||
|
private PrivilegedOperation opTasksInvalid;
|
||||||
|
private PrivilegedOperation opTasks1;
|
||||||
|
private PrivilegedOperation opTasks2;
|
||||||
|
private PrivilegedOperation opTasks3;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
localDataDir = System.getProperty("test.build.data");
|
||||||
|
customExecutorPath = localDataDir + "/bin/container-executor";
|
||||||
|
emptyConf = new YarnConfiguration();
|
||||||
|
confWithExecutorPath = new YarnConfiguration();
|
||||||
|
confWithExecutorPath.set(YarnConfiguration
|
||||||
|
.NM_LINUX_CONTAINER_EXECUTOR_PATH, customExecutorPath);
|
||||||
|
|
||||||
|
cGroupTasksNone = "none";
|
||||||
|
cGroupTasksInvalid = "invalid_string";
|
||||||
|
cGroupTasks1 = "cpu/hadoop_yarn/container_01/tasks";
|
||||||
|
cGroupTasks2 = "net_cls/hadoop_yarn/container_01/tasks";
|
||||||
|
cGroupTasks3 = "blkio/hadoop_yarn/container_01/tasks";
|
||||||
|
opDisallowed = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.DELETE_AS_USER, (String) null);
|
||||||
|
opTasksNone = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP,
|
||||||
|
PrivilegedOperation.CGROUP_ARG_PREFIX + cGroupTasksNone);
|
||||||
|
opTasksInvalid = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP,
|
||||||
|
cGroupTasksInvalid);
|
||||||
|
opTasks1 = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP,
|
||||||
|
PrivilegedOperation.CGROUP_ARG_PREFIX + cGroupTasks1);
|
||||||
|
opTasks2 = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP,
|
||||||
|
PrivilegedOperation.CGROUP_ARG_PREFIX + cGroupTasks2);
|
||||||
|
opTasks3 = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.ADD_PID_TO_CGROUP,
|
||||||
|
PrivilegedOperation.CGROUP_ARG_PREFIX + cGroupTasks3);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testExecutorPath() {
|
||||||
|
String containerExePath = PrivilegedOperationExecutor
|
||||||
|
.getContainerExecutorExecutablePath(nullConf);
|
||||||
|
|
||||||
|
//In case HADOOP_YARN_HOME isn't set, CWD is used. If conf is null or
|
||||||
|
//NM_LINUX_CONTAINER_EXECUTOR_PATH is not set, then a defaultPath is
|
||||||
|
//constructed.
|
||||||
|
String yarnHomeEnvVar = System.getenv("HADOOP_YARN_HOME");
|
||||||
|
String yarnHome = yarnHomeEnvVar != null ? yarnHomeEnvVar
|
||||||
|
: new File("").getAbsolutePath();
|
||||||
|
String expectedPath = yarnHome + "/bin/container-executor";
|
||||||
|
|
||||||
|
Assert.assertEquals(expectedPath, containerExePath);
|
||||||
|
|
||||||
|
containerExePath = PrivilegedOperationExecutor
|
||||||
|
.getContainerExecutorExecutablePath(emptyConf);
|
||||||
|
Assert.assertEquals(expectedPath, containerExePath);
|
||||||
|
|
||||||
|
//if NM_LINUX_CONTAINER_EXECUTOR_PATH is set, this must be returned
|
||||||
|
expectedPath = customExecutorPath;
|
||||||
|
containerExePath = PrivilegedOperationExecutor
|
||||||
|
.getContainerExecutorExecutablePath(confWithExecutorPath);
|
||||||
|
Assert.assertEquals(expectedPath, containerExePath);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testExecutionCommand() {
|
||||||
|
PrivilegedOperationExecutor exec = PrivilegedOperationExecutor
|
||||||
|
.getInstance(confWithExecutorPath);
|
||||||
|
PrivilegedOperation op = new PrivilegedOperation(PrivilegedOperation
|
||||||
|
.OperationType.LAUNCH_CONTAINER, (String) null);
|
||||||
|
String[] cmdArray = exec.getPrivilegedOperationExecutionCommand(null, op);
|
||||||
|
|
||||||
|
//No arguments added - so the resulting array should consist of
|
||||||
|
//1)full path to executor 2) cli switch
|
||||||
|
Assert.assertEquals(2, cmdArray.length);
|
||||||
|
Assert.assertEquals(customExecutorPath, cmdArray[0]);
|
||||||
|
Assert.assertEquals(op.getOperationType().getOption(), cmdArray[1]);
|
||||||
|
|
||||||
|
//other (dummy) arguments to launch container
|
||||||
|
String[] additionalArgs = { "test_user", "yarn", "1", "app_01",
|
||||||
|
"container_01", "workdir", "launch_script.sh", "tokens", "pidfile",
|
||||||
|
"nm-local-dirs", "nm-log-dirs", "resource-spec" };
|
||||||
|
|
||||||
|
op.appendArgs(additionalArgs);
|
||||||
|
cmdArray = exec.getPrivilegedOperationExecutionCommand(null, op);
|
||||||
|
|
||||||
|
//Resulting array should be of length 2 greater than the number of
|
||||||
|
//additional arguments added.
|
||||||
|
|
||||||
|
Assert.assertEquals(2 + additionalArgs.length, cmdArray.length);
|
||||||
|
Assert.assertEquals(customExecutorPath, cmdArray[0]);
|
||||||
|
Assert.assertEquals(op.getOperationType().getOption(), cmdArray[1]);
|
||||||
|
|
||||||
|
//Rest of args should be same as additional args.
|
||||||
|
for (int i = 0; i < additionalArgs.length; ++i) {
|
||||||
|
Assert.assertEquals(additionalArgs[i], cmdArray[2 + i]);
|
||||||
|
}
|
||||||
|
|
||||||
|
//Now test prefix commands
|
||||||
|
List<String> prefixCommands = Arrays.asList("nice", "-10");
|
||||||
|
cmdArray = exec.getPrivilegedOperationExecutionCommand(prefixCommands, op);
|
||||||
|
int prefixLength = prefixCommands.size();
|
||||||
|
//Resulting array should be of length of prefix command args + 2 (exec
|
||||||
|
// path + switch) + length of additional args.
|
||||||
|
Assert.assertEquals(prefixLength + 2 + additionalArgs.length,
|
||||||
|
cmdArray.length);
|
||||||
|
|
||||||
|
//Prefix command array comes first
|
||||||
|
for (int i = 0; i < prefixLength; ++i) {
|
||||||
|
Assert.assertEquals(prefixCommands.get(i), cmdArray[i]);
|
||||||
|
}
|
||||||
|
|
||||||
|
//Followed by the container executor path and the cli switch
|
||||||
|
Assert.assertEquals(customExecutorPath, cmdArray[prefixLength]);
|
||||||
|
Assert.assertEquals(op.getOperationType().getOption(),
|
||||||
|
cmdArray[prefixLength + 1]);
|
||||||
|
|
||||||
|
//Followed by the rest of the args
|
||||||
|
//Rest of args should be same as additional args.
|
||||||
|
for (int i = 0; i < additionalArgs.length; ++i) {
|
||||||
|
Assert.assertEquals(additionalArgs[i], cmdArray[prefixLength + 2 + i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSquashCGroupOperationsWithInvalidOperations() {
|
||||||
|
List<PrivilegedOperation> ops = new ArrayList<>();
|
||||||
|
|
||||||
|
//Ensure that disallowed ops are rejected
|
||||||
|
ops.add(opTasksNone);
|
||||||
|
ops.add(opDisallowed);
|
||||||
|
|
||||||
|
try {
|
||||||
|
PrivilegedOperationExecutor.squashCGroupOperations(ops);
|
||||||
|
Assert.fail("Expected squash operation to fail with an exception!");
|
||||||
|
} catch (PrivilegedOperationException e) {
|
||||||
|
LOG.info("Caught expected exception : " + e);
|
||||||
|
}
|
||||||
|
|
||||||
|
//Ensure that invalid strings are rejected
|
||||||
|
ops.clear();
|
||||||
|
ops.add(opTasksNone);
|
||||||
|
ops.add(opTasksInvalid);
|
||||||
|
|
||||||
|
try {
|
||||||
|
PrivilegedOperationExecutor.squashCGroupOperations(ops);
|
||||||
|
Assert.fail("Expected squash operation to fail with an exception!");
|
||||||
|
} catch (PrivilegedOperationException e) {
|
||||||
|
LOG.info("Caught expected exception : " + e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSquashCGroupOperationsWithValidOperations() {
|
||||||
|
List<PrivilegedOperation> ops = new ArrayList<>();
|
||||||
|
//Test squashing, including 'none'
|
||||||
|
ops.clear();
|
||||||
|
ops.add(opTasks1);
|
||||||
|
//this is expected to be ignored
|
||||||
|
ops.add(opTasksNone);
|
||||||
|
ops.add(opTasks2);
|
||||||
|
ops.add(opTasks3);
|
||||||
|
|
||||||
|
try {
|
||||||
|
PrivilegedOperation op = PrivilegedOperationExecutor
|
||||||
|
.squashCGroupOperations(ops);
|
||||||
|
String expected = new StringBuffer
|
||||||
|
(PrivilegedOperation.CGROUP_ARG_PREFIX)
|
||||||
|
.append(cGroupTasks1).append(',')
|
||||||
|
.append(cGroupTasks2).append(',')
|
||||||
|
.append(cGroupTasks3).toString();
|
||||||
|
|
||||||
|
//We expect axactly one argument
|
||||||
|
Assert.assertEquals(1, op.getArguments().size());
|
||||||
|
//Squashed list of tasks files
|
||||||
|
Assert.assertEquals(expected, op.getArguments().get(0));
|
||||||
|
} catch (PrivilegedOperationException e) {
|
||||||
|
LOG.info("Caught unexpected exception : " + e);
|
||||||
|
Assert.fail("Caught unexpected exception: " + e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,235 @@
|
||||||
|
/*
|
||||||
|
* *
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
* /
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.resources;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileUtil;
|
||||||
|
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperation;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException;
|
||||||
|
import org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationExecutor;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.mockito.ArgumentCaptor;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.file.Files;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
import static org.mockito.Matchers.eq;
|
||||||
|
import static org.mockito.Matchers.isNull;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
import static org.mockito.Mockito.verifyNoMoreInteractions;
|
||||||
|
import static org.mockito.Mockito.verifyZeroInteractions;
|
||||||
|
|
||||||
|
public class TestCGroupsHandlerImpl {
|
||||||
|
private static final Log LOG =
|
||||||
|
LogFactory.getLog(TestCGroupsHandlerImpl.class);
|
||||||
|
|
||||||
|
private PrivilegedOperationExecutor privilegedOperationExecutorMock;
|
||||||
|
private Configuration conf;
|
||||||
|
private String tmpPath;
|
||||||
|
private String hierarchy;
|
||||||
|
private CGroupsHandler.CGroupController controller;
|
||||||
|
private String controllerPath;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
privilegedOperationExecutorMock = mock(PrivilegedOperationExecutor.class);
|
||||||
|
conf = new YarnConfiguration();
|
||||||
|
tmpPath = System.getProperty("test.build.data") + "/cgroups";
|
||||||
|
//no leading or trailing slashes here
|
||||||
|
hierarchy = "test-hadoop-yarn";
|
||||||
|
|
||||||
|
conf.set(YarnConfiguration.NM_LINUX_CONTAINER_CGROUPS_HIERARCHY, hierarchy);
|
||||||
|
conf.setBoolean(YarnConfiguration.NM_LINUX_CONTAINER_CGROUPS_MOUNT, true);
|
||||||
|
conf.set(YarnConfiguration.NM_LINUX_CONTAINER_CGROUPS_MOUNT_PATH, tmpPath);
|
||||||
|
controller = CGroupsHandler.CGroupController.NET_CLS;
|
||||||
|
controllerPath = new StringBuffer(tmpPath).append('/')
|
||||||
|
.append(controller.getName()).append('/').append(hierarchy).toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testMountController() {
|
||||||
|
CGroupsHandler cGroupsHandler = null;
|
||||||
|
//Since we enabled (deferred) cgroup controller mounting, no interactions
|
||||||
|
//should have occurred, with this mock
|
||||||
|
verifyZeroInteractions(privilegedOperationExecutorMock);
|
||||||
|
|
||||||
|
try {
|
||||||
|
cGroupsHandler = new CGroupsHandlerImpl(conf,
|
||||||
|
privilegedOperationExecutorMock);
|
||||||
|
PrivilegedOperation expectedOp = new PrivilegedOperation
|
||||||
|
(PrivilegedOperation.OperationType.MOUNT_CGROUPS, (String) null);
|
||||||
|
//This is expected to be of the form :
|
||||||
|
//net_cls=<mount_path>/net_cls
|
||||||
|
StringBuffer controllerKV = new StringBuffer(controller.getName())
|
||||||
|
.append('=').append(tmpPath).append('/').append(controller.getName());
|
||||||
|
expectedOp.appendArgs(hierarchy, controllerKV.toString());
|
||||||
|
|
||||||
|
cGroupsHandler.mountCGroupController(controller);
|
||||||
|
try {
|
||||||
|
ArgumentCaptor<PrivilegedOperation> opCaptor = ArgumentCaptor.forClass
|
||||||
|
(PrivilegedOperation.class);
|
||||||
|
verify(privilegedOperationExecutorMock)
|
||||||
|
.executePrivilegedOperation(opCaptor.capture(), eq(false));
|
||||||
|
|
||||||
|
//we'll explicitly capture and assert that the
|
||||||
|
//captured op and the expected op are identical.
|
||||||
|
Assert.assertEquals(expectedOp, opCaptor.getValue());
|
||||||
|
verifyNoMoreInteractions(privilegedOperationExecutorMock);
|
||||||
|
|
||||||
|
//Try mounting the same controller again - this should be a no-op
|
||||||
|
cGroupsHandler.mountCGroupController(controller);
|
||||||
|
verifyNoMoreInteractions(privilegedOperationExecutorMock);
|
||||||
|
} catch (PrivilegedOperationException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue("Unexpected PrivilegedOperationException from mock!",
|
||||||
|
false);
|
||||||
|
}
|
||||||
|
} catch (ResourceHandlerException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue("Unexpected ResourceHandler Exception!", false);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCGroupPaths() {
|
||||||
|
//As per junit behavior, we expect a new mock object to be available
|
||||||
|
//in this test.
|
||||||
|
verifyZeroInteractions(privilegedOperationExecutorMock);
|
||||||
|
CGroupsHandler cGroupsHandler = null;
|
||||||
|
try {
|
||||||
|
cGroupsHandler = new CGroupsHandlerImpl(conf,
|
||||||
|
privilegedOperationExecutorMock);
|
||||||
|
cGroupsHandler.mountCGroupController(controller);
|
||||||
|
} catch (ResourceHandlerException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue(
|
||||||
|
"Unexpected ResourceHandlerException when mounting controller!",
|
||||||
|
false);
|
||||||
|
}
|
||||||
|
|
||||||
|
String testCGroup = "container_01";
|
||||||
|
String expectedPath = new StringBuffer(controllerPath).append('/')
|
||||||
|
.append(testCGroup).toString();
|
||||||
|
String path = cGroupsHandler.getPathForCGroup(controller, testCGroup);
|
||||||
|
Assert.assertEquals(expectedPath, path);
|
||||||
|
|
||||||
|
String expectedPathTasks = new StringBuffer(expectedPath).append('/')
|
||||||
|
.append(CGroupsHandler.CGROUP_FILE_TASKS).toString();
|
||||||
|
path = cGroupsHandler.getPathForCGroupTasks(controller, testCGroup);
|
||||||
|
Assert.assertEquals(expectedPathTasks, path);
|
||||||
|
|
||||||
|
String param = CGroupsHandler.CGROUP_PARAM_CLASSID;
|
||||||
|
String expectedPathParam = new StringBuffer(expectedPath).append('/')
|
||||||
|
.append(controller.getName()).append('.').append(param).toString();
|
||||||
|
path = cGroupsHandler.getPathForCGroupParam(controller, testCGroup, param);
|
||||||
|
Assert.assertEquals(expectedPathParam, path);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCGroupOperations() {
|
||||||
|
//As per junit behavior, we expect a new mock object to be available
|
||||||
|
//in this test.
|
||||||
|
verifyZeroInteractions(privilegedOperationExecutorMock);
|
||||||
|
CGroupsHandler cGroupsHandler = null;
|
||||||
|
|
||||||
|
try {
|
||||||
|
cGroupsHandler = new CGroupsHandlerImpl(conf,
|
||||||
|
privilegedOperationExecutorMock);
|
||||||
|
cGroupsHandler.mountCGroupController(controller);
|
||||||
|
} catch (ResourceHandlerException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue(
|
||||||
|
"Unexpected ResourceHandlerException when mounting controller!",
|
||||||
|
false);
|
||||||
|
}
|
||||||
|
//Lets manually create a path to (partially) simulate a mounted controller
|
||||||
|
//this is required because the handler uses a mocked privileged operation
|
||||||
|
//executor
|
||||||
|
new File(controllerPath).mkdirs();
|
||||||
|
|
||||||
|
String testCGroup = "container_01";
|
||||||
|
String expectedPath = new StringBuffer(controllerPath).append('/')
|
||||||
|
.append(testCGroup).toString();
|
||||||
|
try {
|
||||||
|
String path = cGroupsHandler.createCGroup(controller, testCGroup);
|
||||||
|
|
||||||
|
Assert.assertTrue(new File(expectedPath).exists());
|
||||||
|
Assert.assertEquals(expectedPath, path);
|
||||||
|
|
||||||
|
//update param and read param tests.
|
||||||
|
//We don't use net_cls.classid because as a test param here because
|
||||||
|
//cgroups provides very specific read/write semantics for classid (only
|
||||||
|
//numbers can be written - potentially as hex but can be read out only
|
||||||
|
//as decimal)
|
||||||
|
String param = "test_param";
|
||||||
|
String paramValue = "test_param_value";
|
||||||
|
|
||||||
|
cGroupsHandler
|
||||||
|
.updateCGroupParam(controller, testCGroup, param, paramValue);
|
||||||
|
String paramPath = new StringBuffer(expectedPath).append('/')
|
||||||
|
.append(controller.getName()).append('.').append(param).toString();
|
||||||
|
File paramFile = new File(paramPath);
|
||||||
|
|
||||||
|
Assert.assertTrue(paramFile.exists());
|
||||||
|
try {
|
||||||
|
Assert.assertEquals(paramValue, new String(Files.readAllBytes
|
||||||
|
(paramFile
|
||||||
|
.toPath())));
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue("Unexpected IOException trying to read cgroup param!",
|
||||||
|
false);
|
||||||
|
}
|
||||||
|
|
||||||
|
Assert.assertEquals(paramValue, cGroupsHandler.getCGroupParam
|
||||||
|
(controller, testCGroup, param));
|
||||||
|
|
||||||
|
//We can't really do a delete test here. Linux cgroups
|
||||||
|
//implementation provides additional semantics - the cgroup cannot be
|
||||||
|
//deleted if there are any tasks still running in the cgroup even if
|
||||||
|
//the user attempting the delete has the file permissions to do so - we
|
||||||
|
//cannot simulate that here. Even if we create a dummy 'tasks' file, we
|
||||||
|
//wouldn't be able to simulate the delete behavior we need, since a cgroup
|
||||||
|
//can be deleted using using 'rmdir' if the tasks file is empty. Such a
|
||||||
|
//delete is not possible with a regular non-empty directory.
|
||||||
|
} catch (ResourceHandlerException e) {
|
||||||
|
LOG.error("Caught exception: " + e);
|
||||||
|
Assert.assertTrue(
|
||||||
|
"Unexpected ResourceHandlerException during cgroup operations!",
|
||||||
|
false);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void teardown() {
|
||||||
|
FileUtil.fullyDelete(new File(tmpPath));
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue