MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
#!/usr/bin/env bash
|
|
|
|
|
|
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
|
|
# this work for additional information regarding copyright ownership.
|
|
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
|
|
# (the "License"); you may not use this file except in compliance with
|
|
|
|
# the License. You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2015-07-16 19:58:11 -04:00
|
|
|
MYNAME="${BASH_SOURCE-$0}"
|
2016-03-28 12:00:07 -04:00
|
|
|
HADOOP_SHELL_EXECNAME="${MYNAME##*/}"
|
2015-07-16 19:58:11 -04:00
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
## @description build up the yarn command's usage text.
|
|
|
|
## @audience public
|
|
|
|
## @stability stable
|
|
|
|
## @replaceable no
|
2014-08-19 08:11:17 -04:00
|
|
|
function hadoop_usage
|
|
|
|
{
|
2015-07-31 17:32:21 -04:00
|
|
|
hadoop_add_option "--buildpaths" "attempt to add class files from build tree"
|
|
|
|
hadoop_add_option "--daemon (start|status|stop)" "operate on a daemon"
|
2016-06-28 08:53:03 -04:00
|
|
|
hadoop_add_option "--hostnames list[,of,host,names]" "hosts to use in worker mode"
|
2015-07-31 17:32:21 -04:00
|
|
|
hadoop_add_option "--loglevel level" "set the log4j level for this command"
|
2016-06-28 08:53:03 -04:00
|
|
|
hadoop_add_option "--hosts filename" "list of hosts to use in worker mode"
|
|
|
|
hadoop_add_option "--workers" "turn on worker mode"
|
2015-07-31 17:32:21 -04:00
|
|
|
|
2015-07-16 19:58:11 -04:00
|
|
|
hadoop_add_subcommand "application" "prints application(s) report/kill application"
|
|
|
|
hadoop_add_subcommand "applicationattempt" "prints applicationattempt(s) report"
|
|
|
|
hadoop_add_subcommand "classpath" "prints the class path needed to get the hadoop jar and the required libraries"
|
|
|
|
hadoop_add_subcommand "cluster" "prints cluster information"
|
|
|
|
hadoop_add_subcommand "container" "prints container(s) report"
|
|
|
|
hadoop_add_subcommand "daemonlog" "get/set the log level for each daemon"
|
2015-11-07 11:32:56 -05:00
|
|
|
hadoop_add_subcommand "envvars" "display computed Hadoop environment variables"
|
2015-07-16 19:58:11 -04:00
|
|
|
hadoop_add_subcommand "jar <jar>" "run a jar file"
|
|
|
|
hadoop_add_subcommand "logs" "dump container logs"
|
|
|
|
hadoop_add_subcommand "node" "prints node report(s)"
|
2016-06-28 08:53:03 -04:00
|
|
|
hadoop_add_subcommand "nodemanager" "run a nodemanager on each worker"
|
2015-07-16 19:58:11 -04:00
|
|
|
hadoop_add_subcommand "proxyserver" "run the web app proxy server"
|
|
|
|
hadoop_add_subcommand "queue" "prints queue information"
|
|
|
|
hadoop_add_subcommand "resourcemanager" "run the ResourceManager"
|
|
|
|
hadoop_add_subcommand "rmadmin" "admin tools"
|
|
|
|
hadoop_add_subcommand "scmadmin" "SharedCacheManager admin tools"
|
|
|
|
hadoop_add_subcommand "sharedcachemanager" "run the SharedCacheManager daemon"
|
2015-07-08 20:10:10 -04:00
|
|
|
hadoop_add_subcommand "timelinereader" "run the timeline reader server"
|
2016-06-21 19:38:20 -04:00
|
|
|
hadoop_add_subcommand "timelineserver" "run the timeline server"
|
2015-07-16 19:58:11 -04:00
|
|
|
hadoop_add_subcommand "top" "view cluster information"
|
|
|
|
hadoop_add_subcommand "version" "print the version"
|
2016-03-28 12:00:07 -04:00
|
|
|
hadoop_generate_usage "${HADOOP_SHELL_EXECNAME}" true
|
2013-02-22 13:36:24 -05:00
|
|
|
}
|
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
## @description Default command handler for yarn command
|
|
|
|
## @audience public
|
|
|
|
## @stability stable
|
|
|
|
## @replaceable no
|
|
|
|
## @param CLI arguments
|
|
|
|
function yarncmd_case
|
|
|
|
{
|
|
|
|
subcmd=$1
|
|
|
|
shift
|
|
|
|
|
|
|
|
case ${subcmd} in
|
|
|
|
application|applicationattempt|container)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.ApplicationCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
set -- "${subcmd}" "$@"
|
2016-05-26 19:43:08 -04:00
|
|
|
HADOOP_SUBCMD_ARGS=("$@")
|
2016-03-28 12:00:07 -04:00
|
|
|
;;
|
|
|
|
classpath)
|
|
|
|
hadoop_do_classpath_subcommand HADOOP_CLASSNAME "$@"
|
|
|
|
;;
|
|
|
|
cluster)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.ClusterCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
|
|
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
daemonlog)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.log.LogLevel
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
envvars)
|
|
|
|
echo "JAVA_HOME='${JAVA_HOME}'"
|
|
|
|
echo "HADOOP_YARN_HOME='${HADOOP_YARN_HOME}'"
|
|
|
|
echo "YARN_DIR='${YARN_DIR}'"
|
|
|
|
echo "YARN_LIB_JARS_DIR='${YARN_LIB_JARS_DIR}'"
|
|
|
|
echo "HADOOP_CONF_DIR='${HADOOP_CONF_DIR}'"
|
|
|
|
echo "HADOOP_TOOLS_HOME='${HADOOP_TOOLS_HOME}'"
|
|
|
|
echo "HADOOP_TOOLS_DIR='${HADOOP_TOOLS_DIR}'"
|
|
|
|
echo "HADOOP_TOOLS_LIB_JARS_DIR='${HADOOP_TOOLS_LIB_JARS_DIR}'"
|
|
|
|
exit 0
|
|
|
|
;;
|
|
|
|
jar)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.util.RunJar
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
historyserver)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
echo "DEPRECATED: Use of this command to start the timeline server is deprecated." 1>&2
|
|
|
|
echo "Instead use the timelineserver command for it." 1>&2
|
|
|
|
echo "Starting the History Server anyway..." 1>&2
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer'
|
|
|
|
;;
|
|
|
|
logs)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.LogsCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
node)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.NodeCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
nodemanager)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.nodemanager.NodeManager'
|
|
|
|
hadoop_debug "Append YARN_NODEMANAGER_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_NODEMANAGER_OPTS}"
|
|
|
|
# Backwards compatibility
|
|
|
|
if [[ -n "${YARN_NODEMANAGER_HEAPSIZE}" ]]; then
|
|
|
|
HADOOP_HEAPSIZE_MAX="${YARN_NODEMANAGER_HEAPSIZE}"
|
|
|
|
fi
|
|
|
|
;;
|
|
|
|
proxyserver)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.webproxy.WebAppProxyServer'
|
|
|
|
hadoop_debug "Append YARN_PROXYSERVER_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_PROXYSERVER_OPTS}"
|
|
|
|
# Backwards compatibility
|
|
|
|
if [[ -n "${YARN_PROXYSERVER_HEAPSIZE}" ]]; then
|
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_HEAPSIZE_MAX="${YARN_PROXYSERVER_HEAPSIZE}"
|
|
|
|
fi
|
|
|
|
;;
|
|
|
|
queue)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.QueueCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
resourcemanager)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.resourcemanager.ResourceManager'
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_RESOURCEMANAGER_OPTS}"
|
|
|
|
hadoop_debug "Append YARN_RESOURCEMANAGER_OPTS onto HADOOP_OPTS"
|
|
|
|
# Backwards compatibility
|
|
|
|
if [[ -n "${YARN_RESOURCEMANAGER_HEAPSIZE}" ]]; then
|
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_HEAPSIZE_MAX="${YARN_RESOURCEMANAGER_HEAPSIZE}"
|
|
|
|
fi
|
|
|
|
;;
|
|
|
|
rmadmin)
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.client.cli.RMAdminCLI'
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
scmadmin)
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.client.SCMAdmin'
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
sharedcachemanager)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.sharedcachemanager.SharedCacheManager'
|
|
|
|
hadoop_debug "Append YARN_SHAREDCACHEMANAGER_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_SHAREDCACHEMANAGER_OPTS}"
|
|
|
|
;;
|
2016-06-21 19:38:20 -04:00
|
|
|
timelinereader)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderServer'
|
|
|
|
;;
|
2016-03-28 12:00:07 -04:00
|
|
|
timelineserver)
|
|
|
|
HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true"
|
|
|
|
HADOOP_CLASSNAME='org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer'
|
|
|
|
hadoop_debug "Append YARN_TIMELINESERVER_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_TIMELINESERVER_OPTS}"
|
|
|
|
# Backwards compatibility
|
|
|
|
if [[ -n "${YARN_TIMELINESERVER_HEAPSIZE}" ]]; then
|
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_HEAPSIZE_MAX="${YARN_TIMELINESERVER_HEAPSIZE}"
|
|
|
|
fi
|
|
|
|
;;
|
|
|
|
version)
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.util.VersionInfo
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
|
|
|
;;
|
|
|
|
top)
|
|
|
|
doNotSetCols=0
|
|
|
|
doNotSetRows=0
|
|
|
|
for i in "$@"; do
|
|
|
|
if [[ $i == "-cols" ]]; then
|
|
|
|
doNotSetCols=1
|
|
|
|
fi
|
|
|
|
if [[ $i == "-rows" ]]; then
|
|
|
|
doNotSetRows=1
|
|
|
|
fi
|
|
|
|
done
|
|
|
|
if [ $doNotSetCols == 0 ] && [ -n "${TERM}" ]; then
|
|
|
|
cols=$(tput cols)
|
|
|
|
if [ -n "$cols" ]; then
|
|
|
|
args=( $@ )
|
|
|
|
args=("${args[@]}" "-cols" "$cols")
|
|
|
|
set -- "${args[@]}"
|
|
|
|
fi
|
|
|
|
fi
|
|
|
|
if [ $doNotSetRows == 0 ] && [ -n "${TERM}" ]; then
|
|
|
|
rows=$(tput lines)
|
|
|
|
if [ -n "$rows" ]; then
|
|
|
|
args=( $@ )
|
|
|
|
args=("${args[@]}" "-rows" "$rows")
|
|
|
|
set -- "${args[@]}"
|
|
|
|
fi
|
|
|
|
fi
|
|
|
|
HADOOP_CLASSNAME=org.apache.hadoop.yarn.client.cli.TopCLI
|
|
|
|
hadoop_debug "Append YARN_CLIENT_OPTS onto HADOOP_OPTS"
|
|
|
|
HADOOP_OPTS="${HADOOP_OPTS} ${YARN_CLIENT_OPTS}"
|
2016-05-26 19:43:08 -04:00
|
|
|
HADOOP_SUBCMD_ARGS=("$@")
|
2016-03-28 12:00:07 -04:00
|
|
|
;;
|
|
|
|
*)
|
|
|
|
HADOOP_CLASSNAME="${subcmd}"
|
|
|
|
if ! hadoop_validate_classname "${HADOOP_CLASSNAME}"; then
|
|
|
|
hadoop_exit_with_usage 1
|
|
|
|
fi
|
|
|
|
;;
|
|
|
|
esac
|
|
|
|
}
|
2014-08-19 08:11:17 -04:00
|
|
|
|
|
|
|
# let's locate libexec...
|
2016-03-24 11:47:00 -04:00
|
|
|
if [[ -n "${HADOOP_HOME}" ]]; then
|
|
|
|
HADOOP_DEFAULT_LIBEXEC_DIR="${HADOOP_HOME}/libexec"
|
2014-08-19 08:11:17 -04:00
|
|
|
else
|
2015-07-16 19:58:11 -04:00
|
|
|
bin=$(cd -P -- "$(dirname -- "${MYNAME}")" >/dev/null && pwd -P)
|
2015-11-04 05:26:17 -05:00
|
|
|
HADOOP_DEFAULT_LIBEXEC_DIR="${bin}/../libexec"
|
2014-08-19 08:11:17 -04:00
|
|
|
fi
|
|
|
|
|
2015-11-04 05:26:17 -05:00
|
|
|
HADOOP_LIBEXEC_DIR="${HADOOP_LIBEXEC_DIR:-$HADOOP_DEFAULT_LIBEXEC_DIR}"
|
2014-08-19 08:11:17 -04:00
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_NEW_CONFIG=true
|
|
|
|
if [[ -f "${HADOOP_LIBEXEC_DIR}/yarn-config.sh" ]]; then
|
|
|
|
. "${HADOOP_LIBEXEC_DIR}/yarn-config.sh"
|
|
|
|
else
|
|
|
|
echo "ERROR: Cannot execute ${HADOOP_LIBEXEC_DIR}/yarn-config.sh." 2>&1
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
exit 1
|
|
|
|
fi
|
|
|
|
|
2014-08-19 08:11:17 -04:00
|
|
|
# if no args specified, show usage
|
|
|
|
if [[ $# = 0 ]]; then
|
|
|
|
hadoop_exit_with_usage 1
|
|
|
|
fi
|
|
|
|
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
# get arguments
|
2016-03-28 12:00:07 -04:00
|
|
|
HADOOP_SUBCMD=$1
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
shift
|
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
HADOOP_SUBCMD_ARGS=("$@")
|
|
|
|
|
|
|
|
if declare -f yarn_subcommand_"${HADOOP_SUBCMD}" >/dev/null 2>&1; then
|
|
|
|
hadoop_debug "Calling dynamically: yarn_subcommand_${HADOOP_SUBCMD} ${HADOOP_SUBCMD_ARGS[*]}"
|
|
|
|
"yarn_subcommand_${HADOOP_SUBCMD}" "${HADOOP_SUBCMD_ARGS[@]}"
|
|
|
|
else
|
|
|
|
yarncmd_case "${HADOOP_SUBCMD}" "${HADOOP_SUBCMD_ARGS[@]}"
|
|
|
|
fi
|
2013-02-22 13:36:24 -05:00
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
hadoop_verify_user "${HADOOP_SUBCMD}"
|
2014-12-10 16:41:28 -05:00
|
|
|
|
2016-06-28 08:53:03 -04:00
|
|
|
if [[ ${HADOOP_WORKER_MODE} = true ]]; then
|
|
|
|
hadoop_common_worker_mode_execute "${HADOOP_YARN_HOME}/bin/yarn" "${HADOOP_USER_PARAMS[@]}"
|
2015-02-12 21:01:28 -05:00
|
|
|
exit $?
|
|
|
|
fi
|
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
if [[ "${HADOOP_SUBCMD_SECURESERVICE}" = true ]]; then
|
|
|
|
HADOOP_SECURE_USER="${HADOOP_SUBCMD_SECUREUSER}"
|
|
|
|
hadoop_verify_secure_prereq
|
|
|
|
hadoop_setup_secure_service
|
|
|
|
priv_outfile="${HADOOP_LOG_DIR}/privileged-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}-${HOSTNAME}.out"
|
|
|
|
priv_errfile="${HADOOP_LOG_DIR}/privileged-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}-${HOSTNAME}.err"
|
|
|
|
priv_pidfile="${HADOOP_PID_DIR}/privileged-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}.pid"
|
|
|
|
daemon_outfile="${HADOOP_LOG_DIR}/hadoop-${HADOOP_SECURE_USER}-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}-${HOSTNAME}.out"
|
|
|
|
daemon_pidfile="${HADOOP_PID_DIR}/hadoop-${HADOOP_SECURE_USER}-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}.pid"
|
|
|
|
else
|
|
|
|
daemon_outfile="${HADOOP_LOG_DIR}/hadoop-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}-${HOSTNAME}.out"
|
|
|
|
daemon_pidfile="${HADOOP_PID_DIR}/hadoop-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}.pid"
|
|
|
|
fi
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
|
2014-08-19 08:11:17 -04:00
|
|
|
if [[ "${HADOOP_DAEMON_MODE}" != "default" ]]; then
|
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_ROOT_LOGGER="${HADOOP_DAEMON_ROOT_LOGGER}"
|
2016-03-28 12:00:07 -04:00
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_LOGFILE="hadoop-${HADOOP_IDENT_STRING}-${HADOOP_SUBCMD}-${HOSTNAME}.log"
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
fi
|
|
|
|
|
2014-08-19 08:11:17 -04:00
|
|
|
hadoop_finalize
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
|
2016-03-28 12:00:07 -04:00
|
|
|
if [[ "${HADOOP_SUBCMD_SUPPORTDAEMONIZATION}" = true ]]; then
|
|
|
|
if [[ "${HADOOP_SUBCMD_SECURESERVICE}" = true ]]; then
|
|
|
|
hadoop_secure_daemon_handler \
|
|
|
|
"${HADOOP_DAEMON_MODE}" \
|
|
|
|
"${HADOOP_SUBCMD}" \
|
|
|
|
"${HADOOP_CLASSNAME}" \
|
|
|
|
"${daemon_pidfile}" \
|
|
|
|
"${daemon_outfile}" \
|
|
|
|
"${priv_pidfile}" \
|
|
|
|
"${priv_outfile}" \
|
|
|
|
"${priv_errfile}" \
|
|
|
|
"${HADOOP_SUBCMD_ARGS[@]}"
|
2014-02-26 21:18:34 -05:00
|
|
|
else
|
2016-03-28 12:00:07 -04:00
|
|
|
hadoop_daemon_handler \
|
|
|
|
"${HADOOP_DAEMON_MODE}" \
|
|
|
|
"${HADOOP_SUBCMD}" \
|
|
|
|
"${HADOOP_CLASSNAME}" \
|
|
|
|
"${daemon_pidfile}" \
|
|
|
|
"${daemon_outfile}" \
|
|
|
|
"${HADOOP_SUBCMD_ARGS[@]}"
|
2014-02-26 21:18:34 -05:00
|
|
|
fi
|
2014-08-19 08:11:17 -04:00
|
|
|
exit $?
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
else
|
2016-03-28 12:00:07 -04:00
|
|
|
# shellcheck disable=SC2086
|
|
|
|
hadoop_java_exec "${HADOOP_SUBCMD}" "${HADOOP_CLASSNAME}" "${HADOOP_SUBCMD_ARGS[@]}"
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 07:07:10 -04:00
|
|
|
fi
|