HADOOP-11081. Document hadoop properties expected to be set by the shell code in *-env.sh (aw)
This commit is contained in:
parent
0c588904f8
commit
031f980e7e
|
@ -145,6 +145,9 @@ Trunk (Unreleased)
|
||||||
|
|
||||||
HADOOP-10926. Improve smart-apply-patch.sh to apply binary diffs (cmccabe)
|
HADOOP-10926. Improve smart-apply-patch.sh to apply binary diffs (cmccabe)
|
||||||
|
|
||||||
|
HADOOP-11081. Document hadoop properties expected to be set by the shell
|
||||||
|
code in *-env.sh (aw)
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
||||||
HADOOP-9451. Fault single-layer config if node group topology is enabled.
|
HADOOP-9451. Fault single-layer config if node group topology is enabled.
|
||||||
|
|
|
@ -37,31 +37,38 @@
|
||||||
# JAVA_HOME=/usr/java/testing hdfs dfs -ls
|
# JAVA_HOME=/usr/java/testing hdfs dfs -ls
|
||||||
#
|
#
|
||||||
# Therefore, the vast majority (BUT NOT ALL!) of these defaults
|
# Therefore, the vast majority (BUT NOT ALL!) of these defaults
|
||||||
# are configured for substitution and not append. If you would
|
# are configured for substitution and not append. If append
|
||||||
# like append, you'll # need to modify this file accordingly.
|
# is preferable, modify this file accordingly.
|
||||||
|
|
||||||
###
|
###
|
||||||
# Generic settings for HADOOP
|
# Generic settings for HADOOP
|
||||||
###
|
###
|
||||||
|
|
||||||
# Technically, the only required environment variable is JAVA_HOME.
|
# Technically, the only required environment variable is JAVA_HOME.
|
||||||
# All others are optional. However, our defaults are probably not
|
# All others are optional. However, the defaults are probably not
|
||||||
# your defaults. Many sites configure these options outside of Hadoop,
|
# preferred. Many sites configure these options outside of Hadoop,
|
||||||
# such as in /etc/profile.d
|
# such as in /etc/profile.d
|
||||||
|
|
||||||
# The java implementation to use.
|
# The java implementation to use. By default, this environment
|
||||||
export JAVA_HOME=${JAVA_HOME:-"hadoop-env.sh is not configured"}
|
# variable is REQUIRED on ALL platforms except OS X!
|
||||||
|
# export JAVA_HOME=
|
||||||
|
|
||||||
|
# Location of Hadoop. By default, Hadoop will attempt to determine
|
||||||
|
# this location based upon its execution path.
|
||||||
|
# export HADOOP_PREFIX=
|
||||||
|
|
||||||
# Location of Hadoop's configuration information. i.e., where this
|
# Location of Hadoop's configuration information. i.e., where this
|
||||||
# file is probably living. You will almost certainly want to set
|
# file is probably living. Many sites will also set this in the
|
||||||
# this in /etc/profile.d or equivalent.
|
# same location where JAVA_HOME is defined. If this is not defined
|
||||||
|
# Hadoop will attempt to locate it based upon its execution
|
||||||
|
# path.
|
||||||
# export HADOOP_CONF_DIR=$HADOOP_PREFIX/etc/hadoop
|
# export HADOOP_CONF_DIR=$HADOOP_PREFIX/etc/hadoop
|
||||||
|
|
||||||
# The maximum amount of heap to use, in MB. Default is 1024.
|
# The maximum amount of heap to use, in MB. Default is 1024.
|
||||||
# export HADOOP_HEAPSIZE=1024
|
# export HADOOP_HEAPSIZE=1024
|
||||||
|
|
||||||
# Extra Java runtime options for all Hadoop commands. We don't support
|
# Extra Java runtime options for all Hadoop commands. We don't support
|
||||||
# IPv6 yet/still, so by default we set preference to IPv4.
|
# IPv6 yet/still, so by default the preference is set to IPv4.
|
||||||
# export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
|
# export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
|
||||||
|
|
||||||
# Some parts of the shell code may do special things dependent upon
|
# Some parts of the shell code may do special things dependent upon
|
||||||
|
@ -72,8 +79,8 @@ export HADOOP_OS_TYPE=${HADOOP_OS_TYPE:-$(uname -s)}
|
||||||
|
|
||||||
# Under certain conditions, Java on OS X will throw SCDynamicStore errors
|
# Under certain conditions, Java on OS X will throw SCDynamicStore errors
|
||||||
# in the system logs.
|
# in the system logs.
|
||||||
# See HADOOP-8719 for more information. If you need Kerberos
|
# See HADOOP-8719 for more information. If one needs Kerberos
|
||||||
# support on OS X, you'll want to change/remove this extra bit.
|
# support on OS X, one will want to change/remove this extra bit.
|
||||||
case ${HADOOP_OS_TYPE} in
|
case ${HADOOP_OS_TYPE} in
|
||||||
Darwin*)
|
Darwin*)
|
||||||
export HADOOP_OPTS="${HADOOP_OPTS} -Djava.security.krb5.realm= "
|
export HADOOP_OPTS="${HADOOP_OPTS} -Djava.security.krb5.realm= "
|
||||||
|
@ -82,11 +89,11 @@ case ${HADOOP_OS_TYPE} in
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
# Extra Java runtime options for Hadoop clients (i.e., hdfs dfs -blah)
|
# Extra Java runtime options for some Hadoop commands
|
||||||
# These get added to HADOOP_OPTS for such commands. In most cases,
|
# and clients (i.e., hdfs dfs -blah). These get appended to HADOOP_OPTS for
|
||||||
# this should be left empty and let users supply it on the
|
# such commands. In most cases, # this should be left empty and
|
||||||
# command line.
|
# let users supply it on the command line.
|
||||||
# extra HADOOP_CLIENT_OPTS=""
|
# export HADOOP_CLIENT_OPTS=""
|
||||||
|
|
||||||
#
|
#
|
||||||
# A note about classpaths.
|
# A note about classpaths.
|
||||||
|
@ -149,20 +156,22 @@ esac
|
||||||
#
|
#
|
||||||
|
|
||||||
#
|
#
|
||||||
# You can define variables right here and then re-use them later on.
|
# Many options may also be specified as Java properties. It is
|
||||||
# For example, it is common to use the same garbage collection settings
|
# very common, and in many cases, desirable, to hard-set these
|
||||||
# for all the daemons. So we could define:
|
# in daemon _OPTS variables. Where applicable, the appropriate
|
||||||
|
# Java property is also identified. Note that many are re-used
|
||||||
|
# or set differently in certain contexts (e.g., secure vs
|
||||||
|
# non-secure)
|
||||||
#
|
#
|
||||||
# export HADOOP_GC_SETTINGS="-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps"
|
|
||||||
#
|
|
||||||
# .. and then use it as per the b option under the namenode.
|
|
||||||
|
|
||||||
# Where (primarily) daemon log files are stored.
|
# Where (primarily) daemon log files are stored. # $HADOOP_PREFIX/logs
|
||||||
# $HADOOP_PREFIX/logs by default.
|
# by default.
|
||||||
|
# Java property: hadoop.log.dir
|
||||||
# export HADOOP_LOG_DIR=${HADOOP_PREFIX}/logs
|
# export HADOOP_LOG_DIR=${HADOOP_PREFIX}/logs
|
||||||
|
|
||||||
# A string representing this instance of hadoop. $USER by default.
|
# A string representing this instance of hadoop. $USER by default.
|
||||||
# This is used in writing log and pid files, so keep that in mind!
|
# This is used in writing log and pid files, so keep that in mind!
|
||||||
|
# Java property: hadoop.id.str
|
||||||
# export HADOOP_IDENT_STRING=$USER
|
# export HADOOP_IDENT_STRING=$USER
|
||||||
|
|
||||||
# How many seconds to pause after stopping a daemon
|
# How many seconds to pause after stopping a daemon
|
||||||
|
@ -171,23 +180,26 @@ esac
|
||||||
# Where pid files are stored. /tmp by default.
|
# Where pid files are stored. /tmp by default.
|
||||||
# export HADOOP_PID_DIR=/tmp
|
# export HADOOP_PID_DIR=/tmp
|
||||||
|
|
||||||
# Default log level and output location
|
# Default log4j setting for interactive commands
|
||||||
# This sets the hadoop.root.logger property
|
# Java property: hadoop.root.logger
|
||||||
# export HADOOP_ROOT_LOGGER=INFO,console
|
# export HADOOP_ROOT_LOGGER=INFO,console
|
||||||
|
|
||||||
# Default log level for daemons spawned explicitly by
|
# Default log4j setting for daemons spawned explicitly by
|
||||||
# --daemon option of hadoop, hdfs, mapred and yarn command.
|
# --daemon option of hadoop, hdfs, mapred and yarn command.
|
||||||
# This sets the hadoop.root.logger property
|
# Java property: hadoop.root.logger
|
||||||
# export HADOOP_DAEMON_ROOT_LOGGER=INFO,RFA
|
# export HADOOP_DAEMON_ROOT_LOGGER=INFO,RFA
|
||||||
|
|
||||||
# Default log level and output location for security-related messages.
|
# Default log level and output location for security-related messages.
|
||||||
# It sets -Dhadoop.security.logger on the command line.
|
# You will almost certainly want to change this on a per-daemon basis via
|
||||||
# You will almost certainly want to change this on a per-daemon basis!
|
# the Java property (i.e., -Dhadoop.security.logger=foo). (Note that the
|
||||||
|
# defaults for the NN and 2NN override this by default.)
|
||||||
|
# Java property: hadoop.security.logger
|
||||||
# export HADOOP_SECURITY_LOGGER=INFO,NullAppender
|
# export HADOOP_SECURITY_LOGGER=INFO,NullAppender
|
||||||
|
|
||||||
# Default log level for file system audit messages.
|
# Default log level for file system audit messages.
|
||||||
# It sets -Dhdfs.audit.logger on the command line.
|
# Generally, this is specifically set in the namenode-specific
|
||||||
# You will almost certainly want to change this on a per-daemon basis!
|
# options line.
|
||||||
|
# Java property: hdfs.audit.logger
|
||||||
# export HADOOP_AUDIT_LOGGER=INFO,NullAppender
|
# export HADOOP_AUDIT_LOGGER=INFO,NullAppender
|
||||||
|
|
||||||
# Default process priority level
|
# Default process priority level
|
||||||
|
@ -195,8 +207,19 @@ esac
|
||||||
# export HADOOP_NICENESS=0
|
# export HADOOP_NICENESS=0
|
||||||
|
|
||||||
# Default name for the service level authorization file
|
# Default name for the service level authorization file
|
||||||
|
# Java property: hadoop.policy.file
|
||||||
# export HADOOP_POLICYFILE="hadoop-policy.xml"
|
# export HADOOP_POLICYFILE="hadoop-policy.xml"
|
||||||
|
|
||||||
|
#
|
||||||
|
# NOTE: this is not used by default! <-----
|
||||||
|
# You can define variables right here and then re-use them later on.
|
||||||
|
# For example, it is common to use the same garbage collection settings
|
||||||
|
# for all the daemons. So one could define:
|
||||||
|
#
|
||||||
|
# export HADOOP_GC_SETTINGS="-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps"
|
||||||
|
#
|
||||||
|
# .. and then use it as per the b option under the namenode.
|
||||||
|
|
||||||
###
|
###
|
||||||
# Secure/privileged execution
|
# Secure/privileged execution
|
||||||
###
|
###
|
||||||
|
@ -219,18 +242,26 @@ esac
|
||||||
|
|
||||||
#
|
#
|
||||||
# This directory contains the logs for secure and privileged processes.
|
# This directory contains the logs for secure and privileged processes.
|
||||||
|
# Java property: hadoop.log.dir
|
||||||
# export HADOOP_SECURE_LOG=${HADOOP_LOG_DIR}
|
# export HADOOP_SECURE_LOG=${HADOOP_LOG_DIR}
|
||||||
|
|
||||||
#
|
#
|
||||||
# When running a secure daemon, the default value of HADOOP_IDENT_STRING
|
# When running a secure daemon, the default value of HADOOP_IDENT_STRING
|
||||||
# ends up being a bit bogus. Therefore, by default, the code will
|
# ends up being a bit bogus. Therefore, by default, the code will
|
||||||
# replace HADOOP_IDENT_STRING with HADOOP_SECURE_xx_USER. If you want
|
# replace HADOOP_IDENT_STRING with HADOOP_SECURE_xx_USER. If one wants
|
||||||
# to keep HADOOP_IDENT_STRING untouched, then uncomment this line.
|
# to keep HADOOP_IDENT_STRING untouched, then uncomment this line.
|
||||||
# export HADOOP_SECURE_IDENT_PRESERVE="true"
|
# export HADOOP_SECURE_IDENT_PRESERVE="true"
|
||||||
|
|
||||||
###
|
###
|
||||||
# NameNode specific parameters
|
# NameNode specific parameters
|
||||||
###
|
###
|
||||||
|
|
||||||
|
# Default log level and output location for file system related change
|
||||||
|
# messages. For non-namenode daemons, the Java property must be set in
|
||||||
|
# the appropriate _OPTS if one wants something other than INFO,NullAppender
|
||||||
|
# Java property: hdfs.audit.logger
|
||||||
|
# export HDFS_AUDIT_LOGGER=INFO,NullAppender
|
||||||
|
|
||||||
# Specify the JVM options to be used when starting the NameNode.
|
# Specify the JVM options to be used when starting the NameNode.
|
||||||
# These options will be appended to the options specified as HADOOP_OPTS
|
# These options will be appended to the options specified as HADOOP_OPTS
|
||||||
# and therefore may override any similar flags set in HADOOP_OPTS
|
# and therefore may override any similar flags set in HADOOP_OPTS
|
||||||
|
@ -245,7 +276,7 @@ esac
|
||||||
# export HADOOP_NAMENODE_OPTS="-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:${HADOOP_LOG_DIR}/gc-rm.log-$(date +'%Y%m%d%H%M')"
|
# export HADOOP_NAMENODE_OPTS="-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:${HADOOP_LOG_DIR}/gc-rm.log-$(date +'%Y%m%d%H%M')"
|
||||||
|
|
||||||
# this is the default:
|
# this is the default:
|
||||||
# export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender"
|
# export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=INFO,RFAS"
|
||||||
|
|
||||||
###
|
###
|
||||||
# SecondaryNameNode specific parameters
|
# SecondaryNameNode specific parameters
|
||||||
|
@ -255,7 +286,7 @@ esac
|
||||||
# and therefore may override any similar flags set in HADOOP_OPTS
|
# and therefore may override any similar flags set in HADOOP_OPTS
|
||||||
#
|
#
|
||||||
# This is the default:
|
# This is the default:
|
||||||
# export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender"
|
# export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=INFO,RFAS"
|
||||||
|
|
||||||
###
|
###
|
||||||
# DataNode specific parameters
|
# DataNode specific parameters
|
||||||
|
@ -272,14 +303,16 @@ esac
|
||||||
# to provide authentication of data transfer protocol. This **MUST NOT** be
|
# to provide authentication of data transfer protocol. This **MUST NOT** be
|
||||||
# defined if SASL is configured for authentication of data transfer protocol
|
# defined if SASL is configured for authentication of data transfer protocol
|
||||||
# using non-privileged ports.
|
# using non-privileged ports.
|
||||||
|
# This will replace the hadoop.id.str Java property in secure mode.
|
||||||
# export HADOOP_SECURE_DN_USER=hdfs
|
# export HADOOP_SECURE_DN_USER=hdfs
|
||||||
|
|
||||||
# Supplemental options for secure datanodes
|
# Supplemental options for secure datanodes
|
||||||
# By default, we use jsvc which needs to know to launch a
|
# By default, Hadoop uses jsvc which needs to know to launch a
|
||||||
# server jvm.
|
# server jvm.
|
||||||
# export HADOOP_DN_SECURE_EXTRA_OPTS="-jvm server"
|
# export HADOOP_DN_SECURE_EXTRA_OPTS="-jvm server"
|
||||||
|
|
||||||
# Where datanode log files are stored in the secure data environment.
|
# Where datanode log files are stored in the secure data environment.
|
||||||
|
# This will replace the hadoop.log.dir Java property in secure mode.
|
||||||
# export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_SECURE_LOG_DIR}
|
# export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_SECURE_LOG_DIR}
|
||||||
|
|
||||||
# Where datanode pid files are stored in the secure data environment.
|
# Where datanode pid files are stored in the secure data environment.
|
||||||
|
@ -301,11 +334,12 @@ esac
|
||||||
# export HADOOP_PORTMAP_OPTS="-Xmx512m"
|
# export HADOOP_PORTMAP_OPTS="-Xmx512m"
|
||||||
|
|
||||||
# Supplemental options for priviliged gateways
|
# Supplemental options for priviliged gateways
|
||||||
# By default, we use jsvc which needs to know to launch a
|
# By default, Hadoop uses jsvc which needs to know to launch a
|
||||||
# server jvm.
|
# server jvm.
|
||||||
# export HADOOP_NFS3_SECURE_EXTRA_OPTS="-jvm server"
|
# export HADOOP_NFS3_SECURE_EXTRA_OPTS="-jvm server"
|
||||||
|
|
||||||
# On privileged gateways, user to run the gateway as after dropping privileges
|
# On privileged gateways, user to run the gateway as after dropping privileges
|
||||||
|
# This will replace the hadoop.id.str Java property in secure mode.
|
||||||
# export HADOOP_PRIVILEGED_NFS_USER=nfsserver
|
# export HADOOP_PRIVILEGED_NFS_USER=nfsserver
|
||||||
|
|
||||||
###
|
###
|
||||||
|
@ -349,7 +383,7 @@ esac
|
||||||
###
|
###
|
||||||
|
|
||||||
#
|
#
|
||||||
# When building Hadoop, you can add the class paths to your commands
|
# When building Hadoop, one can add the class paths to the commands
|
||||||
# via this special env var:
|
# via this special env var:
|
||||||
# HADOOP_ENABLE_BUILD_PATHS="true"
|
# export HADOOP_ENABLE_BUILD_PATHS="true"
|
||||||
|
|
||||||
|
|
|
@ -180,6 +180,7 @@ case ${COMMAND} in
|
||||||
CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
|
CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
|
||||||
hadoop_debug "Appending HADOOP_NAMENODE_OPTS onto HADOOP_OPTS"
|
hadoop_debug "Appending HADOOP_NAMENODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NAMENODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NAMENODE_OPTS}"
|
||||||
|
hadoop_add_param HADOOP_OPTS hdfs.audit.logger "-Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER}"
|
||||||
;;
|
;;
|
||||||
nfs3)
|
nfs3)
|
||||||
supportdaemonization="true"
|
supportdaemonization="true"
|
||||||
|
@ -221,6 +222,7 @@ case ${COMMAND} in
|
||||||
CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
|
CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
|
||||||
hadoop_debug "Appending HADOOP_SECONDARYNAMENODE_OPTS onto HADOOP_OPTS"
|
hadoop_debug "Appending HADOOP_SECONDARYNAMENODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_SECONDARYNAMENODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_SECONDARYNAMENODE_OPTS}"
|
||||||
|
hadoop_add_param HADOOP_OPTS hdfs.audit.logger "-Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER}"
|
||||||
;;
|
;;
|
||||||
snapshotDiff)
|
snapshotDiff)
|
||||||
CLASS=org.apache.hadoop.hdfs.tools.snapshot.SnapshotDiff
|
CLASS=org.apache.hadoop.hdfs.tools.snapshot.SnapshotDiff
|
||||||
|
|
|
@ -57,10 +57,11 @@ function hadoop_subproject_init
|
||||||
HADOOP_IDENT_STRING="${HADOOP_HDFS_IDENT_STRING:-$HADOOP_IDENT_STRING}"
|
HADOOP_IDENT_STRING="${HADOOP_HDFS_IDENT_STRING:-$HADOOP_IDENT_STRING}"
|
||||||
HADOOP_HDFS_IDENT_STRING="${HADOOP_IDENT_STRING}"
|
HADOOP_HDFS_IDENT_STRING="${HADOOP_IDENT_STRING}"
|
||||||
|
|
||||||
# turn on the defaults
|
|
||||||
|
|
||||||
export HADOOP_NAMENODE_OPTS=${HADOOP_NAMENODE_OPTS:-"-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender"}
|
# turn on the defaults
|
||||||
export HADOOP_SECONDARYNAMENODE_OPTS=${HADOOP_SECONDARYNAMENODE_OPTS:-"-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender"}
|
export HDFS_AUDIT_LOGGER=${HDFS_AUDIT_LOGGER:-INFO,NullAppender}
|
||||||
|
export HADOOP_NAMENODE_OPTS=${HADOOP_NAMENODE_OPTS:-"-Dhadoop.security.logger=INFO,RFAS"}
|
||||||
|
export HADOOP_SECONDARYNAMENODE_OPTS=${HADOOP_SECONDARYNAMENODE_OPTS:-"-Dhadoop.security.logger=INFO,RFAS"}
|
||||||
export HADOOP_DATANODE_OPTS=${HADOOP_DATANODE_OPTS:-"-Dhadoop.security.logger=ERROR,RFAS"}
|
export HADOOP_DATANODE_OPTS=${HADOOP_DATANODE_OPTS:-"-Dhadoop.security.logger=ERROR,RFAS"}
|
||||||
export HADOOP_DN_SECURE_EXTRA_OPTS=${HADOOP_DN_SECURE_EXTRA_OPTS:-"-jvm server"}
|
export HADOOP_DN_SECURE_EXTRA_OPTS=${HADOOP_DN_SECURE_EXTRA_OPTS:-"-jvm server"}
|
||||||
export HADOOP_NFS3_SECURE_EXTRA_OPTS=${HADOOP_NFS3_SECURE_EXTRA_OPTS:-"-jvm server"}
|
export HADOOP_NFS3_SECURE_EXTRA_OPTS=${HADOOP_NFS3_SECURE_EXTRA_OPTS:-"-jvm server"}
|
||||||
|
|
|
@ -29,9 +29,11 @@
|
||||||
###
|
###
|
||||||
|
|
||||||
#Override the log4j settings for all MR apps
|
#Override the log4j settings for all MR apps
|
||||||
|
# Java property: hadoop.root.logger
|
||||||
# export MAPRED_ROOT_LOGGER="INFO,console"
|
# export MAPRED_ROOT_LOGGER="INFO,console"
|
||||||
|
|
||||||
# Override Hadoop's log directory & file
|
# Override Hadoop's log directory & file
|
||||||
|
# Java property: hadoop.log.dir
|
||||||
# export HADOOP_MAPRED_LOG_DIR=""
|
# export HADOOP_MAPRED_LOG_DIR=""
|
||||||
|
|
||||||
# Override Hadoop's pid directory
|
# Override Hadoop's pid directory
|
||||||
|
@ -39,6 +41,7 @@
|
||||||
|
|
||||||
# Override Hadoop's identity string. $USER by default.
|
# Override Hadoop's identity string. $USER by default.
|
||||||
# This is used in writing log and pid files, so keep that in mind!
|
# This is used in writing log and pid files, so keep that in mind!
|
||||||
|
# Java property: hadoop.id.str
|
||||||
# export HADOOP_MAPRED_IDENT_STRING=$USER
|
# export HADOOP_MAPRED_IDENT_STRING=$USER
|
||||||
|
|
||||||
# Override Hadoop's process priority
|
# Override Hadoop's process priority
|
||||||
|
@ -54,8 +57,8 @@
|
||||||
# the value to 1000.
|
# the value to 1000.
|
||||||
# This value will be overridden by an Xmx setting specified in either
|
# This value will be overridden by an Xmx setting specified in either
|
||||||
# MAPRED_OPTS, HADOOP_OPTS, and/or HADOOP_JOB_HISTORYSERVER_OPTS.
|
# MAPRED_OPTS, HADOOP_OPTS, and/or HADOOP_JOB_HISTORYSERVER_OPTS.
|
||||||
# If not specified, the default value will be picked from either YARN_HEAPMAX
|
# If not specified, the default value will be picked from either HADOOP_HEAPSIZE
|
||||||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.
|
# or the built-in default.
|
||||||
#
|
#
|
||||||
#export HADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000
|
#export HADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000
|
||||||
|
|
||||||
|
@ -65,6 +68,7 @@
|
||||||
#export HADOOP_JOB_HISTORYSERVER_OPTS=
|
#export HADOOP_JOB_HISTORYSERVER_OPTS=
|
||||||
|
|
||||||
# Specify the log4j settings for the JobHistoryServer
|
# Specify the log4j settings for the JobHistoryServer
|
||||||
|
# Java property: hadoop.root.logger
|
||||||
#export HADOOP_JHS_LOGGER=INFO,RFA
|
#export HADOOP_JHS_LOGGER=INFO,RFA
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -31,15 +31,18 @@
|
||||||
|
|
||||||
# By default, YARN will use HADOOP_LOG_DIR for YARN logging. Specify a custom
|
# By default, YARN will use HADOOP_LOG_DIR for YARN logging. Specify a custom
|
||||||
# log directory for YARN things here:
|
# log directory for YARN things here:
|
||||||
|
# Java properties: hadoop.log.dir, yarn.log.dir
|
||||||
# export YARN_LOG_DIR="${HADOOP_LOG_DIR}"
|
# export YARN_LOG_DIR="${HADOOP_LOG_DIR}"
|
||||||
|
|
||||||
# By default, YARN will use the value of HADOOP_LOGFILE as the 'fallback' log
|
# By default, YARN will use the value of HADOOP_LOGFILE as the 'fallback' log
|
||||||
# file # when log4j settings are not defined. Specify a custom YARN log file
|
# file # when log4j settings are not defined. Specify a custom YARN log file
|
||||||
# here:
|
# here:
|
||||||
|
# Java properties: hadoop.log.file, yarn.log.file
|
||||||
# export YARN_LOGFILE=${HADOOP_LOGFILE}
|
# export YARN_LOGFILE=${HADOOP_LOGFILE}
|
||||||
|
|
||||||
#Override the log4j settings for all YARN apps By default, YARN will use
|
#Override the log4j settings for all YARN apps By default, YARN will use
|
||||||
# HADOOP_ROOT_LOGGER.
|
# HADOOP_ROOT_LOGGER.
|
||||||
|
# Java properties: hadoop.root.logger, yarn.root.logger
|
||||||
# export YARN_ROOT_LOGGER=${HADOOP_ROOT_LOGGER}
|
# export YARN_ROOT_LOGGER=${HADOOP_ROOT_LOGGER}
|
||||||
|
|
||||||
###
|
###
|
||||||
|
@ -51,8 +54,8 @@
|
||||||
# the value to 1000.
|
# the value to 1000.
|
||||||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
||||||
# HADOOP_OPTS, and/or YARN_RESOURCEMANAGER_OPTS.
|
# HADOOP_OPTS, and/or YARN_RESOURCEMANAGER_OPTS.
|
||||||
# If not specified, the default value will be picked from either YARN_HEAPMAX
|
# If not specified, the default value will be picked from either HADOOP_HEAPSIZE
|
||||||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.
|
# or the built-in default.
|
||||||
#
|
#
|
||||||
#export YARN_RESOURCEMANAGER_HEAPSIZE=1000
|
#export YARN_RESOURCEMANAGER_HEAPSIZE=1000
|
||||||
|
|
||||||
|
@ -85,8 +88,8 @@
|
||||||
# the value to 1000.
|
# the value to 1000.
|
||||||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
||||||
# HADOOP_OPTS, and/or YARN_NODEMANAGER_OPTS.
|
# HADOOP_OPTS, and/or YARN_NODEMANAGER_OPTS.
|
||||||
# If not specified, the default value will be picked from either YARN_HEAPMAX
|
# If not specified, the default value will be picked from either HADOOP_HEAPSIZE
|
||||||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.
|
# or the built-in default.
|
||||||
#
|
#
|
||||||
#export YARN_NODEMANAGER_HEAPSIZE=1000
|
#export YARN_NODEMANAGER_HEAPSIZE=1000
|
||||||
|
|
||||||
|
@ -107,8 +110,8 @@
|
||||||
# the value to 1000.
|
# the value to 1000.
|
||||||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
||||||
# HADOOP_OPTS, and/or YARN_TIMELINESERVER_OPTS.
|
# HADOOP_OPTS, and/or YARN_TIMELINESERVER_OPTS.
|
||||||
# If not specified, the default value will be picked from either YARN_HEAPMAX
|
# If not specified, the default value will be picked from either HADOOP_HEAPSIZE
|
||||||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.
|
# or the built-in default.
|
||||||
#
|
#
|
||||||
#export YARN_TIMELINESERVER_HEAPSIZE=1000
|
#export YARN_TIMELINESERVER_HEAPSIZE=1000
|
||||||
|
|
||||||
|
@ -129,8 +132,8 @@
|
||||||
# the value to 1000.
|
# the value to 1000.
|
||||||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
# This value will be overridden by an Xmx setting specified in either YARN_OPTS,
|
||||||
# HADOOP_OPTS, and/or YARN_PROXYSERVER_OPTS.
|
# HADOOP_OPTS, and/or YARN_PROXYSERVER_OPTS.
|
||||||
# If not specified, the default value will be picked from either YARN_HEAPMAX
|
# If not specified, the default value will be picked from either HADOOP_HEAPSIZE
|
||||||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.
|
# or the built-in default.
|
||||||
#
|
#
|
||||||
#export YARN_PROXYSERVER_HEAPSIZE=1000
|
#export YARN_PROXYSERVER_HEAPSIZE=1000
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue