Merge remote-tracking branch 'apache-commit/trunk' into HDFS-6581
Conflicts: hadoop-common-project/hadoop-common/CHANGES.txt
This commit is contained in:
commit
7e32be8768
|
@ -125,6 +125,8 @@ Trunk (Unreleased)
|
||||||
|
|
||||||
HADOOP-10485. Remove dead classes in hadoop-streaming. (wheat9)
|
HADOOP-10485. Remove dead classes in hadoop-streaming. (wheat9)
|
||||||
|
|
||||||
|
HADOOP-11013. CLASSPATH handling should be consolidated, debuggable (aw)
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
||||||
HADOOP-9451. Fault single-layer config if node group topology is enabled.
|
HADOOP-9451. Fault single-layer config if node group topology is enabled.
|
||||||
|
@ -468,6 +470,9 @@ Release 2.6.0 - UNRELEASED
|
||||||
HADOOP-10880. Move HTTP delegation tokens out of URL querystring to
|
HADOOP-10880. Move HTTP delegation tokens out of URL querystring to
|
||||||
a header. (tucu)
|
a header. (tucu)
|
||||||
|
|
||||||
|
HADOOP-11005. Fix HTTP content type for ReconfigurationServlet.
|
||||||
|
(Lei Xu via wang)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
HADOOP-10838. Byte array native checksumming. (James Thomas via todd)
|
HADOOP-10838. Byte array native checksumming. (James Thomas via todd)
|
||||||
|
|
|
@ -114,6 +114,7 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
archive)
|
archive)
|
||||||
CLASS=org.apache.hadoop.tools.HadoopArchives
|
CLASS=org.apache.hadoop.tools.HadoopArchives
|
||||||
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
hadoop_add_classpath "${TOOL_PATH}"
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
;;
|
;;
|
||||||
checknative)
|
checknative)
|
||||||
|
@ -136,10 +137,12 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
distch)
|
distch)
|
||||||
CLASS=org.apache.hadoop.tools.DistCh
|
CLASS=org.apache.hadoop.tools.DistCh
|
||||||
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
hadoop_add_classpath "${TOOL_PATH}"
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
;;
|
;;
|
||||||
distcp)
|
distcp)
|
||||||
CLASS=org.apache.hadoop.tools.DistCp
|
CLASS=org.apache.hadoop.tools.DistCp
|
||||||
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
hadoop_add_classpath "${TOOL_PATH}"
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
;;
|
;;
|
||||||
fs)
|
fs)
|
||||||
|
@ -168,11 +171,11 @@ case ${COMMAND} in
|
||||||
esac
|
esac
|
||||||
|
|
||||||
# Always respect HADOOP_OPTS and HADOOP_CLIENT_OPTS
|
# Always respect HADOOP_OPTS and HADOOP_CLIENT_OPTS
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
|
|
||||||
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
||||||
|
|
||||||
hadoop_finalize
|
hadoop_finalize
|
||||||
export CLASSPATH
|
|
||||||
hadoop_java_exec "${COMMAND}" "${CLASS}" "$@"
|
hadoop_java_exec "${COMMAND}" "${CLASS}" "$@"
|
||||||
|
|
||||||
|
|
|
@ -129,6 +129,11 @@ while [[ -z "${_hadoop_common_done}" ]]; do
|
||||||
hadoop_exit_with_usage 1
|
hadoop_exit_with_usage 1
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
|
--debug)
|
||||||
|
shift
|
||||||
|
# shellcheck disable=SC2034
|
||||||
|
HADOOP_SHELL_SCRIPT_DEBUG=true
|
||||||
|
;;
|
||||||
--help|-help|-h|help|--h|--\?|-\?|\?)
|
--help|-help|-h|help|--h|--\?|-\?|\?)
|
||||||
hadoop_exit_with_usage 0
|
hadoop_exit_with_usage 0
|
||||||
;;
|
;;
|
||||||
|
|
|
@ -21,6 +21,13 @@ function hadoop_error
|
||||||
echo "$*" 1>&2
|
echo "$*" 1>&2
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function hadoop_debug
|
||||||
|
{
|
||||||
|
if [[ -n "${HADOOP_SHELL_SCRIPT_DEBUG}" ]]; then
|
||||||
|
echo "DEBUG: $*" 1>&2
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
function hadoop_bootstrap_init
|
function hadoop_bootstrap_init
|
||||||
{
|
{
|
||||||
# NOTE: This function is not user replaceable.
|
# NOTE: This function is not user replaceable.
|
||||||
|
@ -62,6 +69,7 @@ function hadoop_bootstrap_init
|
||||||
|
|
||||||
# defaults
|
# defaults
|
||||||
export HADOOP_OPTS=${HADOOP_OPTS:-"-Djava.net.preferIPv4Stack=true"}
|
export HADOOP_OPTS=${HADOOP_OPTS:-"-Djava.net.preferIPv4Stack=true"}
|
||||||
|
hadoop_debug "Initial HADOOP_OPTS=${HADOOP_OPTS}"
|
||||||
}
|
}
|
||||||
|
|
||||||
function hadoop_find_confdir
|
function hadoop_find_confdir
|
||||||
|
@ -80,6 +88,8 @@ function hadoop_find_confdir
|
||||||
conf_dir="etc/hadoop"
|
conf_dir="etc/hadoop"
|
||||||
fi
|
fi
|
||||||
export HADOOP_CONF_DIR="${HADOOP_CONF_DIR:-${HADOOP_PREFIX}/${conf_dir}}"
|
export HADOOP_CONF_DIR="${HADOOP_CONF_DIR:-${HADOOP_PREFIX}/${conf_dir}}"
|
||||||
|
|
||||||
|
hadoop_debug "HADOOP_CONF_DIR=${HADOOP_CONF_DIR}"
|
||||||
}
|
}
|
||||||
|
|
||||||
function hadoop_exec_hadoopenv
|
function hadoop_exec_hadoopenv
|
||||||
|
@ -105,6 +115,7 @@ function hadoop_basic_init
|
||||||
|
|
||||||
# CLASSPATH initially contains $HADOOP_CONF_DIR
|
# CLASSPATH initially contains $HADOOP_CONF_DIR
|
||||||
CLASSPATH="${HADOOP_CONF_DIR}"
|
CLASSPATH="${HADOOP_CONF_DIR}"
|
||||||
|
hadoop_debug "Initial CLASSPATH=${HADOOP_CONF_DIR}"
|
||||||
|
|
||||||
if [[ -z "${HADOOP_COMMON_HOME}" ]] &&
|
if [[ -z "${HADOOP_COMMON_HOME}" ]] &&
|
||||||
[[ -d "${HADOOP_PREFIX}/${HADOOP_COMMON_DIR}" ]]; then
|
[[ -d "${HADOOP_PREFIX}/${HADOOP_COMMON_DIR}" ]]; then
|
||||||
|
@ -116,19 +127,19 @@ function hadoop_basic_init
|
||||||
|
|
||||||
# define HADOOP_HDFS_HOME
|
# define HADOOP_HDFS_HOME
|
||||||
if [[ -z "${HADOOP_HDFS_HOME}" ]] &&
|
if [[ -z "${HADOOP_HDFS_HOME}" ]] &&
|
||||||
[[ -d "${HADOOP_PREFIX}/${HDFS_DIR}" ]]; then
|
[[ -d "${HADOOP_PREFIX}/${HDFS_DIR}" ]]; then
|
||||||
export HADOOP_HDFS_HOME="${HADOOP_PREFIX}"
|
export HADOOP_HDFS_HOME="${HADOOP_PREFIX}"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# define HADOOP_YARN_HOME
|
# define HADOOP_YARN_HOME
|
||||||
if [[ -z "${HADOOP_YARN_HOME}" ]] &&
|
if [[ -z "${HADOOP_YARN_HOME}" ]] &&
|
||||||
[[ -d "${HADOOP_PREFIX}/${YARN_DIR}" ]]; then
|
[[ -d "${HADOOP_PREFIX}/${YARN_DIR}" ]]; then
|
||||||
export HADOOP_YARN_HOME="${HADOOP_PREFIX}"
|
export HADOOP_YARN_HOME="${HADOOP_PREFIX}"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# define HADOOP_MAPRED_HOME
|
# define HADOOP_MAPRED_HOME
|
||||||
if [[ -z "${HADOOP_MAPRED_HOME}" ]] &&
|
if [[ -z "${HADOOP_MAPRED_HOME}" ]] &&
|
||||||
[[ -d "${HADOOP_PREFIX}/${MAPRED_DIR}" ]]; then
|
[[ -d "${HADOOP_PREFIX}/${MAPRED_DIR}" ]]; then
|
||||||
export HADOOP_MAPRED_HOME="${HADOOP_PREFIX}"
|
export HADOOP_MAPRED_HOME="${HADOOP_PREFIX}"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
@ -274,6 +285,9 @@ function hadoop_add_param
|
||||||
if [[ ! ${!1} =~ $2 ]] ; then
|
if [[ ! ${!1} =~ $2 ]] ; then
|
||||||
# shellcheck disable=SC2086
|
# shellcheck disable=SC2086
|
||||||
eval $1="'${!1} $3'"
|
eval $1="'${!1} $3'"
|
||||||
|
hadoop_debug "$1 accepted $3"
|
||||||
|
else
|
||||||
|
hadoop_debug "$1 declined $3"
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -283,8 +297,8 @@ function hadoop_add_classpath
|
||||||
# $1 = directory, file, wildcard, whatever to add
|
# $1 = directory, file, wildcard, whatever to add
|
||||||
# $2 = before or after, which determines where in the
|
# $2 = before or after, which determines where in the
|
||||||
# classpath this object should go. default is after
|
# classpath this object should go. default is after
|
||||||
# return 0 = success
|
# return 0 = success (added or duplicate)
|
||||||
# return 1 = failure (duplicate, doesn't exist, whatever)
|
# return 1 = failure (doesn't exist, whatever)
|
||||||
|
|
||||||
# However, with classpath (& JLP), we can do dedupe
|
# However, with classpath (& JLP), we can do dedupe
|
||||||
# along with some sanity checking (e.g., missing directories)
|
# along with some sanity checking (e.g., missing directories)
|
||||||
|
@ -295,23 +309,29 @@ function hadoop_add_classpath
|
||||||
if [[ $1 =~ ^.*\*$ ]]; then
|
if [[ $1 =~ ^.*\*$ ]]; then
|
||||||
local mp=$(dirname "$1")
|
local mp=$(dirname "$1")
|
||||||
if [[ ! -d "${mp}" ]]; then
|
if [[ ! -d "${mp}" ]]; then
|
||||||
|
hadoop_debug "Rejected CLASSPATH: $1 (not a dir)"
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# no wildcard in the middle, so check existence
|
# no wildcard in the middle, so check existence
|
||||||
# (doesn't matter *what* it is)
|
# (doesn't matter *what* it is)
|
||||||
elif [[ ! $1 =~ ^.*\*.*$ ]] && [[ ! -e "$1" ]]; then
|
elif [[ ! $1 =~ ^.*\*.*$ ]] && [[ ! -e "$1" ]]; then
|
||||||
|
hadoop_debug "Rejected CLASSPATH: $1 (does not exist)"
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ -z "${CLASSPATH}" ]]; then
|
if [[ -z "${CLASSPATH}" ]]; then
|
||||||
CLASSPATH=$1
|
CLASSPATH=$1
|
||||||
|
hadoop_debug "Initial CLASSPATH=$1"
|
||||||
elif [[ ":${CLASSPATH}:" != *":$1:"* ]]; then
|
elif [[ ":${CLASSPATH}:" != *":$1:"* ]]; then
|
||||||
if [[ "$2" = "before" ]]; then
|
if [[ "$2" = "before" ]]; then
|
||||||
CLASSPATH="$1:${CLASSPATH}"
|
CLASSPATH="$1:${CLASSPATH}"
|
||||||
|
hadoop_debug "Prepend CLASSPATH: $1"
|
||||||
else
|
else
|
||||||
CLASSPATH+=:$1
|
CLASSPATH+=:$1
|
||||||
|
hadoop_debug "Append CLASSPATH: $1"
|
||||||
fi
|
fi
|
||||||
|
else
|
||||||
|
hadoop_debug "Dupe CLASSPATH: $1"
|
||||||
fi
|
fi
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
@ -331,14 +351,20 @@ function hadoop_add_colonpath
|
||||||
if [[ -z "${!1}" ]]; then
|
if [[ -z "${!1}" ]]; then
|
||||||
# shellcheck disable=SC2086
|
# shellcheck disable=SC2086
|
||||||
eval $1="'$2'"
|
eval $1="'$2'"
|
||||||
|
hadoop_debug "Initial colonpath($1): $2"
|
||||||
elif [[ "$3" = "before" ]]; then
|
elif [[ "$3" = "before" ]]; then
|
||||||
# shellcheck disable=SC2086
|
# shellcheck disable=SC2086
|
||||||
eval $1="'$2:${!1}'"
|
eval $1="'$2:${!1}'"
|
||||||
|
hadoop_debug "Prepend colonpath($1): $2"
|
||||||
else
|
else
|
||||||
# shellcheck disable=SC2086
|
# shellcheck disable=SC2086
|
||||||
eval $1+="'$2'"
|
eval $1+="'$2'"
|
||||||
|
hadoop_debug "Append colonpath($1): $2"
|
||||||
fi
|
fi
|
||||||
|
return 0
|
||||||
fi
|
fi
|
||||||
|
hadoop_debug "Rejected colonpath($1): $2"
|
||||||
|
return 1
|
||||||
}
|
}
|
||||||
|
|
||||||
function hadoop_add_javalibpath
|
function hadoop_add_javalibpath
|
||||||
|
@ -397,6 +423,7 @@ function hadoop_add_to_classpath_hdfs
|
||||||
|
|
||||||
function hadoop_add_to_classpath_yarn
|
function hadoop_add_to_classpath_yarn
|
||||||
{
|
{
|
||||||
|
local i
|
||||||
#
|
#
|
||||||
# get all of the yarn jars+config in the path
|
# get all of the yarn jars+config in the path
|
||||||
#
|
#
|
||||||
|
@ -459,7 +486,7 @@ function hadoop_add_to_classpath_userpath
|
||||||
local i
|
local i
|
||||||
local j
|
local j
|
||||||
let c=0
|
let c=0
|
||||||
|
|
||||||
if [[ -n "${HADOOP_CLASSPATH}" ]]; then
|
if [[ -n "${HADOOP_CLASSPATH}" ]]; then
|
||||||
# I wonder if Java runs on VMS.
|
# I wonder if Java runs on VMS.
|
||||||
for i in $(echo "${HADOOP_CLASSPATH}" | tr : '\n'); do
|
for i in $(echo "${HADOOP_CLASSPATH}" | tr : '\n'); do
|
||||||
|
@ -715,6 +742,11 @@ function hadoop_java_exec
|
||||||
local command=$1
|
local command=$1
|
||||||
local class=$2
|
local class=$2
|
||||||
shift 2
|
shift 2
|
||||||
|
|
||||||
|
hadoop_debug "Final CLASSPATH: ${CLASSPATH}"
|
||||||
|
hadoop_debug "Final HADOOP_OPTS: ${HADOOP_OPTS}"
|
||||||
|
|
||||||
|
export CLASSPATH
|
||||||
#shellcheck disable=SC2086
|
#shellcheck disable=SC2086
|
||||||
exec "${JAVA}" "-Dproc_${command}" ${HADOOP_OPTS} "${class}" "$@"
|
exec "${JAVA}" "-Dproc_${command}" ${HADOOP_OPTS} "${class}" "$@"
|
||||||
}
|
}
|
||||||
|
@ -727,6 +759,11 @@ function hadoop_start_daemon
|
||||||
local command=$1
|
local command=$1
|
||||||
local class=$2
|
local class=$2
|
||||||
shift 2
|
shift 2
|
||||||
|
|
||||||
|
hadoop_debug "Final CLASSPATH: ${CLASSPATH}"
|
||||||
|
hadoop_debug "Final HADOOP_OPTS: ${HADOOP_OPTS}"
|
||||||
|
|
||||||
|
export CLASSPATH
|
||||||
#shellcheck disable=SC2086
|
#shellcheck disable=SC2086
|
||||||
exec "${JAVA}" "-Dproc_${command}" ${HADOOP_OPTS} "${class}" "$@"
|
exec "${JAVA}" "-Dproc_${command}" ${HADOOP_OPTS} "${class}" "$@"
|
||||||
}
|
}
|
||||||
|
@ -807,6 +844,9 @@ function hadoop_start_secure_daemon
|
||||||
# note that shellcheck will throw a
|
# note that shellcheck will throw a
|
||||||
# bogus for-our-use-case 2086 here.
|
# bogus for-our-use-case 2086 here.
|
||||||
# it doesn't properly support multi-line situations
|
# it doesn't properly support multi-line situations
|
||||||
|
|
||||||
|
hadoop_debug "Final CLASSPATH: ${CLASSPATH}"
|
||||||
|
hadoop_debug "Final HADOOP_OPTS: ${HADOOP_OPTS}"
|
||||||
|
|
||||||
exec "${jsvc}" \
|
exec "${jsvc}" \
|
||||||
"-Dproc_${daemonname}" \
|
"-Dproc_${daemonname}" \
|
||||||
|
|
|
@ -23,6 +23,7 @@ this="$bin/$script"
|
||||||
|
|
||||||
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
||||||
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
||||||
|
# shellcheck disable=SC2034
|
||||||
HADOOP_NEW_CONFIG=true
|
HADOOP_NEW_CONFIG=true
|
||||||
. "$HADOOP_LIBEXEC_DIR/hadoop-config.sh"
|
. "$HADOOP_LIBEXEC_DIR/hadoop-config.sh"
|
||||||
|
|
||||||
|
@ -33,10 +34,10 @@ fi
|
||||||
CLASS='org.apache.hadoop.record.compiler.generated.Rcc'
|
CLASS='org.apache.hadoop.record.compiler.generated.Rcc'
|
||||||
|
|
||||||
# Always respect HADOOP_OPTS and HADOOP_CLIENT_OPTS
|
# Always respect HADOOP_OPTS and HADOOP_CLIENT_OPTS
|
||||||
HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
|
|
||||||
hadoop_add_param HADOOP_OPTS Xmx "$JAVA_HEAP_MAX"
|
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
||||||
|
|
||||||
hadoop_finalize
|
hadoop_finalize
|
||||||
export CLASSPATH
|
|
||||||
hadoop_java_exec rcc "${CLASS}" "$@"
|
hadoop_java_exec rcc "${CLASS}" "$@"
|
||||||
|
|
|
@ -200,6 +200,7 @@ public class ReconfigurationServlet extends HttpServlet {
|
||||||
protected void doGet(HttpServletRequest req, HttpServletResponse resp)
|
protected void doGet(HttpServletRequest req, HttpServletResponse resp)
|
||||||
throws ServletException, IOException {
|
throws ServletException, IOException {
|
||||||
LOG.info("GET");
|
LOG.info("GET");
|
||||||
|
resp.setContentType("text/html");
|
||||||
PrintWriter out = resp.getWriter();
|
PrintWriter out = resp.getWriter();
|
||||||
|
|
||||||
Reconfigurable reconf = getReconfigurable(req);
|
Reconfigurable reconf = getReconfigurable(req);
|
||||||
|
@ -214,6 +215,7 @@ public class ReconfigurationServlet extends HttpServlet {
|
||||||
protected void doPost(HttpServletRequest req, HttpServletResponse resp)
|
protected void doPost(HttpServletRequest req, HttpServletResponse resp)
|
||||||
throws ServletException, IOException {
|
throws ServletException, IOException {
|
||||||
LOG.info("POST");
|
LOG.info("POST");
|
||||||
|
resp.setContentType("text/html");
|
||||||
PrintWriter out = resp.getWriter();
|
PrintWriter out = resp.getWriter();
|
||||||
|
|
||||||
Reconfigurable reconf = getReconfigurable(req);
|
Reconfigurable reconf = getReconfigurable(req);
|
||||||
|
|
|
@ -381,7 +381,8 @@ public abstract class ChecksumFileSystem extends FilterFileSystem {
|
||||||
long blockSize,
|
long blockSize,
|
||||||
Progressable progress)
|
Progressable progress)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
super(DataChecksum.newCrc32(), fs.getBytesPerSum(), 4);
|
super(DataChecksum.newDataChecksum(DataChecksum.Type.CRC32,
|
||||||
|
fs.getBytesPerSum()));
|
||||||
int bytesPerSum = fs.getBytesPerSum();
|
int bytesPerSum = fs.getBytesPerSum();
|
||||||
this.datas = fs.getRawFileSystem().create(file, overwrite, bufferSize,
|
this.datas = fs.getRawFileSystem().create(file, overwrite, bufferSize,
|
||||||
replication, blockSize, progress);
|
replication, blockSize, progress);
|
||||||
|
@ -405,10 +406,11 @@ public abstract class ChecksumFileSystem extends FilterFileSystem {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
protected void writeChunk(byte[] b, int offset, int len, byte[] checksum,
|
||||||
|
int ckoff, int cklen)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
datas.write(b, offset, len);
|
datas.write(b, offset, len);
|
||||||
sums.write(checksum);
|
sums.write(checksum, ckoff, cklen);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
@ -337,7 +337,8 @@ public abstract class ChecksumFs extends FilterFs {
|
||||||
final short replication, final long blockSize,
|
final short replication, final long blockSize,
|
||||||
final Progressable progress, final ChecksumOpt checksumOpt,
|
final Progressable progress, final ChecksumOpt checksumOpt,
|
||||||
final boolean createParent) throws IOException {
|
final boolean createParent) throws IOException {
|
||||||
super(DataChecksum.newCrc32(), fs.getBytesPerSum(), 4);
|
super(DataChecksum.newDataChecksum(DataChecksum.Type.CRC32,
|
||||||
|
fs.getBytesPerSum()));
|
||||||
|
|
||||||
// checksumOpt is passed down to the raw fs. Unless it implements
|
// checksumOpt is passed down to the raw fs. Unless it implements
|
||||||
// checksum impelemts internally, checksumOpt will be ignored.
|
// checksum impelemts internally, checksumOpt will be ignored.
|
||||||
|
@ -370,10 +371,11 @@ public abstract class ChecksumFs extends FilterFs {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
protected void writeChunk(byte[] b, int offset, int len, byte[] checksum,
|
||||||
|
int ckoff, int cklen)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
datas.write(b, offset, len);
|
datas.write(b, offset, len);
|
||||||
sums.write(checksum);
|
sums.write(checksum, ckoff, cklen);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
@ -18,13 +18,14 @@
|
||||||
|
|
||||||
package org.apache.hadoop.fs;
|
package org.apache.hadoop.fs;
|
||||||
|
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.util.DataChecksum;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.OutputStream;
|
import java.io.OutputStream;
|
||||||
import java.util.zip.Checksum;
|
import java.util.zip.Checksum;
|
||||||
|
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This is a generic output stream for generating checksums for
|
* This is a generic output stream for generating checksums for
|
||||||
* data before it is written to the underlying stream
|
* data before it is written to the underlying stream
|
||||||
|
@ -33,7 +34,7 @@ import org.apache.hadoop.classification.InterfaceStability;
|
||||||
@InterfaceStability.Unstable
|
@InterfaceStability.Unstable
|
||||||
abstract public class FSOutputSummer extends OutputStream {
|
abstract public class FSOutputSummer extends OutputStream {
|
||||||
// data checksum
|
// data checksum
|
||||||
private Checksum sum;
|
private final DataChecksum sum;
|
||||||
// internal buffer for storing data before it is checksumed
|
// internal buffer for storing data before it is checksumed
|
||||||
private byte buf[];
|
private byte buf[];
|
||||||
// internal buffer for storing checksum
|
// internal buffer for storing checksum
|
||||||
|
@ -41,18 +42,24 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
// The number of valid bytes in the buffer.
|
// The number of valid bytes in the buffer.
|
||||||
private int count;
|
private int count;
|
||||||
|
|
||||||
protected FSOutputSummer(Checksum sum, int maxChunkSize, int checksumSize) {
|
// We want this value to be a multiple of 3 because the native code checksums
|
||||||
|
// 3 chunks simultaneously. The chosen value of 9 strikes a balance between
|
||||||
|
// limiting the number of JNI calls and flushing to the underlying stream
|
||||||
|
// relatively frequently.
|
||||||
|
private static final int BUFFER_NUM_CHUNKS = 9;
|
||||||
|
|
||||||
|
protected FSOutputSummer(DataChecksum sum) {
|
||||||
this.sum = sum;
|
this.sum = sum;
|
||||||
this.buf = new byte[maxChunkSize];
|
this.buf = new byte[sum.getBytesPerChecksum() * BUFFER_NUM_CHUNKS];
|
||||||
this.checksum = new byte[checksumSize];
|
this.checksum = new byte[sum.getChecksumSize() * BUFFER_NUM_CHUNKS];
|
||||||
this.count = 0;
|
this.count = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* write the data chunk in <code>b</code> staring at <code>offset</code> with
|
/* write the data chunk in <code>b</code> staring at <code>offset</code> with
|
||||||
* a length of <code>len</code>, and its checksum
|
* a length of <code>len > 0</code>, and its checksum
|
||||||
*/
|
*/
|
||||||
protected abstract void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
protected abstract void writeChunk(byte[] b, int bOffset, int bLen,
|
||||||
throws IOException;
|
byte[] checksum, int checksumOffset, int checksumLen) throws IOException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Check if the implementing OutputStream is closed and should no longer
|
* Check if the implementing OutputStream is closed and should no longer
|
||||||
|
@ -66,7 +73,6 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
/** Write one byte */
|
/** Write one byte */
|
||||||
@Override
|
@Override
|
||||||
public synchronized void write(int b) throws IOException {
|
public synchronized void write(int b) throws IOException {
|
||||||
sum.update(b);
|
|
||||||
buf[count++] = (byte)b;
|
buf[count++] = (byte)b;
|
||||||
if(count == buf.length) {
|
if(count == buf.length) {
|
||||||
flushBuffer();
|
flushBuffer();
|
||||||
|
@ -111,18 +117,17 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
*/
|
*/
|
||||||
private int write1(byte b[], int off, int len) throws IOException {
|
private int write1(byte b[], int off, int len) throws IOException {
|
||||||
if(count==0 && len>=buf.length) {
|
if(count==0 && len>=buf.length) {
|
||||||
// local buffer is empty and user data has one chunk
|
// local buffer is empty and user buffer size >= local buffer size, so
|
||||||
// checksum and output data
|
// simply checksum the user buffer and send it directly to the underlying
|
||||||
|
// stream
|
||||||
final int length = buf.length;
|
final int length = buf.length;
|
||||||
sum.update(b, off, length);
|
writeChecksumChunks(b, off, length);
|
||||||
writeChecksumChunk(b, off, length, false);
|
|
||||||
return length;
|
return length;
|
||||||
}
|
}
|
||||||
|
|
||||||
// copy user data to local buffer
|
// copy user data to local buffer
|
||||||
int bytesToCopy = buf.length-count;
|
int bytesToCopy = buf.length-count;
|
||||||
bytesToCopy = (len<bytesToCopy) ? len : bytesToCopy;
|
bytesToCopy = (len<bytesToCopy) ? len : bytesToCopy;
|
||||||
sum.update(b, off, bytesToCopy);
|
|
||||||
System.arraycopy(b, off, buf, count, bytesToCopy);
|
System.arraycopy(b, off, buf, count, bytesToCopy);
|
||||||
count += bytesToCopy;
|
count += bytesToCopy;
|
||||||
if (count == buf.length) {
|
if (count == buf.length) {
|
||||||
|
@ -136,22 +141,45 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
* the underlying output stream.
|
* the underlying output stream.
|
||||||
*/
|
*/
|
||||||
protected synchronized void flushBuffer() throws IOException {
|
protected synchronized void flushBuffer() throws IOException {
|
||||||
flushBuffer(false);
|
flushBuffer(false, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Forces any buffered output bytes to be checksumed and written out to
|
/* Forces buffered output bytes to be checksummed and written out to
|
||||||
* the underlying output stream. If keep is true, then the state of
|
* the underlying output stream. If there is a trailing partial chunk in the
|
||||||
* this object remains intact.
|
* buffer,
|
||||||
|
* 1) flushPartial tells us whether to flush that chunk
|
||||||
|
* 2) if flushPartial is true, keep tells us whether to keep that chunk in the
|
||||||
|
* buffer (if flushPartial is false, it is always kept in the buffer)
|
||||||
|
*
|
||||||
|
* Returns the number of bytes that were flushed but are still left in the
|
||||||
|
* buffer (can only be non-zero if keep is true).
|
||||||
*/
|
*/
|
||||||
protected synchronized void flushBuffer(boolean keep) throws IOException {
|
protected synchronized int flushBuffer(boolean keep,
|
||||||
if (count != 0) {
|
boolean flushPartial) throws IOException {
|
||||||
int chunkLen = count;
|
int bufLen = count;
|
||||||
|
int partialLen = bufLen % sum.getBytesPerChecksum();
|
||||||
|
int lenToFlush = flushPartial ? bufLen : bufLen - partialLen;
|
||||||
|
if (lenToFlush != 0) {
|
||||||
|
writeChecksumChunks(buf, 0, lenToFlush);
|
||||||
|
if (!flushPartial || keep) {
|
||||||
|
count = partialLen;
|
||||||
|
System.arraycopy(buf, bufLen - count, buf, 0, count);
|
||||||
|
} else {
|
||||||
count = 0;
|
count = 0;
|
||||||
writeChecksumChunk(buf, 0, chunkLen, keep);
|
|
||||||
if (keep) {
|
|
||||||
count = chunkLen;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// total bytes left minus unflushed bytes left
|
||||||
|
return count - (bufLen - lenToFlush);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checksums all complete data chunks and flushes them to the underlying
|
||||||
|
* stream. If there is a trailing partial chunk, it is not flushed and is
|
||||||
|
* maintained in the buffer.
|
||||||
|
*/
|
||||||
|
public void flush() throws IOException {
|
||||||
|
flushBuffer(false, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -161,18 +189,18 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
return count;
|
return count;
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Generate checksum for the data chunk and output data chunk & checksum
|
/** Generate checksums for the given data chunks and output chunks & checksums
|
||||||
* to the underlying output stream. If keep is true then keep the
|
* to the underlying output stream.
|
||||||
* current checksum intact, do not reset it.
|
|
||||||
*/
|
*/
|
||||||
private void writeChecksumChunk(byte b[], int off, int len, boolean keep)
|
private void writeChecksumChunks(byte b[], int off, int len)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
int tempChecksum = (int)sum.getValue();
|
sum.calculateChunkedSums(b, off, len, checksum, 0);
|
||||||
if (!keep) {
|
for (int i = 0; i < len; i += sum.getBytesPerChecksum()) {
|
||||||
sum.reset();
|
int chunkLen = Math.min(sum.getBytesPerChecksum(), len - i);
|
||||||
|
int ckOffset = i / sum.getBytesPerChecksum() * sum.getChecksumSize();
|
||||||
|
writeChunk(b, off + i, chunkLen, checksum, ckOffset,
|
||||||
|
sum.getChecksumSize());
|
||||||
}
|
}
|
||||||
int2byte(tempChecksum, checksum);
|
|
||||||
writeChunk(b, off, len, checksum);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -196,9 +224,14 @@ abstract public class FSOutputSummer extends OutputStream {
|
||||||
/**
|
/**
|
||||||
* Resets existing buffer with a new one of the specified size.
|
* Resets existing buffer with a new one of the specified size.
|
||||||
*/
|
*/
|
||||||
protected synchronized void resetChecksumChunk(int size) {
|
protected synchronized void setChecksumBufSize(int size) {
|
||||||
sum.reset();
|
|
||||||
this.buf = new byte[size];
|
this.buf = new byte[size];
|
||||||
|
this.checksum = new byte[((size - 1) / sum.getBytesPerChecksum() + 1) *
|
||||||
|
sum.getChecksumSize()];
|
||||||
this.count = 0;
|
this.count = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
protected synchronized void resetChecksumBufSize() {
|
||||||
|
setChecksumBufSize(sum.getBytesPerChecksum() * BUFFER_NUM_CHUNKS);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -339,6 +339,7 @@ public class DataChecksum implements Checksum {
|
||||||
byte[] data, int dataOff, int dataLen,
|
byte[] data, int dataOff, int dataLen,
|
||||||
byte[] checksums, int checksumsOff, String fileName,
|
byte[] checksums, int checksumsOff, String fileName,
|
||||||
long basePos) throws ChecksumException {
|
long basePos) throws ChecksumException {
|
||||||
|
if (type.size == 0) return;
|
||||||
|
|
||||||
if (NativeCrc32.isAvailable()) {
|
if (NativeCrc32.isAvailable()) {
|
||||||
NativeCrc32.verifyChunkedSumsByteArray(bytesPerChecksum, type.id,
|
NativeCrc32.verifyChunkedSumsByteArray(bytesPerChecksum, type.id,
|
||||||
|
@ -421,6 +422,7 @@ public class DataChecksum implements Checksum {
|
||||||
public void calculateChunkedSums(
|
public void calculateChunkedSums(
|
||||||
byte[] data, int dataOffset, int dataLength,
|
byte[] data, int dataOffset, int dataLength,
|
||||||
byte[] sums, int sumsOffset) {
|
byte[] sums, int sumsOffset) {
|
||||||
|
if (type.size == 0) return;
|
||||||
|
|
||||||
if (NativeCrc32.isAvailable()) {
|
if (NativeCrc32.isAvailable()) {
|
||||||
NativeCrc32.calculateChunkedSumsByteArray(bytesPerChecksum, type.id,
|
NativeCrc32.calculateChunkedSumsByteArray(bytesPerChecksum, type.id,
|
||||||
|
|
|
@ -42,7 +42,7 @@ class NativeCrc32 {
|
||||||
* modified.
|
* modified.
|
||||||
*
|
*
|
||||||
* @param bytesPerSum the chunk size (eg 512 bytes)
|
* @param bytesPerSum the chunk size (eg 512 bytes)
|
||||||
* @param checksumType the DataChecksum type constant
|
* @param checksumType the DataChecksum type constant (NULL is not supported)
|
||||||
* @param sums the DirectByteBuffer pointing at the beginning of the
|
* @param sums the DirectByteBuffer pointing at the beginning of the
|
||||||
* stored checksums
|
* stored checksums
|
||||||
* @param data the DirectByteBuffer pointing at the beginning of the
|
* @param data the DirectByteBuffer pointing at the beginning of the
|
||||||
|
|
|
@ -434,6 +434,9 @@ Release 2.6.0 - UNRELEASED
|
||||||
HDFS-6773. MiniDFSCluster should skip edit log fsync by default (Stephen
|
HDFS-6773. MiniDFSCluster should skip edit log fsync by default (Stephen
|
||||||
Chu via Colin Patrick McCabe)
|
Chu via Colin Patrick McCabe)
|
||||||
|
|
||||||
|
HDFS-6865. Byte array native checksumming on client side
|
||||||
|
(James Thomas via todd)
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
||||||
HDFS-6823. dfs.web.authentication.kerberos.principal shows up in logs for
|
HDFS-6823. dfs.web.authentication.kerberos.principal shows up in logs for
|
||||||
|
|
|
@ -80,6 +80,7 @@ shift
|
||||||
case ${COMMAND} in
|
case ${COMMAND} in
|
||||||
balancer)
|
balancer)
|
||||||
CLASS=org.apache.hadoop.hdfs.server.balancer.Balancer
|
CLASS=org.apache.hadoop.hdfs.server.balancer.Balancer
|
||||||
|
hadoop_debug "Appending HADOOP_BALANCER_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_BALANCER_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_BALANCER_OPTS}"
|
||||||
;;
|
;;
|
||||||
cacheadmin)
|
cacheadmin)
|
||||||
|
@ -105,19 +106,24 @@ case ${COMMAND} in
|
||||||
HADOOP_SECURE_PID_DIR="${HADOOP_SECURE_PID_DIR:-$HADOOP_SECURE_DN_PID_DIR}"
|
HADOOP_SECURE_PID_DIR="${HADOOP_SECURE_PID_DIR:-$HADOOP_SECURE_DN_PID_DIR}"
|
||||||
HADOOP_SECURE_LOG_DIR="${HADOOP_SECURE_LOG_DIR:-$HADOOP_SECURE_DN_LOG_DIR}"
|
HADOOP_SECURE_LOG_DIR="${HADOOP_SECURE_LOG_DIR:-$HADOOP_SECURE_DN_LOG_DIR}"
|
||||||
|
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_DN_SECURE_EXTRA_OPTS} ${HADOOP_DATANODE_OPTS}"
|
hadoop_debug "Appending HADOOP_DATANODE_OPTS onto HADOOP_OPTS"
|
||||||
|
hadoop_debug "Appending HADOOP_DN_SECURE_EXTRA_OPTS onto HADOOP_OPTS"
|
||||||
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_DATANODE_OPTS} ${HADOOP_DN_SECURE_EXTRA_OPTS}"
|
||||||
CLASS="org.apache.hadoop.hdfs.server.datanode.SecureDataNodeStarter"
|
CLASS="org.apache.hadoop.hdfs.server.datanode.SecureDataNodeStarter"
|
||||||
else
|
else
|
||||||
|
hadoop_debug "Appending HADOOP_DATANODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_DATANODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_DATANODE_OPTS}"
|
||||||
CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
|
CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
dfs)
|
dfs)
|
||||||
CLASS=org.apache.hadoop.fs.FsShell
|
CLASS=org.apache.hadoop.fs.FsShell
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
dfsadmin)
|
dfsadmin)
|
||||||
CLASS=org.apache.hadoop.hdfs.tools.DFSAdmin
|
CLASS=org.apache.hadoop.hdfs.tools.DFSAdmin
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
fetchdt)
|
fetchdt)
|
||||||
|
@ -125,6 +131,7 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
fsck)
|
fsck)
|
||||||
CLASS=org.apache.hadoop.hdfs.tools.DFSck
|
CLASS=org.apache.hadoop.hdfs.tools.DFSck
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
getconf)
|
getconf)
|
||||||
|
@ -135,12 +142,15 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
haadmin)
|
haadmin)
|
||||||
CLASS=org.apache.hadoop.hdfs.tools.DFSHAAdmin
|
CLASS=org.apache.hadoop.hdfs.tools.DFSHAAdmin
|
||||||
CLASSPATH="${CLASSPATH}:${TOOL_PATH}"
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
journalnode)
|
journalnode)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.hdfs.qjournal.server.JournalNode'
|
CLASS='org.apache.hadoop.hdfs.qjournal.server.JournalNode'
|
||||||
|
hadoop_debug "Appending HADOOP_JOURNALNODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_JOURNALNODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_JOURNALNODE_OPTS}"
|
||||||
;;
|
;;
|
||||||
jmxget)
|
jmxget)
|
||||||
|
@ -152,6 +162,7 @@ case ${COMMAND} in
|
||||||
namenode)
|
namenode)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
|
CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
|
||||||
|
hadoop_debug "Appending HADOOP_NAMENODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NAMENODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NAMENODE_OPTS}"
|
||||||
;;
|
;;
|
||||||
nfs3)
|
nfs3)
|
||||||
|
@ -164,9 +175,12 @@ case ${COMMAND} in
|
||||||
HADOOP_SECURE_PID_DIR="${HADOOP_SECURE_PID_DIR:-$HADOOP_SECURE_NFS3_PID_DIR}"
|
HADOOP_SECURE_PID_DIR="${HADOOP_SECURE_PID_DIR:-$HADOOP_SECURE_NFS3_PID_DIR}"
|
||||||
HADOOP_SECURE_LOG_DIR="${HADOOP_SECURE_LOG_DIR:-$HADOOP_SECURE_NFS3_LOG_DIR}"
|
HADOOP_SECURE_LOG_DIR="${HADOOP_SECURE_LOG_DIR:-$HADOOP_SECURE_NFS3_LOG_DIR}"
|
||||||
|
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NFS3_SECURE_EXTRA_OPTS} ${HADOOP_NFS3_OPTS}"
|
hadoop_debug "Appending HADOOP_NFS3_OPTS onto HADOOP_OPTS"
|
||||||
|
hadoop_debug "Appending HADOOP_NFS3_SECURE_EXTRA_OPTS onto HADOOP_OPTS"
|
||||||
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NFS3_OPTS} ${HADOOP_NFS3_SECURE_EXTRA_OPTS}"
|
||||||
CLASS=org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter
|
CLASS=org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter
|
||||||
else
|
else
|
||||||
|
hadoop_debug "Appending HADOOP_NFS3_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NFS3_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_NFS3_OPTS}"
|
||||||
CLASS=org.apache.hadoop.hdfs.nfs.nfs3.Nfs3
|
CLASS=org.apache.hadoop.hdfs.nfs.nfs3.Nfs3
|
||||||
fi
|
fi
|
||||||
|
@ -183,11 +197,13 @@ case ${COMMAND} in
|
||||||
portmap)
|
portmap)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS=org.apache.hadoop.portmap.Portmap
|
CLASS=org.apache.hadoop.portmap.Portmap
|
||||||
|
hadoop_debug "Appending HADOOP_PORTMAP_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_PORTMAP_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_PORTMAP_OPTS}"
|
||||||
;;
|
;;
|
||||||
secondarynamenode)
|
secondarynamenode)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
|
CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
|
||||||
|
hadoop_debug "Appending HADOOP_SECONDARYNAMENODE_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_SECONDARYNAMENODE_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_SECONDARYNAMENODE_OPTS}"
|
||||||
;;
|
;;
|
||||||
snapshotDiff)
|
snapshotDiff)
|
||||||
|
@ -196,6 +212,7 @@ case ${COMMAND} in
|
||||||
zkfc)
|
zkfc)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.hdfs.tools.DFSZKFailoverController'
|
CLASS='org.apache.hadoop.hdfs.tools.DFSZKFailoverController'
|
||||||
|
hadoop_debug "Appending HADOOP_ZKFC_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_ZKFC_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_ZKFC_OPTS}"
|
||||||
;;
|
;;
|
||||||
-*)
|
-*)
|
||||||
|
@ -236,8 +253,6 @@ fi
|
||||||
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
||||||
hadoop_finalize
|
hadoop_finalize
|
||||||
|
|
||||||
export CLASSPATH
|
|
||||||
|
|
||||||
if [[ -n "${daemon}" ]]; then
|
if [[ -n "${daemon}" ]]; then
|
||||||
if [[ -n "${secure_service}" ]]; then
|
if [[ -n "${secure_service}" ]]; then
|
||||||
hadoop_secure_daemon_handler \
|
hadoop_secure_daemon_handler \
|
||||||
|
|
|
@ -401,7 +401,7 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
// one chunk that fills up the partial chunk.
|
// one chunk that fills up the partial chunk.
|
||||||
//
|
//
|
||||||
computePacketChunkSize(0, freeInCksum);
|
computePacketChunkSize(0, freeInCksum);
|
||||||
resetChecksumChunk(freeInCksum);
|
setChecksumBufSize(freeInCksum);
|
||||||
appendChunk = true;
|
appendChunk = true;
|
||||||
} else {
|
} else {
|
||||||
// if the remaining space in the block is smaller than
|
// if the remaining space in the block is smaller than
|
||||||
|
@ -1566,7 +1566,7 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
|
|
||||||
private DFSOutputStream(DFSClient dfsClient, String src, Progressable progress,
|
private DFSOutputStream(DFSClient dfsClient, String src, Progressable progress,
|
||||||
HdfsFileStatus stat, DataChecksum checksum) throws IOException {
|
HdfsFileStatus stat, DataChecksum checksum) throws IOException {
|
||||||
super(checksum, checksum.getBytesPerChecksum(), checksum.getChecksumSize());
|
super(checksum);
|
||||||
this.dfsClient = dfsClient;
|
this.dfsClient = dfsClient;
|
||||||
this.src = src;
|
this.src = src;
|
||||||
this.fileId = stat.getFileId();
|
this.fileId = stat.getFileId();
|
||||||
|
@ -1720,22 +1720,21 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
|
|
||||||
// @see FSOutputSummer#writeChunk()
|
// @see FSOutputSummer#writeChunk()
|
||||||
@Override
|
@Override
|
||||||
protected synchronized void writeChunk(byte[] b, int offset, int len, byte[] checksum)
|
protected synchronized void writeChunk(byte[] b, int offset, int len,
|
||||||
throws IOException {
|
byte[] checksum, int ckoff, int cklen) throws IOException {
|
||||||
dfsClient.checkOpen();
|
dfsClient.checkOpen();
|
||||||
checkClosed();
|
checkClosed();
|
||||||
|
|
||||||
int cklen = checksum.length;
|
|
||||||
int bytesPerChecksum = this.checksum.getBytesPerChecksum();
|
int bytesPerChecksum = this.checksum.getBytesPerChecksum();
|
||||||
if (len > bytesPerChecksum) {
|
if (len > bytesPerChecksum) {
|
||||||
throw new IOException("writeChunk() buffer size is " + len +
|
throw new IOException("writeChunk() buffer size is " + len +
|
||||||
" is larger than supported bytesPerChecksum " +
|
" is larger than supported bytesPerChecksum " +
|
||||||
bytesPerChecksum);
|
bytesPerChecksum);
|
||||||
}
|
}
|
||||||
if (checksum.length != this.checksum.getChecksumSize()) {
|
if (cklen != this.checksum.getChecksumSize()) {
|
||||||
throw new IOException("writeChunk() checksum size is supposed to be " +
|
throw new IOException("writeChunk() checksum size is supposed to be " +
|
||||||
this.checksum.getChecksumSize() +
|
this.checksum.getChecksumSize() +
|
||||||
" but found to be " + checksum.length);
|
" but found to be " + cklen);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (currentPacket == null) {
|
if (currentPacket == null) {
|
||||||
|
@ -1751,7 +1750,7 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
currentPacket.writeChecksum(checksum, 0, cklen);
|
currentPacket.writeChecksum(checksum, ckoff, cklen);
|
||||||
currentPacket.writeData(b, offset, len);
|
currentPacket.writeData(b, offset, len);
|
||||||
currentPacket.numChunks++;
|
currentPacket.numChunks++;
|
||||||
bytesCurBlock += len;
|
bytesCurBlock += len;
|
||||||
|
@ -1775,7 +1774,7 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
// crc chunks from now on.
|
// crc chunks from now on.
|
||||||
if (appendChunk && bytesCurBlock%bytesPerChecksum == 0) {
|
if (appendChunk && bytesCurBlock%bytesPerChecksum == 0) {
|
||||||
appendChunk = false;
|
appendChunk = false;
|
||||||
resetChecksumChunk(bytesPerChecksum);
|
resetChecksumBufSize();
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!appendChunk) {
|
if (!appendChunk) {
|
||||||
|
@ -1856,20 +1855,13 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
long lastBlockLength = -1L;
|
long lastBlockLength = -1L;
|
||||||
boolean updateLength = syncFlags.contains(SyncFlag.UPDATE_LENGTH);
|
boolean updateLength = syncFlags.contains(SyncFlag.UPDATE_LENGTH);
|
||||||
synchronized (this) {
|
synchronized (this) {
|
||||||
/* Record current blockOffset. This might be changed inside
|
|
||||||
* flushBuffer() where a partial checksum chunk might be flushed.
|
|
||||||
* After the flush, reset the bytesCurBlock back to its previous value,
|
|
||||||
* any partial checksum chunk will be sent now and in next packet.
|
|
||||||
*/
|
|
||||||
long saveOffset = bytesCurBlock;
|
|
||||||
Packet oldCurrentPacket = currentPacket;
|
|
||||||
// flush checksum buffer, but keep checksum buffer intact
|
// flush checksum buffer, but keep checksum buffer intact
|
||||||
flushBuffer(true);
|
int numKept = flushBuffer(true, true);
|
||||||
// bytesCurBlock potentially incremented if there was buffered data
|
// bytesCurBlock potentially incremented if there was buffered data
|
||||||
|
|
||||||
if (DFSClient.LOG.isDebugEnabled()) {
|
if (DFSClient.LOG.isDebugEnabled()) {
|
||||||
DFSClient.LOG.debug(
|
DFSClient.LOG.debug(
|
||||||
"DFSClient flush() : saveOffset " + saveOffset +
|
"DFSClient flush() :" +
|
||||||
" bytesCurBlock " + bytesCurBlock +
|
" bytesCurBlock " + bytesCurBlock +
|
||||||
" lastFlushOffset " + lastFlushOffset);
|
" lastFlushOffset " + lastFlushOffset);
|
||||||
}
|
}
|
||||||
|
@ -1886,14 +1878,6 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
bytesCurBlock, currentSeqno++, this.checksum.getChecksumSize());
|
bytesCurBlock, currentSeqno++, this.checksum.getChecksumSize());
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// We already flushed up to this offset.
|
|
||||||
// This means that we haven't written anything since the last flush
|
|
||||||
// (or the beginning of the file). Hence, we should not have any
|
|
||||||
// packet queued prior to this call, since the last flush set
|
|
||||||
// currentPacket = null.
|
|
||||||
assert oldCurrentPacket == null :
|
|
||||||
"Empty flush should not occur with a currentPacket";
|
|
||||||
|
|
||||||
if (isSync && bytesCurBlock > 0) {
|
if (isSync && bytesCurBlock > 0) {
|
||||||
// Nothing to send right now,
|
// Nothing to send right now,
|
||||||
// and the block was partially written,
|
// and the block was partially written,
|
||||||
|
@ -1913,7 +1897,7 @@ public class DFSOutputStream extends FSOutputSummer
|
||||||
// Restore state of stream. Record the last flush offset
|
// Restore state of stream. Record the last flush offset
|
||||||
// of the last full chunk that was flushed.
|
// of the last full chunk that was flushed.
|
||||||
//
|
//
|
||||||
bytesCurBlock = saveOffset;
|
bytesCurBlock -= numKept;
|
||||||
toWaitFor = lastQueuedSeqno;
|
toWaitFor = lastQueuedSeqno;
|
||||||
} // end synchronized
|
} // end synchronized
|
||||||
|
|
||||||
|
|
|
@ -261,7 +261,9 @@ public class TestFileAppend{
|
||||||
start += 29;
|
start += 29;
|
||||||
}
|
}
|
||||||
stm.write(fileContents, start, AppendTestUtil.FILE_SIZE -start);
|
stm.write(fileContents, start, AppendTestUtil.FILE_SIZE -start);
|
||||||
|
// need to make sure we completely write out all full blocks before
|
||||||
|
// the checkFile() call (see FSOutputSummer#flush)
|
||||||
|
stm.flush();
|
||||||
// verify that full blocks are sane
|
// verify that full blocks are sane
|
||||||
checkFile(fs, file1, 1);
|
checkFile(fs, file1, 1);
|
||||||
stm.close();
|
stm.close();
|
||||||
|
|
|
@ -394,6 +394,8 @@ public class TestBlockToken {
|
||||||
Path filePath = new Path(fileName);
|
Path filePath = new Path(fileName);
|
||||||
FSDataOutputStream out = fs.create(filePath, (short) 1);
|
FSDataOutputStream out = fs.create(filePath, (short) 1);
|
||||||
out.write(new byte[1000]);
|
out.write(new byte[1000]);
|
||||||
|
// ensure that the first block is written out (see FSOutputSummer#flush)
|
||||||
|
out.flush();
|
||||||
LocatedBlocks locatedBlocks = cluster.getNameNodeRpc().getBlockLocations(
|
LocatedBlocks locatedBlocks = cluster.getNameNodeRpc().getBlockLocations(
|
||||||
fileName, 0, 1000);
|
fileName, 0, 1000);
|
||||||
while (locatedBlocks.getLastLocatedBlock() == null) {
|
while (locatedBlocks.getLastLocatedBlock() == null) {
|
||||||
|
|
|
@ -70,6 +70,9 @@ public class TestBlockUnderConstruction {
|
||||||
long blocksBefore = stm.getPos() / BLOCK_SIZE;
|
long blocksBefore = stm.getPos() / BLOCK_SIZE;
|
||||||
|
|
||||||
TestFileCreation.writeFile(stm, BLOCK_SIZE);
|
TestFileCreation.writeFile(stm, BLOCK_SIZE);
|
||||||
|
// need to make sure the full block is completely flushed to the DataNodes
|
||||||
|
// (see FSOutputSummer#flush)
|
||||||
|
stm.flush();
|
||||||
int blocksAfter = 0;
|
int blocksAfter = 0;
|
||||||
// wait until the block is allocated by DataStreamer
|
// wait until the block is allocated by DataStreamer
|
||||||
BlockLocation[] locatedBlocks;
|
BlockLocation[] locatedBlocks;
|
||||||
|
|
|
@ -141,6 +141,9 @@ public class TestDecommissioningStatus {
|
||||||
Random rand = new Random(seed);
|
Random rand = new Random(seed);
|
||||||
rand.nextBytes(buffer);
|
rand.nextBytes(buffer);
|
||||||
stm.write(buffer);
|
stm.write(buffer);
|
||||||
|
// need to make sure that we actually write out both file blocks
|
||||||
|
// (see FSOutputSummer#flush)
|
||||||
|
stm.flush();
|
||||||
// Do not close stream, return it
|
// Do not close stream, return it
|
||||||
// so that it is not garbage collected
|
// so that it is not garbage collected
|
||||||
return stm;
|
return stm;
|
||||||
|
|
|
@ -64,13 +64,15 @@ shift
|
||||||
|
|
||||||
case ${COMMAND} in
|
case ${COMMAND} in
|
||||||
mradmin|jobtracker|tasktracker|groups)
|
mradmin|jobtracker|tasktracker|groups)
|
||||||
echo "Sorry, the ${COMMAND} command is no longer supported."
|
hadoop_error "Sorry, the ${COMMAND} command is no longer supported."
|
||||||
echo "You may find similar functionality with the \"yarn\" shell command."
|
hadoop_error "You may find similar functionality with the \"yarn\" shell command."
|
||||||
hadoop_exit_with_usage 1
|
hadoop_exit_with_usage 1
|
||||||
;;
|
;;
|
||||||
archive)
|
archive)
|
||||||
CLASS=org.apache.hadoop.tools.HadoopArchives
|
CLASS=org.apache.hadoop.tools.HadoopArchives
|
||||||
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
hadoop_add_classpath "${TOOL_PATH}"
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
classpath)
|
classpath)
|
||||||
|
@ -80,12 +82,15 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
distcp)
|
distcp)
|
||||||
CLASS=org.apache.hadoop.tools.DistCp
|
CLASS=org.apache.hadoop.tools.DistCp
|
||||||
|
hadoop_debug "Injecting TOOL_PATH into CLASSPATH"
|
||||||
hadoop_add_classpath "${TOOL_PATH}"
|
hadoop_add_classpath "${TOOL_PATH}"
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
historyserver)
|
historyserver)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS=org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer
|
CLASS=org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer
|
||||||
|
hadoop_debug "Appending HADOOP_JOB_HISTORYSERVER_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_JOB_HISTORYSERVER_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_JOB_HISTORYSERVER_OPTS}"
|
||||||
if [ -n "${HADOOP_JOB_HISTORYSERVER_HEAPSIZE}" ]; then
|
if [ -n "${HADOOP_JOB_HISTORYSERVER_HEAPSIZE}" ]; then
|
||||||
JAVA_HEAP_MAX="-Xmx${HADOOP_JOB_HISTORYSERVER_HEAPSIZE}m"
|
JAVA_HEAP_MAX="-Xmx${HADOOP_JOB_HISTORYSERVER_HEAPSIZE}m"
|
||||||
|
@ -97,6 +102,7 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
pipes)
|
pipes)
|
||||||
CLASS=org.apache.hadoop.mapred.pipes.Submitter
|
CLASS=org.apache.hadoop.mapred.pipes.Submitter
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
queue)
|
queue)
|
||||||
|
@ -104,10 +110,12 @@ case ${COMMAND} in
|
||||||
;;
|
;;
|
||||||
sampler)
|
sampler)
|
||||||
CLASS=org.apache.hadoop.mapred.lib.InputSampler
|
CLASS=org.apache.hadoop.mapred.lib.InputSampler
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
version)
|
version)
|
||||||
CLASS=org.apache.hadoop.util.VersionInfo
|
CLASS=org.apache.hadoop.util.VersionInfo
|
||||||
|
hadoop_debug "Appending HADOOP_CLIENT_OPTS onto HADOOP_OPTS"
|
||||||
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
HADOOP_OPTS="${HADOOP_OPTS} ${HADOOP_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
-*|*)
|
-*|*)
|
||||||
|
@ -130,8 +138,6 @@ fi
|
||||||
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
hadoop_add_param HADOOP_OPTS Xmx "${JAVA_HEAP_MAX}"
|
||||||
hadoop_finalize
|
hadoop_finalize
|
||||||
|
|
||||||
export CLASSPATH
|
|
||||||
|
|
||||||
if [[ -n "${daemon}" ]]; then
|
if [[ -n "${daemon}" ]]; then
|
||||||
if [[ -n "${secure_service}" ]]; then
|
if [[ -n "${secure_service}" ]]; then
|
||||||
hadoop_secure_daemon_handler "${HADOOP_DAEMON_MODE}" "${COMMAND}"\
|
hadoop_secure_daemon_handler "${HADOOP_DAEMON_MODE}" "${COMMAND}"\
|
||||||
|
|
|
@ -72,6 +72,7 @@ shift
|
||||||
case "${COMMAND}" in
|
case "${COMMAND}" in
|
||||||
application|applicationattempt|container)
|
application|applicationattempt|container)
|
||||||
CLASS=org.apache.hadoop.yarn.client.cli.ApplicationCLI
|
CLASS=org.apache.hadoop.yarn.client.cli.ApplicationCLI
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
set -- "${COMMAND}" "$@"
|
set -- "${COMMAND}" "$@"
|
||||||
;;
|
;;
|
||||||
|
@ -82,10 +83,12 @@ case "${COMMAND}" in
|
||||||
;;
|
;;
|
||||||
daemonlog)
|
daemonlog)
|
||||||
CLASS=org.apache.hadoop.log.LogLevel
|
CLASS=org.apache.hadoop.log.LogLevel
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
jar)
|
jar)
|
||||||
CLASS=org.apache.hadoop.util.RunJar
|
CLASS=org.apache.hadoop.util.RunJar
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
historyserver)
|
historyserver)
|
||||||
|
@ -97,15 +100,18 @@ case "${COMMAND}" in
|
||||||
;;
|
;;
|
||||||
logs)
|
logs)
|
||||||
CLASS=org.apache.hadoop.yarn.logaggregation.LogDumper
|
CLASS=org.apache.hadoop.yarn.logaggregation.LogDumper
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
node)
|
node)
|
||||||
CLASS=org.apache.hadoop.yarn.client.cli.NodeCLI
|
CLASS=org.apache.hadoop.yarn.client.cli.NodeCLI
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
nodemanager)
|
nodemanager)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.yarn.server.nodemanager.NodeManager'
|
CLASS='org.apache.hadoop.yarn.server.nodemanager.NodeManager'
|
||||||
|
hadoop_debug "Append YARN_NODEMANAGER_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_NODEMANAGER_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_NODEMANAGER_OPTS}"
|
||||||
if [[ -n "${YARN_NODEMANAGER_HEAPSIZE}" ]]; then
|
if [[ -n "${YARN_NODEMANAGER_HEAPSIZE}" ]]; then
|
||||||
JAVA_HEAP_MAX="-Xmx${YARN_NODEMANAGER_HEAPSIZE}m"
|
JAVA_HEAP_MAX="-Xmx${YARN_NODEMANAGER_HEAPSIZE}m"
|
||||||
|
@ -114,6 +120,7 @@ case "${COMMAND}" in
|
||||||
proxyserver)
|
proxyserver)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.yarn.server.webproxy.WebAppProxyServer'
|
CLASS='org.apache.hadoop.yarn.server.webproxy.WebAppProxyServer'
|
||||||
|
hadoop_debug "Append YARN_PROXYSERVER_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_PROXYSERVER_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_PROXYSERVER_OPTS}"
|
||||||
if [[ -n "${YARN_PROXYSERVER_HEAPSIZE}" ]]; then
|
if [[ -n "${YARN_PROXYSERVER_HEAPSIZE}" ]]; then
|
||||||
JAVA_HEAP_MAX="-Xmx${YARN_PROXYSERVER_HEAPSIZE}m"
|
JAVA_HEAP_MAX="-Xmx${YARN_PROXYSERVER_HEAPSIZE}m"
|
||||||
|
@ -123,17 +130,20 @@ case "${COMMAND}" in
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.yarn.server.resourcemanager.ResourceManager'
|
CLASS='org.apache.hadoop.yarn.server.resourcemanager.ResourceManager'
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_RESOURCEMANAGER_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_RESOURCEMANAGER_OPTS}"
|
||||||
|
hadoop_debug "Append YARN_RESOURCEMANAGER_OPTS onto YARN_OPTS"
|
||||||
if [[ -n "${YARN_RESOURCEMANAGER_HEAPSIZE}" ]]; then
|
if [[ -n "${YARN_RESOURCEMANAGER_HEAPSIZE}" ]]; then
|
||||||
JAVA_HEAP_MAX="-Xmx${YARN_RESOURCEMANAGER_HEAPSIZE}m"
|
JAVA_HEAP_MAX="-Xmx${YARN_RESOURCEMANAGER_HEAPSIZE}m"
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
rmadmin)
|
rmadmin)
|
||||||
CLASS='org.apache.hadoop.yarn.client.cli.RMAdminCLI'
|
CLASS='org.apache.hadoop.yarn.client.cli.RMAdminCLI'
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
timelineserver)
|
timelineserver)
|
||||||
daemon="true"
|
daemon="true"
|
||||||
CLASS='org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer'
|
CLASS='org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer'
|
||||||
|
hadoop_debug "Append YARN_TIMELINESERVER_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_TIMELINESERVER_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_TIMELINESERVER_OPTS}"
|
||||||
if [[ -n "${YARN_TIMELINESERVER_HEAPSIZE}" ]]; then
|
if [[ -n "${YARN_TIMELINESERVER_HEAPSIZE}" ]]; then
|
||||||
JAVA_HEAP_MAX="-Xmx${YARN_TIMELINESERVER_HEAPSIZE}m"
|
JAVA_HEAP_MAX="-Xmx${YARN_TIMELINESERVER_HEAPSIZE}m"
|
||||||
|
@ -141,6 +151,7 @@ case "${COMMAND}" in
|
||||||
;;
|
;;
|
||||||
version)
|
version)
|
||||||
CLASS=org.apache.hadoop.util.VersionInfo
|
CLASS=org.apache.hadoop.util.VersionInfo
|
||||||
|
hadoop_debug "Append YARN_CLIENT_OPTS onto YARN_OPTS"
|
||||||
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
YARN_OPTS="${YARN_OPTS} ${YARN_CLIENT_OPTS}"
|
||||||
;;
|
;;
|
||||||
-*)
|
-*)
|
||||||
|
@ -153,6 +164,8 @@ esac
|
||||||
|
|
||||||
# set HADOOP_OPTS to YARN_OPTS so that we can use
|
# set HADOOP_OPTS to YARN_OPTS so that we can use
|
||||||
# finalize, etc, without doing anything funky
|
# finalize, etc, without doing anything funky
|
||||||
|
hadoop_debug "Resetting HADOOP_OPTS=YARN_OPTS"
|
||||||
|
# shellcheck disable=SC2034
|
||||||
HADOOP_OPTS="${YARN_OPTS}"
|
HADOOP_OPTS="${YARN_OPTS}"
|
||||||
|
|
||||||
daemon_outfile="${HADOOP_LOG_DIR}/hadoop-${HADOOP_IDENT_STRING}-${COMMAND}-${HOSTNAME}.out"
|
daemon_outfile="${HADOOP_LOG_DIR}/hadoop-${HADOOP_IDENT_STRING}-${COMMAND}-${HOSTNAME}.out"
|
||||||
|
@ -180,8 +193,6 @@ hadoop_add_param HADOOP_OPTS yarn.root.logger "-Dyarn.root.logger=${YARN_ROOT_LO
|
||||||
|
|
||||||
hadoop_finalize
|
hadoop_finalize
|
||||||
|
|
||||||
export CLASSPATH
|
|
||||||
|
|
||||||
if [[ -n "${daemon}" ]]; then
|
if [[ -n "${daemon}" ]]; then
|
||||||
if [[ -n "${secure_service}" ]]; then
|
if [[ -n "${secure_service}" ]]; then
|
||||||
hadoop_secure_daemon_handler "${HADOOP_DAEMON_MODE}" "${COMMAND}" \
|
hadoop_secure_daemon_handler "${HADOOP_DAEMON_MODE}" "${COMMAND}" \
|
||||||
|
|
Loading…
Reference in New Issue