Merge r1236386 through r1237583 from 0.23.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23-PB@1237590 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
commit
ab824d9e9b
|
@ -18,4 +18,4 @@
|
||||||
|
|
||||||
OK_RELEASEAUDIT_WARNINGS=0
|
OK_RELEASEAUDIT_WARNINGS=0
|
||||||
OK_FINDBUGS_WARNINGS=0
|
OK_FINDBUGS_WARNINGS=0
|
||||||
OK_JAVADOC_WARNINGS=6
|
OK_JAVADOC_WARNINGS=13
|
||||||
|
|
|
@ -142,10 +142,21 @@ Release 0.23.1 - Unreleased
|
||||||
HADOOP-7919. Remove the unused hadoop.logfile.* properties from the
|
HADOOP-7919. Remove the unused hadoop.logfile.* properties from the
|
||||||
core-default.xml file. (harsh)
|
core-default.xml file. (harsh)
|
||||||
|
|
||||||
|
HADOOP-7939. Improve Hadoop subcomponent integration in Hadoop 0.23. (rvs via tucu)
|
||||||
|
|
||||||
|
HADOOP-7988. Upper case in hostname part of the principals doesn't work with
|
||||||
|
kerberos. (jitendra)
|
||||||
|
|
||||||
|
HADOOP-8002. SecurityUtil acquired token message should be a debug rather than info.
|
||||||
|
(Arpit Gupta via mahadev)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
||||||
|
HADOOP-7998. CheckFileSystem does not correctly honor setVerifyChecksum
|
||||||
|
(Daryn Sharp via bobby)
|
||||||
|
|
||||||
HADOOP-7811. TestUserGroupInformation#testGetServerSideGroups test fails in chroot.
|
HADOOP-7811. TestUserGroupInformation#testGetServerSideGroups test fails in chroot.
|
||||||
(Jonathan Eagles via mahadev)
|
(Jonathan Eagles via mahadev)
|
||||||
|
|
||||||
|
@ -227,6 +238,15 @@ Release 0.23.1 - Unreleased
|
||||||
HADOOP-7981. Improve documentation for org.apache.hadoop.io.compress.
|
HADOOP-7981. Improve documentation for org.apache.hadoop.io.compress.
|
||||||
Decompressor.getRemaining (Jonathan Eagles via mahadev)
|
Decompressor.getRemaining (Jonathan Eagles via mahadev)
|
||||||
|
|
||||||
|
HADOOP-7997. SequenceFile.createWriter(...createParent...) no
|
||||||
|
longer works on existing file. (Gregory Chanan via eli)
|
||||||
|
|
||||||
|
HADOOP-7993. Hadoop ignores old-style config options for enabling compressed
|
||||||
|
output. (Anupam Seth via mahadev)
|
||||||
|
|
||||||
|
HADOOP-8000. fetchdt command not available in bin/hadoop.
|
||||||
|
(Arpit Gupta via mahadev)
|
||||||
|
|
||||||
Release 0.23.0 - 2011-11-01
|
Release 0.23.0 - 2011-11-01
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -50,7 +50,7 @@ fi
|
||||||
COMMAND=$1
|
COMMAND=$1
|
||||||
case $COMMAND in
|
case $COMMAND in
|
||||||
#hdfs commands
|
#hdfs commands
|
||||||
namenode|secondarynamenode|datanode|dfs|dfsadmin|fsck|balancer)
|
namenode|secondarynamenode|datanode|dfs|dfsadmin|fsck|balancer|fetchdt)
|
||||||
echo "DEPRECATED: Use of this script to execute hdfs command is deprecated." 1>&2
|
echo "DEPRECATED: Use of this script to execute hdfs command is deprecated." 1>&2
|
||||||
echo "Instead use the hdfs command for it." 1>&2
|
echo "Instead use the hdfs command for it." 1>&2
|
||||||
echo "" 1>&2
|
echo "" 1>&2
|
||||||
|
|
|
@ -25,9 +25,21 @@ common_bin=$(cd -P -- "$(dirname -- "$this")" && pwd -P)
|
||||||
script="$(basename -- "$this")"
|
script="$(basename -- "$this")"
|
||||||
this="$common_bin/$script"
|
this="$common_bin/$script"
|
||||||
|
|
||||||
|
[ -f "$common_bin/hadoop-layout.sh" ] && . "$common_bin/hadoop-layout.sh"
|
||||||
|
|
||||||
|
HADOOP_COMMON_DIR=${HADOOP_COMMON_DIR:-"share/hadoop/common"}
|
||||||
|
HADOOP_COMMON_LIB_JARS_DIR=${HADOOP_COMMON_LIB_JARS_DIR:-"share/hadoop/common/lib"}
|
||||||
|
HADOOP_COMMON_LIB_NATIVE_DIR=${HADOOP_COMMON_LIB_NATIVE_DIR:-"lib/native"}
|
||||||
|
HDFS_DIR=${HDFS_DIR:-"share/hadoop/hdfs"}
|
||||||
|
HDFS_LIB_JARS_DIR=${HDFS_LIB_JARS_DIR:-"share/hadoop/hdfs/lib"}
|
||||||
|
YARN_DIR=${YARN_DIR:-"share/hadoop/mapreduce"}
|
||||||
|
YARN_LIB_JARS_DIR=${YARN_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
|
||||||
|
MAPRED_DIR=${MAPRED_DIR:-"share/hadoop/mapreduce"}
|
||||||
|
MAPRED_LIB_JARS_DIR=${MAPRED_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
|
||||||
|
|
||||||
# the root of the Hadoop installation
|
# the root of the Hadoop installation
|
||||||
# See HADOOP-6255 for directory structure layout
|
# See HADOOP-6255 for directory structure layout
|
||||||
HADOOP_DEFAULT_PREFIX=`dirname "$this"`/..
|
HADOOP_DEFAULT_PREFIX=$(cd -P -- "$common_bin"/.. && pwd -P)
|
||||||
HADOOP_PREFIX=${HADOOP_PREFIX:-$HADOOP_DEFAULT_PREFIX}
|
HADOOP_PREFIX=${HADOOP_PREFIX:-$HADOOP_DEFAULT_PREFIX}
|
||||||
export HADOOP_PREFIX
|
export HADOOP_PREFIX
|
||||||
|
|
||||||
|
@ -144,16 +156,22 @@ CLASSPATH="${HADOOP_CONF_DIR}"
|
||||||
# so that filenames w/ spaces are handled correctly in loops below
|
# so that filenames w/ spaces are handled correctly in loops below
|
||||||
IFS=
|
IFS=
|
||||||
|
|
||||||
|
if [ "$HADOOP_COMMON_HOME" = "" ]; then
|
||||||
|
if [ -d "${HADOOP_PREFIX}/$HADOOP_COMMON_DIR" ]; then
|
||||||
|
HADOOP_COMMON_HOME=$HADOOP_PREFIX
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
# for releases, add core hadoop jar & webapps to CLASSPATH
|
# for releases, add core hadoop jar & webapps to CLASSPATH
|
||||||
if [ -d "$HADOOP_PREFIX/share/hadoop/common/webapps" ]; then
|
if [ -d "$HADOOP_COMMON_HOME/$HADOOP_COMMON_DIR/webapps" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_PREFIX/share/hadoop/common/webapps
|
CLASSPATH=${CLASSPATH}:$HADOOP_COMMON_HOME/$HADOOP_COMMON_DIR
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ -d "$HADOOP_PREFIX/share/hadoop/common/lib" ]; then
|
if [ -d "$HADOOP_COMMON_HOME/$HADOOP_COMMON_LIB_JARS_DIR" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_PREFIX/share/hadoop/common/lib'/*'
|
CLASSPATH=${CLASSPATH}:$HADOOP_COMMON_HOME/$HADOOP_COMMON_LIB_JARS_DIR'/*'
|
||||||
fi
|
fi
|
||||||
|
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_PREFIX/share/hadoop/common'/*'
|
CLASSPATH=${CLASSPATH}:$HADOOP_COMMON_HOME/$HADOOP_COMMON_DIR'/*'
|
||||||
|
|
||||||
# add user-specified CLASSPATH last
|
# add user-specified CLASSPATH last
|
||||||
if [ "$HADOOP_CLASSPATH" != "" ]; then
|
if [ "$HADOOP_CLASSPATH" != "" ]; then
|
||||||
|
@ -185,13 +203,13 @@ fi
|
||||||
|
|
||||||
# setup 'java.library.path' for native-hadoop code if necessary
|
# setup 'java.library.path' for native-hadoop code if necessary
|
||||||
|
|
||||||
if [ -d "${HADOOP_PREFIX}/build/native" -o -d "${HADOOP_PREFIX}/lib/native" ]; then
|
if [ -d "${HADOOP_PREFIX}/build/native" -o -d "${HADOOP_PREFIX}/$HADOOP_COMMON_LIB_NATIVE_DIR" ]; then
|
||||||
|
|
||||||
if [ -d "${HADOOP_PREFIX}/lib/native" ]; then
|
if [ -d "${HADOOP_PREFIX}/$HADOOP_COMMON_LIB_NATIVE_DIR" ]; then
|
||||||
if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
|
if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
|
||||||
JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:${HADOOP_PREFIX}/lib/native
|
JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:${HADOOP_PREFIX}/$HADOOP_COMMON_LIB_NATIVE_DIR
|
||||||
else
|
else
|
||||||
JAVA_LIBRARY_PATH=${HADOOP_PREFIX}/lib/native
|
JAVA_LIBRARY_PATH=${HADOOP_PREFIX}/$HADOOP_COMMON_LIB_NATIVE_DIR
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
@ -216,37 +234,56 @@ HADOOP_OPTS="$HADOOP_OPTS -Djava.net.preferIPv4Stack=true"
|
||||||
|
|
||||||
# put hdfs in classpath if present
|
# put hdfs in classpath if present
|
||||||
if [ "$HADOOP_HDFS_HOME" = "" ]; then
|
if [ "$HADOOP_HDFS_HOME" = "" ]; then
|
||||||
if [ -d "${HADOOP_PREFIX}/share/hadoop/hdfs" ]; then
|
if [ -d "${HADOOP_PREFIX}/$HDFS_DIR" ]; then
|
||||||
HADOOP_HDFS_HOME=$HADOOP_PREFIX
|
HADOOP_HDFS_HOME=$HADOOP_PREFIX
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ -d "$HADOOP_HDFS_HOME/share/hadoop/hdfs/webapps" ]; then
|
if [ -d "$HADOOP_HDFS_HOME/$HDFS_DIR/webapps" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/share/hadoop/hdfs
|
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/$HDFS_DIR
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ -d "$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib" ]; then
|
if [ -d "$HADOOP_HDFS_HOME/$HDFS_LIB_JARS_DIR" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib'/*'
|
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/$HDFS_LIB_JARS_DIR'/*'
|
||||||
fi
|
fi
|
||||||
|
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/share/hadoop/hdfs'/*'
|
CLASSPATH=${CLASSPATH}:$HADOOP_HDFS_HOME/$HDFS_DIR'/*'
|
||||||
|
|
||||||
# put yarn in classpath if present
|
# put yarn in classpath if present
|
||||||
if [ "$YARN_HOME" = "" ]; then
|
if [ "$YARN_HOME" = "" ]; then
|
||||||
if [ -d "${HADOOP_PREFIX}/share/hadoop/mapreduce" ]; then
|
if [ -d "${HADOOP_PREFIX}/$YARN_DIR" ]; then
|
||||||
YARN_HOME=$HADOOP_PREFIX
|
YARN_HOME=$HADOOP_PREFIX
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ -d "$YARN_HOME/share/hadoop/mapreduce/webapps" ]; then
|
if [ -d "$YARN_HOME/$YARN_DIR/webapps" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/share/hadoop/mapreduce
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/$YARN_DIR
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ -d "$YARN_HOME/share/hadoop/mapreduce/lib" ]; then
|
if [ -d "$YARN_HOME/$YARN_LIB_JARS_DIR" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/share/hadoop/mapreduce/lib'/*'
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/$YARN_LIB_JARS_DIR'/*'
|
||||||
fi
|
fi
|
||||||
|
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/share/hadoop/mapreduce'/*'
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/$YARN_DIR'/*'
|
||||||
|
|
||||||
|
# put mapred in classpath if present AND different from YARN
|
||||||
|
if [ "$HADOOP_MAPRED_HOME" = "" ]; then
|
||||||
|
if [ -d "${HADOOP_PREFIX}/$MAPRED_DIR" ]; then
|
||||||
|
HADOOP_MAPRED_HOME=$HADOOP_PREFIX
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ "$HADOOP_MAPRED_HOME/$MAPRED_DIR" != "$YARN_HOME/$YARN_DIR" ] ; then
|
||||||
|
if [ -d "$HADOOP_MAPRED_HOME/$MAPRED_DIR/webapps" ]; then
|
||||||
|
CLASSPATH=${CLASSPATH}:$HADOOP_MAPRED_HOME/$MAPRED_DIR
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -d "$HADOOP_MAPRED_HOME/$MAPRED_LIB_JARS_DIR" ]; then
|
||||||
|
CLASSPATH=${CLASSPATH}:$HADOOP_MAPRED_HOME/$MAPRED_LIB_JARS_DIR'/*'
|
||||||
|
fi
|
||||||
|
|
||||||
|
CLASSPATH=${CLASSPATH}:$HADOOP_MAPRED_HOME/$MAPRED_DIR'/*'
|
||||||
|
fi
|
||||||
|
|
||||||
# cygwin path translation
|
# cygwin path translation
|
||||||
if $cygwin; then
|
if $cygwin; then
|
||||||
|
|
|
@ -346,6 +346,16 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
|
||||||
return name;
|
return name;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private void handleDeprecation() {
|
||||||
|
LOG.debug("Handling deprecation for all properties in config...");
|
||||||
|
Set<Object> keys = new HashSet<Object>();
|
||||||
|
keys.addAll(getProps().keySet());
|
||||||
|
for (Object item: keys) {
|
||||||
|
LOG.debug("Handling deprecation for " + (String)item);
|
||||||
|
handleDeprecation((String)item);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static{
|
static{
|
||||||
//print deprecation warning if hadoop-site.xml is found in classpath
|
//print deprecation warning if hadoop-site.xml is found in classpath
|
||||||
ClassLoader cL = Thread.currentThread().getContextClassLoader();
|
ClassLoader cL = Thread.currentThread().getContextClassLoader();
|
||||||
|
@ -1665,7 +1675,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
|
||||||
Element conf = doc.createElement("configuration");
|
Element conf = doc.createElement("configuration");
|
||||||
doc.appendChild(conf);
|
doc.appendChild(conf);
|
||||||
conf.appendChild(doc.createTextNode("\n"));
|
conf.appendChild(doc.createTextNode("\n"));
|
||||||
getProps(); // ensure properties is set
|
handleDeprecation(); //ensure properties is set and deprecation is handled
|
||||||
for (Enumeration e = properties.keys(); e.hasMoreElements();) {
|
for (Enumeration e = properties.keys(); e.hasMoreElements();) {
|
||||||
String name = (String)e.nextElement();
|
String name = (String)e.nextElement();
|
||||||
Object object = properties.get(name);
|
Object object = properties.get(name);
|
||||||
|
|
|
@ -304,8 +304,9 @@ public abstract class ChecksumFileSystem extends FilterFileSystem {
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
public FSDataInputStream open(Path f, int bufferSize) throws IOException {
|
public FSDataInputStream open(Path f, int bufferSize) throws IOException {
|
||||||
return new FSDataInputStream(
|
return verifyChecksum
|
||||||
new ChecksumFSInputChecker(this, f, bufferSize));
|
? new FSDataInputStream(new ChecksumFSInputChecker(this, f, bufferSize))
|
||||||
|
: getRawFileSystem().open(f, bufferSize);
|
||||||
}
|
}
|
||||||
|
|
||||||
/** {@inheritDoc} */
|
/** {@inheritDoc} */
|
||||||
|
|
|
@ -467,7 +467,7 @@ public class SequenceFile {
|
||||||
Metadata metadata) throws IOException {
|
Metadata metadata) throws IOException {
|
||||||
return createWriter(FileContext.getFileContext(fs.getUri(), conf),
|
return createWriter(FileContext.getFileContext(fs.getUri(), conf),
|
||||||
conf, name, keyClass, valClass, compressionType, codec,
|
conf, name, keyClass, valClass, compressionType, codec,
|
||||||
metadata, EnumSet.of(CreateFlag.CREATE),
|
metadata, EnumSet.of(CreateFlag.CREATE,CreateFlag.OVERWRITE),
|
||||||
CreateOpts.bufferSize(bufferSize),
|
CreateOpts.bufferSize(bufferSize),
|
||||||
createParent ? CreateOpts.createParent()
|
createParent ? CreateOpts.createParent()
|
||||||
: CreateOpts.donotCreateParent(),
|
: CreateOpts.donotCreateParent(),
|
||||||
|
|
|
@ -236,7 +236,7 @@ public class SecurityUtil {
|
||||||
if (fqdn == null || fqdn.equals("") || fqdn.equals("0.0.0.0")) {
|
if (fqdn == null || fqdn.equals("") || fqdn.equals("0.0.0.0")) {
|
||||||
fqdn = getLocalHostName();
|
fqdn = getLocalHostName();
|
||||||
}
|
}
|
||||||
return components[0] + "/" + fqdn + "@" + components[2];
|
return components[0] + "/" + fqdn.toLowerCase() + "@" + components[2];
|
||||||
}
|
}
|
||||||
|
|
||||||
static String getLocalHostName() throws UnknownHostException {
|
static String getLocalHostName() throws UnknownHostException {
|
||||||
|
@ -409,7 +409,9 @@ public class SecurityUtil {
|
||||||
Text service = buildTokenService(addr);
|
Text service = buildTokenService(addr);
|
||||||
if (token != null) {
|
if (token != null) {
|
||||||
token.setService(service);
|
token.setService(service);
|
||||||
LOG.info("Acquired token "+token); // Token#toString() prints service
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Acquired token "+token); // Token#toString() prints service
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
LOG.warn("Failed to get token for service "+service);
|
LOG.warn("Failed to get token for service "+service);
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,6 +18,8 @@
|
||||||
|
|
||||||
package org.apache.hadoop.conf;
|
package org.apache.hadoop.conf;
|
||||||
|
|
||||||
|
import java.io.ByteArrayOutputStream;
|
||||||
|
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||||
|
|
||||||
|
@ -32,4 +34,22 @@ public class TestDeprecatedKeys extends TestCase {
|
||||||
String scriptFile = conf.get(CommonConfigurationKeys.NET_TOPOLOGY_SCRIPT_FILE_NAME_KEY);
|
String scriptFile = conf.get(CommonConfigurationKeys.NET_TOPOLOGY_SCRIPT_FILE_NAME_KEY);
|
||||||
assertTrue(scriptFile.equals("xyz")) ;
|
assertTrue(scriptFile.equals("xyz")) ;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//Tests reading / writing a conf file with deprecation after setting
|
||||||
|
public void testReadWriteWithDeprecatedKeys() throws Exception {
|
||||||
|
Configuration conf = new Configuration();
|
||||||
|
conf.setBoolean("old.config.yet.to.be.deprecated", true);
|
||||||
|
Configuration.addDeprecation("old.config.yet.to.be.deprecated",
|
||||||
|
new String[]{"new.conf.to.replace.deprecated.conf"});
|
||||||
|
ByteArrayOutputStream out=new ByteArrayOutputStream();
|
||||||
|
String fileContents;
|
||||||
|
try {
|
||||||
|
conf.writeXml(out);
|
||||||
|
fileContents = out.toString();
|
||||||
|
} finally {
|
||||||
|
out.close();
|
||||||
|
}
|
||||||
|
assertTrue(fileContents.contains("old.config.yet.to.be.deprecated"));
|
||||||
|
assertTrue(fileContents.contains("new.conf.to.replace.deprecated.conf"));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -22,12 +22,22 @@ import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.FSDataOutputStream;
|
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
import static org.apache.hadoop.fs.FileSystemTestHelper.*;
|
import static org.apache.hadoop.fs.FileSystemTestHelper.*;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import junit.framework.TestCase;
|
import org.junit.*;
|
||||||
|
import static org.junit.Assert.*;
|
||||||
|
|
||||||
public class TestChecksumFileSystem extends TestCase {
|
public class TestChecksumFileSystem {
|
||||||
static final String TEST_ROOT_DIR
|
static final String TEST_ROOT_DIR
|
||||||
= System.getProperty("test.build.data","build/test/data/work-dir/localfs");
|
= System.getProperty("test.build.data","build/test/data/work-dir/localfs");
|
||||||
|
|
||||||
|
static LocalFileSystem localFs;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void resetLocalFs() throws Exception {
|
||||||
|
localFs = FileSystem.getLocal(new Configuration());
|
||||||
|
localFs.setVerifyChecksum(true);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testgetChecksumLength() throws Exception {
|
public void testgetChecksumLength() throws Exception {
|
||||||
assertEquals(8, ChecksumFileSystem.getChecksumLength(0L, 512));
|
assertEquals(8, ChecksumFileSystem.getChecksumLength(0L, 512));
|
||||||
assertEquals(12, ChecksumFileSystem.getChecksumLength(1L, 512));
|
assertEquals(12, ChecksumFileSystem.getChecksumLength(1L, 512));
|
||||||
|
@ -40,9 +50,8 @@ public class TestChecksumFileSystem extends TestCase {
|
||||||
ChecksumFileSystem.getChecksumLength(10000000000000L, 10));
|
ChecksumFileSystem.getChecksumLength(10000000000000L, 10));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testVerifyChecksum() throws Exception {
|
public void testVerifyChecksum() throws Exception {
|
||||||
Configuration conf = new Configuration();
|
|
||||||
LocalFileSystem localFs = FileSystem.getLocal(conf);
|
|
||||||
Path testPath = new Path(TEST_ROOT_DIR, "testPath");
|
Path testPath = new Path(TEST_ROOT_DIR, "testPath");
|
||||||
Path testPath11 = new Path(TEST_ROOT_DIR, "testPath11");
|
Path testPath11 = new Path(TEST_ROOT_DIR, "testPath11");
|
||||||
FSDataOutputStream fout = localFs.create(testPath);
|
FSDataOutputStream fout = localFs.create(testPath);
|
||||||
|
@ -68,7 +77,7 @@ public class TestChecksumFileSystem extends TestCase {
|
||||||
|
|
||||||
//copying the wrong checksum file
|
//copying the wrong checksum file
|
||||||
FileUtil.copy(localFs, localFs.getChecksumFile(testPath11), localFs,
|
FileUtil.copy(localFs, localFs.getChecksumFile(testPath11), localFs,
|
||||||
localFs.getChecksumFile(testPath),false,true,conf);
|
localFs.getChecksumFile(testPath),false,true,localFs.getConf());
|
||||||
assertTrue("checksum exists", localFs.exists(localFs.getChecksumFile(testPath)));
|
assertTrue("checksum exists", localFs.exists(localFs.getChecksumFile(testPath)));
|
||||||
|
|
||||||
boolean errorRead = false;
|
boolean errorRead = false;
|
||||||
|
@ -80,20 +89,13 @@ public class TestChecksumFileSystem extends TestCase {
|
||||||
assertTrue("error reading", errorRead);
|
assertTrue("error reading", errorRead);
|
||||||
|
|
||||||
//now setting verify false, the read should succeed
|
//now setting verify false, the read should succeed
|
||||||
try {
|
localFs.setVerifyChecksum(false);
|
||||||
localFs.setVerifyChecksum(false);
|
String str = readFile(localFs, testPath, 1024).toString();
|
||||||
String str = readFile(localFs, testPath, 1024).toString();
|
assertTrue("read", "testing".equals(str));
|
||||||
assertTrue("read", "testing".equals(str));
|
|
||||||
} finally {
|
|
||||||
// reset for other tests
|
|
||||||
localFs.setVerifyChecksum(true);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testMultiChunkFile() throws Exception {
|
public void testMultiChunkFile() throws Exception {
|
||||||
Configuration conf = new Configuration();
|
|
||||||
LocalFileSystem localFs = FileSystem.getLocal(conf);
|
|
||||||
Path testPath = new Path(TEST_ROOT_DIR, "testMultiChunk");
|
Path testPath = new Path(TEST_ROOT_DIR, "testMultiChunk");
|
||||||
FSDataOutputStream fout = localFs.create(testPath);
|
FSDataOutputStream fout = localFs.create(testPath);
|
||||||
for (int i = 0; i < 1000; i++) {
|
for (int i = 0; i < 1000; i++) {
|
||||||
|
@ -116,9 +118,8 @@ public class TestChecksumFileSystem extends TestCase {
|
||||||
* Test to ensure that if the checksum file is truncated, a
|
* Test to ensure that if the checksum file is truncated, a
|
||||||
* ChecksumException is thrown
|
* ChecksumException is thrown
|
||||||
*/
|
*/
|
||||||
|
@Test
|
||||||
public void testTruncatedChecksum() throws Exception {
|
public void testTruncatedChecksum() throws Exception {
|
||||||
Configuration conf = new Configuration();
|
|
||||||
LocalFileSystem localFs = FileSystem.getLocal(conf);
|
|
||||||
Path testPath = new Path(TEST_ROOT_DIR, "testtruncatedcrc");
|
Path testPath = new Path(TEST_ROOT_DIR, "testtruncatedcrc");
|
||||||
FSDataOutputStream fout = localFs.create(testPath);
|
FSDataOutputStream fout = localFs.create(testPath);
|
||||||
fout.write("testing truncation".getBytes());
|
fout.write("testing truncation".getBytes());
|
||||||
|
@ -146,14 +147,60 @@ public class TestChecksumFileSystem extends TestCase {
|
||||||
}
|
}
|
||||||
|
|
||||||
// telling it not to verify checksums, should avoid issue.
|
// telling it not to verify checksums, should avoid issue.
|
||||||
|
localFs.setVerifyChecksum(false);
|
||||||
|
String str = readFile(localFs, testPath, 1024).toString();
|
||||||
|
assertTrue("read", "testing truncation".equals(str));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testStreamType() throws Exception {
|
||||||
|
Path testPath = new Path(TEST_ROOT_DIR, "testStreamType");
|
||||||
|
localFs.create(testPath).close();
|
||||||
|
FSDataInputStream in = null;
|
||||||
|
|
||||||
|
localFs.setVerifyChecksum(true);
|
||||||
|
in = localFs.open(testPath);
|
||||||
|
assertTrue("stream is input checker",
|
||||||
|
in.getWrappedStream() instanceof FSInputChecker);
|
||||||
|
|
||||||
|
localFs.setVerifyChecksum(false);
|
||||||
|
in = localFs.open(testPath);
|
||||||
|
assertFalse("stream is not input checker",
|
||||||
|
in.getWrappedStream() instanceof FSInputChecker);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCorruptedChecksum() throws Exception {
|
||||||
|
Path testPath = new Path(TEST_ROOT_DIR, "testCorruptChecksum");
|
||||||
|
Path checksumPath = localFs.getChecksumFile(testPath);
|
||||||
|
|
||||||
|
// write a file to generate checksum
|
||||||
|
FSDataOutputStream out = localFs.create(testPath, true);
|
||||||
|
out.write("testing 1 2 3".getBytes());
|
||||||
|
out.close();
|
||||||
|
assertTrue(localFs.exists(checksumPath));
|
||||||
|
FileStatus stat = localFs.getFileStatus(checksumPath);
|
||||||
|
|
||||||
|
// alter file directly so checksum is invalid
|
||||||
|
out = localFs.getRawFileSystem().create(testPath, true);
|
||||||
|
out.write("testing stale checksum".getBytes());
|
||||||
|
out.close();
|
||||||
|
assertTrue(localFs.exists(checksumPath));
|
||||||
|
// checksum didn't change on disk
|
||||||
|
assertEquals(stat, localFs.getFileStatus(checksumPath));
|
||||||
|
|
||||||
|
Exception e = null;
|
||||||
try {
|
try {
|
||||||
localFs.setVerifyChecksum(false);
|
|
||||||
String str = readFile(localFs, testPath, 1024).toString();
|
|
||||||
assertTrue("read", "testing truncation".equals(str));
|
|
||||||
} finally {
|
|
||||||
// reset for other tests
|
|
||||||
localFs.setVerifyChecksum(true);
|
localFs.setVerifyChecksum(true);
|
||||||
|
readFile(localFs, testPath, 1024);
|
||||||
|
} catch (ChecksumException ce) {
|
||||||
|
e = ce;
|
||||||
|
} finally {
|
||||||
|
assertNotNull("got checksum error", e);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
localFs.setVerifyChecksum(false);
|
||||||
|
String str = readFile(localFs, testPath, 1024);
|
||||||
|
assertEquals("testing stale checksum", str);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -517,6 +517,23 @@ public class TestSequenceFile extends TestCase {
|
||||||
assertTrue("InputStream for " + path + " should have been closed.", openedFile[0].isClosed());
|
assertTrue("InputStream for " + path + " should have been closed.", openedFile[0].isClosed());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test that makes sure createWriter succeeds on a file that was
|
||||||
|
* already created
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
public void testCreateWriterOnExistingFile() throws IOException {
|
||||||
|
Configuration conf = new Configuration();
|
||||||
|
FileSystem fs = FileSystem.getLocal(conf);
|
||||||
|
Path name = new Path(new Path(System.getProperty("test.build.data","."),
|
||||||
|
"createWriterOnExistingFile") , "file");
|
||||||
|
|
||||||
|
fs.create(name);
|
||||||
|
SequenceFile.createWriter(fs, conf, name, RandomDatum.class,
|
||||||
|
RandomDatum.class, 512, (short) 1, 4096, false,
|
||||||
|
CompressionType.NONE, null, new Metadata());
|
||||||
|
}
|
||||||
|
|
||||||
public void testRecursiveSeqFileCreate() throws IOException {
|
public void testRecursiveSeqFileCreate() throws IOException {
|
||||||
Configuration conf = new Configuration();
|
Configuration conf = new Configuration();
|
||||||
FileSystem fs = FileSystem.getLocal(conf);
|
FileSystem fs = FileSystem.getLocal(conf);
|
||||||
|
|
|
@ -89,6 +89,16 @@ public class TestSecurityUtil {
|
||||||
Mockito.verify(notUsed, Mockito.never()).getCanonicalHostName();
|
Mockito.verify(notUsed, Mockito.never()).getCanonicalHostName();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testPrincipalsWithLowerCaseHosts() throws IOException {
|
||||||
|
String service = "xyz/";
|
||||||
|
String realm = "@REALM";
|
||||||
|
String principalInConf = service + SecurityUtil.HOSTNAME_PATTERN + realm;
|
||||||
|
String hostname = "FooHost";
|
||||||
|
String principal = service + hostname.toLowerCase() + realm;
|
||||||
|
verify(principalInConf, hostname, principal);
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testLocalHostNameForNullOrWild() throws Exception {
|
public void testLocalHostNameForNullOrWild() throws Exception {
|
||||||
String local = SecurityUtil.getLocalHostName();
|
String local = SecurityUtil.getLocalHostName();
|
||||||
|
|
|
@ -55,8 +55,8 @@ if [ "${1}" = "stop" ]; then
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ "${HTTPFS_SILENT}" != "true" ]; then
|
if [ "${HTTPFS_SILENT}" != "true" ]; then
|
||||||
${BASEDIR}/share/hadoop/httpfs/tomcat/bin/catalina.sh "$@"
|
${CATALINA_BASE:-"${BASEDIR}/share/hadoop/httpfs/tomcat"}/bin/catalina.sh "$@"
|
||||||
else
|
else
|
||||||
${BASEDIR}/share/hadoop/httpfs/tomcat/bin/catalina.sh "$@" > /dev/null
|
${CATALINA_BASE:-"${BASEDIR}/share/hadoop/httpfs/tomcat"}/bin/catalina.sh "$@" > /dev/null
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
|
@ -47,7 +47,7 @@ public class TestHostnameFilter extends HTestCase {
|
||||||
@Override
|
@Override
|
||||||
public void doFilter(ServletRequest servletRequest, ServletResponse servletResponse)
|
public void doFilter(ServletRequest servletRequest, ServletResponse servletResponse)
|
||||||
throws IOException, ServletException {
|
throws IOException, ServletException {
|
||||||
Assert.assertEquals(HostnameFilter.get(), "localhost");
|
Assert.assertTrue(HostnameFilter.get().contains("localhost"));
|
||||||
invoked.set(true);
|
invoked.set(true);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
|
@ -295,6 +295,11 @@ Release 0.23.1 - UNRELEASED
|
||||||
|
|
||||||
HDFS-2837. mvn javadoc:javadoc not seeing LimitedPrivate class (revans2 via tucu)
|
HDFS-2837. mvn javadoc:javadoc not seeing LimitedPrivate class (revans2 via tucu)
|
||||||
|
|
||||||
|
HDFS-2840. TestHostnameFilter should work with localhost or localhost.localdomain (tucu)
|
||||||
|
|
||||||
|
HDFS-2791. If block report races with closing of file, replica is
|
||||||
|
incorrectly marked corrupt. (todd)
|
||||||
|
|
||||||
Release 0.23.0 - 2011-11-01
|
Release 0.23.0 - 2011-11-01
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -22,8 +22,6 @@ bin=`which "$0"`
|
||||||
bin=`dirname "${bin}"`
|
bin=`dirname "${bin}"`
|
||||||
bin=`cd "$bin"; pwd`
|
bin=`cd "$bin"; pwd`
|
||||||
|
|
||||||
export HADOOP_PREFIX="${HADOOP_PREFIX:-$bin/..}"
|
|
||||||
|
|
||||||
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
||||||
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
||||||
if [ -e "${HADOOP_LIBEXEC_DIR}/hadoop-config.sh" ]; then
|
if [ -e "${HADOOP_LIBEXEC_DIR}/hadoop-config.sh" ]; then
|
||||||
|
|
|
@ -1571,7 +1571,24 @@ public class BlockManager {
|
||||||
}
|
}
|
||||||
case RBW:
|
case RBW:
|
||||||
case RWR:
|
case RWR:
|
||||||
return storedBlock.isComplete();
|
if (!storedBlock.isComplete()) {
|
||||||
|
return false;
|
||||||
|
} else if (storedBlock.getGenerationStamp() != iblk.getGenerationStamp()) {
|
||||||
|
return true;
|
||||||
|
} else { // COMPLETE block, same genstamp
|
||||||
|
if (reportedState == ReplicaState.RBW) {
|
||||||
|
// If it's a RBW report for a COMPLETE block, it may just be that
|
||||||
|
// the block report got a little bit delayed after the pipeline
|
||||||
|
// closed. So, ignore this report, assuming we will get a
|
||||||
|
// FINALIZED replica later. See HDFS-2791
|
||||||
|
LOG.info("Received an RBW replica for block " + storedBlock +
|
||||||
|
" on " + dn.getName() + ": ignoring it, since the block is " +
|
||||||
|
"complete with the same generation stamp.");
|
||||||
|
return false;
|
||||||
|
} else {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
||||||
case RUR: // should not be reported
|
case RUR: // should not be reported
|
||||||
case TEMPORARY: // should not be reported
|
case TEMPORARY: // should not be reported
|
||||||
default:
|
default:
|
||||||
|
|
|
@ -781,4 +781,13 @@ class BPOfferService implements Runnable {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
DatanodeProtocol getBpNamenode() {
|
||||||
|
return bpNamenode;
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
void setBpNamenode(DatanodeProtocol bpNamenode) {
|
||||||
|
this.bpNamenode = bpNamenode;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -101,7 +101,7 @@ public class AppendTestUtil {
|
||||||
return DFSTestUtil.getFileSystemAs(ugi, conf);
|
return DFSTestUtil.getFileSystemAs(ugi, conf);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void write(OutputStream out, int offset, int length) throws IOException {
|
public static void write(OutputStream out, int offset, int length) throws IOException {
|
||||||
final byte[] bytes = new byte[length];
|
final byte[] bytes = new byte[length];
|
||||||
for(int i = 0; i < length; i++) {
|
for(int i = 0; i < length; i++) {
|
||||||
bytes[i] = (byte)(offset + i);
|
bytes[i] = (byte)(offset + i);
|
||||||
|
|
|
@ -17,6 +17,15 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.datanode;
|
package org.apache.hadoop.hdfs.server.datanode;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.DatanodeProtocol;
|
||||||
|
import org.apache.hadoop.hdfs.server.namenode.NameNode;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.mockito.Mockito;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* WARNING!! This is TEST ONLY class: it never has to be used
|
* WARNING!! This is TEST ONLY class: it never has to be used
|
||||||
* for ANY development purposes.
|
* for ANY development purposes.
|
||||||
|
@ -42,4 +51,34 @@ public class DataNodeAdapter {
|
||||||
boolean heartbeatsDisabledForTests) {
|
boolean heartbeatsDisabledForTests) {
|
||||||
dn.setHeartbeatsDisabledForTests(heartbeatsDisabledForTests);
|
dn.setHeartbeatsDisabledForTests(heartbeatsDisabledForTests);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Insert a Mockito spy object between the given DataNode and
|
||||||
|
* the given NameNode. This can be used to delay or wait for
|
||||||
|
* RPC calls on the datanode->NN path.
|
||||||
|
*/
|
||||||
|
public static DatanodeProtocol spyOnBposToNN(
|
||||||
|
DataNode dn, NameNode nn) {
|
||||||
|
String bpid = nn.getNamesystem().getBlockPoolId();
|
||||||
|
|
||||||
|
BPOfferService bpos = null;
|
||||||
|
for (BPOfferService thisBpos : dn.getAllBpOs()) {
|
||||||
|
if (thisBpos.getBlockPoolId().equals(bpid)) {
|
||||||
|
bpos = thisBpos;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Preconditions.checkArgument(bpos != null,
|
||||||
|
"No such bpid: %s", bpid);
|
||||||
|
|
||||||
|
// When protobufs are merged, the following can be converted
|
||||||
|
// to a simple spy. Because you can't spy on proxy objects,
|
||||||
|
// we have to use the DelegateAnswer trick.
|
||||||
|
DatanodeProtocol origNN = bpos.getBpNamenode();
|
||||||
|
DatanodeProtocol spy = Mockito.mock(DatanodeProtocol.class,
|
||||||
|
new GenericTestUtils.DelegateAnswer(origNN));
|
||||||
|
|
||||||
|
bpos.setBpNamenode(spy);
|
||||||
|
return spy;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -21,7 +21,9 @@ import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.commons.logging.impl.Log4JLogger;
|
import org.apache.commons.logging.impl.Log4JLogger;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hdfs.AppendTestUtil;
|
||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.DFSTestUtil;
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||||
|
@ -35,14 +37,19 @@ import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.NameNode;
|
import org.apache.hadoop.hdfs.server.namenode.NameNode;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.DatanodeCommand;
|
import org.apache.hadoop.hdfs.server.protocol.DatanodeCommand;
|
||||||
|
import org.apache.hadoop.hdfs.server.protocol.DatanodeProtocol;
|
||||||
import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration;
|
import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration;
|
||||||
|
import org.apache.hadoop.io.IOUtils;
|
||||||
import org.apache.hadoop.test.GenericTestUtils;
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils.DelayAnswer;
|
||||||
import org.apache.log4j.Level;
|
import org.apache.log4j.Level;
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import org.junit.Before;
|
import org.junit.Before;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
import org.mockito.Mockito;
|
||||||
|
import org.mockito.invocation.InvocationOnMock;
|
||||||
|
|
||||||
import java.io.File;
|
import java.io.File;
|
||||||
import java.io.FilenameFilter;
|
import java.io.FilenameFilter;
|
||||||
|
@ -50,6 +57,7 @@ import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Random;
|
import java.util.Random;
|
||||||
|
import java.util.concurrent.CountDownLatch;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This test simulates a variety of situations when blocks are being
|
* This test simulates a variety of situations when blocks are being
|
||||||
|
@ -492,6 +500,84 @@ public class TestBlockReport {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test for the case where one of the DNs in the pipeline is in the
|
||||||
|
* process of doing a block report exactly when the block is closed.
|
||||||
|
* In this case, the block report becomes delayed until after the
|
||||||
|
* block is marked completed on the NN, and hence it reports an RBW
|
||||||
|
* replica for a COMPLETE block. Such a report should not be marked
|
||||||
|
* corrupt.
|
||||||
|
* This is a regression test for HDFS-2791.
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testOneReplicaRbwReportArrivesAfterBlockCompleted() throws Exception {
|
||||||
|
final CountDownLatch brFinished = new CountDownLatch(1);
|
||||||
|
DelayAnswer delayer = new GenericTestUtils.DelayAnswer(LOG) {
|
||||||
|
@Override
|
||||||
|
protected Object passThrough(InvocationOnMock invocation)
|
||||||
|
throws Throwable {
|
||||||
|
try {
|
||||||
|
return super.passThrough(invocation);
|
||||||
|
} finally {
|
||||||
|
// inform the test that our block report went through.
|
||||||
|
brFinished.countDown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
final String METHOD_NAME = GenericTestUtils.getMethodName();
|
||||||
|
Path filePath = new Path("/" + METHOD_NAME + ".dat");
|
||||||
|
|
||||||
|
// Start a second DN for this test -- we're checking
|
||||||
|
// what happens when one of the DNs is slowed for some reason.
|
||||||
|
REPL_FACTOR = 2;
|
||||||
|
startDNandWait(null, false);
|
||||||
|
|
||||||
|
NameNode nn = cluster.getNameNode();
|
||||||
|
|
||||||
|
FSDataOutputStream out = fs.create(filePath, REPL_FACTOR);
|
||||||
|
try {
|
||||||
|
AppendTestUtil.write(out, 0, 10);
|
||||||
|
out.hflush();
|
||||||
|
|
||||||
|
// Set up a spy so that we can delay the block report coming
|
||||||
|
// from this node.
|
||||||
|
DataNode dn = cluster.getDataNodes().get(0);
|
||||||
|
DatanodeProtocol spy =
|
||||||
|
DataNodeAdapter.spyOnBposToNN(dn, nn);
|
||||||
|
|
||||||
|
Mockito.doAnswer(delayer)
|
||||||
|
.when(spy).blockReport(
|
||||||
|
Mockito.<DatanodeRegistration>anyObject(),
|
||||||
|
Mockito.anyString(),
|
||||||
|
Mockito.<long[]>anyObject());
|
||||||
|
|
||||||
|
// Force a block report to be generated. The block report will have
|
||||||
|
// an RBW replica in it. Wait for the RPC to be sent, but block
|
||||||
|
// it before it gets to the NN.
|
||||||
|
dn.scheduleAllBlockReport(0);
|
||||||
|
delayer.waitForCall();
|
||||||
|
|
||||||
|
} finally {
|
||||||
|
IOUtils.closeStream(out);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now that the stream is closed, the NN will have the block in COMPLETE
|
||||||
|
// state.
|
||||||
|
delayer.proceed();
|
||||||
|
brFinished.await();
|
||||||
|
|
||||||
|
// Verify that no replicas are marked corrupt, and that the
|
||||||
|
// file is still readable.
|
||||||
|
BlockManagerTestUtil.updateState(nn.getNamesystem().getBlockManager());
|
||||||
|
assertEquals(0, nn.getNamesystem().getCorruptReplicaBlocks());
|
||||||
|
DFSTestUtil.readFile(fs, filePath);
|
||||||
|
|
||||||
|
// Ensure that the file is readable even from the DN that we futzed with.
|
||||||
|
cluster.stopDataNode(1);
|
||||||
|
DFSTestUtil.readFile(fs, filePath);
|
||||||
|
}
|
||||||
|
|
||||||
private void waitForTempReplica(Block bl, int DN_N1) throws IOException {
|
private void waitForTempReplica(Block bl, int DN_N1) throws IOException {
|
||||||
final boolean tooLongWait = false;
|
final boolean tooLongWait = false;
|
||||||
final int TIMEOUT = 40000;
|
final int TIMEOUT = 40000;
|
||||||
|
|
|
@ -36,6 +36,8 @@ Release 0.23.1 - Unreleased
|
||||||
MAPREDUCE-778. Rumen Anonymizer. (Amar Kamat and Chris Douglas via amarrk)
|
MAPREDUCE-778. Rumen Anonymizer. (Amar Kamat and Chris Douglas via amarrk)
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
|
MAPREDUCE-3481. [Gridmix] Improve Gridmix STRESS mode. (amarrk)
|
||||||
|
|
||||||
MAPREDUCE-3597. [Rumen] Rumen should provide APIs to access all the
|
MAPREDUCE-3597. [Rumen] Rumen should provide APIs to access all the
|
||||||
job-history related information.
|
job-history related information.
|
||||||
|
|
||||||
|
@ -153,6 +155,15 @@ Release 0.23.1 - Unreleased
|
||||||
MAPREDUCE-3737. The Web Application Proxy's is not documented very well.
|
MAPREDUCE-3737. The Web Application Proxy's is not documented very well.
|
||||||
(Robert Evans via mahadev)
|
(Robert Evans via mahadev)
|
||||||
|
|
||||||
|
MAPREDUCE-3699. Increased RPC handlers for all YARN servers to reasonable
|
||||||
|
values for working at scale. (Hitesh Shah via vinodkv)
|
||||||
|
|
||||||
|
MAPREDUCE-3693. Added mapreduce.admin.user.env to mapred-default.xml.
|
||||||
|
(Roman Shapshonik via acmurthy)
|
||||||
|
|
||||||
|
MAPREDUCE-3732. Modified CapacityScheduler to use only users with pending
|
||||||
|
requests for computing user-limits. (Arun C Murthy via vinodkv)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
MAPREDUCE-3567. Extraneous JobConf objects in AM heap. (Vinod Kumar
|
MAPREDUCE-3567. Extraneous JobConf objects in AM heap. (Vinod Kumar
|
||||||
|
@ -179,7 +190,13 @@ Release 0.23.1 - Unreleased
|
||||||
MAPREDUCE-3718. Change default AM heartbeat interval to 1 second. (Hitesh
|
MAPREDUCE-3718. Change default AM heartbeat interval to 1 second. (Hitesh
|
||||||
Shah via sseth)
|
Shah via sseth)
|
||||||
|
|
||||||
|
MAPREDUCE-3360. Added information about lost/rebooted/decommissioned nodes
|
||||||
|
on the webapps. (Bhallamudi Venkata Siva Kamesh and Jason Lowe via vinodkv)
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
MAPREDUCE-2784. [Gridmix] Bug fixes in ExecutionSummarizer and
|
||||||
|
ResourceUsageMatcher. (amarrk)
|
||||||
|
|
||||||
MAPREDUCE-3194. "mapred mradmin" command is broken in mrv2
|
MAPREDUCE-3194. "mapred mradmin" command is broken in mrv2
|
||||||
(Jason Lowe via bobby)
|
(Jason Lowe via bobby)
|
||||||
|
|
||||||
|
@ -558,6 +575,12 @@ Release 0.23.1 - Unreleased
|
||||||
MAPREDUCE-3735. Add distcp jar to the distribution (tar).
|
MAPREDUCE-3735. Add distcp jar to the distribution (tar).
|
||||||
(mahadev)
|
(mahadev)
|
||||||
|
|
||||||
|
MAPREDUCE-3720. Changed bin/mapred job -list to not print job-specific
|
||||||
|
information not available at RM. (vinodkv via acmurthy)
|
||||||
|
|
||||||
|
MAPREDUCE-3742. "yarn logs" command fails with ClassNotFoundException.
|
||||||
|
(Jason Lowe via mahadev)
|
||||||
|
|
||||||
Release 0.23.0 - 2011-11-01
|
Release 0.23.0 - 2011-11-01
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -91,15 +91,15 @@ if [ -d "$HADOOP_MAPRED_HOME/build/tools" ]; then
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# for releases, add core mapred jar & webapps to CLASSPATH
|
# for releases, add core mapred jar & webapps to CLASSPATH
|
||||||
if [ -d "$HADOOP_PREFIX/share/hadoop/mapreduce/webapps" ]; then
|
if [ -d "$HADOOP_PREFIX/${MAPRED_DIR}/webapps" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$HADOOP_PREFIX/share/hadoop/mapreduce
|
CLASSPATH=${CLASSPATH}:$HADOOP_PREFIX/${MAPRED_DIR}
|
||||||
fi
|
fi
|
||||||
for f in $HADOOP_MAPRED_HOME/share/hadoop-mapreduce/*.jar; do
|
for f in $HADOOP_MAPRED_HOME/${MAPRED_DIR}/*.jar; do
|
||||||
CLASSPATH=${CLASSPATH}:$f;
|
CLASSPATH=${CLASSPATH}:$f;
|
||||||
done
|
done
|
||||||
|
|
||||||
# add libs to CLASSPATH
|
# add libs to CLASSPATH
|
||||||
for f in $HADOOP_MAPRED_HOME/lib/*.jar; do
|
for f in $HADOOP_MAPRED_HOME/${MAPRED_LIB_JARS_DIR}/*.jar; do
|
||||||
CLASSPATH=${CLASSPATH}:$f;
|
CLASSPATH=${CLASSPATH}:$f;
|
||||||
done
|
done
|
||||||
|
|
||||||
|
|
|
@ -455,10 +455,14 @@ public class Job extends JobContextImpl implements JobContext {
|
||||||
public String toString() {
|
public String toString() {
|
||||||
ensureState(JobState.RUNNING);
|
ensureState(JobState.RUNNING);
|
||||||
String reasonforFailure = " ";
|
String reasonforFailure = " ";
|
||||||
|
int numMaps = 0;
|
||||||
|
int numReduces = 0;
|
||||||
try {
|
try {
|
||||||
updateStatus();
|
updateStatus();
|
||||||
if (status.getState().equals(JobStatus.State.FAILED))
|
if (status.getState().equals(JobStatus.State.FAILED))
|
||||||
reasonforFailure = getTaskFailureEventString();
|
reasonforFailure = getTaskFailureEventString();
|
||||||
|
numMaps = getTaskReports(TaskType.MAP).length;
|
||||||
|
numReduces = getTaskReports(TaskType.REDUCE).length;
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
} catch (InterruptedException ie) {
|
} catch (InterruptedException ie) {
|
||||||
}
|
}
|
||||||
|
@ -468,6 +472,8 @@ public class Job extends JobContextImpl implements JobContext {
|
||||||
sb.append("Job Tracking URL : ").append(status.getTrackingUrl());
|
sb.append("Job Tracking URL : ").append(status.getTrackingUrl());
|
||||||
sb.append("\n");
|
sb.append("\n");
|
||||||
sb.append("Uber job : ").append(status.isUber()).append("\n");
|
sb.append("Uber job : ").append(status.isUber()).append("\n");
|
||||||
|
sb.append("Number of maps: ").append(numMaps);
|
||||||
|
sb.append("Number of reduces: ").append(numReduces);
|
||||||
sb.append("map() completion: ");
|
sb.append("map() completion: ");
|
||||||
sb.append(status.getMapProgress()).append("\n");
|
sb.append(status.getMapProgress()).append("\n");
|
||||||
sb.append("reduce() completion: ");
|
sb.append("reduce() completion: ");
|
||||||
|
|
|
@ -412,7 +412,7 @@ public interface MRJobConfig {
|
||||||
/** The number of threads used to handle task RPC calls.*/
|
/** The number of threads used to handle task RPC calls.*/
|
||||||
public static final String MR_AM_TASK_LISTENER_THREAD_COUNT =
|
public static final String MR_AM_TASK_LISTENER_THREAD_COUNT =
|
||||||
MR_AM_PREFIX + "job.task.listener.thread-count";
|
MR_AM_PREFIX + "job.task.listener.thread-count";
|
||||||
public static final int DEFAULT_MR_AM_TASK_LISTENER_THREAD_COUNT = 10;
|
public static final int DEFAULT_MR_AM_TASK_LISTENER_THREAD_COUNT = 30;
|
||||||
|
|
||||||
/** How often the AM should send heartbeats to the RM.*/
|
/** How often the AM should send heartbeats to the RM.*/
|
||||||
public static final String MR_AM_TO_RM_HEARTBEAT_INTERVAL_MS =
|
public static final String MR_AM_TO_RM_HEARTBEAT_INTERVAL_MS =
|
||||||
|
|
|
@ -18,6 +18,7 @@
|
||||||
package org.apache.hadoop.mapreduce.tools;
|
package org.apache.hadoop.mapreduce.tools;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.io.PrintWriter;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
|
@ -25,6 +26,7 @@ import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.conf.Configured;
|
import org.apache.hadoop.conf.Configured;
|
||||||
import org.apache.hadoop.ipc.RemoteException;
|
import org.apache.hadoop.ipc.RemoteException;
|
||||||
|
@ -563,22 +565,25 @@ public class CLI extends Configured implements Tool {
|
||||||
|
|
||||||
public void displayJobList(JobStatus[] jobs)
|
public void displayJobList(JobStatus[] jobs)
|
||||||
throws IOException, InterruptedException {
|
throws IOException, InterruptedException {
|
||||||
System.out.println("Total jobs:" + jobs.length);
|
displayJobList(jobs, new PrintWriter(System.out));
|
||||||
System.out.println("JobId\tState\tStartTime\t" +
|
}
|
||||||
"UserName\tQueue\tPriority\tMaps\tReduces\tUsedContainers\t" +
|
|
||||||
"RsvdContainers\tUsedMem\tRsvdMem\tNeededMem\tAM info");
|
|
||||||
for (JobStatus job : jobs) {
|
|
||||||
TaskReport[] mapReports =
|
|
||||||
cluster.getJob(job.getJobID()).getTaskReports(TaskType.MAP);
|
|
||||||
TaskReport[] reduceReports =
|
|
||||||
cluster.getJob(job.getJobID()).getTaskReports(TaskType.REDUCE);
|
|
||||||
|
|
||||||
System.out.printf("%s\t%s\t%d\t%s\t%s\t%s\t%d\t%d\t%d\t%d\t%dM\t%dM\t%dM\t%s\n",
|
@Private
|
||||||
|
public static String headerPattern = "%23s\t%10s\t%14s\t%12s\t%12s\t%10s\t%15s\t%15s\t%8s\t%8s\t%10s\t%10s\n";
|
||||||
|
@Private
|
||||||
|
public static String dataPattern = "%23s\t%10s\t%14d\t%12s\t%12s\t%10s\t%14d\t%14d\t%7dM\t%7sM\t%9dM\t%10s\n";
|
||||||
|
|
||||||
|
@Private
|
||||||
|
public void displayJobList(JobStatus[] jobs, PrintWriter writer) {
|
||||||
|
writer.println("Total jobs:" + jobs.length);
|
||||||
|
writer.printf(headerPattern, "JobId", "State", "StartTime", "UserName",
|
||||||
|
"Queue", "Priority", "UsedContainers",
|
||||||
|
"RsvdContainers", "UsedMem", "RsvdMem", "NeededMem", "AM info");
|
||||||
|
for (JobStatus job : jobs) {
|
||||||
|
writer.printf(dataPattern,
|
||||||
job.getJobID().toString(), job.getState(), job.getStartTime(),
|
job.getJobID().toString(), job.getState(), job.getStartTime(),
|
||||||
job.getUsername(), job.getQueue(),
|
job.getUsername(), job.getQueue(),
|
||||||
job.getPriority().name(),
|
job.getPriority().name(),
|
||||||
mapReports.length,
|
|
||||||
reduceReports.length,
|
|
||||||
job.getNumUsedSlots(),
|
job.getNumUsedSlots(),
|
||||||
job.getNumReservedSlots(),
|
job.getNumReservedSlots(),
|
||||||
job.getUsedMem(),
|
job.getUsedMem(),
|
||||||
|
@ -586,6 +591,7 @@ public class CLI extends Configured implements Tool {
|
||||||
job.getNeededMem(),
|
job.getNeededMem(),
|
||||||
job.getSchedulingInfo());
|
job.getSchedulingInfo());
|
||||||
}
|
}
|
||||||
|
writer.flush();
|
||||||
}
|
}
|
||||||
|
|
||||||
public static void main(String[] argv) throws Exception {
|
public static void main(String[] argv) throws Exception {
|
||||||
|
|
|
@ -440,6 +440,16 @@
|
||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>mapreduce.admin.user.env</name>
|
||||||
|
<value>LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native</value>
|
||||||
|
<description>Expert: Additional execution environment entries for
|
||||||
|
map and reduce task processes. This is not an additive property.
|
||||||
|
You must preserve the original value if you want your map and
|
||||||
|
reduce tasks to have access to native libraries (compression, etc).
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>mapreduce.task.tmp.dir</name>
|
<name>mapreduce.task.tmp.dir</name>
|
||||||
<value>./tmp</value>
|
<value>./tmp</value>
|
||||||
|
@ -1224,4 +1234,18 @@
|
||||||
mapreduce.job.end-notification.max.retry.interval</description>
|
mapreduce.job.end-notification.max.retry.interval</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>yarn.app.mapreduce.am.job.task.listener.thread-count</name>
|
||||||
|
<value>30</value>
|
||||||
|
<description>The number of threads used to handle RPC calls in the
|
||||||
|
MR AppMaster from remote tasks</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms</name>
|
||||||
|
<value>1000</value>
|
||||||
|
<description>The interval in ms at which the MR AppMaster should send
|
||||||
|
heartbeats to the ResourceManager</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
</configuration>
|
</configuration>
|
||||||
|
|
|
@ -22,19 +22,24 @@ import static org.junit.Assert.assertEquals;
|
||||||
import static org.mockito.Matchers.isA;
|
import static org.mockito.Matchers.isA;
|
||||||
import static org.mockito.Mockito.atLeastOnce;
|
import static org.mockito.Mockito.atLeastOnce;
|
||||||
import static org.mockito.Mockito.mock;
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.never;
|
||||||
import static org.mockito.Mockito.verify;
|
import static org.mockito.Mockito.verify;
|
||||||
import static org.mockito.Mockito.when;
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.io.ByteArrayOutputStream;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import org.apache.hadoop.mapred.JobConf;
|
import java.io.PrintWriter;
|
||||||
|
|
||||||
import org.apache.hadoop.mapreduce.Cluster;
|
import org.apache.hadoop.mapreduce.Cluster;
|
||||||
import org.apache.hadoop.mapreduce.Job;
|
import org.apache.hadoop.mapreduce.Job;
|
||||||
import org.apache.hadoop.mapreduce.JobPriority;
|
import org.apache.hadoop.mapreduce.JobPriority;
|
||||||
import org.apache.hadoop.mapreduce.JobStatus;
|
import org.apache.hadoop.mapreduce.JobStatus;
|
||||||
import org.apache.hadoop.mapreduce.TaskType;
|
|
||||||
import org.apache.hadoop.mapreduce.TaskReport;
|
import org.apache.hadoop.mapreduce.TaskReport;
|
||||||
|
import org.apache.hadoop.mapreduce.TaskType;
|
||||||
|
import org.junit.Assert;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
|
@SuppressWarnings("deprecation")
|
||||||
public class JobClientUnitTest {
|
public class JobClientUnitTest {
|
||||||
|
|
||||||
public class TestJobClient extends JobClient {
|
public class TestJobClient extends JobClient {
|
||||||
|
@ -48,7 +53,6 @@ public class JobClientUnitTest {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@SuppressWarnings("deprecation")
|
|
||||||
@Test
|
@Test
|
||||||
public void testMapTaskReportsWithNullJob() throws Exception {
|
public void testMapTaskReportsWithNullJob() throws Exception {
|
||||||
TestJobClient client = new TestJobClient(new JobConf());
|
TestJobClient client = new TestJobClient(new JobConf());
|
||||||
|
@ -64,7 +68,6 @@ public class JobClientUnitTest {
|
||||||
verify(mockCluster).getJob(id);
|
verify(mockCluster).getJob(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@SuppressWarnings("deprecation")
|
|
||||||
@Test
|
@Test
|
||||||
public void testReduceTaskReportsWithNullJob() throws Exception {
|
public void testReduceTaskReportsWithNullJob() throws Exception {
|
||||||
TestJobClient client = new TestJobClient(new JobConf());
|
TestJobClient client = new TestJobClient(new JobConf());
|
||||||
|
@ -80,7 +83,6 @@ public class JobClientUnitTest {
|
||||||
verify(mockCluster).getJob(id);
|
verify(mockCluster).getJob(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@SuppressWarnings("deprecation")
|
|
||||||
@Test
|
@Test
|
||||||
public void testSetupTaskReportsWithNullJob() throws Exception {
|
public void testSetupTaskReportsWithNullJob() throws Exception {
|
||||||
TestJobClient client = new TestJobClient(new JobConf());
|
TestJobClient client = new TestJobClient(new JobConf());
|
||||||
|
@ -96,7 +98,6 @@ public class JobClientUnitTest {
|
||||||
verify(mockCluster).getJob(id);
|
verify(mockCluster).getJob(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@SuppressWarnings("deprecation")
|
|
||||||
@Test
|
@Test
|
||||||
public void testCleanupTaskReportsWithNullJob() throws Exception {
|
public void testCleanupTaskReportsWithNullJob() throws Exception {
|
||||||
TestJobClient client = new TestJobClient(new JobConf());
|
TestJobClient client = new TestJobClient(new JobConf());
|
||||||
|
@ -115,12 +116,15 @@ public class JobClientUnitTest {
|
||||||
@Test
|
@Test
|
||||||
public void testShowJob() throws Exception {
|
public void testShowJob() throws Exception {
|
||||||
TestJobClient client = new TestJobClient(new JobConf());
|
TestJobClient client = new TestJobClient(new JobConf());
|
||||||
JobID jobID = new JobID("test", 0);
|
|
||||||
|
long startTime = System.currentTimeMillis();
|
||||||
|
|
||||||
|
JobID jobID = new JobID(String.valueOf(startTime), 12345);
|
||||||
|
|
||||||
JobStatus mockJobStatus = mock(JobStatus.class);
|
JobStatus mockJobStatus = mock(JobStatus.class);
|
||||||
when(mockJobStatus.getJobID()).thenReturn(jobID);
|
when(mockJobStatus.getJobID()).thenReturn(jobID);
|
||||||
when(mockJobStatus.getState()).thenReturn(JobStatus.State.RUNNING);
|
when(mockJobStatus.getState()).thenReturn(JobStatus.State.RUNNING);
|
||||||
when(mockJobStatus.getStartTime()).thenReturn(0L);
|
when(mockJobStatus.getStartTime()).thenReturn(startTime);
|
||||||
when(mockJobStatus.getUsername()).thenReturn("mockuser");
|
when(mockJobStatus.getUsername()).thenReturn("mockuser");
|
||||||
when(mockJobStatus.getQueue()).thenReturn("mockqueue");
|
when(mockJobStatus.getQueue()).thenReturn("mockqueue");
|
||||||
when(mockJobStatus.getPriority()).thenReturn(JobPriority.NORMAL);
|
when(mockJobStatus.getPriority()).thenReturn(JobPriority.NORMAL);
|
||||||
|
@ -132,18 +136,21 @@ public class JobClientUnitTest {
|
||||||
when(mockJobStatus.getSchedulingInfo()).thenReturn("NA");
|
when(mockJobStatus.getSchedulingInfo()).thenReturn("NA");
|
||||||
|
|
||||||
Job mockJob = mock(Job.class);
|
Job mockJob = mock(Job.class);
|
||||||
when(mockJob.getTaskReports(isA(TaskType.class))).thenReturn(new TaskReport[0]);
|
when(mockJob.getTaskReports(isA(TaskType.class))).thenReturn(
|
||||||
|
new TaskReport[5]);
|
||||||
|
|
||||||
Cluster mockCluster = mock(Cluster.class);
|
Cluster mockCluster = mock(Cluster.class);
|
||||||
when(mockCluster.getJob(jobID)).thenReturn(mockJob);
|
when(mockCluster.getJob(jobID)).thenReturn(mockJob);
|
||||||
|
|
||||||
client.setCluster(mockCluster);
|
client.setCluster(mockCluster);
|
||||||
|
|
||||||
|
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||||
|
client.displayJobList(new JobStatus[] {mockJobStatus}, new PrintWriter(out));
|
||||||
|
String commandLineOutput = out.toString();
|
||||||
|
System.out.println(commandLineOutput);
|
||||||
|
Assert.assertTrue(commandLineOutput.contains("Total jobs:1"));
|
||||||
|
|
||||||
client.displayJobList(new JobStatus[] {mockJobStatus});
|
|
||||||
verify(mockJobStatus, atLeastOnce()).getJobID();
|
verify(mockJobStatus, atLeastOnce()).getJobID();
|
||||||
verify(mockJob, atLeastOnce()).getTaskReports(isA(TaskType.class));
|
|
||||||
verify(mockCluster, atLeastOnce()).getJob(jobID);
|
|
||||||
verify(mockJobStatus).getState();
|
verify(mockJobStatus).getState();
|
||||||
verify(mockJobStatus).getStartTime();
|
verify(mockJobStatus).getStartTime();
|
||||||
verify(mockJobStatus).getUsername();
|
verify(mockJobStatus).getUsername();
|
||||||
|
@ -155,5 +162,9 @@ public class JobClientUnitTest {
|
||||||
verify(mockJobStatus).getReservedMem();
|
verify(mockJobStatus).getReservedMem();
|
||||||
verify(mockJobStatus).getNeededMem();
|
verify(mockJobStatus).getNeededMem();
|
||||||
verify(mockJobStatus).getSchedulingInfo();
|
verify(mockJobStatus).getSchedulingInfo();
|
||||||
|
|
||||||
|
// This call should not go to each AM.
|
||||||
|
verify(mockCluster, never()).getJob(jobID);
|
||||||
|
verify(mockJob, never()).getTaskReports(isA(TaskType.class));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -140,8 +140,8 @@ if [ -d "$YARN_HOME/build/tools" ]; then
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/build/tools
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/build/tools
|
||||||
fi
|
fi
|
||||||
|
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/share/hadoop/mapreduce/*
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/${YARN_DIR}/*
|
||||||
CLASSPATH=${CLASSPATH}:$YARN_HOME/share/hadoop/mapreduce/lib/*
|
CLASSPATH=${CLASSPATH}:$YARN_HOME/${YARN_LIB_JARS_DIR}/*
|
||||||
|
|
||||||
# so that filenames w/ spaces are handled correctly in loops below
|
# so that filenames w/ spaces are handled correctly in loops below
|
||||||
IFS=
|
IFS=
|
||||||
|
@ -194,7 +194,7 @@ elif [ "$COMMAND" = "jar" ] ; then
|
||||||
CLASS=org.apache.hadoop.util.RunJar
|
CLASS=org.apache.hadoop.util.RunJar
|
||||||
YARN_OPTS="$YARN_OPTS $YARN_CLIENT_OPTS"
|
YARN_OPTS="$YARN_OPTS $YARN_CLIENT_OPTS"
|
||||||
elif [ "$COMMAND" = "logs" ] ; then
|
elif [ "$COMMAND" = "logs" ] ; then
|
||||||
CLASS=org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.LogDumper
|
CLASS=org.apache.hadoop.yarn.logaggregation.LogDumper
|
||||||
YARN_OPTS="$YARN_OPTS $YARN_CLIENT_OPTS"
|
YARN_OPTS="$YARN_OPTS $YARN_CLIENT_OPTS"
|
||||||
elif [ "$COMMAND" = "daemonlog" ] ; then
|
elif [ "$COMMAND" = "daemonlog" ] ; then
|
||||||
CLASS=org.apache.hadoop.log.LogLevel
|
CLASS=org.apache.hadoop.log.LogLevel
|
||||||
|
|
|
@ -19,8 +19,6 @@ bin=`which "$0"`
|
||||||
bin=`dirname "${bin}"`
|
bin=`dirname "${bin}"`
|
||||||
bin=`cd "$bin"; pwd`
|
bin=`cd "$bin"; pwd`
|
||||||
|
|
||||||
export HADOOP_PREFIX="${HADOOP_PREFIX:-$bin/..}"
|
|
||||||
|
|
||||||
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
DEFAULT_LIBEXEC_DIR="$bin"/../libexec
|
||||||
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
|
||||||
if [ -e "${HADOOP_LIBEXEC_DIR}/hadoop-config.sh" ]; then
|
if [ -e "${HADOOP_LIBEXEC_DIR}/hadoop-config.sh" ]; then
|
||||||
|
|
|
@ -90,7 +90,7 @@ public class YarnConfiguration extends Configuration {
|
||||||
/** The number of threads used to handle applications manager requests.*/
|
/** The number of threads used to handle applications manager requests.*/
|
||||||
public static final String RM_CLIENT_THREAD_COUNT =
|
public static final String RM_CLIENT_THREAD_COUNT =
|
||||||
RM_PREFIX + "client.thread-count";
|
RM_PREFIX + "client.thread-count";
|
||||||
public static final int DEFAULT_RM_CLIENT_THREAD_COUNT = 10;
|
public static final int DEFAULT_RM_CLIENT_THREAD_COUNT = 50;
|
||||||
|
|
||||||
/** The Kerberos principal for the resource manager.*/
|
/** The Kerberos principal for the resource manager.*/
|
||||||
public static final String RM_PRINCIPAL =
|
public static final String RM_PRINCIPAL =
|
||||||
|
@ -106,7 +106,7 @@ public class YarnConfiguration extends Configuration {
|
||||||
/** Number of threads to handle scheduler interface.*/
|
/** Number of threads to handle scheduler interface.*/
|
||||||
public static final String RM_SCHEDULER_CLIENT_THREAD_COUNT =
|
public static final String RM_SCHEDULER_CLIENT_THREAD_COUNT =
|
||||||
RM_PREFIX + "scheduler.client.thread-count";
|
RM_PREFIX + "scheduler.client.thread-count";
|
||||||
public static final int DEFAULT_RM_SCHEDULER_CLIENT_THREAD_COUNT = 10;
|
public static final int DEFAULT_RM_SCHEDULER_CLIENT_THREAD_COUNT = 50;
|
||||||
|
|
||||||
/** The address of the RM web application.*/
|
/** The address of the RM web application.*/
|
||||||
public static final String RM_WEBAPP_ADDRESS =
|
public static final String RM_WEBAPP_ADDRESS =
|
||||||
|
@ -184,7 +184,7 @@ public class YarnConfiguration extends Configuration {
|
||||||
/** Number of threads to handle resource tracker calls.*/
|
/** Number of threads to handle resource tracker calls.*/
|
||||||
public static final String RM_RESOURCE_TRACKER_CLIENT_THREAD_COUNT =
|
public static final String RM_RESOURCE_TRACKER_CLIENT_THREAD_COUNT =
|
||||||
RM_PREFIX + "resource-tracker.client.thread-count";
|
RM_PREFIX + "resource-tracker.client.thread-count";
|
||||||
public static final int DEFAULT_RM_RESOURCE_TRACKER_CLIENT_THREAD_COUNT = 10;
|
public static final int DEFAULT_RM_RESOURCE_TRACKER_CLIENT_THREAD_COUNT = 50;
|
||||||
|
|
||||||
/** The class to use as the resource scheduler.*/
|
/** The class to use as the resource scheduler.*/
|
||||||
public static final String RM_SCHEDULER =
|
public static final String RM_SCHEDULER =
|
||||||
|
@ -257,7 +257,7 @@ public class YarnConfiguration extends Configuration {
|
||||||
/** Number of threads container manager uses.*/
|
/** Number of threads container manager uses.*/
|
||||||
public static final String NM_CONTAINER_MGR_THREAD_COUNT =
|
public static final String NM_CONTAINER_MGR_THREAD_COUNT =
|
||||||
NM_PREFIX + "container-manager.thread-count";
|
NM_PREFIX + "container-manager.thread-count";
|
||||||
public static final int DEFAULT_NM_CONTAINER_MGR_THREAD_COUNT = 5;
|
public static final int DEFAULT_NM_CONTAINER_MGR_THREAD_COUNT = 20;
|
||||||
|
|
||||||
/** Number of threads used in cleanup.*/
|
/** Number of threads used in cleanup.*/
|
||||||
public static final String NM_DELETE_THREAD_COUNT =
|
public static final String NM_DELETE_THREAD_COUNT =
|
||||||
|
|
|
@ -67,7 +67,7 @@
|
||||||
<property>
|
<property>
|
||||||
<description>The number of threads used to handle applications manager requests.</description>
|
<description>The number of threads used to handle applications manager requests.</description>
|
||||||
<name>yarn.resourcemanager.client.thread-count</name>
|
<name>yarn.resourcemanager.client.thread-count</name>
|
||||||
<value>10</value>
|
<value>50</value>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
|
@ -90,7 +90,7 @@
|
||||||
<property>
|
<property>
|
||||||
<description>Number of threads to handle scheduler interface.</description>
|
<description>Number of threads to handle scheduler interface.</description>
|
||||||
<name>yarn.resourcemanager.scheduler.client.thread-count</name>
|
<name>yarn.resourcemanager.scheduler.client.thread-count</name>
|
||||||
<value>10</value>
|
<value>50</value>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
|
@ -179,7 +179,7 @@
|
||||||
<property>
|
<property>
|
||||||
<description>Number of threads to handle resource tracker calls.</description>
|
<description>Number of threads to handle resource tracker calls.</description>
|
||||||
<name>yarn.resourcemanager.resource-tracker.client.thread-count</name>
|
<name>yarn.resourcemanager.resource-tracker.client.thread-count</name>
|
||||||
<value>10</value>
|
<value>50</value>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
|
@ -244,7 +244,7 @@
|
||||||
<property>
|
<property>
|
||||||
<description>Number of threads container manager uses.</description>
|
<description>Number of threads container manager uses.</description>
|
||||||
<name>yarn.nodemanager.container-manager.thread-count</name>
|
<name>yarn.nodemanager.container-manager.thread-count</name>
|
||||||
<value>5</value>
|
<value>20</value>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
|
|
|
@ -29,7 +29,6 @@ import org.apache.hadoop.metrics2.annotation.Metric;
|
||||||
import org.apache.hadoop.metrics2.annotation.Metrics;
|
import org.apache.hadoop.metrics2.annotation.Metrics;
|
||||||
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
|
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
|
||||||
import org.apache.hadoop.metrics2.lib.MetricsRegistry;
|
import org.apache.hadoop.metrics2.lib.MetricsRegistry;
|
||||||
import org.apache.hadoop.metrics2.lib.MutableCounterInt;
|
|
||||||
import org.apache.hadoop.metrics2.lib.MutableGaugeInt;
|
import org.apache.hadoop.metrics2.lib.MutableGaugeInt;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeEventType;
|
||||||
|
|
||||||
|
@ -39,9 +38,9 @@ public class ClusterMetrics {
|
||||||
|
|
||||||
private static AtomicBoolean isInitialized = new AtomicBoolean(false);
|
private static AtomicBoolean isInitialized = new AtomicBoolean(false);
|
||||||
|
|
||||||
@Metric("# of NMs") MutableGaugeInt numNMs;
|
@Metric("# of active NMs") MutableGaugeInt numNMs;
|
||||||
@Metric("# of decommissioned NMs") MutableCounterInt numDecommissionedNMs;
|
@Metric("# of decommissioned NMs") MutableGaugeInt numDecommissionedNMs;
|
||||||
@Metric("# of lost NMs") MutableCounterInt numLostNMs;
|
@Metric("# of lost NMs") MutableGaugeInt numLostNMs;
|
||||||
@Metric("# of unhealthy NMs") MutableGaugeInt numUnhealthyNMs;
|
@Metric("# of unhealthy NMs") MutableGaugeInt numUnhealthyNMs;
|
||||||
@Metric("# of Rebooted NMs") MutableGaugeInt numRebootedNMs;
|
@Metric("# of Rebooted NMs") MutableGaugeInt numRebootedNMs;
|
||||||
|
|
||||||
|
@ -73,8 +72,8 @@ public class ClusterMetrics {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
//Total Nodemanagers
|
//Active Nodemanagers
|
||||||
public int getNumNMs() {
|
public int getNumActiveNMs() {
|
||||||
return numNMs.value();
|
return numNMs.value();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -87,6 +86,10 @@ public class ClusterMetrics {
|
||||||
numDecommissionedNMs.incr();
|
numDecommissionedNMs.incr();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void decrDecommisionedNMs() {
|
||||||
|
numDecommissionedNMs.decr();
|
||||||
|
}
|
||||||
|
|
||||||
//Lost NMs
|
//Lost NMs
|
||||||
public int getNumLostNMs() {
|
public int getNumLostNMs() {
|
||||||
return numLostNMs.value();
|
return numLostNMs.value();
|
||||||
|
@ -96,6 +99,10 @@ public class ClusterMetrics {
|
||||||
numLostNMs.incr();
|
numLostNMs.incr();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void decrNumLostNMs() {
|
||||||
|
numLostNMs.decr();
|
||||||
|
}
|
||||||
|
|
||||||
//Unhealthy NMs
|
//Unhealthy NMs
|
||||||
public int getUnhealthyNMs() {
|
public int getUnhealthyNMs() {
|
||||||
return numUnhealthyNMs.value();
|
return numUnhealthyNMs.value();
|
||||||
|
@ -118,6 +125,10 @@ public class ClusterMetrics {
|
||||||
numRebootedNMs.incr();
|
numRebootedNMs.incr();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void decrNumRebootedNMs() {
|
||||||
|
numRebootedNMs.decr();
|
||||||
|
}
|
||||||
|
|
||||||
public void removeNode(RMNodeEventType nodeEventType) {
|
public void removeNode(RMNodeEventType nodeEventType) {
|
||||||
numNMs.decr();
|
numNMs.decr();
|
||||||
switch(nodeEventType){
|
switch(nodeEventType){
|
||||||
|
|
|
@ -44,6 +44,8 @@ public interface RMContext {
|
||||||
|
|
||||||
ConcurrentMap<ApplicationId, RMApp> getRMApps();
|
ConcurrentMap<ApplicationId, RMApp> getRMApps();
|
||||||
|
|
||||||
|
ConcurrentMap<String, RMNode> getInactiveRMNodes();
|
||||||
|
|
||||||
ConcurrentMap<NodeId, RMNode> getRMNodes();
|
ConcurrentMap<NodeId, RMNode> getRMNodes();
|
||||||
|
|
||||||
AMLivelinessMonitor getAMLivelinessMonitor();
|
AMLivelinessMonitor getAMLivelinessMonitor();
|
||||||
|
|
|
@ -44,6 +44,9 @@ public class RMContextImpl implements RMContext {
|
||||||
private final ConcurrentMap<NodeId, RMNode> nodes
|
private final ConcurrentMap<NodeId, RMNode> nodes
|
||||||
= new ConcurrentHashMap<NodeId, RMNode>();
|
= new ConcurrentHashMap<NodeId, RMNode>();
|
||||||
|
|
||||||
|
private final ConcurrentMap<String, RMNode> inactiveNodes
|
||||||
|
= new ConcurrentHashMap<String, RMNode>();
|
||||||
|
|
||||||
private AMLivelinessMonitor amLivelinessMonitor;
|
private AMLivelinessMonitor amLivelinessMonitor;
|
||||||
private ContainerAllocationExpirer containerAllocationExpirer;
|
private ContainerAllocationExpirer containerAllocationExpirer;
|
||||||
private final DelegationTokenRenewer tokenRenewer;
|
private final DelegationTokenRenewer tokenRenewer;
|
||||||
|
@ -84,6 +87,11 @@ public class RMContextImpl implements RMContext {
|
||||||
return this.nodes;
|
return this.nodes;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ConcurrentMap<String, RMNode> getInactiveRMNodes() {
|
||||||
|
return this.inactiveNodes;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ContainerAllocationExpirer getContainerAllocationExpirer() {
|
public ContainerAllocationExpirer getContainerAllocationExpirer() {
|
||||||
return this.containerAllocationExpirer;
|
return this.containerAllocationExpirer;
|
||||||
|
|
|
@ -220,10 +220,6 @@ public class ResourceTrackerService extends AbstractService implements
|
||||||
if (rmNode == null) {
|
if (rmNode == null) {
|
||||||
/* node does not exist */
|
/* node does not exist */
|
||||||
LOG.info("Node not found rebooting " + remoteNodeStatus.getNodeId());
|
LOG.info("Node not found rebooting " + remoteNodeStatus.getNodeId());
|
||||||
|
|
||||||
// Updating the metrics directly as reboot event cannot be
|
|
||||||
// triggered on a null rmNode
|
|
||||||
ClusterMetrics.getMetrics().incrNumRebootedNMs();
|
|
||||||
return reboot;
|
return reboot;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -119,7 +119,7 @@ public class RMNodeImpl implements RMNode, EventHandler<RMNodeEvent> {
|
||||||
RMNodeEventType.DECOMMISSION, new RemoveNodeTransition())
|
RMNodeEventType.DECOMMISSION, new RemoveNodeTransition())
|
||||||
.addTransition(RMNodeState.RUNNING, RMNodeState.LOST,
|
.addTransition(RMNodeState.RUNNING, RMNodeState.LOST,
|
||||||
RMNodeEventType.EXPIRE, new RemoveNodeTransition())
|
RMNodeEventType.EXPIRE, new RemoveNodeTransition())
|
||||||
.addTransition(RMNodeState.RUNNING, RMNodeState.LOST,
|
.addTransition(RMNodeState.RUNNING, RMNodeState.REBOOTED,
|
||||||
RMNodeEventType.REBOOTING, new RemoveNodeTransition())
|
RMNodeEventType.REBOOTING, new RemoveNodeTransition())
|
||||||
.addTransition(RMNodeState.RUNNING, RMNodeState.RUNNING,
|
.addTransition(RMNodeState.RUNNING, RMNodeState.RUNNING,
|
||||||
RMNodeEventType.CLEANUP_APP, new CleanUpAppTransition())
|
RMNodeEventType.CLEANUP_APP, new CleanUpAppTransition())
|
||||||
|
@ -308,6 +308,21 @@ public class RMNodeImpl implements RMNode, EventHandler<RMNodeEvent> {
|
||||||
public static class AddNodeTransition implements
|
public static class AddNodeTransition implements
|
||||||
SingleArcTransition<RMNodeImpl, RMNodeEvent> {
|
SingleArcTransition<RMNodeImpl, RMNodeEvent> {
|
||||||
|
|
||||||
|
private void updateMetrics(RMNodeState nodeState) {
|
||||||
|
ClusterMetrics metrics = ClusterMetrics.getMetrics();
|
||||||
|
switch (nodeState) {
|
||||||
|
case LOST:
|
||||||
|
metrics.decrNumLostNMs();
|
||||||
|
break;
|
||||||
|
case REBOOTED:
|
||||||
|
metrics.decrNumRebootedNMs();
|
||||||
|
break;
|
||||||
|
case DECOMMISSIONED:
|
||||||
|
metrics.decrDecommisionedNMs();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@SuppressWarnings("unchecked")
|
@SuppressWarnings("unchecked")
|
||||||
@Override
|
@Override
|
||||||
public void transition(RMNodeImpl rmNode, RMNodeEvent event) {
|
public void transition(RMNodeImpl rmNode, RMNodeEvent event) {
|
||||||
|
@ -316,6 +331,13 @@ public class RMNodeImpl implements RMNode, EventHandler<RMNodeEvent> {
|
||||||
rmNode.context.getDispatcher().getEventHandler().handle(
|
rmNode.context.getDispatcher().getEventHandler().handle(
|
||||||
new NodeAddedSchedulerEvent(rmNode));
|
new NodeAddedSchedulerEvent(rmNode));
|
||||||
|
|
||||||
|
String host = rmNode.nodeId.getHost();
|
||||||
|
if (rmNode.context.getInactiveRMNodes().containsKey(host)) {
|
||||||
|
RMNode node = rmNode.context.getInactiveRMNodes().get(host);
|
||||||
|
rmNode.context.getInactiveRMNodes().remove(host);
|
||||||
|
updateMetrics(node.getState());
|
||||||
|
}
|
||||||
|
|
||||||
ClusterMetrics.getMetrics().addNode();
|
ClusterMetrics.getMetrics().addNode();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -353,7 +375,7 @@ public class RMNodeImpl implements RMNode, EventHandler<RMNodeEvent> {
|
||||||
// Remove the node from the system.
|
// Remove the node from the system.
|
||||||
rmNode.context.getRMNodes().remove(rmNode.nodeId);
|
rmNode.context.getRMNodes().remove(rmNode.nodeId);
|
||||||
LOG.info("Removed Node " + rmNode.nodeId);
|
LOG.info("Removed Node " + rmNode.nodeId);
|
||||||
|
rmNode.context.getInactiveRMNodes().put(rmNode.nodeId.getHost(), rmNode);
|
||||||
//Update the metrics
|
//Update the metrics
|
||||||
ClusterMetrics.getMetrics().removeNode(event.getType());
|
ClusterMetrics.getMetrics().removeNode(event.getType());
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,109 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.yarn.server.resourcemanager.scheduler;
|
||||||
|
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.HashSet;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.Set;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.classification.InterfaceAudience.Private;
|
||||||
|
import org.apache.hadoop.yarn.Lock;
|
||||||
|
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* {@link ActiveUsersManager} tracks active users in the system.
|
||||||
|
* A user is deemed to be active if he has any running applications with
|
||||||
|
* outstanding resource requests.
|
||||||
|
*
|
||||||
|
* An active user is defined as someone with outstanding resource requests.
|
||||||
|
*/
|
||||||
|
@Private
|
||||||
|
public class ActiveUsersManager {
|
||||||
|
|
||||||
|
private static final Log LOG = LogFactory.getLog(ActiveUsersManager.class);
|
||||||
|
|
||||||
|
private final QueueMetrics metrics;
|
||||||
|
|
||||||
|
private int activeUsers = 0;
|
||||||
|
private Map<String, Set<ApplicationId>> usersApplications =
|
||||||
|
new HashMap<String, Set<ApplicationId>>();
|
||||||
|
|
||||||
|
public ActiveUsersManager(QueueMetrics metrics) {
|
||||||
|
this.metrics = metrics;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* An application has new outstanding requests.
|
||||||
|
*
|
||||||
|
* @param user application user
|
||||||
|
* @param applicationId activated application
|
||||||
|
*/
|
||||||
|
@Lock({Queue.class, SchedulerApp.class})
|
||||||
|
synchronized public void activateApplication(
|
||||||
|
String user, ApplicationId applicationId) {
|
||||||
|
Set<ApplicationId> userApps = usersApplications.get(user);
|
||||||
|
if (userApps == null) {
|
||||||
|
userApps = new HashSet<ApplicationId>();
|
||||||
|
usersApplications.put(user, userApps);
|
||||||
|
++activeUsers;
|
||||||
|
metrics.incrActiveUsers();
|
||||||
|
LOG.debug("User " + user + " added to activeUsers, currently: " +
|
||||||
|
activeUsers);
|
||||||
|
}
|
||||||
|
if (userApps.add(applicationId)) {
|
||||||
|
metrics.activateApp(user);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* An application has no more outstanding requests.
|
||||||
|
*
|
||||||
|
* @param user application user
|
||||||
|
* @param applicationId deactivated application
|
||||||
|
*/
|
||||||
|
@Lock({Queue.class, SchedulerApp.class})
|
||||||
|
synchronized public void deactivateApplication(
|
||||||
|
String user, ApplicationId applicationId) {
|
||||||
|
Set<ApplicationId> userApps = usersApplications.get(user);
|
||||||
|
if (userApps != null) {
|
||||||
|
if (userApps.remove(applicationId)) {
|
||||||
|
metrics.deactivateApp(user);
|
||||||
|
}
|
||||||
|
if (userApps.isEmpty()) {
|
||||||
|
usersApplications.remove(user);
|
||||||
|
--activeUsers;
|
||||||
|
metrics.decrActiveUsers();
|
||||||
|
LOG.debug("User " + user + " removed from activeUsers, currently: " +
|
||||||
|
activeUsers);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get number of active users i.e. users with applications which have pending
|
||||||
|
* resource requests.
|
||||||
|
* @return number of active users
|
||||||
|
*/
|
||||||
|
@Lock({Queue.class, SchedulerApp.class})
|
||||||
|
synchronized public int getNumActiveUsers() {
|
||||||
|
return activeUsers;
|
||||||
|
}
|
||||||
|
}
|
|
@ -36,12 +36,11 @@ import org.apache.hadoop.yarn.api.records.Container;
|
||||||
import org.apache.hadoop.yarn.api.records.Priority;
|
import org.apache.hadoop.yarn.api.records.Priority;
|
||||||
import org.apache.hadoop.yarn.api.records.Resource;
|
import org.apache.hadoop.yarn.api.records.Resource;
|
||||||
import org.apache.hadoop.yarn.api.records.ResourceRequest;
|
import org.apache.hadoop.yarn.api.records.ResourceRequest;
|
||||||
import org.apache.hadoop.yarn.factories.RecordFactory;
|
|
||||||
import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.recovery.ApplicationsStore.ApplicationStore;
|
import org.apache.hadoop.yarn.server.resourcemanager.recovery.ApplicationsStore.ApplicationStore;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.resource.Resources;
|
import org.apache.hadoop.yarn.server.resourcemanager.resource.Resources;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptState;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptState;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This class keeps track of all the consumption of an application. This also
|
* This class keeps track of all the consumption of an application. This also
|
||||||
|
@ -59,27 +58,27 @@ public class AppSchedulingInfo {
|
||||||
final String user;
|
final String user;
|
||||||
private final AtomicInteger containerIdCounter = new AtomicInteger(0);
|
private final AtomicInteger containerIdCounter = new AtomicInteger(0);
|
||||||
|
|
||||||
private final RecordFactory recordFactory = RecordFactoryProvider
|
|
||||||
.getRecordFactory(null);
|
|
||||||
|
|
||||||
final Set<Priority> priorities = new TreeSet<Priority>(
|
final Set<Priority> priorities = new TreeSet<Priority>(
|
||||||
new org.apache.hadoop.yarn.server.resourcemanager.resource.Priority.Comparator());
|
new org.apache.hadoop.yarn.server.resourcemanager.resource.Priority.Comparator());
|
||||||
final Map<Priority, Map<String, ResourceRequest>> requests =
|
final Map<Priority, Map<String, ResourceRequest>> requests =
|
||||||
new HashMap<Priority, Map<String, ResourceRequest>>();
|
new HashMap<Priority, Map<String, ResourceRequest>>();
|
||||||
|
|
||||||
private final ApplicationStore store;
|
//private final ApplicationStore store;
|
||||||
|
private final ActiveUsersManager activeUsersManager;
|
||||||
|
|
||||||
/* Allocated by scheduler */
|
/* Allocated by scheduler */
|
||||||
boolean pending = true; // for app metrics
|
boolean pending = true; // for app metrics
|
||||||
|
|
||||||
public AppSchedulingInfo(ApplicationAttemptId appAttemptId,
|
public AppSchedulingInfo(ApplicationAttemptId appAttemptId,
|
||||||
String user, Queue queue, ApplicationStore store) {
|
String user, Queue queue, ActiveUsersManager activeUsersManager,
|
||||||
|
ApplicationStore store) {
|
||||||
this.applicationAttemptId = appAttemptId;
|
this.applicationAttemptId = appAttemptId;
|
||||||
this.applicationId = appAttemptId.getApplicationId();
|
this.applicationId = appAttemptId.getApplicationId();
|
||||||
this.queue = queue;
|
this.queue = queue;
|
||||||
this.queueName = queue.getQueueName();
|
this.queueName = queue.getQueueName();
|
||||||
this.user = user;
|
this.user = user;
|
||||||
this.store = store;
|
//this.store = store;
|
||||||
|
this.activeUsersManager = activeUsersManager;
|
||||||
}
|
}
|
||||||
|
|
||||||
public ApplicationId getApplicationId() {
|
public ApplicationId getApplicationId() {
|
||||||
|
@ -123,7 +122,8 @@ public class AppSchedulingInfo {
|
||||||
* @param requests
|
* @param requests
|
||||||
* resources to be acquired
|
* resources to be acquired
|
||||||
*/
|
*/
|
||||||
synchronized public void updateResourceRequests(List<ResourceRequest> requests) {
|
synchronized public void updateResourceRequests(
|
||||||
|
List<ResourceRequest> requests) {
|
||||||
QueueMetrics metrics = queue.getMetrics();
|
QueueMetrics metrics = queue.getMetrics();
|
||||||
// Update resource requests
|
// Update resource requests
|
||||||
for (ResourceRequest request : requests) {
|
for (ResourceRequest request : requests) {
|
||||||
|
@ -138,6 +138,16 @@ public class AppSchedulingInfo {
|
||||||
+ request);
|
+ request);
|
||||||
}
|
}
|
||||||
updatePendingResources = true;
|
updatePendingResources = true;
|
||||||
|
|
||||||
|
// Premature optimization?
|
||||||
|
// Assumes that we won't see more than one priority request updated
|
||||||
|
// in one call, reasonable assumption... however, it's totally safe
|
||||||
|
// to activate same application more than once.
|
||||||
|
// Thus we don't need another loop ala the one in decrementOutstanding()
|
||||||
|
// which is needed during deactivate.
|
||||||
|
if (request.getNumContainers() > 0) {
|
||||||
|
activeUsersManager.activateApplication(user, applicationId);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
Map<String, ResourceRequest> asks = this.requests.get(priority);
|
Map<String, ResourceRequest> asks = this.requests.get(priority);
|
||||||
|
@ -246,10 +256,7 @@ public class AppSchedulingInfo {
|
||||||
this.requests.get(priority).remove(node.getRackName());
|
this.requests.get(priority).remove(node.getRackName());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Do not remove ANY
|
decrementOutstanding(requests.get(priority).get(RMNode.ANY));
|
||||||
ResourceRequest offSwitchRequest = requests.get(priority).get(
|
|
||||||
RMNode.ANY);
|
|
||||||
offSwitchRequest.setNumContainers(offSwitchRequest.getNumContainers() - 1);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -271,10 +278,7 @@ public class AppSchedulingInfo {
|
||||||
this.requests.get(priority).remove(node.getRackName());
|
this.requests.get(priority).remove(node.getRackName());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Do not remove ANY
|
decrementOutstanding(requests.get(priority).get(RMNode.ANY));
|
||||||
ResourceRequest offSwitchRequest = requests.get(priority).get(
|
|
||||||
RMNode.ANY);
|
|
||||||
offSwitchRequest.setNumContainers(offSwitchRequest.getNumContainers() - 1);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -291,11 +295,32 @@ public class AppSchedulingInfo {
|
||||||
allocate(container);
|
allocate(container);
|
||||||
|
|
||||||
// Update future requirements
|
// Update future requirements
|
||||||
|
decrementOutstanding(offSwitchRequest);
|
||||||
// Do not remove ANY
|
|
||||||
offSwitchRequest.setNumContainers(offSwitchRequest.getNumContainers() - 1);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
synchronized private void decrementOutstanding(
|
||||||
|
ResourceRequest offSwitchRequest) {
|
||||||
|
int numOffSwitchContainers = offSwitchRequest.getNumContainers() - 1;
|
||||||
|
|
||||||
|
// Do not remove ANY
|
||||||
|
offSwitchRequest.setNumContainers(numOffSwitchContainers);
|
||||||
|
|
||||||
|
// Do we have any outstanding requests?
|
||||||
|
// If there is nothing, we need to deactivate this application
|
||||||
|
if (numOffSwitchContainers == 0) {
|
||||||
|
boolean deactivate = true;
|
||||||
|
for (Priority priority : getPriorities()) {
|
||||||
|
ResourceRequest request = getResourceRequest(priority, RMNodeImpl.ANY);
|
||||||
|
if (request.getNumContainers() > 0) {
|
||||||
|
deactivate = false;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (deactivate) {
|
||||||
|
activeUsersManager.deactivateApplication(user, applicationId);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
synchronized private void allocate(Container container) {
|
synchronized private void allocate(Container container) {
|
||||||
// Update consumption and track allocations
|
// Update consumption and track allocations
|
||||||
//TODO: fixme sharad
|
//TODO: fixme sharad
|
||||||
|
|
|
@ -60,6 +60,8 @@ public class QueueMetrics {
|
||||||
@Metric("# of pending containers") MutableGaugeInt pendingContainers;
|
@Metric("# of pending containers") MutableGaugeInt pendingContainers;
|
||||||
@Metric("# of reserved memory in GiB") MutableGaugeInt reservedGB;
|
@Metric("# of reserved memory in GiB") MutableGaugeInt reservedGB;
|
||||||
@Metric("# of reserved containers") MutableGaugeInt reservedContainers;
|
@Metric("# of reserved containers") MutableGaugeInt reservedContainers;
|
||||||
|
@Metric("# of active users") MutableGaugeInt activeUsers;
|
||||||
|
@Metric("# of active users") MutableGaugeInt activeApplications;
|
||||||
|
|
||||||
static final Logger LOG = LoggerFactory.getLogger(QueueMetrics.class);
|
static final Logger LOG = LoggerFactory.getLogger(QueueMetrics.class);
|
||||||
static final int GB = 1024; // resource.memory is in MB
|
static final int GB = 1024; // resource.memory is in MB
|
||||||
|
@ -287,6 +289,36 @@ public class QueueMetrics {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void incrActiveUsers() {
|
||||||
|
activeUsers.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void decrActiveUsers() {
|
||||||
|
activeUsers.decr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void activateApp(String user) {
|
||||||
|
activeApplications.incr();
|
||||||
|
QueueMetrics userMetrics = getUserMetrics(user);
|
||||||
|
if (userMetrics != null) {
|
||||||
|
userMetrics.activateApp(user);
|
||||||
|
}
|
||||||
|
if (parent != null) {
|
||||||
|
parent.activateApp(user);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void deactivateApp(String user) {
|
||||||
|
activeApplications.decr();
|
||||||
|
QueueMetrics userMetrics = getUserMetrics(user);
|
||||||
|
if (userMetrics != null) {
|
||||||
|
userMetrics.deactivateApp(user);
|
||||||
|
}
|
||||||
|
if (parent != null) {
|
||||||
|
parent.deactivateApp(user);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
public int getAppsSubmitted() {
|
public int getAppsSubmitted() {
|
||||||
return appsSubmitted.value();
|
return appsSubmitted.value();
|
||||||
}
|
}
|
||||||
|
@ -338,4 +370,12 @@ public class QueueMetrics {
|
||||||
public int getReservedContainers() {
|
public int getReservedContainers() {
|
||||||
return reservedContainers.value();
|
return reservedContainers.value();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public int getActiveUsers() {
|
||||||
|
return activeUsers.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
public int getActiveApps() {
|
||||||
|
return activeApplications.value();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -102,11 +102,12 @@ public class SchedulerApp {
|
||||||
|
|
||||||
private final RMContext rmContext;
|
private final RMContext rmContext;
|
||||||
public SchedulerApp(ApplicationAttemptId applicationAttemptId,
|
public SchedulerApp(ApplicationAttemptId applicationAttemptId,
|
||||||
String user, Queue queue,
|
String user, Queue queue, ActiveUsersManager activeUsersManager,
|
||||||
RMContext rmContext, ApplicationStore store) {
|
RMContext rmContext, ApplicationStore store) {
|
||||||
this.rmContext = rmContext;
|
this.rmContext = rmContext;
|
||||||
this.appSchedulingInfo =
|
this.appSchedulingInfo =
|
||||||
new AppSchedulingInfo(applicationAttemptId, user, queue, store);
|
new AppSchedulingInfo(applicationAttemptId, user, queue,
|
||||||
|
activeUsersManager, store);
|
||||||
this.queue = queue;
|
this.queue = queue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -32,6 +32,7 @@ import org.apache.hadoop.yarn.api.records.QueueState;
|
||||||
import org.apache.hadoop.yarn.api.records.Resource;
|
import org.apache.hadoop.yarn.api.records.Resource;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode;
|
||||||
|
|
||||||
|
@ -197,6 +198,12 @@ extends org.apache.hadoop.yarn.server.resourcemanager.scheduler.Queue {
|
||||||
*/
|
*/
|
||||||
public void updateClusterResource(Resource clusterResource);
|
public void updateClusterResource(Resource clusterResource);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the {@link ActiveUsersManager} for the queue.
|
||||||
|
* @return the <code>ActiveUsersManager</code> for the queue
|
||||||
|
*/
|
||||||
|
public ActiveUsersManager getActiveUsersManager();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Recover the state of the queue
|
* Recover the state of the queue
|
||||||
* @param clusterResource the resource of the cluster
|
* @param clusterResource the resource of the cluster
|
||||||
|
|
|
@ -355,7 +355,8 @@ implements ResourceScheduler, CapacitySchedulerContext {
|
||||||
|
|
||||||
// TODO: Fix store
|
// TODO: Fix store
|
||||||
SchedulerApp SchedulerApp =
|
SchedulerApp SchedulerApp =
|
||||||
new SchedulerApp(applicationAttemptId, user, queue, rmContext, null);
|
new SchedulerApp(applicationAttemptId, user, queue,
|
||||||
|
queue.getActiveUsersManager(), rmContext, null);
|
||||||
|
|
||||||
// Submit to the queue
|
// Submit to the queue
|
||||||
try {
|
try {
|
||||||
|
|
|
@ -37,6 +37,8 @@ import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
||||||
import org.apache.hadoop.security.AccessControlException;
|
import org.apache.hadoop.security.AccessControlException;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
import org.apache.hadoop.security.authorize.AccessControlList;
|
import org.apache.hadoop.security.authorize.AccessControlList;
|
||||||
|
import org.apache.hadoop.yarn.Lock;
|
||||||
|
import org.apache.hadoop.yarn.Lock.NoLock;
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
||||||
import org.apache.hadoop.yarn.api.records.Container;
|
import org.apache.hadoop.yarn.api.records.Container;
|
||||||
import org.apache.hadoop.yarn.api.records.ContainerId;
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
|
@ -58,6 +60,7 @@ import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerState;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerState;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.QueueMetrics;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.QueueMetrics;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
||||||
|
@ -120,6 +123,8 @@ public class LeafQueue implements CSQueue {
|
||||||
|
|
||||||
private CapacitySchedulerContext scheduler;
|
private CapacitySchedulerContext scheduler;
|
||||||
|
|
||||||
|
private final ActiveUsersManager activeUsersManager;
|
||||||
|
|
||||||
final static int DEFAULT_AM_RESOURCE = 2 * 1024;
|
final static int DEFAULT_AM_RESOURCE = 2 * 1024;
|
||||||
|
|
||||||
public LeafQueue(CapacitySchedulerContext cs,
|
public LeafQueue(CapacitySchedulerContext cs,
|
||||||
|
@ -132,7 +137,7 @@ public class LeafQueue implements CSQueue {
|
||||||
this.metrics = old != null ? old.getMetrics() :
|
this.metrics = old != null ? old.getMetrics() :
|
||||||
QueueMetrics.forQueue(getQueuePath(), parent,
|
QueueMetrics.forQueue(getQueuePath(), parent,
|
||||||
cs.getConfiguration().getEnableUserMetrics());
|
cs.getConfiguration().getEnableUserMetrics());
|
||||||
|
this.activeUsersManager = new ActiveUsersManager(metrics);
|
||||||
this.minimumAllocation = cs.getMinimumResourceCapability();
|
this.minimumAllocation = cs.getMinimumResourceCapability();
|
||||||
this.maximumAllocation = cs.getMaximumResourceCapability();
|
this.maximumAllocation = cs.getMaximumResourceCapability();
|
||||||
this.minimumAllocationFactor =
|
this.minimumAllocationFactor =
|
||||||
|
@ -348,6 +353,11 @@ public class LeafQueue implements CSQueue {
|
||||||
return maxActiveApplicationsPerUser;
|
return maxActiveApplicationsPerUser;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ActiveUsersManager getActiveUsersManager() {
|
||||||
|
return activeUsersManager;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public synchronized float getUsedCapacity() {
|
public synchronized float getUsedCapacity() {
|
||||||
return usedCapacity;
|
return usedCapacity;
|
||||||
|
@ -674,6 +684,12 @@ public class LeafQueue implements CSQueue {
|
||||||
// Check if we can activate more applications
|
// Check if we can activate more applications
|
||||||
activateApplications();
|
activateApplications();
|
||||||
|
|
||||||
|
// Inform the activeUsersManager
|
||||||
|
synchronized (application) {
|
||||||
|
activeUsersManager.deactivateApplication(
|
||||||
|
application.getUser(), application.getApplicationId());
|
||||||
|
}
|
||||||
|
|
||||||
LOG.info("Application removed -" +
|
LOG.info("Application removed -" +
|
||||||
" appId: " + application.getApplicationId() +
|
" appId: " + application.getApplicationId() +
|
||||||
" user: " + application.getUser() +
|
" user: " + application.getUser() +
|
||||||
|
@ -837,6 +853,7 @@ public class LeafQueue implements CSQueue {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Lock({LeafQueue.class, SchedulerApp.class})
|
||||||
private Resource computeAndSetUserResourceLimit(SchedulerApp application,
|
private Resource computeAndSetUserResourceLimit(SchedulerApp application,
|
||||||
Resource clusterResource, Resource required) {
|
Resource clusterResource, Resource required) {
|
||||||
String user = application.getUser();
|
String user = application.getUser();
|
||||||
|
@ -853,6 +870,7 @@ public class LeafQueue implements CSQueue {
|
||||||
minimumAllocation.getMemory();
|
minimumAllocation.getMemory();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Lock(NoLock.class)
|
||||||
private Resource computeUserLimit(SchedulerApp application,
|
private Resource computeUserLimit(SchedulerApp application,
|
||||||
Resource clusterResource, Resource required) {
|
Resource clusterResource, Resource required) {
|
||||||
// What is our current capacity?
|
// What is our current capacity?
|
||||||
|
@ -878,10 +896,7 @@ public class LeafQueue implements CSQueue {
|
||||||
// Also, the queue's configured capacity should be higher than
|
// Also, the queue's configured capacity should be higher than
|
||||||
// queue-hard-limit * ulMin
|
// queue-hard-limit * ulMin
|
||||||
|
|
||||||
String userName = application.getUser();
|
final int activeUsers = activeUsersManager.getNumActiveUsers();
|
||||||
|
|
||||||
final int activeUsers = users.size();
|
|
||||||
User user = getUser(userName);
|
|
||||||
|
|
||||||
int limit =
|
int limit =
|
||||||
roundUp(
|
roundUp(
|
||||||
|
@ -893,12 +908,13 @@ public class LeafQueue implements CSQueue {
|
||||||
);
|
);
|
||||||
|
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
|
String userName = application.getUser();
|
||||||
LOG.debug("User limit computation for " + userName +
|
LOG.debug("User limit computation for " + userName +
|
||||||
" in queue " + getQueueName() +
|
" in queue " + getQueueName() +
|
||||||
" userLimit=" + userLimit +
|
" userLimit=" + userLimit +
|
||||||
" userLimitFactor=" + userLimitFactor +
|
" userLimitFactor=" + userLimitFactor +
|
||||||
" required: " + required +
|
" required: " + required +
|
||||||
" consumed: " + user.getConsumedResources() +
|
" consumed: " + getUser(userName).getConsumedResources() +
|
||||||
" limit: " + limit +
|
" limit: " + limit +
|
||||||
" queueCapacity: " + queueCapacity +
|
" queueCapacity: " + queueCapacity +
|
||||||
" qconsumed: " + consumed +
|
" qconsumed: " + consumed +
|
||||||
|
@ -1308,8 +1324,10 @@ public class LeafQueue implements CSQueue {
|
||||||
|
|
||||||
// Update application properties
|
// Update application properties
|
||||||
for (SchedulerApp application : activeApplications) {
|
for (SchedulerApp application : activeApplications) {
|
||||||
computeAndSetUserResourceLimit(
|
synchronized (application) {
|
||||||
application, clusterResource, Resources.none());
|
computeAndSetUserResourceLimit(
|
||||||
|
application, clusterResource, Resources.none());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -48,6 +48,7 @@ import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.resource.Resources;
|
import org.apache.hadoop.yarn.server.resourcemanager.resource.Resources;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.QueueMetrics;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.QueueMetrics;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
||||||
|
@ -240,6 +241,12 @@ public class ParentQueue implements CSQueue {
|
||||||
return maximumCapacity;
|
return maximumCapacity;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ActiveUsersManager getActiveUsersManager() {
|
||||||
|
// Should never be called since all applications are submitted to LeafQueues
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public synchronized float getUsedCapacity() {
|
public synchronized float getUsedCapacity() {
|
||||||
return usedCapacity;
|
return usedCapacity;
|
||||||
|
|
|
@ -66,6 +66,7 @@ import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeCleanContainerEvent;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeCleanContainerEvent;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.Allocation;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.Allocation;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.Queue;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.Queue;
|
||||||
|
@ -125,9 +126,10 @@ public class FifoScheduler implements ResourceScheduler {
|
||||||
private Map<ApplicationAttemptId, SchedulerApp> applications
|
private Map<ApplicationAttemptId, SchedulerApp> applications
|
||||||
= new TreeMap<ApplicationAttemptId, SchedulerApp>();
|
= new TreeMap<ApplicationAttemptId, SchedulerApp>();
|
||||||
|
|
||||||
|
private final ActiveUsersManager activeUsersManager;
|
||||||
|
|
||||||
private static final String DEFAULT_QUEUE_NAME = "default";
|
private static final String DEFAULT_QUEUE_NAME = "default";
|
||||||
private final QueueMetrics metrics =
|
private final QueueMetrics metrics;
|
||||||
QueueMetrics.forQueue(DEFAULT_QUEUE_NAME, null, false);
|
|
||||||
|
|
||||||
private final Queue DEFAULT_QUEUE = new Queue() {
|
private final Queue DEFAULT_QUEUE = new Queue() {
|
||||||
@Override
|
@Override
|
||||||
|
@ -174,6 +176,11 @@ public class FifoScheduler implements ResourceScheduler {
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
public FifoScheduler() {
|
||||||
|
metrics = QueueMetrics.forQueue(DEFAULT_QUEUE_NAME, null, false);
|
||||||
|
activeUsersManager = new ActiveUsersManager(metrics);
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Resource getMinimumResourceCapability() {
|
public Resource getMinimumResourceCapability() {
|
||||||
return minimumAllocation;
|
return minimumAllocation;
|
||||||
|
@ -288,7 +295,7 @@ public class FifoScheduler implements ResourceScheduler {
|
||||||
String user) {
|
String user) {
|
||||||
// TODO: Fix store
|
// TODO: Fix store
|
||||||
SchedulerApp schedulerApp =
|
SchedulerApp schedulerApp =
|
||||||
new SchedulerApp(appAttemptId, user, DEFAULT_QUEUE,
|
new SchedulerApp(appAttemptId, user, DEFAULT_QUEUE, activeUsersManager,
|
||||||
this.rmContext, null);
|
this.rmContext, null);
|
||||||
applications.put(appAttemptId, schedulerApp);
|
applications.put(appAttemptId, schedulerApp);
|
||||||
metrics.submitApp(user);
|
metrics.submitApp(user);
|
||||||
|
@ -318,6 +325,12 @@ public class FifoScheduler implements ResourceScheduler {
|
||||||
RMContainerEventType.KILL);
|
RMContainerEventType.KILL);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Inform the activeUsersManager
|
||||||
|
synchronized (application) {
|
||||||
|
activeUsersManager.deactivateApplication(
|
||||||
|
application.getUser(), application.getApplicationId());
|
||||||
|
}
|
||||||
|
|
||||||
// Clean up pending requests, metrics etc.
|
// Clean up pending requests, metrics etc.
|
||||||
application.stop(rmAppAttemptFinalState);
|
application.stop(rmAppAttemptFinalState);
|
||||||
|
|
||||||
|
|
|
@ -68,7 +68,7 @@ public class MetricsOverviewTable extends HtmlBlock {
|
||||||
th().$class("ui-state-default")._("Memory Used")._().
|
th().$class("ui-state-default")._("Memory Used")._().
|
||||||
th().$class("ui-state-default")._("Memory Total")._().
|
th().$class("ui-state-default")._("Memory Total")._().
|
||||||
th().$class("ui-state-default")._("Memory Reserved")._().
|
th().$class("ui-state-default")._("Memory Reserved")._().
|
||||||
th().$class("ui-state-default")._("Total Nodes")._().
|
th().$class("ui-state-default")._("Active Nodes")._().
|
||||||
th().$class("ui-state-default")._("Decommissioned Nodes")._().
|
th().$class("ui-state-default")._("Decommissioned Nodes")._().
|
||||||
th().$class("ui-state-default")._("Lost Nodes")._().
|
th().$class("ui-state-default")._("Lost Nodes")._().
|
||||||
th().$class("ui-state-default")._("Unhealthy Nodes")._().
|
th().$class("ui-state-default")._("Unhealthy Nodes")._().
|
||||||
|
@ -82,7 +82,7 @@ public class MetricsOverviewTable extends HtmlBlock {
|
||||||
td(StringUtils.byteDesc(clusterMetrics.getAllocatedMB() * BYTES_IN_MB)).
|
td(StringUtils.byteDesc(clusterMetrics.getAllocatedMB() * BYTES_IN_MB)).
|
||||||
td(StringUtils.byteDesc(clusterMetrics.getTotalMB() * BYTES_IN_MB)).
|
td(StringUtils.byteDesc(clusterMetrics.getTotalMB() * BYTES_IN_MB)).
|
||||||
td(StringUtils.byteDesc(clusterMetrics.getReservedMB() * BYTES_IN_MB)).
|
td(StringUtils.byteDesc(clusterMetrics.getReservedMB() * BYTES_IN_MB)).
|
||||||
td().a(url("nodes"),String.valueOf(clusterMetrics.getTotalNodes()))._().
|
td().a(url("nodes"),String.valueOf(clusterMetrics.getActiveNodes()))._().
|
||||||
td().a(url("nodes/decommissioned"),String.valueOf(clusterMetrics.getDecommissionedNodes()))._().
|
td().a(url("nodes/decommissioned"),String.valueOf(clusterMetrics.getDecommissionedNodes()))._().
|
||||||
td().a(url("nodes/lost"),String.valueOf(clusterMetrics.getLostNodes()))._().
|
td().a(url("nodes/lost"),String.valueOf(clusterMetrics.getLostNodes()))._().
|
||||||
td().a(url("nodes/unhealthy"),String.valueOf(clusterMetrics.getUnhealthyNodes()))._().
|
td().a(url("nodes/unhealthy"),String.valueOf(clusterMetrics.getUnhealthyNodes()))._().
|
||||||
|
|
|
@ -24,6 +24,8 @@ import static org.apache.hadoop.yarn.webapp.view.JQueryUI.DATATABLES_ID;
|
||||||
import static org.apache.hadoop.yarn.webapp.view.JQueryUI.initID;
|
import static org.apache.hadoop.yarn.webapp.view.JQueryUI.initID;
|
||||||
import static org.apache.hadoop.yarn.webapp.view.JQueryUI.tableInit;
|
import static org.apache.hadoop.yarn.webapp.view.JQueryUI.tableInit;
|
||||||
|
|
||||||
|
import java.util.Collection;
|
||||||
|
|
||||||
import org.apache.hadoop.util.StringUtils;
|
import org.apache.hadoop.util.StringUtils;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.RMContext;
|
import org.apache.hadoop.yarn.server.resourcemanager.RMContext;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.ResourceManager;
|
import org.apache.hadoop.yarn.server.resourcemanager.ResourceManager;
|
||||||
|
@ -36,6 +38,7 @@ import org.apache.hadoop.yarn.webapp.SubView;
|
||||||
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet;
|
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet;
|
||||||
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.TABLE;
|
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.TABLE;
|
||||||
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.TBODY;
|
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.TBODY;
|
||||||
|
import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.TR;
|
||||||
import org.apache.hadoop.yarn.webapp.view.HtmlBlock;
|
import org.apache.hadoop.yarn.webapp.view.HtmlBlock;
|
||||||
|
|
||||||
import com.google.inject.Inject;
|
import com.google.inject.Inject;
|
||||||
|
@ -79,7 +82,19 @@ class NodesPage extends RmView {
|
||||||
if(type != null && !type.isEmpty()) {
|
if(type != null && !type.isEmpty()) {
|
||||||
stateFilter = RMNodeState.valueOf(type.toUpperCase());
|
stateFilter = RMNodeState.valueOf(type.toUpperCase());
|
||||||
}
|
}
|
||||||
for (RMNode ni : this.rmContext.getRMNodes().values()) {
|
Collection<RMNode> rmNodes = this.rmContext.getRMNodes().values();
|
||||||
|
boolean isInactive = false;
|
||||||
|
if (stateFilter != null) {
|
||||||
|
switch (stateFilter) {
|
||||||
|
case DECOMMISSIONED:
|
||||||
|
case LOST:
|
||||||
|
case REBOOTED:
|
||||||
|
rmNodes = this.rmContext.getInactiveRMNodes().values();
|
||||||
|
isInactive = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (RMNode ni : rmNodes) {
|
||||||
if(stateFilter != null) {
|
if(stateFilter != null) {
|
||||||
RMNodeState state = ni.getState();
|
RMNodeState state = ni.getState();
|
||||||
if(!stateFilter.equals(state)) {
|
if(!stateFilter.equals(state)) {
|
||||||
|
@ -89,12 +104,17 @@ class NodesPage extends RmView {
|
||||||
NodeInfo info = new NodeInfo(ni, sched);
|
NodeInfo info = new NodeInfo(ni, sched);
|
||||||
int usedMemory = (int)info.getUsedMemory();
|
int usedMemory = (int)info.getUsedMemory();
|
||||||
int availableMemory = (int)info.getAvailableMemory();
|
int availableMemory = (int)info.getAvailableMemory();
|
||||||
tbody.tr().
|
TR<TBODY<TABLE<Hamlet>>> row = tbody.tr().
|
||||||
td(info.getRack()).
|
td(info.getRack()).
|
||||||
td(info.getState()).
|
td(info.getState()).
|
||||||
td(info.getNodeId()).
|
td(info.getNodeId());
|
||||||
td().a("http://" + info.getNodeHTTPAddress(), info.getNodeHTTPAddress())._().
|
if (isInactive) {
|
||||||
td(info.getHealthStatus()).
|
row.td()._("N/A")._();
|
||||||
|
} else {
|
||||||
|
String httpAddress = info.getNodeHTTPAddress();
|
||||||
|
row.td().a("http://" + httpAddress, httpAddress)._();
|
||||||
|
}
|
||||||
|
row.td(info.getHealthStatus()).
|
||||||
td(Times.format(info.getLastHealthUpdate())).
|
td(Times.format(info.getLastHealthUpdate())).
|
||||||
td(info.getHealthReport()).
|
td(info.getHealthReport()).
|
||||||
td(String.valueOf(info.getNumContainers())).
|
td(String.valueOf(info.getNumContainers())).
|
||||||
|
|
|
@ -19,6 +19,7 @@
|
||||||
package org.apache.hadoop.yarn.server.resourcemanager.webapp;
|
package org.apache.hadoop.yarn.server.resourcemanager.webapp;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.util.Collection;
|
||||||
import java.util.concurrent.ConcurrentMap;
|
import java.util.concurrent.ConcurrentMap;
|
||||||
|
|
||||||
import javax.servlet.http.HttpServletRequest;
|
import javax.servlet.http.HttpServletRequest;
|
||||||
|
@ -68,6 +69,7 @@ import com.google.inject.Singleton;
|
||||||
@Singleton
|
@Singleton
|
||||||
@Path("/ws/v1/cluster")
|
@Path("/ws/v1/cluster")
|
||||||
public class RMWebServices {
|
public class RMWebServices {
|
||||||
|
private static final String EMPTY = "";
|
||||||
private static final Log LOG = LogFactory.getLog(RMWebServices.class);
|
private static final Log LOG = LogFactory.getLog(RMWebServices.class);
|
||||||
private final ResourceManager rm;
|
private final ResourceManager rm;
|
||||||
private static RecordFactory recordFactory = RecordFactoryProvider
|
private static RecordFactory recordFactory = RecordFactoryProvider
|
||||||
|
@ -144,12 +146,23 @@ public class RMWebServices {
|
||||||
if (sched == null) {
|
if (sched == null) {
|
||||||
throw new NotFoundException("Null ResourceScheduler instance");
|
throw new NotFoundException("Null ResourceScheduler instance");
|
||||||
}
|
}
|
||||||
|
Collection<RMNode> rmNodes = this.rm.getRMContext().getRMNodes().values();
|
||||||
|
boolean isInactive = false;
|
||||||
|
if (filterState != null && !filterState.isEmpty()) {
|
||||||
|
RMNodeState nodeState = RMNodeState.valueOf(filterState.toUpperCase());
|
||||||
|
switch (nodeState) {
|
||||||
|
case DECOMMISSIONED:
|
||||||
|
case LOST:
|
||||||
|
case REBOOTED:
|
||||||
|
rmNodes = this.rm.getRMContext().getInactiveRMNodes().values();
|
||||||
|
isInactive = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
NodesInfo allNodes = new NodesInfo();
|
NodesInfo allNodes = new NodesInfo();
|
||||||
for (RMNode ni : this.rm.getRMContext().getRMNodes().values()) {
|
for (RMNode ni : rmNodes) {
|
||||||
NodeInfo nodeInfo = new NodeInfo(ni, sched);
|
NodeInfo nodeInfo = new NodeInfo(ni, sched);
|
||||||
if (filterState != null) {
|
if (filterState != null) {
|
||||||
RMNodeState.valueOf(filterState);
|
|
||||||
if (!(nodeInfo.getState().equalsIgnoreCase(filterState))) {
|
if (!(nodeInfo.getState().equalsIgnoreCase(filterState))) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
@ -165,6 +178,9 @@ public class RMWebServices {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if (isInactive) {
|
||||||
|
nodeInfo.setNodeHTTPAddress(EMPTY);
|
||||||
|
}
|
||||||
allNodes.add(nodeInfo);
|
allNodes.add(nodeInfo);
|
||||||
}
|
}
|
||||||
return allNodes;
|
return allNodes;
|
||||||
|
@ -183,10 +199,19 @@ public class RMWebServices {
|
||||||
}
|
}
|
||||||
NodeId nid = ConverterUtils.toNodeId(nodeId);
|
NodeId nid = ConverterUtils.toNodeId(nodeId);
|
||||||
RMNode ni = this.rm.getRMContext().getRMNodes().get(nid);
|
RMNode ni = this.rm.getRMContext().getRMNodes().get(nid);
|
||||||
|
boolean isInactive = false;
|
||||||
if (ni == null) {
|
if (ni == null) {
|
||||||
throw new NotFoundException("nodeId, " + nodeId + ", is not found");
|
ni = this.rm.getRMContext().getInactiveRMNodes().get(nid.getHost());
|
||||||
|
if (ni == null) {
|
||||||
|
throw new NotFoundException("nodeId, " + nodeId + ", is not found");
|
||||||
|
}
|
||||||
|
isInactive = true;
|
||||||
}
|
}
|
||||||
return new NodeInfo(ni, sched);
|
NodeInfo nodeInfo = new NodeInfo(ni, sched);
|
||||||
|
if (isInactive) {
|
||||||
|
nodeInfo.setNodeHTTPAddress(EMPTY);
|
||||||
|
}
|
||||||
|
return nodeInfo;
|
||||||
}
|
}
|
||||||
|
|
||||||
@GET
|
@GET
|
||||||
|
|
|
@ -44,6 +44,7 @@ public class ClusterMetricsInfo {
|
||||||
protected int unhealthyNodes;
|
protected int unhealthyNodes;
|
||||||
protected int decommissionedNodes;
|
protected int decommissionedNodes;
|
||||||
protected int rebootedNodes;
|
protected int rebootedNodes;
|
||||||
|
protected int activeNodes;
|
||||||
|
|
||||||
public ClusterMetricsInfo() {
|
public ClusterMetricsInfo() {
|
||||||
} // JAXB needs this
|
} // JAXB needs this
|
||||||
|
@ -59,12 +60,13 @@ public class ClusterMetricsInfo {
|
||||||
this.allocatedMB = metrics.getAllocatedGB() * MB_IN_GB;
|
this.allocatedMB = metrics.getAllocatedGB() * MB_IN_GB;
|
||||||
this.containersAllocated = metrics.getAllocatedContainers();
|
this.containersAllocated = metrics.getAllocatedContainers();
|
||||||
this.totalMB = availableMB + reservedMB + allocatedMB;
|
this.totalMB = availableMB + reservedMB + allocatedMB;
|
||||||
this.totalNodes = clusterMetrics.getNumNMs();
|
this.activeNodes = clusterMetrics.getNumActiveNMs();
|
||||||
this.lostNodes = clusterMetrics.getNumLostNMs();
|
this.lostNodes = clusterMetrics.getNumLostNMs();
|
||||||
this.unhealthyNodes = clusterMetrics.getUnhealthyNMs();
|
this.unhealthyNodes = clusterMetrics.getUnhealthyNMs();
|
||||||
this.decommissionedNodes = clusterMetrics.getNumDecommisionedNMs();
|
this.decommissionedNodes = clusterMetrics.getNumDecommisionedNMs();
|
||||||
this.rebootedNodes = clusterMetrics.getNumRebootedNMs();
|
this.rebootedNodes = clusterMetrics.getNumRebootedNMs();
|
||||||
|
this.totalNodes = activeNodes + lostNodes + decommissionedNodes
|
||||||
|
+ rebootedNodes;
|
||||||
}
|
}
|
||||||
|
|
||||||
public int getAppsSubmitted() {
|
public int getAppsSubmitted() {
|
||||||
|
@ -95,6 +97,10 @@ public class ClusterMetricsInfo {
|
||||||
return this.totalNodes;
|
return this.totalNodes;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public int getActiveNodes() {
|
||||||
|
return this.activeNodes;
|
||||||
|
}
|
||||||
|
|
||||||
public int getLostNodes() {
|
public int getLostNodes() {
|
||||||
return this.lostNodes;
|
return this.lostNodes;
|
||||||
}
|
}
|
||||||
|
|
|
@ -95,6 +95,10 @@ public class NodeInfo {
|
||||||
return this.nodeHTTPAddress;
|
return this.nodeHTTPAddress;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void setNodeHTTPAddress(String nodeHTTPAddress) {
|
||||||
|
this.nodeHTTPAddress = nodeHTTPAddress;
|
||||||
|
}
|
||||||
|
|
||||||
public String getHealthStatus() {
|
public String getHealthStatus() {
|
||||||
return this.healthStatus;
|
return this.healthStatus;
|
||||||
}
|
}
|
||||||
|
|
|
@ -81,13 +81,20 @@ public class MockNM {
|
||||||
}
|
}
|
||||||
|
|
||||||
public HeartbeatResponse nodeHeartbeat(boolean b) throws Exception {
|
public HeartbeatResponse nodeHeartbeat(boolean b) throws Exception {
|
||||||
return nodeHeartbeat(new HashMap<ApplicationId, List<ContainerStatus>>(), b);
|
return nodeHeartbeat(new HashMap<ApplicationId, List<ContainerStatus>>(),
|
||||||
|
b, ++responseId);
|
||||||
}
|
}
|
||||||
|
|
||||||
public HeartbeatResponse nodeHeartbeat(Map<ApplicationId,
|
public HeartbeatResponse nodeHeartbeat(Map<ApplicationId,
|
||||||
List<ContainerStatus>> conts, boolean isHealthy) throws Exception {
|
List<ContainerStatus>> conts, boolean isHealthy) throws Exception {
|
||||||
|
return nodeHeartbeat(conts, isHealthy, ++responseId);
|
||||||
|
}
|
||||||
|
|
||||||
|
public HeartbeatResponse nodeHeartbeat(Map<ApplicationId,
|
||||||
|
List<ContainerStatus>> conts, boolean isHealthy, int resId) throws Exception {
|
||||||
NodeHeartbeatRequest req = Records.newRecord(NodeHeartbeatRequest.class);
|
NodeHeartbeatRequest req = Records.newRecord(NodeHeartbeatRequest.class);
|
||||||
NodeStatus status = Records.newRecord(NodeStatus.class);
|
NodeStatus status = Records.newRecord(NodeStatus.class);
|
||||||
|
status.setResponseId(resId);
|
||||||
status.setNodeId(nodeId);
|
status.setNodeId(nodeId);
|
||||||
for (Map.Entry<ApplicationId, List<ContainerStatus>> entry : conts.entrySet()) {
|
for (Map.Entry<ApplicationId, List<ContainerStatus>> entry : conts.entrySet()) {
|
||||||
status.setContainersStatuses(entry.getValue());
|
status.setContainersStatuses(entry.getValue());
|
||||||
|
@ -97,7 +104,6 @@ public class MockNM {
|
||||||
healthStatus.setIsNodeHealthy(isHealthy);
|
healthStatus.setIsNodeHealthy(isHealthy);
|
||||||
healthStatus.setLastHealthReportTime(1);
|
healthStatus.setLastHealthReportTime(1);
|
||||||
status.setNodeHealthStatus(healthStatus);
|
status.setNodeHealthStatus(healthStatus);
|
||||||
status.setResponseId(++responseId);
|
|
||||||
req.setNodeStatus(status);
|
req.setNodeStatus(status);
|
||||||
return resourceTracker.nodeHeartbeat(req).getHeartbeatResponse();
|
return resourceTracker.nodeHeartbeat(req).getHeartbeatResponse();
|
||||||
}
|
}
|
||||||
|
|
|
@ -57,6 +57,17 @@ public class MockNodes {
|
||||||
return list;
|
return list;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static List<RMNode> lostNodes(int racks, int nodesPerRack,
|
||||||
|
Resource perNode) {
|
||||||
|
List<RMNode> list = Lists.newArrayList();
|
||||||
|
for (int i = 0; i < racks; ++i) {
|
||||||
|
for (int j = 0; j < nodesPerRack; ++j) {
|
||||||
|
list.add(lostNodeInfo(i, perNode, RMNodeState.LOST));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return list;
|
||||||
|
}
|
||||||
|
|
||||||
public static NodeId newNodeID(String host, int port) {
|
public static NodeId newNodeID(String host, int port) {
|
||||||
NodeId nid = recordFactory.newRecordInstance(NodeId.class);
|
NodeId nid = recordFactory.newRecordInstance(NodeId.class);
|
||||||
nid.setHost(host);
|
nid.setHost(host);
|
||||||
|
@ -82,92 +93,120 @@ public class MockNodes {
|
||||||
return rs;
|
return rs;
|
||||||
}
|
}
|
||||||
|
|
||||||
public static RMNode newNodeInfo(int rack, final Resource perNode) {
|
private static class MockRMNodeImpl implements RMNode {
|
||||||
|
private NodeId nodeId;
|
||||||
|
private String hostName;
|
||||||
|
private String nodeAddr;
|
||||||
|
private String httpAddress;
|
||||||
|
private int cmdPort;
|
||||||
|
private Resource perNode;
|
||||||
|
private String rackName;
|
||||||
|
private NodeHealthStatus nodeHealthStatus;
|
||||||
|
private RMNodeState state;
|
||||||
|
|
||||||
|
public MockRMNodeImpl(NodeId nodeId, String nodeAddr, String httpAddress,
|
||||||
|
Resource perNode, String rackName, NodeHealthStatus nodeHealthStatus,
|
||||||
|
int cmdPort, String hostName, RMNodeState state) {
|
||||||
|
this.nodeId = nodeId;
|
||||||
|
this.nodeAddr = nodeAddr;
|
||||||
|
this.httpAddress = httpAddress;
|
||||||
|
this.perNode = perNode;
|
||||||
|
this.rackName = rackName;
|
||||||
|
this.nodeHealthStatus = nodeHealthStatus;
|
||||||
|
this.cmdPort = cmdPort;
|
||||||
|
this.hostName = hostName;
|
||||||
|
this.state = state;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public NodeId getNodeID() {
|
||||||
|
return this.nodeId;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getHostName() {
|
||||||
|
return this.hostName;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getCommandPort() {
|
||||||
|
return this.cmdPort;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getHttpPort() {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getNodeAddress() {
|
||||||
|
return this.nodeAddr;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getHttpAddress() {
|
||||||
|
return this.httpAddress;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public NodeHealthStatus getNodeHealthStatus() {
|
||||||
|
return this.nodeHealthStatus;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Resource getTotalCapability() {
|
||||||
|
return this.perNode;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getRackName() {
|
||||||
|
return this.rackName;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Node getNode() {
|
||||||
|
throw new UnsupportedOperationException("Not supported yet.");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public RMNodeState getState() {
|
||||||
|
return this.state;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<ContainerId> getContainersToCleanUp() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public List<ApplicationId> getAppsToCleanup() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public HeartbeatResponse getLastHeartBeatResponse() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
private static RMNode buildRMNode(int rack, final Resource perNode, RMNodeState state, String httpAddr) {
|
||||||
final String rackName = "rack"+ rack;
|
final String rackName = "rack"+ rack;
|
||||||
final int nid = NODE_ID++;
|
final int nid = NODE_ID++;
|
||||||
final String hostName = "host"+ nid;
|
final String hostName = "host"+ nid;
|
||||||
final int port = 123;
|
final int port = 123;
|
||||||
final NodeId nodeID = newNodeID(hostName, port);
|
final NodeId nodeID = newNodeID(hostName, port);
|
||||||
final String httpAddress = "localhost:0";
|
final String httpAddress = httpAddr;
|
||||||
final NodeHealthStatus nodeHealthStatus =
|
final NodeHealthStatus nodeHealthStatus =
|
||||||
recordFactory.newRecordInstance(NodeHealthStatus.class);
|
recordFactory.newRecordInstance(NodeHealthStatus.class);
|
||||||
final Resource used = newUsedResource(perNode);
|
return new MockRMNodeImpl(nodeID, hostName, httpAddress, perNode, rackName,
|
||||||
final Resource avail = newAvailResource(perNode, used);
|
nodeHealthStatus, nid, hostName, state);
|
||||||
return new RMNode() {
|
}
|
||||||
@Override
|
|
||||||
public NodeId getNodeID() {
|
|
||||||
return nodeID;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
public static RMNode lostNodeInfo(int rack, final Resource perNode, RMNodeState state) {
|
||||||
public String getNodeAddress() {
|
return buildRMNode(rack, perNode, state, "N/A");
|
||||||
return hostName;
|
}
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
public static RMNode newNodeInfo(int rack, final Resource perNode) {
|
||||||
public String getHttpAddress() {
|
return buildRMNode(rack, perNode, null, "localhost:0");
|
||||||
return httpAddress;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Resource getTotalCapability() {
|
|
||||||
return perNode;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public String getRackName() {
|
|
||||||
return rackName;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Node getNode() {
|
|
||||||
throw new UnsupportedOperationException("Not supported yet.");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public NodeHealthStatus getNodeHealthStatus() {
|
|
||||||
return nodeHealthStatus;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int getCommandPort() {
|
|
||||||
return nid;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int getHttpPort() {
|
|
||||||
// TODO Auto-generated method stub
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public String getHostName() {
|
|
||||||
return hostName;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public RMNodeState getState() {
|
|
||||||
// TODO Auto-generated method stub
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public List<ApplicationId> getAppsToCleanup() {
|
|
||||||
// TODO Auto-generated method stub
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public List<ContainerId> getContainersToCleanUp() {
|
|
||||||
// TODO Auto-generated method stub
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public HeartbeatResponse getLastHeartBeatResponse() {
|
|
||||||
// TODO Auto-generated method stub
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -131,6 +131,12 @@ public class MockRM extends ResourceManager {
|
||||||
node.handle(new RMNodeEvent(nm.getNodeId(), RMNodeEventType.STARTED));
|
node.handle(new RMNodeEvent(nm.getNodeId(), RMNodeEventType.STARTED));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void sendNodeLost(MockNM nm) throws Exception {
|
||||||
|
RMNodeImpl node = (RMNodeImpl) getRMContext().getRMNodes().get(
|
||||||
|
nm.getNodeId());
|
||||||
|
node.handle(new RMNodeEvent(nm.getNodeId(), RMNodeEventType.EXPIRE));
|
||||||
|
}
|
||||||
|
|
||||||
public void NMwaitForState(NodeId nodeid, RMNodeState finalState)
|
public void NMwaitForState(NodeId nodeid, RMNodeState finalState)
|
||||||
throws Exception {
|
throws Exception {
|
||||||
RMNode node = getRMContext().getRMNodes().get(nodeid);
|
RMNode node = getRMContext().getRMNodes().get(nodeid);
|
||||||
|
|
|
@ -31,6 +31,7 @@ import junit.framework.Assert;
|
||||||
import org.apache.hadoop.yarn.api.records.ContainerId;
|
import org.apache.hadoop.yarn.api.records.ContainerId;
|
||||||
import org.apache.hadoop.yarn.api.records.ContainerStatus;
|
import org.apache.hadoop.yarn.api.records.ContainerStatus;
|
||||||
import org.apache.hadoop.yarn.api.records.NodeHealthStatus;
|
import org.apache.hadoop.yarn.api.records.NodeHealthStatus;
|
||||||
|
import org.apache.hadoop.yarn.api.records.NodeId;
|
||||||
import org.apache.hadoop.yarn.event.EventHandler;
|
import org.apache.hadoop.yarn.event.EventHandler;
|
||||||
import org.apache.hadoop.yarn.server.api.records.HeartbeatResponse;
|
import org.apache.hadoop.yarn.server.api.records.HeartbeatResponse;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.recovery.MemStore;
|
import org.apache.hadoop.yarn.server.resourcemanager.recovery.MemStore;
|
||||||
|
@ -100,8 +101,8 @@ public class TestRMNodeTransitions {
|
||||||
rmDispatcher.register(SchedulerEventType.class,
|
rmDispatcher.register(SchedulerEventType.class,
|
||||||
new TestSchedulerEventDispatcher());
|
new TestSchedulerEventDispatcher());
|
||||||
|
|
||||||
|
NodeId nodeId = BuilderUtils.newNodeId("localhost", 0);
|
||||||
node = new RMNodeImpl(null, rmContext, null, 0, 0, null, null);
|
node = new RMNodeImpl(nodeId, rmContext, null, 0, 0, null, null);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -157,14 +157,14 @@ public class TestResourceTrackerService {
|
||||||
rm.start();
|
rm.start();
|
||||||
|
|
||||||
MockNM nm1 = rm.registerNode("host1:1234", 5120);
|
MockNM nm1 = rm.registerNode("host1:1234", 5120);
|
||||||
MockNM nm2 = new MockNM("host2:1234", 2048, rm.getResourceTrackerService());
|
MockNM nm2 = rm.registerNode("host2:1234", 2048);
|
||||||
|
|
||||||
int initialMetricCount = ClusterMetrics.getMetrics().getNumRebootedNMs();
|
int initialMetricCount = ClusterMetrics.getMetrics().getNumRebootedNMs();
|
||||||
HeartbeatResponse nodeHeartbeat = nm1.nodeHeartbeat(true);
|
HeartbeatResponse nodeHeartbeat = nm1.nodeHeartbeat(true);
|
||||||
Assert.assertTrue(NodeAction.NORMAL.equals(nodeHeartbeat.getNodeAction()));
|
Assert.assertTrue(NodeAction.NORMAL.equals(nodeHeartbeat.getNodeAction()));
|
||||||
|
|
||||||
nodeHeartbeat = nm2.nodeHeartbeat(
|
nodeHeartbeat = nm2.nodeHeartbeat(
|
||||||
new HashMap<ApplicationId, List<ContainerStatus>>(), true);
|
new HashMap<ApplicationId, List<ContainerStatus>>(), true, -100);
|
||||||
Assert.assertTrue(NodeAction.REBOOT.equals(nodeHeartbeat.getNodeAction()));
|
Assert.assertTrue(NodeAction.REBOOT.equals(nodeHeartbeat.getNodeAction()));
|
||||||
checkRebootedNMCount(rm, ++initialMetricCount);
|
checkRebootedNMCount(rm, ++initialMetricCount);
|
||||||
}
|
}
|
||||||
|
|
|
@ -302,7 +302,8 @@ public class TestApplicationLimits {
|
||||||
final ApplicationAttemptId appAttemptId_0_0 =
|
final ApplicationAttemptId appAttemptId_0_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0_0 =
|
SchedulerApp app_0_0 =
|
||||||
spy(new SchedulerApp(appAttemptId_0_0, user_0, queue, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_0_0, user_0, queue,
|
||||||
|
queue.getActiveUsersManager(), rmContext, null));
|
||||||
queue.submitApplication(app_0_0, user_0, A);
|
queue.submitApplication(app_0_0, user_0, A);
|
||||||
|
|
||||||
List<ResourceRequest> app_0_0_requests = new ArrayList<ResourceRequest>();
|
List<ResourceRequest> app_0_0_requests = new ArrayList<ResourceRequest>();
|
||||||
|
@ -320,7 +321,8 @@ public class TestApplicationLimits {
|
||||||
final ApplicationAttemptId appAttemptId_0_1 =
|
final ApplicationAttemptId appAttemptId_0_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_0_1 =
|
SchedulerApp app_0_1 =
|
||||||
spy(new SchedulerApp(appAttemptId_0_1, user_0, queue, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_0_1, user_0, queue,
|
||||||
|
queue.getActiveUsersManager(), rmContext, null));
|
||||||
queue.submitApplication(app_0_1, user_0, A);
|
queue.submitApplication(app_0_1, user_0, A);
|
||||||
|
|
||||||
List<ResourceRequest> app_0_1_requests = new ArrayList<ResourceRequest>();
|
List<ResourceRequest> app_0_1_requests = new ArrayList<ResourceRequest>();
|
||||||
|
@ -338,7 +340,8 @@ public class TestApplicationLimits {
|
||||||
final ApplicationAttemptId appAttemptId_1_0 =
|
final ApplicationAttemptId appAttemptId_1_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(2, 0);
|
TestUtils.getMockApplicationAttemptId(2, 0);
|
||||||
SchedulerApp app_1_0 =
|
SchedulerApp app_1_0 =
|
||||||
spy(new SchedulerApp(appAttemptId_1_0, user_1, queue, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_1_0, user_1, queue,
|
||||||
|
queue.getActiveUsersManager(), rmContext, null));
|
||||||
queue.submitApplication(app_1_0, user_1, A);
|
queue.submitApplication(app_1_0, user_1, A);
|
||||||
|
|
||||||
List<ResourceRequest> app_1_0_requests = new ArrayList<ResourceRequest>();
|
List<ResourceRequest> app_1_0_requests = new ArrayList<ResourceRequest>();
|
||||||
|
|
|
@ -18,8 +18,18 @@
|
||||||
|
|
||||||
package org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity;
|
package org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity;
|
||||||
|
|
||||||
import static org.junit.Assert.*;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.mockito.Mockito.*;
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.mockito.Matchers.any;
|
||||||
|
import static org.mockito.Matchers.eq;
|
||||||
|
import static org.mockito.Mockito.doAnswer;
|
||||||
|
import static org.mockito.Mockito.doNothing;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.never;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
|
@ -28,9 +38,6 @@ import java.util.HashMap;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
|
||||||
import org.apache.commons.logging.LogFactory;
|
|
||||||
|
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
import org.apache.hadoop.yarn.api.records.ApplicationAttemptId;
|
||||||
import org.apache.hadoop.yarn.api.records.Container;
|
import org.apache.hadoop.yarn.api.records.Container;
|
||||||
|
@ -48,19 +55,17 @@ import org.apache.hadoop.yarn.server.resourcemanager.resource.Resources;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerEventType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl;
|
import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl;
|
||||||
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.NodeType;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApp;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode;
|
import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode;
|
||||||
|
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
import org.junit.Before;
|
import org.junit.Before;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
import org.mockito.invocation.InvocationOnMock;
|
import org.mockito.invocation.InvocationOnMock;
|
||||||
import org.mockito.stubbing.Answer;
|
import org.mockito.stubbing.Answer;
|
||||||
|
|
||||||
public class TestLeafQueue {
|
public class TestLeafQueue {
|
||||||
private static final Log LOG = LogFactory.getLog(TestLeafQueue.class);
|
|
||||||
|
|
||||||
private final RecordFactory recordFactory =
|
private final RecordFactory recordFactory =
|
||||||
RecordFactoryProvider.getRecordFactory(null);
|
RecordFactoryProvider.getRecordFactory(null);
|
||||||
|
@ -136,7 +141,6 @@ public class TestLeafQueue {
|
||||||
final String Q_C1 = Q_C + "." + C1;
|
final String Q_C1 = Q_C + "." + C1;
|
||||||
conf.setCapacity(Q_C1, 100);
|
conf.setCapacity(Q_C1, 100);
|
||||||
|
|
||||||
LOG.info("Setup top-level queues a and b");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static LeafQueue stubLeafQueue(LeafQueue queue) {
|
static LeafQueue stubLeafQueue(LeafQueue queue) {
|
||||||
|
@ -217,13 +221,15 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_0, user_0, B);
|
a.submitApplication(app_0, user_0, B);
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_1 =
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_1 =
|
SchedulerApp app_1 =
|
||||||
new SchedulerApp(appAttemptId_1, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_1, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_1, user_0, B); // same user
|
a.submitApplication(app_1, user_0, B); // same user
|
||||||
|
|
||||||
|
|
||||||
|
@ -264,13 +270,15 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_1 =
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_1 =
|
SchedulerApp app_1 =
|
||||||
new SchedulerApp(appAttemptId_1, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_1, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_1, user_0, A); // same user
|
a.submitApplication(app_1, user_0, A); // same user
|
||||||
|
|
||||||
|
|
||||||
|
@ -371,6 +379,99 @@ public class TestLeafQueue {
|
||||||
assertEquals(1, a.getMetrics().getAvailableGB());
|
assertEquals(1, a.getMetrics().getAvailableGB());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testUserLimits() throws Exception {
|
||||||
|
// Mock the queue
|
||||||
|
LeafQueue a = stubLeafQueue((LeafQueue)queues.get(A));
|
||||||
|
//unset maxCapacity
|
||||||
|
a.setMaxCapacity(1.0f);
|
||||||
|
|
||||||
|
// Users
|
||||||
|
final String user_0 = "user_0";
|
||||||
|
final String user_1 = "user_1";
|
||||||
|
|
||||||
|
// Submit applications
|
||||||
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
|
SchedulerApp app_0 =
|
||||||
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
|
SchedulerApp app_1 =
|
||||||
|
new SchedulerApp(appAttemptId_1, user_0, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
|
a.submitApplication(app_1, user_0, A); // same user
|
||||||
|
|
||||||
|
final ApplicationAttemptId appAttemptId_2 =
|
||||||
|
TestUtils.getMockApplicationAttemptId(2, 0);
|
||||||
|
SchedulerApp app_2 =
|
||||||
|
new SchedulerApp(appAttemptId_2, user_1, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
|
a.submitApplication(app_2, user_1, A);
|
||||||
|
|
||||||
|
// Setup some nodes
|
||||||
|
String host_0 = "host_0";
|
||||||
|
SchedulerNode node_0 = TestUtils.getMockNode(host_0, DEFAULT_RACK, 0, 8*GB);
|
||||||
|
String host_1 = "host_1";
|
||||||
|
SchedulerNode node_1 = TestUtils.getMockNode(host_1, DEFAULT_RACK, 0, 8*GB);
|
||||||
|
|
||||||
|
final int numNodes = 2;
|
||||||
|
Resource clusterResource = Resources.createResource(numNodes * (8*GB));
|
||||||
|
when(csContext.getNumClusterNodes()).thenReturn(numNodes);
|
||||||
|
|
||||||
|
// Setup resource-requests
|
||||||
|
Priority priority = TestUtils.createMockPriority(1);
|
||||||
|
app_0.updateResourceRequests(Collections.singletonList(
|
||||||
|
TestUtils.createResourceRequest(RMNodeImpl.ANY, 2*GB, 1, priority,
|
||||||
|
recordFactory)));
|
||||||
|
|
||||||
|
app_1.updateResourceRequests(Collections.singletonList(
|
||||||
|
TestUtils.createResourceRequest(RMNodeImpl.ANY, 1*GB, 2, priority,
|
||||||
|
recordFactory)));
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Start testing...
|
||||||
|
*/
|
||||||
|
|
||||||
|
// Set user-limit
|
||||||
|
a.setUserLimit(50);
|
||||||
|
a.setUserLimitFactor(2);
|
||||||
|
|
||||||
|
// Now, only user_0 should be active since he is the only one with
|
||||||
|
// outstanding requests
|
||||||
|
assertEquals("There should only be 1 active user!",
|
||||||
|
1, a.getActiveUsersManager().getNumActiveUsers());
|
||||||
|
|
||||||
|
// This commented code is key to test 'activeUsers'.
|
||||||
|
// It should fail the test if uncommented since
|
||||||
|
// it would increase 'activeUsers' to 2 and stop user_2
|
||||||
|
// Pre MAPREDUCE-3732 this test should fail without this block too
|
||||||
|
// app_2.updateResourceRequests(Collections.singletonList(
|
||||||
|
// TestUtils.createResourceRequest(RMNodeImpl.ANY, 1*GB, 1, priority,
|
||||||
|
// recordFactory)));
|
||||||
|
|
||||||
|
// 1 container to user_0
|
||||||
|
a.assignContainers(clusterResource, node_0);
|
||||||
|
assertEquals(2*GB, a.getUsedResources().getMemory());
|
||||||
|
assertEquals(2*GB, app_0.getCurrentConsumption().getMemory());
|
||||||
|
assertEquals(0*GB, app_1.getCurrentConsumption().getMemory());
|
||||||
|
|
||||||
|
// Again one to user_0 since he hasn't exceeded user limit yet
|
||||||
|
a.assignContainers(clusterResource, node_0);
|
||||||
|
assertEquals(3*GB, a.getUsedResources().getMemory());
|
||||||
|
assertEquals(2*GB, app_0.getCurrentConsumption().getMemory());
|
||||||
|
assertEquals(1*GB, app_1.getCurrentConsumption().getMemory());
|
||||||
|
|
||||||
|
// One more to user_0 since he is the only active user
|
||||||
|
a.assignContainers(clusterResource, node_1);
|
||||||
|
assertEquals(4*GB, a.getUsedResources().getMemory());
|
||||||
|
assertEquals(2*GB, app_0.getCurrentConsumption().getMemory());
|
||||||
|
assertEquals(2*GB, app_1.getCurrentConsumption().getMemory());
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testSingleQueueWithMultipleUsers() throws Exception {
|
public void testSingleQueueWithMultipleUsers() throws Exception {
|
||||||
|
|
||||||
|
@ -388,15 +489,31 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_1 =
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_1 =
|
SchedulerApp app_1 =
|
||||||
new SchedulerApp(appAttemptId_1, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_1, user_0, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
a.submitApplication(app_1, user_0, A); // same user
|
a.submitApplication(app_1, user_0, A); // same user
|
||||||
|
|
||||||
|
final ApplicationAttemptId appAttemptId_2 =
|
||||||
|
TestUtils.getMockApplicationAttemptId(2, 0);
|
||||||
|
SchedulerApp app_2 =
|
||||||
|
new SchedulerApp(appAttemptId_2, user_1, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
|
a.submitApplication(app_2, user_1, A);
|
||||||
|
|
||||||
|
final ApplicationAttemptId appAttemptId_3 =
|
||||||
|
TestUtils.getMockApplicationAttemptId(3, 0);
|
||||||
|
SchedulerApp app_3 =
|
||||||
|
new SchedulerApp(appAttemptId_3, user_2, a,
|
||||||
|
a.getActiveUsersManager(), rmContext, null);
|
||||||
|
a.submitApplication(app_3, user_2, A);
|
||||||
|
|
||||||
// Setup some nodes
|
// Setup some nodes
|
||||||
String host_0 = "host_0";
|
String host_0 = "host_0";
|
||||||
SchedulerNode node_0 = TestUtils.getMockNode(host_0, DEFAULT_RACK, 0, 8*GB);
|
SchedulerNode node_0 = TestUtils.getMockNode(host_0, DEFAULT_RACK, 0, 8*GB);
|
||||||
|
@ -439,18 +556,7 @@ public class TestLeafQueue {
|
||||||
assertEquals(2*GB, app_0.getCurrentConsumption().getMemory());
|
assertEquals(2*GB, app_0.getCurrentConsumption().getMemory());
|
||||||
assertEquals(0*GB, app_1.getCurrentConsumption().getMemory());
|
assertEquals(0*GB, app_1.getCurrentConsumption().getMemory());
|
||||||
|
|
||||||
// Submit more apps
|
// Submit resource requests for other apps now to 'activate' them
|
||||||
final ApplicationAttemptId appAttemptId_2 =
|
|
||||||
TestUtils.getMockApplicationAttemptId(2, 0);
|
|
||||||
SchedulerApp app_2 =
|
|
||||||
new SchedulerApp(appAttemptId_2, user_1, a, rmContext, null);
|
|
||||||
a.submitApplication(app_2, user_1, A);
|
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_3 =
|
|
||||||
TestUtils.getMockApplicationAttemptId(3, 0);
|
|
||||||
SchedulerApp app_3 =
|
|
||||||
new SchedulerApp(appAttemptId_3, user_2, a, rmContext, null);
|
|
||||||
a.submitApplication(app_3, user_2, A);
|
|
||||||
|
|
||||||
app_2.updateResourceRequests(Collections.singletonList(
|
app_2.updateResourceRequests(Collections.singletonList(
|
||||||
TestUtils.createResourceRequest(RMNodeImpl.ANY, 3*GB, 1, priority,
|
TestUtils.createResourceRequest(RMNodeImpl.ANY, 3*GB, 1, priority,
|
||||||
|
@ -558,13 +664,15 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_1 =
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_1 =
|
SchedulerApp app_1 =
|
||||||
new SchedulerApp(appAttemptId_1, user_1, a, rmContext, null);
|
new SchedulerApp(appAttemptId_1, user_1, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_1, user_1, A);
|
a.submitApplication(app_1, user_1, A);
|
||||||
|
|
||||||
// Setup some nodes
|
// Setup some nodes
|
||||||
|
@ -657,13 +765,15 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null);
|
new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
final ApplicationAttemptId appAttemptId_1 =
|
final ApplicationAttemptId appAttemptId_1 =
|
||||||
TestUtils.getMockApplicationAttemptId(1, 0);
|
TestUtils.getMockApplicationAttemptId(1, 0);
|
||||||
SchedulerApp app_1 =
|
SchedulerApp app_1 =
|
||||||
new SchedulerApp(appAttemptId_1, user_1, a, rmContext, null);
|
new SchedulerApp(appAttemptId_1, user_1, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null);
|
||||||
a.submitApplication(app_1, user_1, A);
|
a.submitApplication(app_1, user_1, A);
|
||||||
|
|
||||||
// Setup some nodes
|
// Setup some nodes
|
||||||
|
@ -770,7 +880,8 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
spy(new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null));
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
// Setup some nodes and racks
|
// Setup some nodes and racks
|
||||||
|
@ -899,7 +1010,8 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
spy(new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null));
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
// Setup some nodes and racks
|
// Setup some nodes and racks
|
||||||
|
@ -1028,7 +1140,8 @@ public class TestLeafQueue {
|
||||||
final ApplicationAttemptId appAttemptId_0 =
|
final ApplicationAttemptId appAttemptId_0 =
|
||||||
TestUtils.getMockApplicationAttemptId(0, 0);
|
TestUtils.getMockApplicationAttemptId(0, 0);
|
||||||
SchedulerApp app_0 =
|
SchedulerApp app_0 =
|
||||||
spy(new SchedulerApp(appAttemptId_0, user_0, a, rmContext, null));
|
spy(new SchedulerApp(appAttemptId_0, user_0, a,
|
||||||
|
mock(ActiveUsersManager.class), rmContext, null));
|
||||||
a.submitApplication(app_0, user_0, A);
|
a.submitApplication(app_0, user_0, A);
|
||||||
|
|
||||||
// Setup some nodes and racks
|
// Setup some nodes and racks
|
||||||
|
|
|
@ -24,6 +24,7 @@ import org.apache.hadoop.yarn.server.resourcemanager.RMContext;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.ResourceManager;
|
import org.apache.hadoop.yarn.server.resourcemanager.ResourceManager;
|
||||||
import org.apache.hadoop.yarn.server.resourcemanager.webapp.NodesPage.NodesBlock;
|
import org.apache.hadoop.yarn.server.resourcemanager.webapp.NodesPage.NodesBlock;
|
||||||
import org.apache.hadoop.yarn.webapp.test.WebAppTests;
|
import org.apache.hadoop.yarn.webapp.test.WebAppTests;
|
||||||
|
import org.junit.Before;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.mockito.Mockito;
|
import org.mockito.Mockito;
|
||||||
|
|
||||||
|
@ -37,38 +38,64 @@ import com.google.inject.Module;
|
||||||
*/
|
*/
|
||||||
public class TestNodesPage {
|
public class TestNodesPage {
|
||||||
|
|
||||||
@Test
|
final int numberOfRacks = 2;
|
||||||
public void testNodesBlockRender() throws Exception {
|
final int numberOfNodesPerRack = 2;
|
||||||
final int numberOfRacks = 2;
|
// Number of Actual Table Headers for NodesPage.NodesBlock might change in
|
||||||
final int numberOfNodesPerRack = 2;
|
// future. In that case this value should be adjusted to the new value.
|
||||||
// Number of Actual Table Headers for NodesPage.NodesBlock might change in
|
final int numberOfThInMetricsTable = 10;
|
||||||
// future. In that case this value should be adjusted to the new value.
|
final int numberOfActualTableHeaders = 10;
|
||||||
final int numberOfThInMetricsTable = 10;
|
|
||||||
final int numberOfActualTableHeaders = 10;
|
|
||||||
|
|
||||||
Injector injector = WebAppTests.createMockInjector(RMContext.class,
|
private Injector injector;
|
||||||
TestRMWebApp.mockRMContext(3, numberOfRacks, numberOfNodesPerRack, 8*TestRMWebApp.GiB),
|
|
||||||
new Module() {
|
@Before
|
||||||
|
public void setUp() throws Exception {
|
||||||
|
injector = WebAppTests.createMockInjector(RMContext.class, TestRMWebApp
|
||||||
|
.mockRMContext(3, numberOfRacks, numberOfNodesPerRack,
|
||||||
|
8 * TestRMWebApp.GiB), new Module() {
|
||||||
@Override
|
@Override
|
||||||
public void configure(Binder binder) {
|
public void configure(Binder binder) {
|
||||||
try {
|
try {
|
||||||
binder.bind(ResourceManager.class).toInstance(TestRMWebApp.mockRm(3,
|
binder.bind(ResourceManager.class).toInstance(
|
||||||
numberOfRacks, numberOfNodesPerRack, 8*TestRMWebApp.GiB));
|
TestRMWebApp.mockRm(3, numberOfRacks, numberOfNodesPerRack,
|
||||||
|
8 * TestRMWebApp.GiB));
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new IllegalStateException(e);
|
throw new IllegalStateException(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testNodesBlockRender() throws Exception {
|
||||||
injector.getInstance(NodesBlock.class).render();
|
injector.getInstance(NodesBlock.class).render();
|
||||||
PrintWriter writer = injector.getInstance(PrintWriter.class);
|
PrintWriter writer = injector.getInstance(PrintWriter.class);
|
||||||
WebAppTests.flushOutput(injector);
|
WebAppTests.flushOutput(injector);
|
||||||
|
|
||||||
Mockito.verify(writer, Mockito.times(numberOfActualTableHeaders +
|
Mockito.verify(writer,
|
||||||
numberOfThInMetricsTable)).print(
|
Mockito.times(numberOfActualTableHeaders + numberOfThInMetricsTable))
|
||||||
"<th");
|
.print("<th");
|
||||||
Mockito.verify(
|
Mockito.verify(
|
||||||
writer,
|
writer,
|
||||||
Mockito.times(numberOfRacks * numberOfNodesPerRack
|
Mockito.times(numberOfRacks * numberOfNodesPerRack
|
||||||
* numberOfActualTableHeaders + numberOfThInMetricsTable)).print("<td");
|
* numberOfActualTableHeaders + numberOfThInMetricsTable)).print(
|
||||||
|
"<td");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testNodesBlockRenderForLostNodes() {
|
||||||
|
NodesBlock nodesBlock = injector.getInstance(NodesBlock.class);
|
||||||
|
nodesBlock.set("node.state", "lost");
|
||||||
|
nodesBlock.render();
|
||||||
|
PrintWriter writer = injector.getInstance(PrintWriter.class);
|
||||||
|
WebAppTests.flushOutput(injector);
|
||||||
|
|
||||||
|
Mockito.verify(writer,
|
||||||
|
Mockito.times(numberOfActualTableHeaders + numberOfThInMetricsTable))
|
||||||
|
.print("<th");
|
||||||
|
Mockito.verify(
|
||||||
|
writer,
|
||||||
|
Mockito.times(numberOfRacks * numberOfNodesPerRack
|
||||||
|
* numberOfActualTableHeaders + numberOfThInMetricsTable)).print(
|
||||||
|
"<td");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -120,12 +120,23 @@ public class TestRMWebApp {
|
||||||
for (RMNode node : nodes) {
|
for (RMNode node : nodes) {
|
||||||
nodesMap.put(node.getNodeID(), node);
|
nodesMap.put(node.getNodeID(), node);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
final List<RMNode> lostNodes = MockNodes.lostNodes(racks, numNodes,
|
||||||
|
newResource(mbsPerNode));
|
||||||
|
final ConcurrentMap<String, RMNode> lostNodesMap = Maps.newConcurrentMap();
|
||||||
|
for (RMNode node : lostNodes) {
|
||||||
|
lostNodesMap.put(node.getHostName(), node);
|
||||||
|
}
|
||||||
return new RMContextImpl(new MemStore(), null, null, null, null) {
|
return new RMContextImpl(new MemStore(), null, null, null, null) {
|
||||||
@Override
|
@Override
|
||||||
public ConcurrentMap<ApplicationId, RMApp> getRMApps() {
|
public ConcurrentMap<ApplicationId, RMApp> getRMApps() {
|
||||||
return applicationsMaps;
|
return applicationsMaps;
|
||||||
}
|
}
|
||||||
@Override
|
@Override
|
||||||
|
public ConcurrentMap<String, RMNode> getInactiveRMNodes() {
|
||||||
|
return lostNodesMap;
|
||||||
|
}
|
||||||
|
@Override
|
||||||
public ConcurrentMap<NodeId, RMNode> getRMNodes() {
|
public ConcurrentMap<NodeId, RMNode> getRMNodes() {
|
||||||
return nodesMap;
|
return nodesMap;
|
||||||
}
|
}
|
||||||
|
|
|
@ -370,7 +370,8 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
WebServicesTestUtils.getXmlInt(element, "lostNodes"),
|
WebServicesTestUtils.getXmlInt(element, "lostNodes"),
|
||||||
WebServicesTestUtils.getXmlInt(element, "unhealthyNodes"),
|
WebServicesTestUtils.getXmlInt(element, "unhealthyNodes"),
|
||||||
WebServicesTestUtils.getXmlInt(element, "decommissionedNodes"),
|
WebServicesTestUtils.getXmlInt(element, "decommissionedNodes"),
|
||||||
WebServicesTestUtils.getXmlInt(element, "rebootedNodes"));
|
WebServicesTestUtils.getXmlInt(element, "rebootedNodes"),
|
||||||
|
WebServicesTestUtils.getXmlInt(element, "activeNodes"));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -378,7 +379,7 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
Exception {
|
Exception {
|
||||||
assertEquals("incorrect number of elements", 1, json.length());
|
assertEquals("incorrect number of elements", 1, json.length());
|
||||||
JSONObject clusterinfo = json.getJSONObject("clusterMetrics");
|
JSONObject clusterinfo = json.getJSONObject("clusterMetrics");
|
||||||
assertEquals("incorrect number of elements", 11, clusterinfo.length());
|
assertEquals("incorrect number of elements", 12, clusterinfo.length());
|
||||||
verifyClusterMetrics(clusterinfo.getInt("appsSubmitted"),
|
verifyClusterMetrics(clusterinfo.getInt("appsSubmitted"),
|
||||||
clusterinfo.getInt("reservedMB"), clusterinfo.getInt("availableMB"),
|
clusterinfo.getInt("reservedMB"), clusterinfo.getInt("availableMB"),
|
||||||
clusterinfo.getInt("allocatedMB"),
|
clusterinfo.getInt("allocatedMB"),
|
||||||
|
@ -386,13 +387,13 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
clusterinfo.getInt("totalMB"), clusterinfo.getInt("totalNodes"),
|
clusterinfo.getInt("totalMB"), clusterinfo.getInt("totalNodes"),
|
||||||
clusterinfo.getInt("lostNodes"), clusterinfo.getInt("unhealthyNodes"),
|
clusterinfo.getInt("lostNodes"), clusterinfo.getInt("unhealthyNodes"),
|
||||||
clusterinfo.getInt("decommissionedNodes"),
|
clusterinfo.getInt("decommissionedNodes"),
|
||||||
clusterinfo.getInt("rebootedNodes"));
|
clusterinfo.getInt("rebootedNodes"),clusterinfo.getInt("activeNodes"));
|
||||||
}
|
}
|
||||||
|
|
||||||
public void verifyClusterMetrics(int sub, int reservedMB, int availableMB,
|
public void verifyClusterMetrics(int sub, int reservedMB, int availableMB,
|
||||||
int allocMB, int containersAlloc, int totalMB, int totalNodes,
|
int allocMB, int containersAlloc, int totalMB, int totalNodes,
|
||||||
int lostNodes, int unhealthyNodes, int decommissionedNodes,
|
int lostNodes, int unhealthyNodes, int decommissionedNodes,
|
||||||
int rebootedNodes) throws JSONException, Exception {
|
int rebootedNodes, int activeNodes) throws JSONException, Exception {
|
||||||
|
|
||||||
ResourceScheduler rs = rm.getResourceScheduler();
|
ResourceScheduler rs = rm.getResourceScheduler();
|
||||||
QueueMetrics metrics = rs.getRootQueueMetrics();
|
QueueMetrics metrics = rs.getRootQueueMetrics();
|
||||||
|
@ -412,8 +413,11 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
* MB_IN_GB, allocMB);
|
* MB_IN_GB, allocMB);
|
||||||
assertEquals("containersAllocated doesn't match", 0, containersAlloc);
|
assertEquals("containersAllocated doesn't match", 0, containersAlloc);
|
||||||
assertEquals("totalMB doesn't match", totalMBExpect, totalMB);
|
assertEquals("totalMB doesn't match", totalMBExpect, totalMB);
|
||||||
assertEquals("totalNodes doesn't match", clusterMetrics.getNumNMs(),
|
assertEquals(
|
||||||
totalNodes);
|
"totalNodes doesn't match",
|
||||||
|
clusterMetrics.getNumActiveNMs() + clusterMetrics.getNumLostNMs()
|
||||||
|
+ clusterMetrics.getNumDecommisionedNMs()
|
||||||
|
+ clusterMetrics.getNumRebootedNMs(), totalNodes);
|
||||||
assertEquals("lostNodes doesn't match", clusterMetrics.getNumLostNMs(),
|
assertEquals("lostNodes doesn't match", clusterMetrics.getNumLostNMs(),
|
||||||
lostNodes);
|
lostNodes);
|
||||||
assertEquals("unhealthyNodes doesn't match",
|
assertEquals("unhealthyNodes doesn't match",
|
||||||
|
@ -422,6 +426,8 @@ public class TestRMWebServices extends JerseyTest {
|
||||||
clusterMetrics.getNumDecommisionedNMs(), decommissionedNodes);
|
clusterMetrics.getNumDecommisionedNMs(), decommissionedNodes);
|
||||||
assertEquals("rebootedNodes doesn't match",
|
assertEquals("rebootedNodes doesn't match",
|
||||||
clusterMetrics.getNumRebootedNMs(), rebootedNodes);
|
clusterMetrics.getNumRebootedNMs(), rebootedNodes);
|
||||||
|
assertEquals("activeNodes doesn't match", clusterMetrics.getNumActiveNMs(),
|
||||||
|
activeNodes);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|
|
@ -203,6 +203,69 @@ public class TestRMWebServicesNodes extends JerseyTest {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testNodesQueryStateLost() throws JSONException, Exception {
|
||||||
|
WebResource r = resource();
|
||||||
|
MockNM nm1 = rm.registerNode("h1:1234", 5120);
|
||||||
|
MockNM nm2 = rm.registerNode("h2:1234", 5120);
|
||||||
|
rm.sendNodeStarted(nm1);
|
||||||
|
rm.sendNodeStarted(nm2);
|
||||||
|
rm.NMwaitForState(nm1.getNodeId(), RMNodeState.RUNNING);
|
||||||
|
rm.NMwaitForState(nm2.getNodeId(), RMNodeState.RUNNING);
|
||||||
|
rm.sendNodeLost(nm1);
|
||||||
|
rm.sendNodeLost(nm2);
|
||||||
|
|
||||||
|
ClientResponse response = r.path("ws").path("v1").path("cluster")
|
||||||
|
.path("nodes").queryParam("state", RMNodeState.LOST.toString())
|
||||||
|
.accept(MediaType.APPLICATION_JSON).get(ClientResponse.class);
|
||||||
|
|
||||||
|
assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType());
|
||||||
|
JSONObject json = response.getEntity(JSONObject.class);
|
||||||
|
JSONObject nodes = json.getJSONObject("nodes");
|
||||||
|
assertEquals("incorrect number of elements", 1, nodes.length());
|
||||||
|
JSONArray nodeArray = nodes.getJSONArray("node");
|
||||||
|
assertEquals("incorrect number of elements", 2, nodeArray.length());
|
||||||
|
for (int i = 0; i < nodeArray.length(); ++i) {
|
||||||
|
JSONObject info = nodeArray.getJSONObject(i);
|
||||||
|
String host = info.get("id").toString().split(":")[0];
|
||||||
|
RMNode rmNode = rm.getRMContext().getInactiveRMNodes().get(host);
|
||||||
|
WebServicesTestUtils.checkStringMatch("nodeHTTPAddress", "",
|
||||||
|
info.getString("nodeHTTPAddress"));
|
||||||
|
WebServicesTestUtils.checkStringMatch("state", rmNode.getState()
|
||||||
|
.toString(), info.getString("state"));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSingleNodeQueryStateLost() throws JSONException, Exception {
|
||||||
|
WebResource r = resource();
|
||||||
|
MockNM nm1 = rm.registerNode("h1:1234", 5120);
|
||||||
|
MockNM nm2 = rm.registerNode("h2:1234", 5120);
|
||||||
|
rm.sendNodeStarted(nm1);
|
||||||
|
rm.sendNodeStarted(nm2);
|
||||||
|
rm.NMwaitForState(nm1.getNodeId(), RMNodeState.RUNNING);
|
||||||
|
rm.NMwaitForState(nm2.getNodeId(), RMNodeState.RUNNING);
|
||||||
|
rm.sendNodeLost(nm1);
|
||||||
|
rm.sendNodeLost(nm2);
|
||||||
|
|
||||||
|
ClientResponse response = r.path("ws").path("v1").path("cluster")
|
||||||
|
.path("nodes").path("h2:1234").accept(MediaType.APPLICATION_JSON)
|
||||||
|
.get(ClientResponse.class);
|
||||||
|
|
||||||
|
assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType());
|
||||||
|
JSONObject json = response.getEntity(JSONObject.class);
|
||||||
|
JSONObject info = json.getJSONObject("node");
|
||||||
|
String id = info.get("id").toString();
|
||||||
|
|
||||||
|
assertEquals("Incorrect Node Information.", "h2:1234", id);
|
||||||
|
|
||||||
|
RMNode rmNode = rm.getRMContext().getInactiveRMNodes().get("h2");
|
||||||
|
WebServicesTestUtils.checkStringMatch("nodeHTTPAddress", "",
|
||||||
|
info.getString("nodeHTTPAddress"));
|
||||||
|
WebServicesTestUtils.checkStringMatch("state",
|
||||||
|
rmNode.getState().toString(), info.getString("state"));
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testNodesQueryHealthy() throws JSONException, Exception {
|
public void testNodesQueryHealthy() throws JSONException, Exception {
|
||||||
WebResource r = resource();
|
WebResource r = resource();
|
||||||
|
|
|
@ -149,10 +149,15 @@ class ExecutionSummarizer implements StatListener<JobStats> {
|
||||||
throws IOException {
|
throws IOException {
|
||||||
numJobsInInputTrace = factory.numJobsInTrace;
|
numJobsInInputTrace = factory.numJobsInTrace;
|
||||||
endTime = System.currentTimeMillis();
|
endTime = System.currentTimeMillis();
|
||||||
Path inputTracePath = new Path(inputPath);
|
if ("-".equals(inputPath)) {
|
||||||
FileSystem fs = inputTracePath.getFileSystem(conf);
|
inputTraceLocation = Summarizer.NA;
|
||||||
inputTraceLocation = fs.makeQualified(inputTracePath).toString();
|
inputTraceSignature = Summarizer.NA;
|
||||||
inputTraceSignature = getTraceSignature(inputTraceLocation);
|
} else {
|
||||||
|
Path inputTracePath = new Path(inputPath);
|
||||||
|
FileSystem fs = inputTracePath.getFileSystem(conf);
|
||||||
|
inputTraceLocation = fs.makeQualified(inputTracePath).toString();
|
||||||
|
inputTraceSignature = getTraceSignature(inputPath);
|
||||||
|
}
|
||||||
jobSubmissionPolicy = Gridmix.getJobSubmissionPolicy(conf).name();
|
jobSubmissionPolicy = Gridmix.getJobSubmissionPolicy(conf).name();
|
||||||
resolver = userResolver.getClass().getName();
|
resolver = userResolver.getClass().getName();
|
||||||
if (dataSize > 0) {
|
if (dataSize > 0) {
|
||||||
|
|
|
@ -314,9 +314,13 @@ public class Gridmix extends Configured implements Tool {
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
// print the run summary
|
// print the gridmix summary if the run was successful
|
||||||
System.out.print("\n\n");
|
if (val == 0) {
|
||||||
System.out.println(summarizer.toString());
|
// print the run summary
|
||||||
|
System.out.print("\n\n");
|
||||||
|
System.out.println(summarizer.toString());
|
||||||
|
}
|
||||||
|
|
||||||
return val;
|
return val;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -101,13 +101,15 @@ public class Statistics implements Component<Job> {
|
||||||
}
|
}
|
||||||
|
|
||||||
int maps = 0;
|
int maps = 0;
|
||||||
|
int reds = 0;
|
||||||
if (jobdesc == null) {
|
if (jobdesc == null) {
|
||||||
throw new IllegalArgumentException(
|
throw new IllegalArgumentException(
|
||||||
" JobStory not available for job " + job.getJobName());
|
" JobStory not available for job " + job.getJobName());
|
||||||
} else {
|
} else {
|
||||||
maps = jobdesc.getNumberMaps();
|
maps = jobdesc.getNumberMaps();
|
||||||
|
reds = jobdesc.getNumberReduces();
|
||||||
}
|
}
|
||||||
JobStats stats = new JobStats(maps,job);
|
JobStats stats = new JobStats(maps, reds, job);
|
||||||
jobMaps.put(seq,stats);
|
jobMaps.put(seq,stats);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -258,15 +260,20 @@ public class Statistics implements Component<Job> {
|
||||||
*/
|
*/
|
||||||
static class JobStats {
|
static class JobStats {
|
||||||
private int noOfMaps;
|
private int noOfMaps;
|
||||||
|
private int noOfReds;
|
||||||
private Job job;
|
private Job job;
|
||||||
|
|
||||||
public JobStats(int noOfMaps,Job job){
|
public JobStats(int noOfMaps,int numOfReds, Job job){
|
||||||
this.job = job;
|
this.job = job;
|
||||||
this.noOfMaps = noOfMaps;
|
this.noOfMaps = noOfMaps;
|
||||||
|
this.noOfReds = numOfReds;
|
||||||
}
|
}
|
||||||
public int getNoOfMaps() {
|
public int getNoOfMaps() {
|
||||||
return noOfMaps;
|
return noOfMaps;
|
||||||
}
|
}
|
||||||
|
public int getNoOfReds() {
|
||||||
|
return noOfReds;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns the job ,
|
* Returns the job ,
|
||||||
|
|
|
@ -31,13 +31,12 @@ import org.apache.hadoop.tools.rumen.JobStoryProducer;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.concurrent.CountDownLatch;
|
import java.util.concurrent.CountDownLatch;
|
||||||
import java.util.concurrent.locks.Condition;
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
|
||||||
public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
public static final Log LOG = LogFactory.getLog(StressJobFactory.class);
|
public static final Log LOG = LogFactory.getLog(StressJobFactory.class);
|
||||||
|
|
||||||
private final LoadStatus loadStatus = new LoadStatus();
|
private final LoadStatus loadStatus = new LoadStatus();
|
||||||
private final Condition condUnderloaded = this.lock.newCondition();
|
|
||||||
/**
|
/**
|
||||||
* The minimum ratio between pending+running map tasks (aka. incomplete map
|
* The minimum ratio between pending+running map tasks (aka. incomplete map
|
||||||
* tasks) and cluster map slot capacity for us to consider the cluster is
|
* tasks) and cluster map slot capacity for us to consider the cluster is
|
||||||
|
@ -150,23 +149,32 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
}
|
}
|
||||||
LOG.info("START STRESS @ " + System.currentTimeMillis());
|
LOG.info("START STRESS @ " + System.currentTimeMillis());
|
||||||
while (!Thread.currentThread().isInterrupted()) {
|
while (!Thread.currentThread().isInterrupted()) {
|
||||||
lock.lock();
|
|
||||||
try {
|
try {
|
||||||
while (loadStatus.overloaded()) {
|
while (loadStatus.overloaded()) {
|
||||||
//Wait while JT is overloaded.
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Cluster overloaded in run! Sleeping...");
|
||||||
|
}
|
||||||
|
// sleep
|
||||||
try {
|
try {
|
||||||
condUnderloaded.await();
|
Thread.sleep(1000);
|
||||||
} catch (InterruptedException ie) {
|
} catch (InterruptedException ie) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
while (!loadStatus.overloaded()) {
|
while (!loadStatus.overloaded()) {
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Cluster underloaded in run! Stressing...");
|
||||||
|
}
|
||||||
try {
|
try {
|
||||||
|
//TODO This in-line read can block submission for large jobs.
|
||||||
final JobStory job = getNextJobFiltered();
|
final JobStory job = getNextJobFiltered();
|
||||||
if (null == job) {
|
if (null == job) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
if (LOG.isDebugEnabled()) {
|
||||||
|
LOG.debug("Job Selected: " + job.getJobID());
|
||||||
|
}
|
||||||
submitter.add(
|
submitter.add(
|
||||||
jobCreator.createGridmixJob(
|
jobCreator.createGridmixJob(
|
||||||
conf, 0L, job, scratch,
|
conf, 0L, job, scratch,
|
||||||
|
@ -175,14 +183,20 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
sequence.getAndIncrement()));
|
sequence.getAndIncrement()));
|
||||||
// TODO: We need to take care of scenario when one map/reduce
|
// TODO: We need to take care of scenario when one map/reduce
|
||||||
// takes more than 1 slot.
|
// takes more than 1 slot.
|
||||||
loadStatus.mapSlotsBackfill -=
|
|
||||||
calcEffectiveIncompleteMapTasks(
|
// Lock the loadjob as we are making updates
|
||||||
loadStatus.mapSlotCapacity, job.getNumberMaps(), 0.0f);
|
int incompleteMapTasks = (int) calcEffectiveIncompleteMapTasks(
|
||||||
loadStatus.reduceSlotsBackfill -=
|
loadStatus.getMapCapacity(),
|
||||||
calcEffectiveIncompleteReduceTasks(
|
job.getNumberMaps(), 0.0f);
|
||||||
loadStatus.reduceSlotCapacity, job.getNumberReduces(),
|
loadStatus.decrementMapLoad(incompleteMapTasks);
|
||||||
0.0f);
|
|
||||||
--loadStatus.numJobsBackfill;
|
int incompleteReduceTasks =
|
||||||
|
(int) calcEffectiveIncompleteReduceTasks(
|
||||||
|
loadStatus.getReduceCapacity(),
|
||||||
|
job.getNumberReduces(), 0.0f);
|
||||||
|
loadStatus.decrementReduceLoad(incompleteReduceTasks);
|
||||||
|
|
||||||
|
loadStatus.decrementJobLoad(1);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
LOG.error("Error while submitting the job ", e);
|
LOG.error("Error while submitting the job ", e);
|
||||||
error = e;
|
error = e;
|
||||||
|
@ -191,7 +205,7 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
|
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
lock.unlock();
|
// do nothing
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} catch (InterruptedException e) {
|
} catch (InterruptedException e) {
|
||||||
|
@ -210,19 +224,11 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
public void update(Statistics.ClusterStats item) {
|
public void update(Statistics.ClusterStats item) {
|
||||||
lock.lock();
|
ClusterStatus clusterMetrics = item.getStatus();
|
||||||
try {
|
try {
|
||||||
ClusterStatus clusterMetrics = item.getStatus();
|
checkLoadAndGetSlotsToBackfill(item, clusterMetrics);
|
||||||
try {
|
} catch (Exception e) {
|
||||||
checkLoadAndGetSlotsToBackfill(item,clusterMetrics);
|
LOG.error("Couldn't get the new Status",e);
|
||||||
} catch (Exception e) {
|
|
||||||
LOG.error("Couldn't get the new Status",e);
|
|
||||||
}
|
|
||||||
if (!loadStatus.overloaded()) {
|
|
||||||
condUnderloaded.signalAll();
|
|
||||||
}
|
|
||||||
} finally {
|
|
||||||
lock.unlock();
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -254,18 +260,25 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
*/
|
*/
|
||||||
private void checkLoadAndGetSlotsToBackfill(
|
private void checkLoadAndGetSlotsToBackfill(
|
||||||
ClusterStats stats, ClusterStatus clusterStatus) throws IOException, InterruptedException {
|
ClusterStats stats, ClusterStatus clusterStatus) throws IOException, InterruptedException {
|
||||||
loadStatus.mapSlotCapacity = clusterStatus.getMaxMapTasks();
|
|
||||||
loadStatus.reduceSlotCapacity = clusterStatus.getMaxReduceTasks();
|
|
||||||
|
|
||||||
|
// update the max cluster capacity incase its updated
|
||||||
|
int mapCapacity = clusterStatus.getMaxMapTasks();
|
||||||
|
loadStatus.updateMapCapacity(mapCapacity);
|
||||||
|
|
||||||
loadStatus.numJobsBackfill =
|
int reduceCapacity = clusterStatus.getMaxReduceTasks();
|
||||||
(int) (maxJobTrackerRatio * clusterStatus.getTaskTrackers())
|
|
||||||
- stats.getNumRunningJob();
|
loadStatus.updateReduceCapacity(reduceCapacity);
|
||||||
if (loadStatus.numJobsBackfill <= 0) {
|
|
||||||
|
int numTrackers = clusterStatus.getTaskTrackers();
|
||||||
|
|
||||||
|
int jobLoad =
|
||||||
|
(int) (maxJobTrackerRatio * numTrackers) - stats.getNumRunningJob();
|
||||||
|
loadStatus.updateJobLoad(jobLoad);
|
||||||
|
if (loadStatus.getJobLoad() <= 0) {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug(System.currentTimeMillis() + " Overloaded is "
|
LOG.debug(System.currentTimeMillis() + " [JobLoad] Overloaded is "
|
||||||
+ Boolean.TRUE.toString() + " NumJobsBackfill is "
|
+ Boolean.TRUE.toString() + " NumJobsBackfill is "
|
||||||
+ loadStatus.numJobsBackfill);
|
+ loadStatus.getJobLoad());
|
||||||
}
|
}
|
||||||
return; // stop calculation because we know it is overloaded.
|
return; // stop calculation because we know it is overloaded.
|
||||||
}
|
}
|
||||||
|
@ -275,56 +288,84 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
float mapProgress = job.getJob().mapProgress();
|
float mapProgress = job.getJob().mapProgress();
|
||||||
int noOfMaps = job.getNoOfMaps();
|
int noOfMaps = job.getNoOfMaps();
|
||||||
incompleteMapTasks +=
|
incompleteMapTasks +=
|
||||||
calcEffectiveIncompleteMapTasks(
|
calcEffectiveIncompleteMapTasks(mapCapacity, noOfMaps, mapProgress);
|
||||||
clusterStatus.getMaxMapTasks(), noOfMaps, mapProgress);
|
|
||||||
}
|
}
|
||||||
loadStatus.mapSlotsBackfill =
|
|
||||||
(int) ((overloadMapTaskMapSlotRatio * clusterStatus.getMaxMapTasks())
|
int mapSlotsBackFill =
|
||||||
- incompleteMapTasks);
|
(int) ((overloadMapTaskMapSlotRatio * mapCapacity) - incompleteMapTasks);
|
||||||
if (loadStatus.mapSlotsBackfill <= 0) {
|
loadStatus.updateMapLoad(mapSlotsBackFill);
|
||||||
|
|
||||||
|
if (loadStatus.getMapLoad() <= 0) {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug(System.currentTimeMillis() + " Overloaded is "
|
LOG.debug(System.currentTimeMillis() + " [MAP-LOAD] Overloaded is "
|
||||||
+ Boolean.TRUE.toString() + " MapSlotsBackfill is "
|
+ Boolean.TRUE.toString() + " MapSlotsBackfill is "
|
||||||
+ loadStatus.mapSlotsBackfill);
|
+ loadStatus.getMapLoad());
|
||||||
}
|
}
|
||||||
return; // stop calculation because we know it is overloaded.
|
return; // stop calculation because we know it is overloaded.
|
||||||
}
|
}
|
||||||
|
|
||||||
float incompleteReduceTasks = 0; // include pending & running reduce tasks.
|
float incompleteReduceTasks = 0; // include pending & running reduce tasks.
|
||||||
for (JobStats job : ClusterStats.getRunningJobStats()) {
|
for (JobStats job : ClusterStats.getRunningJobStats()) {
|
||||||
int noOfReduces = job.getJob().getNumReduceTasks();
|
// Cached the num-reds value in JobStats
|
||||||
|
int noOfReduces = job.getNoOfReds();
|
||||||
if (noOfReduces > 0) {
|
if (noOfReduces > 0) {
|
||||||
float reduceProgress = job.getJob().reduceProgress();
|
float reduceProgress = job.getJob().reduceProgress();
|
||||||
incompleteReduceTasks +=
|
incompleteReduceTasks +=
|
||||||
calcEffectiveIncompleteReduceTasks(
|
calcEffectiveIncompleteReduceTasks(reduceCapacity, noOfReduces,
|
||||||
clusterStatus.getMaxReduceTasks(), noOfReduces, reduceProgress);
|
reduceProgress);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
loadStatus.reduceSlotsBackfill =
|
|
||||||
(int) ((overloadReduceTaskReduceSlotRatio * clusterStatus.getMaxReduceTasks())
|
int reduceSlotsBackFill =
|
||||||
|
(int)((overloadReduceTaskReduceSlotRatio * reduceCapacity)
|
||||||
- incompleteReduceTasks);
|
- incompleteReduceTasks);
|
||||||
if (loadStatus.reduceSlotsBackfill <= 0) {
|
loadStatus.updateReduceLoad(reduceSlotsBackFill);
|
||||||
|
if (loadStatus.getReduceLoad() <= 0) {
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug(System.currentTimeMillis() + " Overloaded is "
|
LOG.debug(System.currentTimeMillis() + " [REDUCE-LOAD] Overloaded is "
|
||||||
+ Boolean.TRUE.toString() + " ReduceSlotsBackfill is "
|
+ Boolean.TRUE.toString() + " ReduceSlotsBackfill is "
|
||||||
+ loadStatus.reduceSlotsBackfill);
|
+ loadStatus.getReduceLoad());
|
||||||
}
|
}
|
||||||
return; // stop calculation because we know it is overloaded.
|
return; // stop calculation because we know it is overloaded.
|
||||||
}
|
}
|
||||||
|
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug(System.currentTimeMillis() + " Overloaded is "
|
LOG.debug(System.currentTimeMillis() + " [OVERALL] Overloaded is "
|
||||||
+ Boolean.FALSE.toString() + "Current load Status is "
|
+ Boolean.FALSE.toString() + "Current load Status is "
|
||||||
+ loadStatus);
|
+ loadStatus);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static class LoadStatus {
|
static class LoadStatus {
|
||||||
int mapSlotsBackfill;
|
/**
|
||||||
int mapSlotCapacity;
|
* Additional number of map slots that can be requested before
|
||||||
int reduceSlotsBackfill;
|
* declaring (by Gridmix STRESS mode) the cluster as overloaded.
|
||||||
int reduceSlotCapacity;
|
*/
|
||||||
int numJobsBackfill;
|
private volatile int mapSlotsBackfill;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Determines the total map slot capacity of the cluster.
|
||||||
|
*/
|
||||||
|
private volatile int mapSlotCapacity;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Additional number of reduce slots that can be requested before
|
||||||
|
* declaring (by Gridmix STRESS mode) the cluster as overloaded.
|
||||||
|
*/
|
||||||
|
private volatile int reduceSlotsBackfill;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Determines the total reduce slot capacity of the cluster.
|
||||||
|
*/
|
||||||
|
private volatile int reduceSlotCapacity;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Determines the max count of running jobs in the cluster.
|
||||||
|
*/
|
||||||
|
private volatile int numJobsBackfill;
|
||||||
|
|
||||||
|
// set the default to true
|
||||||
|
private AtomicBoolean overloaded = new AtomicBoolean(true);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Construct the LoadStatus in an unknown state - assuming the cluster is
|
* Construct the LoadStatus in an unknown state - assuming the cluster is
|
||||||
|
@ -339,12 +380,76 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
|
||||||
reduceSlotCapacity = -1;
|
reduceSlotCapacity = -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
public boolean overloaded() {
|
public synchronized int getMapLoad() {
|
||||||
return (mapSlotsBackfill <= 0) || (reduceSlotsBackfill <= 0)
|
return mapSlotsBackfill;
|
||||||
|| (numJobsBackfill <= 0);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
public String toString() {
|
public synchronized int getMapCapacity() {
|
||||||
|
return mapSlotCapacity;
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized int getReduceLoad() {
|
||||||
|
return reduceSlotsBackfill;
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized int getReduceCapacity() {
|
||||||
|
return reduceSlotCapacity;
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized int getJobLoad() {
|
||||||
|
return numJobsBackfill;
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void decrementMapLoad(int mapSlotsConsumed) {
|
||||||
|
this.mapSlotsBackfill -= mapSlotsConsumed;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void decrementReduceLoad(int reduceSlotsConsumed) {
|
||||||
|
this.reduceSlotsBackfill -= reduceSlotsConsumed;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void decrementJobLoad(int numJobsConsumed) {
|
||||||
|
this.numJobsBackfill -= numJobsConsumed;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void updateMapCapacity(int mapSlotsCapacity) {
|
||||||
|
this.mapSlotCapacity = mapSlotsCapacity;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void updateReduceCapacity(int reduceSlotsCapacity) {
|
||||||
|
this.reduceSlotCapacity = reduceSlotsCapacity;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void updateMapLoad(int mapSlotsBackfill) {
|
||||||
|
this.mapSlotsBackfill = mapSlotsBackfill;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void updateReduceLoad(int reduceSlotsBackfill) {
|
||||||
|
this.reduceSlotsBackfill = reduceSlotsBackfill;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void updateJobLoad(int numJobsBackfill) {
|
||||||
|
this.numJobsBackfill = numJobsBackfill;
|
||||||
|
updateOverloadStatus();
|
||||||
|
}
|
||||||
|
|
||||||
|
private synchronized void updateOverloadStatus() {
|
||||||
|
overloaded.set((mapSlotsBackfill <= 0) || (reduceSlotsBackfill <= 0)
|
||||||
|
|| (numJobsBackfill <= 0));
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized boolean overloaded() {
|
||||||
|
return overloaded.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized String toString() {
|
||||||
// TODO Use StringBuilder instead
|
// TODO Use StringBuilder instead
|
||||||
return " Overloaded = " + overloaded()
|
return " Overloaded = " + overloaded()
|
||||||
+ ", MapSlotBackfill = " + mapSlotsBackfill
|
+ ", MapSlotBackfill = " + mapSlotsBackfill
|
||||||
|
|
|
@ -52,15 +52,23 @@ public class ResourceUsageMatcher {
|
||||||
@SuppressWarnings("unchecked")
|
@SuppressWarnings("unchecked")
|
||||||
public void configure(Configuration conf, ResourceCalculatorPlugin monitor,
|
public void configure(Configuration conf, ResourceCalculatorPlugin monitor,
|
||||||
ResourceUsageMetrics metrics, Progressive progress) {
|
ResourceUsageMetrics metrics, Progressive progress) {
|
||||||
Class[] plugins =
|
Class[] plugins = conf.getClasses(RESOURCE_USAGE_EMULATION_PLUGINS);
|
||||||
conf.getClasses(RESOURCE_USAGE_EMULATION_PLUGINS,
|
|
||||||
ResourceUsageEmulatorPlugin.class);
|
|
||||||
if (plugins == null) {
|
if (plugins == null) {
|
||||||
System.out.println("No resource usage emulator plugins configured.");
|
System.out.println("No resource usage emulator plugins configured.");
|
||||||
} else {
|
} else {
|
||||||
for (Class<? extends ResourceUsageEmulatorPlugin> plugin : plugins) {
|
for (Class clazz : plugins) {
|
||||||
if (plugin != null) {
|
if (clazz != null) {
|
||||||
emulationPlugins.add(ReflectionUtils.newInstance(plugin, conf));
|
if (ResourceUsageEmulatorPlugin.class.isAssignableFrom(clazz)) {
|
||||||
|
ResourceUsageEmulatorPlugin plugin =
|
||||||
|
(ResourceUsageEmulatorPlugin) ReflectionUtils.newInstance(clazz,
|
||||||
|
conf);
|
||||||
|
emulationPlugins.add(plugin);
|
||||||
|
} else {
|
||||||
|
throw new RuntimeException("Misconfigured resource usage plugins. "
|
||||||
|
+ "Class " + clazz.getClass().getName() + " is not a resource "
|
||||||
|
+ "usage plugin as it does not extend "
|
||||||
|
+ ResourceUsageEmulatorPlugin.class.getName());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -101,10 +101,17 @@ public class TestGridmixSubmission {
|
||||||
retiredJobs = new LinkedBlockingQueue<Job>();
|
retiredJobs = new LinkedBlockingQueue<Job>();
|
||||||
}
|
}
|
||||||
|
|
||||||
public void verify(ArrayList<JobStory> submitted) throws Exception {
|
public void verify(ArrayList<JobStory> submitted, Configuration clientConf)
|
||||||
|
throws Exception {
|
||||||
final ArrayList<Job> succeeded = new ArrayList<Job>();
|
final ArrayList<Job> succeeded = new ArrayList<Job>();
|
||||||
assertEquals("Bad job count", expected, retiredJobs.drainTo(succeeded));
|
assertEquals("Bad job count", expected, retiredJobs.drainTo(succeeded));
|
||||||
final HashMap<String,JobStory> sub = new HashMap<String,JobStory>();
|
final HashMap<String,JobStory> sub = new HashMap<String,JobStory>();
|
||||||
|
|
||||||
|
// define the input and output path for the run
|
||||||
|
final Path in = new Path("foo").makeQualified(GridmixTestUtils.dfs);
|
||||||
|
final Path out =
|
||||||
|
new Path(in, clientConf.get(Gridmix.GRIDMIX_OUT_DIR, "gridmix"));
|
||||||
|
|
||||||
for (JobStory spec : submitted) {
|
for (JobStory spec : submitted) {
|
||||||
sub.put(spec.getJobID().toString(), spec);
|
sub.put(spec.getJobID().toString(), spec);
|
||||||
}
|
}
|
||||||
|
@ -115,8 +122,7 @@ public class TestGridmixSubmission {
|
||||||
Configuration conf = job.getConfiguration();
|
Configuration conf = job.getConfiguration();
|
||||||
if (GenerateData.JOB_NAME.equals(jobName)) {
|
if (GenerateData.JOB_NAME.equals(jobName)) {
|
||||||
verifyQueue(conf, jobName);
|
verifyQueue(conf, jobName);
|
||||||
final Path in = new Path("foo").makeQualified(GridmixTestUtils.dfs);
|
|
||||||
final Path out = new Path("/gridmix").makeQualified(GridmixTestUtils.dfs);
|
|
||||||
final ContentSummary generated = GridmixTestUtils.dfs.getContentSummary(in);
|
final ContentSummary generated = GridmixTestUtils.dfs.getContentSummary(in);
|
||||||
assertTrue("Mismatched data gen", // +/- 100k for logs
|
assertTrue("Mismatched data gen", // +/- 100k for logs
|
||||||
(GENDATA << 20) < generated.getLength() + GENSLOP ||
|
(GENDATA << 20) < generated.getLength() + GENSLOP ||
|
||||||
|
@ -164,7 +170,7 @@ public class TestGridmixSubmission {
|
||||||
|
|
||||||
final FileStatus stat =
|
final FileStatus stat =
|
||||||
GridmixTestUtils.dfs.getFileStatus(
|
GridmixTestUtils.dfs.getFileStatus(
|
||||||
new Path(GridmixTestUtils.DEST, "" + Integer.valueOf(jobSeqNum)));
|
new Path(out, "" + Integer.valueOf(jobSeqNum)));
|
||||||
assertEquals("Wrong owner for " + jobName, spec.getUser(),
|
assertEquals("Wrong owner for " + jobName, spec.getUser(),
|
||||||
stat.getOwner());
|
stat.getOwner());
|
||||||
|
|
||||||
|
@ -337,8 +343,9 @@ public class TestGridmixSubmission {
|
||||||
private JobFactory factory;
|
private JobFactory factory;
|
||||||
private TestMonitor monitor;
|
private TestMonitor monitor;
|
||||||
|
|
||||||
public void checkMonitor() throws Exception {
|
public void checkMonitor(Configuration conf) throws Exception {
|
||||||
monitor.verify(((DebugJobFactory.Debuggable)factory).getSubmitted());
|
monitor.verify(((DebugJobFactory.Debuggable)factory).getSubmitted(),
|
||||||
|
conf);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -534,9 +541,11 @@ public class TestGridmixSubmission {
|
||||||
GridmixTestUtils.dfs.setPermission(root, new FsPermission((short)0777));
|
GridmixTestUtils.dfs.setPermission(root, new FsPermission((short)0777));
|
||||||
int res = ToolRunner.run(conf, client, argv);
|
int res = ToolRunner.run(conf, client, argv);
|
||||||
assertEquals("Client exited with nonzero status", 0, res);
|
assertEquals("Client exited with nonzero status", 0, res);
|
||||||
client.checkMonitor();
|
client.checkMonitor(conf);
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
e.printStackTrace();
|
e.printStackTrace();
|
||||||
|
// fail the test if there is an exception
|
||||||
|
throw new RuntimeException(e);
|
||||||
} finally {
|
} finally {
|
||||||
in.getFileSystem(conf).delete(in, true);
|
in.getFileSystem(conf).delete(in, true);
|
||||||
out.getFileSystem(conf).delete(out, true);
|
out.getFileSystem(conf).delete(out, true);
|
||||||
|
|
|
@ -159,7 +159,7 @@ public class TestGridmixSummary {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected Thread createReaderThread() {
|
protected Thread createReaderThread() {
|
||||||
return null;
|
return new Thread();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -243,7 +243,7 @@ public class TestGridmixSummary {
|
||||||
tid, es.getInputTraceSignature());
|
tid, es.getInputTraceSignature());
|
||||||
// test trace location
|
// test trace location
|
||||||
Path qPath = fs.makeQualified(testTraceFile);
|
Path qPath = fs.makeQualified(testTraceFile);
|
||||||
assertEquals("Mismatch in trace signature",
|
assertEquals("Mismatch in trace filename",
|
||||||
qPath.toString(), es.getInputTraceLocation());
|
qPath.toString(), es.getInputTraceLocation());
|
||||||
// test expected data size
|
// test expected data size
|
||||||
assertEquals("Mismatch in expected data size",
|
assertEquals("Mismatch in expected data size",
|
||||||
|
@ -275,7 +275,7 @@ public class TestGridmixSummary {
|
||||||
es.finalize(factory, testTraceFile.toString(), 0L, resolver, dataStats,
|
es.finalize(factory, testTraceFile.toString(), 0L, resolver, dataStats,
|
||||||
conf);
|
conf);
|
||||||
// test missing expected data size
|
// test missing expected data size
|
||||||
assertEquals("Mismatch in trace signature",
|
assertEquals("Mismatch in trace data size",
|
||||||
Summarizer.NA, es.getExpectedDataSize());
|
Summarizer.NA, es.getExpectedDataSize());
|
||||||
assertFalse("Mismatch in trace signature",
|
assertFalse("Mismatch in trace signature",
|
||||||
tid.equals(es.getInputTraceSignature()));
|
tid.equals(es.getInputTraceSignature()));
|
||||||
|
@ -295,6 +295,12 @@ public class TestGridmixSummary {
|
||||||
assertEquals("Mismatch in trace signature",
|
assertEquals("Mismatch in trace signature",
|
||||||
tid, es.getInputTraceSignature());
|
tid, es.getInputTraceSignature());
|
||||||
|
|
||||||
|
// finalize trace identifier '-' input
|
||||||
|
es.finalize(factory, "-", 0L, resolver, dataStats, conf);
|
||||||
|
assertEquals("Mismatch in trace signature",
|
||||||
|
Summarizer.NA, es.getInputTraceSignature());
|
||||||
|
assertEquals("Mismatch in trace file location",
|
||||||
|
Summarizer.NA, es.getInputTraceLocation());
|
||||||
}
|
}
|
||||||
|
|
||||||
// test the ExecutionSummarizer
|
// test the ExecutionSummarizer
|
||||||
|
@ -332,7 +338,7 @@ public class TestGridmixSummary {
|
||||||
return isSuccessful;
|
return isSuccessful;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
return new JobStats(numMaps, fakeJob);
|
return new JobStats(numMaps, numReds, fakeJob);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
Loading…
Reference in New Issue