hadoop/hadoop-mapreduce-project/lib/jdiff/hadoop-mapred_0.22.0.xml

31741 lines
1.3 MiB

<?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
<!-- Generated by the JDiff Javadoc doclet -->
<!-- (http://www.jdiff.org) -->
<!-- on Sun Dec 04 01:04:32 UTC 2011 -->
<api
xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance'
xsi:noNamespaceSchemaLocation='api.xsd'
name="hadoop-mapred 0.22.0"
jdversion="1.0.9">
<!-- Command line arguments = -doclet org.apache.hadoop.classification.tools.ExcludePrivateAnnotationsJDiffDoclet -docletpath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/common/hadoop-common-0.22.0-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/jdiff/jdiff-1.0.9.jar:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/jdiff/xerces-1.4.4.jar -classpath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/classes:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/conf:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-common/jars/hadoop-common-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/commons-cli/commons-cli/jars/commons-cli-1.2.jar:/home/jenkins/.ivy2/cache/xmlenc/xmlenc/jars/xmlenc-0.52.jar:/home/jenkins/.ivy2/cache/commons-codec/commons-codec/jars/commons-codec-1.4.jar:/home/jenkins/.ivy2/cache/commons-logging/commons-logging/jars/commons-logging-1.1.1.jar:/home/jenkins/.ivy2/cache/org.slf4j/slf4j-api/jars/slf4j-api-1.6.1.jar:/home/jenkins/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.6.1.jar:/home/jenkins/.ivy2/cache/log4j/log4j/bundles/log4j-1.2.16.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jetty/jars/jetty-6.1.26.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jetty-util/jars/jetty-util-6.1.26.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/servlet-api/jars/servlet-api-2.5-20081211.jar:/home/jenkins/.ivy2/cache/tomcat/jasper-runtime/jars/jasper-runtime-5.5.12.jar:/home/jenkins/.ivy2/cache/tomcat/jasper-compiler/jars/jasper-compiler-5.5.12.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-2.1-jetty/jars/jsp-2.1-jetty-6.1.26.jar:/home/jenkins/.ivy2/cache/org.eclipse.jdt/core/jars/core-3.1.1.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-api-2.1-glassfish/jars/jsp-api-2.1-glassfish-2.1.v20091210.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-2.1-glassfish/jars/jsp-2.1-glassfish-2.1.v20091210.jar:/home/jenkins/.ivy2/cache/org.eclipse.jdt.core.compiler/ecj/jars/ecj-3.5.1.jar:/home/jenkins/.ivy2/cache/commons-el/commons-el/jars/commons-el-1.0.jar:/home/jenkins/.ivy2/cache/net.java.dev.jets3t/jets3t/jars/jets3t-0.7.1.jar:/home/jenkins/.ivy2/cache/commons-httpclient/commons-httpclient/jars/commons-httpclient-3.1.jar:/home/jenkins/.ivy2/cache/commons-net/commons-net/jars/commons-net-1.4.1.jar:/home/jenkins/.ivy2/cache/oro/oro/jars/oro-2.0.8.jar:/home/jenkins/.ivy2/cache/net.sf.kosmosfs/kfs/jars/kfs-0.3.jar:/home/jenkins/.ivy2/cache/junit/junit/jars/junit-4.8.1.jar:/home/jenkins/.ivy2/cache/hsqldb/hsqldb/jars/hsqldb-1.8.0.10.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro/jars/avro-1.5.3.jar:/home/jenkins/.ivy2/cache/org.codehaus.jackson/jackson-mapper-asl/jars/jackson-mapper-asl-1.7.3.jar:/home/jenkins/.ivy2/cache/org.codehaus.jackson/jackson-core-asl/jars/jackson-core-asl-1.7.3.jar:/home/jenkins/.ivy2/cache/com.thoughtworks.paranamer/paranamer/jars/paranamer-2.3.jar:/home/jenkins/.ivy2/cache/org.xerial.snappy/snappy-java/bundles/snappy-java-1.0.3.2.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro-ipc/jars/avro-ipc-1.5.3.jar:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-common-test/jars/hadoop-common-test-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/jars/hadoop-hdfs-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/com.google.guava/guava/jars/guava-r09.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro-compiler/jars/avro-compiler-1.5.3.jar:/home/jenkins/.ivy2/cache/commons-lang/commons-lang/jars/commons-lang-2.5.jar:/home/jenkins/.ivy2/cache/org.apache.velocity/velocity/jars/velocity-1.6.4.jar:/home/jenkins/.ivy2/cache/commons-collections/commons-collections/jars/commons-collections-3.2.1.jar:/home/jenkins/.ivy2/cache/org.aspectj/aspectjrt/jars/aspectjrt-1.6.5.jar:/home/jenkins/.ivy2/cache/org.aspectj/aspectjtools/jars/aspectjtools-1.6.5.jar:/home/jenkins/.ivy2/cache/jdiff/jdiff/jars/jdiff-1.0.9.jar:/home/jenkins/.ivy2/cache/xerces/xerces/jars/xerces-1.4.4.jar:/home/jenkins/tools/ant/latest/lib/ant-launcher.jar:/usr/share/java/xmlParserAPIs.jar:/usr/share/java/xercesImpl.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-resolver.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-bcel.jar:/home/jenkins/tools/ant/latest/lib/ant-jsch.jar:/home/jenkins/tools/ant/latest/lib/ant-jmf.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-oro.jar:/home/jenkins/tools/ant/latest/lib/ant-netrexx.jar:/home/jenkins/tools/ant/latest/lib/ant-testutil.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-xalan2.jar:/home/jenkins/tools/ant/latest/lib/ant-javamail.jar:/home/jenkins/tools/ant/latest/lib/ant.jar:/home/jenkins/tools/ant/latest/lib/ant-junit.jar:/home/jenkins/tools/ant/latest/lib/ant-swing.jar:/home/jenkins/tools/ant/latest/lib/ant-commons-net.jar:/home/jenkins/tools/ant/latest/lib/ant-jdepend.jar:/home/jenkins/tools/ant/latest/lib/ant-junit4.jar:/home/jenkins/tools/ant/latest/lib/ant-commons-logging.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-bsf.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-log4j.jar:/home/jenkins/tools/ant/latest/lib/ant-jai.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-regexp.jar:/home/jenkins/tools/ant/latest/lib/ant-antlr.jar:/tmp/jdk1.6.0_29/lib/tools.jar -sourcepath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/src/java:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/src/tools -apidir /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/lib/jdiff -apiname hadoop-mapred 0.22.0 -->
<package name="org.apache.hadoop.filecache">
<!-- start class org.apache.hadoop.filecache.DistributedCache -->
<class name="DistributedCache" extends="org.apache.hadoop.mapreduce.filecache.DistributedCache"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use methods on {@link Job}.">
<constructor name="DistributedCache"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Distribute application-specific large, read-only files efficiently.
<p><code>DistributedCache</code> is a facility provided by the Map-Reduce
framework to cache files (text, archives, jars etc.) needed by applications.
</p>
<p>Applications specify the files, via urls (hdfs:// or http://) to be cached
via the {@link org.apache.hadoop.mapred.JobConf}. The
<code>DistributedCache</code> assumes that the files specified via urls are
already present on the {@link FileSystem} at the path specified by the url
and are accessible by every machine in the cluster.</p>
<p>The framework will copy the necessary files on to the slave node before
any tasks for the job are executed on that node. Its efficiency stems from
the fact that the files are only copied once per job and the ability to
cache archives which are un-archived on the slaves.</p>
<p><code>DistributedCache</code> can be used to distribute simple, read-only
data/text files and/or more complex types such as archives, jars etc.
Archives (zip, tar and tgz/tar.gz files) are un-archived at the slave nodes.
Jars may be optionally added to the classpath of the tasks, a rudimentary
software distribution mechanism. Files have execution permissions.
Optionally users can also direct it to symlink the distributed cache file(s)
into the working directory of the task.</p>
<p><code>DistributedCache</code> tracks modification timestamps of the cache
files. Clearly the cache files should not be modified by the application
or externally while the job is executing.</p>
<p>Here is an illustrative example on how to use the
<code>DistributedCache</code>:</p>
<p><blockquote><pre>
// Setting up the cache for the application
1. Copy the requisite files to the <code>FileSystem</code>:
$ bin/hadoop fs -copyFromLocal lookup.dat /myapp/lookup.dat
$ bin/hadoop fs -copyFromLocal map.zip /myapp/map.zip
$ bin/hadoop fs -copyFromLocal mylib.jar /myapp/mylib.jar
$ bin/hadoop fs -copyFromLocal mytar.tar /myapp/mytar.tar
$ bin/hadoop fs -copyFromLocal mytgz.tgz /myapp/mytgz.tgz
$ bin/hadoop fs -copyFromLocal mytargz.tar.gz /myapp/mytargz.tar.gz
2. Setup the application's <code>JobConf</code>:
JobConf job = new JobConf();
DistributedCache.addCacheFile(new URI("/myapp/lookup.dat#lookup.dat"),
job);
DistributedCache.addCacheArchive(new URI("/myapp/map.zip", job);
DistributedCache.addFileToClassPath(new Path("/myapp/mylib.jar"), job);
DistributedCache.addCacheArchive(new URI("/myapp/mytar.tar", job);
DistributedCache.addCacheArchive(new URI("/myapp/mytgz.tgz", job);
DistributedCache.addCacheArchive(new URI("/myapp/mytargz.tar.gz", job);
3. Use the cached files in the {@link org.apache.hadoop.mapred.Mapper}
or {@link org.apache.hadoop.mapred.Reducer}:
public static class MapClass extends MapReduceBase
implements Mapper&lt;K, V, K, V&gt; {
private Path[] localArchives;
private Path[] localFiles;
public void configure(JobConf job) {
// Get the cached archives/files
localArchives = DistributedCache.getLocalCacheArchives(job);
localFiles = DistributedCache.getLocalCacheFiles(job);
}
public void map(K key, V value,
OutputCollector&lt;K, V&gt; output, Reporter reporter)
throws IOException {
// Use data from the cached archives/files here
// ...
// ...
output.collect(k, v);
}
}
</pre></blockquote></p>
It is also very common to use the DistributedCache by using
{@link org.apache.hadoop.util.GenericOptionsParser}.
This class includes methods that should be used by users
(specifically those mentioned in the example above, as well
as {@link DistributedCache#addArchiveToClassPath(Path, Configuration)}),
as well as methods intended for use by the MapReduce framework
(e.g., {@link org.apache.hadoop.mapred.JobClient}). For implementation
details, see {@link TrackerDistributedCacheManager} and
{@link TaskDistributedCacheManager}.
@see org.apache.hadoop.mapred.JobConf
@see org.apache.hadoop.mapred.JobClient
@see org.apache.hadoop.mapreduce.Job
@deprecated Use methods on {@link Job}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.filecache.DistributedCache -->
</package>
<package name="org.apache.hadoop.mapred">
<!-- start class org.apache.hadoop.mapred.ClusterStatus -->
<class name="ClusterStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link ClusterMetrics} or {@link TaskTrackerInfo} instead">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getTaskTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of task trackers in the cluster.
@return the number of task trackers in the cluster.]]>
</doc>
</method>
<method name="getActiveTrackerNames" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the names of task trackers in the cluster.
@return the active task trackers in the cluster.]]>
</doc>
</method>
<method name="getBlacklistedTrackerNames" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the names of task trackers in the cluster.
@return the blacklisted task trackers in the cluster.]]>
</doc>
</method>
<method name="getBlacklistedTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of blacklisted task trackers in the cluster.
@return the number of blacklisted task trackers in the cluster.]]>
</doc>
</method>
<method name="getNumExcludedNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of excluded hosts in the cluster.
@return the number of excluded hosts in the cluster.]]>
</doc>
</method>
<method name="getTTExpiryInterval" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the tasktracker expiry interval for the cluster
@return the expiry interval in msec]]>
</doc>
</method>
<method name="getMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running map tasks in the cluster.
@return the number of currently running map tasks in the cluster.]]>
</doc>
</method>
<method name="getReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running reduce tasks in the cluster.
@return the number of currently running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running map tasks in the cluster.
@return the maximum capacity for running map tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running reduce tasks in the cluster.
@return the maximum capacity for running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getJobTrackerState" return="org.apache.hadoop.mapred.JobTracker.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #getJobTrackerStatus()} instead.">
<doc>
<![CDATA[Get the current state of the <code>JobTracker</code>,
as {@link JobTracker.State}
@return the current state of the <code>JobTracker</code>.
@deprecated Use {@link #getJobTrackerStatus()} instead.]]>
</doc>
</method>
<method name="getJobTrackerStatus" return="org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the JobTracker's status.
@return {@link JobTrackerStatus} of the JobTracker]]>
</doc>
</method>
<method name="getBlackListedTrackersInfo" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the list of blacklisted trackers along with reasons for blacklisting.
@return the collection of {@link BlackListInfo} objects.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Status information on the current state of the Map-Reduce cluster.
<p><code>ClusterStatus</code> provides clients with information such as:
<ol>
<li>
Size of the cluster.
</li>
<li>
Name of the trackers.
</li>
<li>
Task capacity of the cluster.
</li>
<li>
The number of currently running map & reduce tasks.
</li>
<li>
State of the <code>JobTracker</code>.
</li>
<li>
Details regarding black listed trackers.
</li>
</ol></p>
<p>Clients can query for the latest <code>ClusterStatus</code>, via
{@link JobClient#getClusterStatus()}.</p>
@see JobClient
@deprecated Use {@link ClusterMetrics} or {@link TaskTrackerInfo} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ClusterStatus -->
<!-- start class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo -->
<class name="ClusterStatus.BlackListInfo" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getTrackerName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the blacklisted tasktracker's name.
@return tracker's name.]]>
</doc>
</method>
<method name="getReasonForBlackListing" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the reason for which the tasktracker was blacklisted.
@return reason which tracker was blacklisted]]>
</doc>
</method>
<method name="getBlackListReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets a descriptive report about why the tasktracker was blacklisted.
@return report describing why the tasktracker was blacklisted.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Class which encapsulates information about a blacklisted tasktracker.
The information includes the tasktracker's name and reasons for
getting blacklisted. The toString method of the class will print
the information in a whitespace separated fashion to enable parsing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo -->
<!-- start class org.apache.hadoop.mapred.Counters -->
<class name="Counters" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Counters} instead.">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<constructor name="Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getGroupNames" return="java.util.Collection"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the names of all counter classes.
@return Set of counter names.]]>
</doc>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroup" return="org.apache.hadoop.mapred.Counters.Group"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the named counter group, or an empty group if there is none
with the specified name.]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Find the counter for the given enum. The same enum will always return the
same counter.
@param key the counter key
@return the matching counter object]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter given the group and the name.
@param group the name of the group
@param name the internal name of the counter
@return the counter for that name]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<param name="group" type="java.lang.String"/>
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter by using strings
@param group the name of the group
@param id the id of the counter within the group (0 to N-1)
@param name the internal name of the counter
@return the counter for that name
@deprecated]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the specified counter by the specified amount, creating it if
it didn't already exist.
@param key identifies a counter
@param amount amount by which counter is to be incremented]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="counter" type="java.lang.String"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the specified counter by the specified amount, creating it if
it didn't already exist.
@param group the name of the group
@param counter the internal name of the counter
@param amount amount by which counter is to be incremented]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Returns current value of the specified counter, or 0 if the counter
does not exist.]]>
</doc>
</method>
<method name="incrAllCounters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Increments multiple counters by their amounts in another Counters
instance.
@param other the other Counters instance]]>
</doc>
</method>
<method name="sum" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.mapred.Counters"/>
<param name="b" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Convenience method for computing the sum of two sets of counters.]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of counters, by summing the number of counters
in each group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the set of groups.
The external format is:
#groups (groupName group)*
i.e. the number of groups followed by 0 or more groups, where each
group is of the form:
groupDisplayName #counters (false | true counter)*
where each counter is of the form:
name (false | true displayName) value]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a set of groups.]]>
</doc>
</method>
<method name="log"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="log" type="org.apache.commons.logging.Log"/>
<doc>
<![CDATA[Logs the current counter values.
@param log The log to use.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return textual representation of the counter values.]]>
</doc>
</method>
<method name="makeCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert a counters object into a single line that is easy to parse.
@return the string with "name=value" for each counter and separated by ","]]>
</doc>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Represent the counter in a textual format that can be converted back to
its object form
@return the string in the following format
{(groupname)(group-displayname)[(countername)(displayname)(value)][][]}{}{}]]>
</doc>
</method>
<method name="fromEscapedCompactString" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="compactString" type="java.lang.String"/>
<exception name="ParseException" type="java.text.ParseException"/>
<doc>
<![CDATA[Convert a stringified counter representation into a counter object. Note
that the counter can be recovered if its stringified using
{@link #makeEscapedCompactString()}.
@return a Counter]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<doc>
<![CDATA[A set of named counters.
<p><code>Counters</code> represent global counters, defined either by the
Map-Reduce framework or applications. Each <code>Counter</code> can be of
any {@link Enum} type.</p>
<p><code>Counters</code> are bunched into {@link Group}s, each comprising of
counters from a particular <code>Enum</code> class.
@deprecated Use {@link org.apache.hadoop.mapreduce.Counters} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters -->
<!-- start class org.apache.hadoop.mapred.Counters.Counter -->
<class name="Counters.Counter" extends="org.apache.hadoop.mapreduce.Counter"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="setDisplayName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newName" type="java.lang.String"/>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compact stringified version of the counter in the format
[(actual-name)(display-name)(value)]]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[What is the current value of this counter?
@return the current value]]>
</doc>
</method>
<doc>
<![CDATA[A counter record, comprising its name and value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Counter -->
<!-- start class org.apache.hadoop.mapred.Counters.Group -->
<class name="Counters.Group" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns raw name of the group. This is the name of the enum class
for this group of counters.]]>
</doc>
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns localized name of the group. This is the same as getName() by
default, but different if an appropriate ResourceBundle is found.]]>
</doc>
</method>
<method name="setDisplayName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="displayName" type="java.lang.String"/>
<doc>
<![CDATA[Set the display name]]>
</doc>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compact stringified version of the group in the format
{(actual-name)(display-name)(value)[][][]} where [] are compact strings for the
counters within.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
<doc>
<![CDATA[Checks for (content) equality of Groups]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the value of the specified counter, or 0 if the counter does
not exist.]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="use {@link #getCounter(String)} instead">
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the counter for the given id and create it if it doesn't exist.
@param id the numeric id of the counter within the group
@param name the internal counter name
@return the counter
@deprecated use {@link #getCounter(String)} instead]]>
</doc>
</method>
<method name="getCounterForName" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the counter for the given name and create it if it doesn't exist.
@param name the internal counter name
@return the counter]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of counters in this group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[<code>Group</code> of counters, comprising of counters from a particular
counter {@link Enum} class.
<p><code>Group</code>handles localization of the class name and the
counter names.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Group -->
<!-- start class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<class name="FileAlreadyExistsException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileAlreadyExistsException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileAlreadyExistsException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when target file already exists for any operation and
is not configured to be overwritten.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<!-- start class org.apache.hadoop.mapred.FileInputFormat -->
<class name="FileInputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
instead.">
<implements name="org.apache.hadoop.mapred.InputFormat"/>
<constructor name="FileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setMinSplitSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="minSplitSize" type="long"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="filename" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
<code>FileInputFormat</code> implementations can override this and return
<code>false</code> to ensure that individual input files are never split-up
so that {@link Mapper}s process entire files.
@param fs the file system that the file is on
@param filename the file name to check
@return is this file splitable?]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setInputPathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="filter" type="java.lang.Class"/>
<doc>
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
@param filter the PathFilter class use for filtering the input paths.]]>
</doc>
</method>
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
@return the PathFilter instance set for the job, NULL if none has been set.]]>
</doc>
</method>
<method name="addInputPathRecursively"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="result" type="java.util.List"/>
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFilter" type="org.apache.hadoop.fs.PathFilter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add files in the input path recursively into the results.
@param result
The List to store all files.
@param fs
The FileSystem.
@param path
The input path.
@param inputFilter
The input filter that can be used to filter files/dirs.
@throws IOException]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List input directories.
Subclasses may override to, e.g., select only files matching a regular
expression.
@param job the job to list input paths for
@return array of FileStatus objects
@throws IOException if zero items.]]>
</doc>
</method>
<method name="makeSplit" return="org.apache.hadoop.mapred.FileSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="length" type="long"/>
<param name="hosts" type="java.lang.String[]"/>
<doc>
<![CDATA[A factory that makes the split for this class. It can be overridden
by sub-classes to make sub-types]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Splits files returned by {@link #listStatus(JobConf)} when
they're too big.]]>
</doc>
</method>
<method name="computeSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="goalSize" type="long"/>
<param name="minSize" type="long"/>
<param name="blockSize" type="long"/>
</method>
<method name="getBlockIndex" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Sets the given comma separated paths as the list of inputs
for the map-reduce job.
@param conf Configuration of the job
@param commaSeparatedPaths Comma separated paths to be set as
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Add the given comma separated paths to the list of inputs for
the map-reduce job.
@param conf The configuration of the job
@param commaSeparatedPaths Comma separated paths to be added to
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Set the array of {@link Path}s as the list of inputs
for the map-reduce job.
@param conf Configuration of the job.
@param inputPaths the {@link Path}s of the input directories/files
for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
@param conf The configuration of the job
@param path {@link Path} to be added to the list of inputs for
the map-reduce job.]]>
</doc>
</method>
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
@param conf The configuration of the job
@return the list of input {@link Path}s for the map-reduce job.]]>
</doc>
</method>
<method name="getSplitHosts" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
<param name="splitSize" type="long"/>
<param name="clusterMap" type="org.apache.hadoop.net.NetworkTopology"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This function identifies and returns the hosts that contribute
most for a given split. For calculating the contribution, rack
locality is treated on par with host locality, so hosts from racks
that contribute the most are preferred over hosts on racks that
contribute less
@param blkLocations The list of block locations
@param offset
@param splitSize
@return array of hosts that contribute most to this split
@throws IOException]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NUM_INPUT_FILES" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class for file-based {@link InputFormat}.
<p><code>FileInputFormat</code> is the base class for all file-based
<code>InputFormat</code>s. This provides a generic implementation of
{@link #getSplits(JobConf, int)}.
Subclasses of <code>FileInputFormat</code> can also override the
{@link #isSplitable(FileSystem, Path)} method to ensure input-files are
not split-up and are processed as a whole by {@link Mapper}s.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileInputFormat -->
<!-- start class org.apache.hadoop.mapred.FileOutputCommitter -->
<class name="FileOutputCommitter" extends="org.apache.hadoop.mapred.OutputCommitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileOutputCommitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<param name="runState" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setupTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="commitTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abortTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="TEMP_DIR_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Temporary directory name]]>
</doc>
</field>
<field name="SUCCEEDED_FILE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link OutputCommitter} that commits files specified
in job output directory i.e. ${mapreduce.output.fileoutputformat.outputdir}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileOutputCommitter -->
<!-- start class org.apache.hadoop.mapred.FileOutputFormat -->
<class name="FileOutputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
<constructor name="FileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setCompressOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Set whether the output of the job is compressed.
@param conf the {@link JobConf} to modify
@param compress should the output of the job be compressed?]]>
</doc>
</method>
<method name="getCompressOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Is the job output compressed?
@param conf the {@link JobConf} to look in
@return <code>true</code> if the job output should be compressed,
<code>false</code> otherwise]]>
</doc>
</method>
<method name="setOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="codecClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
@param conf the {@link JobConf} to modify
@param codecClass the {@link CompressionCodec} to be used to
compress the job outputs]]>
</doc>
</method>
<method name="getOutputCompressorClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="defaultValue" type="java.lang.Class"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
@param conf the {@link JobConf} to look in
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} to be used to compress the
job outputs
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOutputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
@param conf The configuration of the job.
@param outputDir the {@link Path} of the output directory for
the map-reduce job.]]>
</doc>
</method>
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
@return the {@link Path} to the output directory for the map-reduce job.
@see FileOutputFormat#getWorkOutputPath(JobConf)]]>
</doc>
</method>
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the task's temporary output directory
for the map-reduce job
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
<p><i>Note:</i> The following is valid only if the {@link OutputCommitter}
is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not
a <code>FileOutputCommitter</code>, the task's temporary output
directory is same as {@link #getOutputPath(JobConf)} i.e.
<tt>${mapreduce.output.fileoutputformat.outputdir}$</tt></p>
<p>Some applications need to create/write-to side-files, which differ from
the actual job-outputs.
<p>In such cases there could be issues with 2 instances of the same TIP
(running simultaneously e.g. speculative tasks) trying to open/write-to the
same file (path) on HDFS. Hence the application-writer will have to pick
unique names per task-attempt (e.g. using the attemptid, say
<tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
<p>To get around this the Map-Reduce framework helps the application-writer
out by maintaining a special
<tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt>
sub-directory for each task-attempt on HDFS where the output of the
task-attempt goes. On successful completion of the task-attempt the files
in the <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> (only)
are <i>promoted</i> to <tt>${mapreduce.output.fileoutputformat.outputdir}</tt>. Of course, the
framework discards the sub-directory of unsuccessful task-attempts. This
is completely transparent to the application.</p>
<p>The application-writer can take advantage of this by creating any
side-files required in <tt>${mapreduce.task.output.dir}</tt> during execution
of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the
framework will move them out similarly - thus she doesn't have to pick
unique paths per task-attempt.</p>
<p><i>Note</i>: the value of <tt>${mapreduce.task.output.dir}</tt> during
execution of a particular task-attempt is actually
<tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_{$taskid}</tt>, and this value is
set by the map-reduce framework. So, just create any side-files in the
path returned by {@link #getWorkOutputPath(JobConf)} from map/reduce
task to take advantage of this feature.</p>
<p>The entire discussion holds true for maps of jobs with
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
goes directly to HDFS.</p>
@return the {@link Path} to the task's temporary output directory
for the map-reduce job.]]>
</doc>
</method>
<method name="getTaskOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Helper function to create the task's temporary output directory and
return the path to the task's output file.
@param conf job-configuration
@param name temporary task-output filename
@return path to the task's temporary output file
@throws IOException]]>
</doc>
</method>
<method name="getUniqueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Helper function to generate a name that is unique for the task.
<p>The generated name can be used to create custom files from within the
different tasks for the job, the names for different tasks will not collide
with each other.</p>
<p>The given name is postfixed with the task type, 'm' for maps, 'r' for
reduces and the task partition number. For example, give a name 'test'
running on the first map o the job the generated name will be
'test-m-00000'.</p>
@param conf the configuration for the job.
@param name the name to make unique.
@return a unique name accross all tasks of the job.]]>
</doc>
</method>
<method name="getPathForCustomFile" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Helper function to generate a {@link Path} for a file that is unique for
the task within the job output directory.
<p>The path can be used to create custom files from within the map and
reduce tasks. The path name will be unique for each task. The path parent
will be the job output directory.</p>ls
<p>This method uses the {@link #getUniqueName} method to make the file name
unique for the task.</p>
@param conf the configuration for the job.
@param name the name for the file.
@return a unique path accross all tasks of the job.]]>
</doc>
</method>
<doc>
<![CDATA[A base class for {@link OutputFormat}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.FileSplit -->
<class name="FileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
instead.">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="FileSplit"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[Constructs a split.
@deprecated
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process]]>
</doc>
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a split with host information
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process
@param hosts the list of hosts containing the block, possibly null]]>
</doc>
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.mapreduce.lib.input.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The file containing this split's data.]]>
</doc>
</method>
<method name="getStart" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The position of the first byte in the file to process.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes in the file to process.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A section of an input file. Returned by {@link
InputFormat#getSplits(JobConf, int)} and passed to
{@link InputFormat#getRecordReader(InputSplit,JobConf,Reporter)}.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileSplit -->
<!-- start class org.apache.hadoop.mapred.ID -->
<class name="ID" extends="org.apache.hadoop.mapreduce.ID"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ID" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructs an ID object from the given int]]>
</doc>
</constructor>
<constructor name="ID"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A general identifier, which internally stores the id
as an integer. This is the super class of {@link JobID},
{@link TaskID} and {@link TaskAttemptID}.
@see JobID
@see TaskID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ID -->
<!-- start interface org.apache.hadoop.mapred.InputFormat -->
<interface name="InputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.">
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Logically split the set of input files for the job.
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
for processing.</p>
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
input files are not physically split into chunks. For e.g. a split could
be <i>&lt;input-file-path, start, offset&gt;</i> tuple.
@param job job configuration.
@param numSplits the desired number of splits, a hint.
@return an array of {@link InputSplit}s for the job.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the {@link RecordReader} for the given {@link InputSplit}.
<p>It is the responsibility of the <code>RecordReader</code> to respect
record boundaries while processing the logical split to present a
record-oriented view to the individual task.</p>
@param split the {@link InputSplit}
@param job the job that this split belongs to
@return a {@link RecordReader}]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputFormat</code> describes the input-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the input-specification of the job.
<li>
Split-up the input file(s) into logical {@link InputSplit}s, each of
which is then assigned to an individual {@link Mapper}.
</li>
<li>
Provide the {@link RecordReader} implementation to be used to glean
input records from the logical <code>InputSplit</code> for processing by
the {@link Mapper}.
</li>
</ol>
<p>The default behavior of file-based {@link InputFormat}s, typically
sub-classes of {@link FileInputFormat}, is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of the input files. However, the {@link FileSystem} blocksize of
the input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../mapred-default.html#mapreduce.input.fileinputformat.split.minsize">
mapreduce.input.fileinputformat.split.minsize</a>.</p>
<p>Clearly, logical splits based on input-size is insufficient for many
applications since record boundaries are to respected. In such cases, the
application has to also implement a {@link RecordReader} on whom lies the
responsibilty to respect record-boundaries and present a record-oriented
view of the logical <code>InputSplit</code> to the individual task.
@see InputSplit
@see RecordReader
@see JobClient
@see FileInputFormat
@deprecated Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputFormat -->
<!-- start interface org.apache.hadoop.mapred.InputSplit -->
<interface name="InputSplit" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the total number of bytes in the data of the <code>InputSplit</code>.
@return the number of bytes in the input split.
@throws IOException]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the list of hostnames where the input split is located.
@return list of hostnames where data of the <code>InputSplit</code> is
located as an array of <code>String</code>s.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
individual {@link Mapper}.
<p>Typically, it presents a byte-oriented view on the input and is the
responsibility of {@link RecordReader} of the job to process this and present
a record-oriented view.
@see InputFormat
@see RecordReader
@deprecated Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputSplit -->
<!-- start class org.apache.hadoop.mapred.InvalidFileTypeException -->
<class name="InvalidFileTypeException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidFileTypeException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidFileTypeException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when file type differs from the desired file type. like
getting a file when a directory is expected. Or a wrong file type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidFileTypeException -->
<!-- start class org.apache.hadoop.mapred.InvalidInputException -->
<class name="InvalidInputException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidInputException" type="java.util.List"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create the exception with the given list.
@param probs the list of problems to report. this list is not copied.]]>
</doc>
</constructor>
<method name="getProblems" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the complete list of the problems reported.
@return the list of problems, which must not be modified]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a summary message of the problems found.
@return the concatenated messages from all of the problems.]]>
</doc>
</method>
<doc>
<![CDATA[This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidInputException -->
<!-- start class org.apache.hadoop.mapred.InvalidJobConfException -->
<class name="InvalidJobConfException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidJobConfException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidJobConfException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This exception is thrown when jobconf misses some mendatory attributes
or value of some attributes is invalid.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidJobConfException -->
<!-- start class org.apache.hadoop.mapred.IsolationRunner -->
<class name="IsolationRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IsolationRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Run a single task.
@param args the first argument is the task directory]]>
</doc>
</method>
<doc>
<![CDATA[IsolationRunner is intended to facilitate debugging by re-running a specific
task, given left-over task files for a (typically failed) past job.
Currently, it is limited to re-running map tasks.
Users may coerce MapReduce to keep task files around by setting
mapreduce.task.files.preserve.failedtasks. See mapred_tutorial.xml for more documentation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.IsolationRunner -->
<!-- start class org.apache.hadoop.mapred.JobClient -->
<class name="JobClient" extends="org.apache.hadoop.mapreduce.tools.CLI"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link Job} and {@link Cluster} instead">
<constructor name="JobClient"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job client.]]>
</doc>
</constructor>
<constructor name="JobClient" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client with the given {@link JobConf}, and connect to the
default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobClient" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client with the given {@link Configuration},
and connect to the default {@link JobTracker}.
@param conf the configuration.
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobClient" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client, connect to the indicated job tracker.
@param jobTrackAddr the job tracker to connect to.
@param conf configuration.]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Connect to the default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the <code>JobClient</code>.]]>
</doc>
</method>
<method name="getFs" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a filesystem handle. We need this to prepare jobs
for submission to the MapReduce system.
@return the filesystem handle.]]>
</doc>
</method>
<method name="getClusterHandle" return="org.apache.hadoop.mapreduce.Cluster"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a handle to the Cluster]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobFile" type="java.lang.String"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param jobFile the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws InvalidJobConfException
@throws IOException]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param conf the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws IOException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an {@link RunningJob} object to track an ongoing job. Returns
null if the id does not correspond to any known job.
@param jobid the jobid of the job.
@return the {@link RunningJob} handle to track the job, null if the
<code>jobid</code> doesn't correspond to any known job.
@throws IOException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getJob(JobID)}.">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getJob(JobID)}.]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the map tasks of a job.
@param jobId the job to query.
@return the list of all of the map tips.
@throws IOException]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getMapTaskReports(JobID)}">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getMapTaskReports(JobID)}]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the reduce tasks of a job.
@param jobId the job to query.
@return the list of all of the reduce tips.
@throws IOException]]>
</doc>
</method>
<method name="getCleanupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the cleanup tasks of a job.
@param jobId the job to query.
@return the list of all of the cleanup tips.
@throws IOException]]>
</doc>
</method>
<method name="getSetupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the setup tasks of a job.
@param jobId the job to query.
@return the list of all of the setup tips.
@throws IOException]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getReduceTaskReports(JobID)}">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getReduceTaskReports(JobID)}]]>
</doc>
</method>
<method name="displayTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<param name="type" type="java.lang.String"/>
<param name="state" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Display the information about a job's tasks, of a particular type and
in a particular state
@param jobId the ID of the job
@param type the type of the task (map/reduce/setup/cleanup)
@param state the state of the task
(pending/running/completed/failed/killed)]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the Map-Reduce cluster.
@return the status information about the Map-Reduce cluster as an object
of {@link ClusterStatus}.
@throws IOException]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="detailed" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the Map-Reduce cluster.
@param detailed if true then get a detailed status including the
tracker names
@return the status information about the Map-Reduce cluster as an object
of {@link ClusterStatus}.
@throws IOException]]>
</doc>
</method>
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are not completed and not failed.
@return array of {@link JobStatus} for the running/to-be-run jobs.
@throws IOException]]>
</doc>
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are submitted.
@return array of {@link JobStatus} for the submitted jobs.
@throws IOException]]>
</doc>
</method>
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Utility that submits a job, then polls for progress until the job is
complete.
@param job the job configuration.
@throws IOException if the job fails]]>
</doc>
</method>
<method name="monitorAndPrintJob" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="job" type="org.apache.hadoop.mapred.RunningJob"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Monitor a job and print status in real-time as progress is made and tasks
fail.
@param conf the job's configuration
@param job the job to track
@return true if the job succeeded
@throws IOException if communication to the JobTracker fails]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Sets the output filter for tasks. only those tasks are printed whose
output matches the filter.
@param newValue task filter.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the task output filter out of the JobConf.
@param job the JobConf to examine.
@return the filter level.]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Modify the JobConf to set the task output filter.
@param job the JobConf to modify.
@param newValue the value to set.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task output filter.
@return task filter.]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="cntrs" type="org.apache.hadoop.mapreduce.Counters"/>
<param name="counterGroupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDefaultMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the max available Maps in the cluster.
@return the max available Maps in the cluster
@throws IOException]]>
</doc>
</method>
<method name="getDefaultReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the max available Reduces in the cluster.
@return the max available Reduces in the cluster
@throws IOException]]>
</doc>
</method>
<method name="getSystemDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Grab the jobtracker system directory path where job-specific files are to be placed.
@return the system directory where job-specific files are to be placed.]]>
</doc>
</method>
<method name="getRootQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an array of queue information objects about root level queues
configured
@return the array of root level JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getChildQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an array of queue information objects about immediate children
of queue queueName.
@param queueName
@return the array of immediate children JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return an array of queue information objects about all the Job Queues
configured.
@return Array of JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getJobsFromQueue" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets all the jobs which were added to particular Job Queue
@param queueName name of the Job Queue
@return Array of jobs present in the job queue
@throws IOException]]>
</doc>
</method>
<method name="getQueueInfo" return="org.apache.hadoop.mapred.JobQueueInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the queue information associated to a particular Job Queue
@param queueName name of the job queue.
@return Queue information associated to particular queue.
@throws IOException]]>
</doc>
</method>
<method name="getQueueAclsForCurrentUser" return="org.apache.hadoop.mapred.QueueAclsInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the Queue ACLs for current user
@return array of QueueAclsInfo object for current user.
@throws IOException]]>
</doc>
</method>
<method name="getDelegationToken" return="org.apache.hadoop.security.token.Token"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="renewer" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get a delegation token for the user from the JobTracker.
@param renewer the user who can renew the token
@return the new token
@throws IOException]]>
</doc>
</method>
<method name="renewDelegationToken" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="SecretManager.InvalidToken" type="org.apache.hadoop.security.token.SecretManager.InvalidToken"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Renew a delegation token
@param token the token to renew
@return true if the renewal went well
@throws InvalidToken
@throws IOException]]>
</doc>
</method>
<method name="cancelDelegationToken"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="SecretManager.InvalidToken" type="org.apache.hadoop.security.token.SecretManager.InvalidToken"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Cancel a delegation token from the JobTracker
@param token the token to cancel
@throws IOException]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[<code>JobClient</code> is the primary interface for the user-job to interact
with the {@link JobTracker}.
<code>JobClient</code> provides facilities to submit jobs, track their
progress, access component-tasks' reports/logs, get the Map-Reduce cluster
status information etc.
<p>The job submission process involves:
<ol>
<li>
Checking the input and output specifications of the job.
</li>
<li>
Computing the {@link InputSplit}s for the job.
</li>
<li>
Setup the requisite accounting information for the {@link DistributedCache}
of the job, if necessary.
</li>
<li>
Copying the job's jar and configuration to the map-reduce system directory
on the distributed file-system.
</li>
<li>
Submitting the job to the <code>JobTracker</code> and optionally monitoring
it's status.
</li>
</ol></p>
Normally the user creates the application, describes various facets of the
job via {@link JobConf} and then uses the <code>JobClient</code> to submit
the job and monitor its progress.
<p>Here is an example on how to use <code>JobClient</code>:</p>
<p><blockquote><pre>
// Create a new JobConf
JobConf job = new JobConf(new Configuration(), MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
job.setInputPath(new Path("in"));
job.setOutputPath(new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setReducerClass(MyJob.MyReducer.class);
// Submit the job, then poll for progress until the job is complete
JobClient.runJob(job);
</pre></blockquote></p>
<h4 id="JobControl">Job Control</h4>
<p>At times clients would chain map-reduce jobs to accomplish complex tasks
which cannot be done via a single map-reduce job. This is fairly easy since
the output of the job, typically, goes to distributed file-system and that
can be used as the input for the next job.</p>
<p>However, this also means that the onus on ensuring jobs are complete
(success/failure) lies squarely on the clients. In such situations the
various job-control options are:
<ol>
<li>
{@link #runJob(JobConf)} : submits the job and returns only after
the job has completed.
</li>
<li>
{@link #submitJob(JobConf)} : only submits the job, then poll the
returned handle to the {@link RunningJob} to query status and make
scheduling decisions.
</li>
<li>
{@link JobConf#setJobEndNotificationURI(String)} : setup a notification
on job-completion, thus avoiding polling.
</li>
</ol></p>
@see JobConf
@see ClusterStatus
@see Tool
@see DistributedCache
@deprecated Use {@link Job} and {@link Cluster} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient -->
<!-- start class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<class name="JobClient.TaskStatusFilter" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<!-- start class org.apache.hadoop.mapred.JobConf -->
<class name="JobConf" extends="org.apache.hadoop.conf.Configuration"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link Configuration} instead">
<constructor name="JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<constructor name="JobConf" type="boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A new map/reduce configuration where the behavior of reading from the
default resources can be turned off.
<p/>
If the parameter {@code loadDefaults} is false, the new instance
will not load resources from the default files.
@param loadDefaults specifies whether to load from the default files]]>
</doc>
</constructor>
<method name="getCredentials" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get credentials for the job.
@return credentials for the job]]>
</doc>
</method>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user jar for the map-reduce job.
@return the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="setJar"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jar" type="java.lang.String"/>
<doc>
<![CDATA[Set the user jar for the map-reduce job.
@param jar the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="getJarUnpackPattern" return="java.util.regex.Pattern"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the pattern for jar contents to unpack on the tasktracker]]>
</doc>
</method>
<method name="setJarByClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<doc>
<![CDATA[Set the job's jar file by finding an example class location.
@param cls the example class.]]>
</doc>
</method>
<method name="getLocalDirs" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Use MRAsyncDiskService.moveAndDeleteAllVolumes instead.
@see org.apache.hadoop.mapreduce.util.MRAsyncDiskService#cleanupAllVolumes()]]>
</doc>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="subdir" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathString" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructs a local file name. Files are distributed among configured
local directories.]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reported username for this job.
@return the username]]>
</doc>
</method>
<method name="setUser"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="user" type="java.lang.String"/>
<doc>
<![CDATA[Set the reported username for this job.
@param user the username for this job.]]>
</doc>
</method>
<method name="setKeepFailedTaskFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keep" type="boolean"/>
<doc>
<![CDATA[Set whether the framework should keep the intermediate files for
failed tasks.
@param keep <code>true</code> if framework should keep the intermediate files
for failed tasks, <code>false</code> otherwise.]]>
</doc>
</method>
<method name="getKeepFailedTaskFiles" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should the temporary files for failed tasks be kept?
@return should the files be kept?]]>
</doc>
</method>
<method name="setKeepTaskFilesPattern"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pattern" type="java.lang.String"/>
<doc>
<![CDATA[Set a regular expression for task names that should be kept.
The regular expression ".*_m_000123_0" would keep the files
for the first instance of map 123 that ran.
@param pattern the java.util.regex.Pattern to match against the
task names.]]>
</doc>
</method>
<method name="getKeepTaskFilesPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the regular expression that is matched against the task names
to see if we need to keep the files.
@return the pattern as a string, if it was set, othewise null.]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the current working directory for the default file system.
@param dir the new current working directory.]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current working directory for the default file system.
@return the directory name.]]>
</doc>
</method>
<method name="setNumTasksToExecutePerJvm"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="numTasks" type="int"/>
<doc>
<![CDATA[Sets the number of tasks that a spawned task JVM should run
before it exits
@param numTasks the number of tasks to execute; defaults to 1;
-1 signifies no limit]]>
</doc>
</method>
<method name="getNumTasksToExecutePerJvm" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of tasks that a spawned JVM should execute]]>
</doc>
</method>
<method name="getInputFormat" return="org.apache.hadoop.mapred.InputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link InputFormat} implementation for the map-reduce job,
defaults to {@link TextInputFormat} if not specified explicity.
@return the {@link InputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setInputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link InputFormat} implementation for the map-reduce job.
@param theClass the {@link InputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="getOutputFormat" return="org.apache.hadoop.mapred.OutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputFormat} implementation for the map-reduce job,
defaults to {@link TextOutputFormat} if not specified explicity.
@return the {@link OutputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapred.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputCommitter} implementation for the map-reduce job,
defaults to {@link FileOutputCommitter} if not specified explicitly.
@return the {@link OutputCommitter} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setOutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link OutputCommitter} implementation for the map-reduce job.
@param theClass the {@link OutputCommitter} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="setOutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link OutputFormat} implementation for the map-reduce job.
@param theClass the {@link OutputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="setCompressMapOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Should the map outputs be compressed before transfer?
Uses the SequenceFile compression.
@param compress should the map outputs be compressed?]]>
</doc>
</method>
<method name="getCompressMapOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Are the outputs of the maps be compressed?
@return <code>true</code> if the outputs of the maps are to be compressed,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setMapOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="codecClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the given class as the {@link CompressionCodec} for the map outputs.
@param codecClass the {@link CompressionCodec} class that will compress
the map outputs.]]>
</doc>
</method>
<method name="getMapOutputCompressorClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="defaultValue" type="java.lang.Class"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the map outputs.
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} class that should be used to compress the
map outputs.
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the map output data. If it is not set, use the
(final) output key class. This allows the map output key class to be
different than the final output key class.
@return the map output key class.]]>
</doc>
</method>
<method name="setMapOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the map output data. This allows the user to
specify the map output key class to be different than the final output
value class.
@param theClass the map output key class.]]>
</doc>
</method>
<method name="getMapOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for the map output data. If it is not set, use the
(final) output value class This allows the map output value class to be
different than the final output value class.
@return the map output value class.]]>
</doc>
</method>
<method name="setMapOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the value class for the map output data. This allows the user to
specify the map output value class to be different than the final output
value class.
@param theClass the map output value class.]]>
</doc>
</method>
<method name="getOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the job output data.
@return the key class for the job output data.]]>
</doc>
</method>
<method name="setOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the job output data.
@param theClass the key class for the job output data.]]>
</doc>
</method>
<method name="getOutputKeyComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link RawComparator} comparator used to compare keys.
@return the {@link RawComparator} comparator used to compare keys.]]>
</doc>
</method>
<method name="setOutputKeyComparatorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link RawComparator} comparator used to compare keys.
@param theClass the {@link RawComparator} comparator used to
compare keys.
@see #setOutputValueGroupingComparator(Class)]]>
</doc>
</method>
<method name="setKeyFieldComparatorOptions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keySpec" type="java.lang.String"/>
<doc>
<![CDATA[Set the {@link KeyFieldBasedComparator} options used to compare keys.
@param keySpec the key specification of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field). opts are ordering options. The supported options
are:
-n, (Sort numerically)
-r, (Reverse the result of comparison)]]>
</doc>
</method>
<method name="getKeyFieldComparatorOption" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link KeyFieldBasedComparator} options]]>
</doc>
</method>
<method name="setKeyFieldPartitionerOptions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keySpec" type="java.lang.String"/>
<doc>
<![CDATA[Set the {@link KeyFieldBasedPartitioner} options used for
{@link Partitioner}
@param keySpec the key specification of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field).]]>
</doc>
</method>
<method name="getKeyFieldPartitionerOption" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link KeyFieldBasedPartitioner} options]]>
</doc>
</method>
<method name="getOutputValueGroupingComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user defined {@link WritableComparable} comparator for
grouping keys of inputs to the reduce.
@return comparator set by the user for grouping values.
@see #setOutputValueGroupingComparator(Class) for details.]]>
</doc>
</method>
<method name="setOutputValueGroupingComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the user defined {@link RawComparator} comparator for
grouping keys in the input to the reduce.
<p>This comparator should be provided if the equivalence rules for keys
for sorting the intermediates are different from those for grouping keys
before each call to
{@link Reducer#reduce(Object, java.util.Iterator, OutputCollector, Reporter)}.</p>
<p>For key-value pairs (K1,V1) and (K2,V2), the values (V1, V2) are passed
in a single call to the reduce function if K1 and K2 compare as equal.</p>
<p>Since {@link #setOutputKeyComparatorClass(Class)} can be used to control
how keys are sorted, this can be used in conjunction to simulate
<i>secondary sort on values</i>.</p>
<p><i>Note</i>: This is not a guarantee of the reduce sort being
<i>stable</i> in any sense. (In any case, with the order of available
map-outputs to the reduce being non-deterministic, it wouldn't make
that much sense.)</p>
@param theClass the comparator class to be used for grouping keys.
It should implement <code>RawComparator</code>.
@see #setOutputKeyComparatorClass(Class)]]>
</doc>
</method>
<method name="getUseNewMapper" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should the framework use the new context-object code for running
the mapper?
@return true, if the new api should be used]]>
</doc>
</method>
<method name="setUseNewMapper"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="flag" type="boolean"/>
<doc>
<![CDATA[Set whether the framework should use the new api for the mapper.
This is the default for jobs submitted with the new Job api.
@param flag true, if the new api should be used]]>
</doc>
</method>
<method name="getUseNewReducer" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should the framework use the new context-object code for running
the reducer?
@return true, if the new api should be used]]>
</doc>
</method>
<method name="setUseNewReducer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="flag" type="boolean"/>
<doc>
<![CDATA[Set whether the framework should use the new api for the reducer.
This is the default for jobs submitted with the new Job api.
@param flag true, if the new api should be used]]>
</doc>
</method>
<method name="getOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for job outputs.
@return the value class for job outputs.]]>
</doc>
</method>
<method name="setOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the value class for job outputs.
@param theClass the value class for job outputs.]]>
</doc>
</method>
<method name="getMapperClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Mapper} class for the job.
@return the {@link Mapper} class for the job.]]>
</doc>
</method>
<method name="setMapperClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link Mapper} class for the job.
@param theClass the {@link Mapper} class for the job.]]>
</doc>
</method>
<method name="getMapRunnerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link MapRunnable} class for the job.
@return the {@link MapRunnable} class for the job.]]>
</doc>
</method>
<method name="setMapRunnerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Expert: Set the {@link MapRunnable} class for the job.
Typically used to exert greater control on {@link Mapper}s.
@param theClass the {@link MapRunnable} class for the job.]]>
</doc>
</method>
<method name="getPartitionerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Partitioner} used to partition {@link Mapper}-outputs
to be sent to the {@link Reducer}s.
@return the {@link Partitioner} used to partition map-outputs.]]>
</doc>
</method>
<method name="setPartitionerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link Partitioner} class used to partition
{@link Mapper}-outputs to be sent to the {@link Reducer}s.
@param theClass the {@link Partitioner} used to partition map-outputs.]]>
</doc>
</method>
<method name="getReducerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Reducer} class for the job.
@return the {@link Reducer} class for the job.]]>
</doc>
</method>
<method name="setReducerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link Reducer} class for the job.
@param theClass the {@link Reducer} class for the job.]]>
</doc>
</method>
<method name="getCombinerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-defined <i>combiner</i> class used to combine map-outputs
before being sent to the reducers. Typically the combiner is same as the
the {@link Reducer} for the job i.e. {@link #getReducerClass()}.
@return the user-defined combiner class used to combine map-outputs.]]>
</doc>
</method>
<method name="setCombinerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the user-defined <i>combiner</i> class used to combine map-outputs
before being sent to the reducers.
<p>The combiner is an application-specified aggregation operation, which
can help cut down the amount of data transferred between the
{@link Mapper} and the {@link Reducer}, leading to better performance.</p>
<p>The framework may invoke the combiner 0, 1, or multiple times, in both
the mapper and reducer tasks. In general, the combiner is called as the
sort/merge result is written to disk. The combiner must:
<ul>
<li> be side-effect free</li>
<li> have the same input and output key types and the same input and
output value types</li>
</ul></p>
<p>Typically the combiner is same as the <code>Reducer</code> for the
job i.e. {@link #setReducerClass(Class)}.</p>
@param theClass the user-defined combiner class used to combine
map-outputs.]]>
</doc>
</method>
<method name="getSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be used for this job,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on, else <code>false</code>.]]>
</doc>
</method>
<method name="getMapSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job for map tasks?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be
used for this job for map tasks,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setMapSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for map tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for map tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="getReduceSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job for reduce tasks?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be used
for reduce tasks for this job,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setReduceSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for reduce tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for reduce tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="getNumMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get configured the number of reduce tasks for this job.
Defaults to <code>1</code>.
@return the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="setNumMapTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Set the number of map tasks for this job.
<p><i>Note</i>: This is only a <i>hint</i> to the framework. The actual
number of spawned map tasks depends on the number of {@link InputSplit}s
generated by the job's {@link InputFormat#getSplits(JobConf, int)}.
A custom {@link InputFormat} is typically used to accurately control
the number of map tasks for the job.</p>
<h4 id="NoOfMaps">How many maps?</h4>
<p>The number of maps is usually driven by the total size of the inputs
i.e. total number of blocks of the input files.</p>
<p>The right level of parallelism for maps seems to be around 10-100 maps
per-node, although it has been set up to 300 or so for very cpu-light map
tasks. Task setup takes awhile, so it is best if the maps take at least a
minute to execute.</p>
<p>The default behavior of file-based {@link InputFormat}s is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of input files. However, the {@link FileSystem} blocksize of the
input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../mapred-default.html#mapreduce.input.fileinputformat.split.minsize">
mapreduce.input.fileinputformat.split.minsize</a>.</p>
<p>Thus, if you expect 10TB of input data and have a blocksize of 128MB,
you'll end up with 82,000 maps, unless {@link #setNumMapTasks(int)} is
used to set it even higher.</p>
@param n the number of map tasks for this job.
@see InputFormat#getSplits(JobConf, int)
@see FileInputFormat
@see FileSystem#getDefaultBlockSize()
@see FileStatus#getBlockSize()]]>
</doc>
</method>
<method name="getNumReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get configured the number of reduce tasks for this job. Defaults to
<code>1</code>.
@return the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="setNumReduceTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Set the requisite number of reduce tasks for this job.
<h4 id="NoOfReduces">How many reduces?</h4>
<p>The right number of reduces seems to be <code>0.95</code> or
<code>1.75</code> multiplied by (&lt;<i>no. of nodes</i>&gt; *
<a href="{@docRoot}/../mapred-default.html#mapreduce.tasktracker.reduce.tasks.maximum">
mapreduce.tasktracker.reduce.tasks.maximum</a>).
</p>
<p>With <code>0.95</code> all of the reduces can launch immediately and
start transfering map outputs as the maps finish. With <code>1.75</code>
the faster nodes will finish their first round of reduces and launch a
second wave of reduces doing a much better job of load balancing.</p>
<p>Increasing the number of reduces increases the framework overhead, but
increases load balancing and lowers the cost of failures.</p>
<p>The scaling factors above are slightly less than whole numbers to
reserve a few reduce slots in the framework for speculative-tasks, failures
etc.</p>
<h4 id="ReducerNone">Reducer NONE</h4>
<p>It is legal to set the number of reduce-tasks to <code>zero</code>.</p>
<p>In this case the output of the map-tasks directly go to distributed
file-system, to the path set by
{@link FileOutputFormat#setOutputPath(JobConf, Path)}. Also, the
framework doesn't sort the map-outputs before writing it out to HDFS.</p>
@param n the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="getMaxMapAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
map task, as specified by the <code>mapreduce.map.maxattempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per map task.]]>
</doc>
</method>
<method name="setMaxMapAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
map task.
@param n the number of attempts per map task.]]>
</doc>
</method>
<method name="getMaxReduceAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
reduce task, as specified by the <code>mapreduce.reduce.maxattempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per reduce task.]]>
</doc>
</method>
<method name="setMaxReduceAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
reduce task.
@param n the number of attempts per reduce task.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified job name. This is only used to identify the
job to the user.
@return the job's name, defaulting to "".]]>
</doc>
</method>
<method name="setJobName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Set the user-specified job name.
@param name the job's new name.]]>
</doc>
</method>
<method name="getSessionId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified session identifier. The default is the empty string.
The session identifier is used to tag metric data that is reported to some
performance metrics system via the org.apache.hadoop.metrics API. The
session identifier is intended, in particular, for use by Hadoop-On-Demand
(HOD) which allocates a virtual Hadoop cluster dynamically and transiently.
HOD will set the session identifier by modifying the mapred-site.xml file
before starting the cluster.
When not running under HOD, this identifer is expected to remain set to
the empty string.
@return the session identifier, defaulting to "".]]>
</doc>
</method>
<method name="setSessionId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="sessionId" type="java.lang.String"/>
<doc>
<![CDATA[Set the user-specified session identifier.
@param sessionId the new session id.]]>
</doc>
</method>
<method name="setMaxTaskFailuresPerTracker"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="noFailures" type="int"/>
<doc>
<![CDATA[Set the maximum no. of failures of a given job per tasktracker.
If the no. of task failures exceeds <code>noFailures</code>, the
tasktracker is <i>blacklisted</i> for this job.
@param noFailures maximum no. of failures of a given job per tasktracker.]]>
</doc>
</method>
<method name="getMaxTaskFailuresPerTracker" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Expert: Get the maximum no. of failures of a given job per tasktracker.
If the no. of task failures exceeds this, the tasktracker is
<i>blacklisted</i> for this job.
@return the maximum no. of failures of a given job per tasktracker.]]>
</doc>
</method>
<method name="getMaxMapTaskFailuresPercent" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum percentage of map tasks that can fail without
the job being aborted.
Each map task is executed a minimum of {@link #getMaxMapAttempts()}
attempts before being declared as <i>failed</i>.
Defaults to <code>zero</code>, i.e. <i>any</i> failed map-task results in
the job being declared as {@link JobStatus#FAILED}.
@return the maximum percentage of map tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setMaxMapTaskFailuresPercent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="percent" type="int"/>
<doc>
<![CDATA[Expert: Set the maximum percentage of map tasks that can fail without the
job being aborted.
Each map task is executed a minimum of {@link #getMaxMapAttempts} attempts
before being declared as <i>failed</i>.
@param percent the maximum percentage of map tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="getMaxReduceTaskFailuresPercent" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum percentage of reduce tasks that can fail without
the job being aborted.
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
attempts before being declared as <i>failed</i>.
Defaults to <code>zero</code>, i.e. <i>any</i> failed reduce-task results
in the job being declared as {@link JobStatus#FAILED}.
@return the maximum percentage of reduce tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setMaxReduceTaskFailuresPercent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="percent" type="int"/>
<doc>
<![CDATA[Set the maximum percentage of reduce tasks that can fail without the job
being aborted.
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
attempts before being declared as <i>failed</i>.
@param percent the maximum percentage of reduce tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setJobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="prio" type="org.apache.hadoop.mapred.JobPriority"/>
<doc>
<![CDATA[Set {@link JobPriority} for this job.
@param prio the {@link JobPriority} for this job.]]>
</doc>
</method>
<method name="getJobPriority" return="org.apache.hadoop.mapred.JobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link JobPriority} for this job.
@return the {@link JobPriority} for this job.]]>
</doc>
</method>
<method name="getProfileEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get whether the task profiling is enabled.
@return true if some tasks will be profiled]]>
</doc>
</method>
<method name="setProfileEnabled"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="boolean"/>
<doc>
<![CDATA[Set whether the system should collect profiler information for some of
the tasks in this job? The information is stored in the user log
directory.
@param newValue true means it should be gathered]]>
</doc>
</method>
<method name="getProfileParams" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the profiler configuration arguments.
The default value for this property is
"-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s"
@return the parameters to pass to the task child to configure profiling]]>
</doc>
</method>
<method name="setProfileParams"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="java.lang.String"/>
<doc>
<![CDATA[Set the profiler configuration arguments. If the string contains a '%s' it
will be replaced with the name of the profiling output file when the task
runs.
This value is passed to the task child JVM on the command line.
@param value the configuration string]]>
</doc>
</method>
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<doc>
<![CDATA[Get the range of maps or reduces to profile.
@param isMap is the task a map?
@return the task ranges]]>
</doc>
</method>
<method name="setProfileTaskRange"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<param name="newValue" type="java.lang.String"/>
<doc>
<![CDATA[Set the ranges of maps or reduces to profile. setProfileEnabled(true)
must also be called.
@param newValue a set of integer ranges of the map ids]]>
</doc>
</method>
<method name="setMapDebugScript"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mDbgScript" type="java.lang.String"/>
<doc>
<![CDATA[Set the debug script to run when the map tasks fail.
<p>The debug script can aid debugging of failed map tasks. The script is
given task's stdout, stderr, syslog, jobconf files as arguments.</p>
<p>The debug command, run on the node where the map failed, is:</p>
<p><pre><blockquote>
$script $stdout $stderr $syslog $jobconf.
</blockquote></pre></p>
<p> The script file is distributed through {@link DistributedCache}
APIs. The script needs to be symlinked. </p>
<p>Here is an example on how to submit a script
<p><blockquote><pre>
job.setMapDebugScript("./myscript");
DistributedCache.createSymlink(job);
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
</pre></blockquote></p>
@param mDbgScript the script name]]>
</doc>
</method>
<method name="getMapDebugScript" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the map task's debug script.
@return the debug Script for the mapred job for failed map tasks.
@see #setMapDebugScript(String)]]>
</doc>
</method>
<method name="setReduceDebugScript"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rDbgScript" type="java.lang.String"/>
<doc>
<![CDATA[Set the debug script to run when the reduce tasks fail.
<p>The debug script can aid debugging of failed reduce tasks. The script
is given task's stdout, stderr, syslog, jobconf files as arguments.</p>
<p>The debug command, run on the node where the map failed, is:</p>
<p><pre><blockquote>
$script $stdout $stderr $syslog $jobconf.
</blockquote></pre></p>
<p> The script file is distributed through {@link DistributedCache}
APIs. The script file needs to be symlinked </p>
<p>Here is an example on how to submit a script
<p><blockquote><pre>
job.setReduceDebugScript("./myscript");
DistributedCache.createSymlink(job);
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
</pre></blockquote></p>
@param rDbgScript the script name]]>
</doc>
</method>
<method name="getReduceDebugScript" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reduce task's debug Script
@return the debug script for the mapred job for failed reduce tasks.
@see #setReduceDebugScript(String)]]>
</doc>
</method>
<method name="getJobEndNotificationURI" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
@return the job end notification uri, <code>null</code> if it hasn't
been set.
@see #setJobEndNotificationURI(String)]]>
</doc>
</method>
<method name="setJobEndNotificationURI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.lang.String"/>
<doc>
<![CDATA[Set the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
<p>The uri can contain 2 special parameters: <tt>$jobId</tt> and
<tt>$jobStatus</tt>. Those, if present, are replaced by the job's
identifier and completion-status respectively.</p>
<p>This is typically used by application-writers to implement chaining of
Map-Reduce jobs in an <i>asynchronous manner</i>.</p>
@param uri the job end notification uri
@see JobStatus
@see <a href="{@docRoot}/org/apache/hadoop/mapred/JobClient.html#
JobCompletionAndChaining">Job Completion and Chaining</a>]]>
</doc>
</method>
<method name="getJobLocalDir" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get job-specific shared directory for use as scratch space
<p>
When a job starts, a shared directory is created at location
<code>
${mapreduce.cluster.local.dir}/taskTracker/$user/jobcache/$jobid/work/ </code>.
This directory is exposed to the users through
<code>mapreduce.job.local.dir </code>.
So, the tasks can use this space
as scratch space and share files among them. </p>
This value is available as System property also.
@return The localized job specific shared directory]]>
</doc>
</method>
<method name="getMemoryForMapTask" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get memory required to run a map task of the job, in MB.
If a value is specified in the configuration, it is returned.
Else, it returns {@link #DISABLED_MEMORY_LIMIT}.
<p/>
For backward compatibility, if the job configuration sets the
key {@link #MAPRED_TASK_MAXVMEM_PROPERTY} to a value different
from {@link #DISABLED_MEMORY_LIMIT}, that value will be used
after converting it from bytes to MB.
@return memory required to run a map task of the job, in MB,
or {@link #DISABLED_MEMORY_LIMIT} if unset.]]>
</doc>
</method>
<method name="setMemoryForMapTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mem" type="long"/>
</method>
<method name="getMemoryForReduceTask" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get memory required to run a reduce task of the job, in MB.
If a value is specified in the configuration, it is returned.
Else, it returns {@link #DISABLED_MEMORY_LIMIT}.
<p/>
For backward compatibility, if the job configuration sets the
key {@link #MAPRED_TASK_MAXVMEM_PROPERTY} to a value different
from {@link #DISABLED_MEMORY_LIMIT}, that value will be used
after converting it from bytes to MB.
@return memory required to run a reduce task of the job, in MB,
or {@link #DISABLED_MEMORY_LIMIT} if unset.]]>
</doc>
</method>
<method name="setMemoryForReduceTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mem" type="long"/>
</method>
<method name="getQueueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the name of the queue to which this job is submitted.
Defaults to 'default'.
@return name of the queue]]>
</doc>
</method>
<method name="setQueueName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<doc>
<![CDATA[Set the name of the queue to which this job should be submitted.
@param queueName Name of the queue]]>
</doc>
</method>
<method name="normalizeMemoryConfigValue" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="long"/>
<doc>
<![CDATA[Normalize the negative values in configuration
@param val
@return normalized value]]>
</doc>
</method>
<method name="getMaxVirtualMemoryForTask" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #getMemoryForMapTask()} and
{@link #getMemoryForReduceTask()}">
<doc>
<![CDATA[Get the memory required to run a task of this job, in bytes. See
{@link #MAPRED_TASK_MAXVMEM_PROPERTY}
<p/>
This method is deprecated. Now, different memory limits can be
set for map and reduce tasks of a job, in MB.
<p/>
For backward compatibility, if the job configuration sets the
key {@link #MAPRED_TASK_MAXVMEM_PROPERTY} to a value different
from {@link #DISABLED_MEMORY_LIMIT}, that value is returned.
Otherwise, this method will return the larger of the values returned by
{@link #getMemoryForMapTask()} and {@link #getMemoryForReduceTask()}
after converting them into bytes.
@return Memory required to run a task of this job, in bytes,
or {@link #DISABLED_MEMORY_LIMIT}, if unset.
@see #setMaxVirtualMemoryForTask(long)
@deprecated Use {@link #getMemoryForMapTask()} and
{@link #getMemoryForReduceTask()}]]>
</doc>
</method>
<method name="setMaxVirtualMemoryForTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #setMemoryForMapTask(long mem)} and
Use {@link #setMemoryForReduceTask(long mem)}">
<param name="vmem" type="long"/>
<doc>
<![CDATA[Set the maximum amount of memory any task of this job can use. See
{@link #MAPRED_TASK_MAXVMEM_PROPERTY}
<p/>
mapred.task.maxvmem is split into
mapreduce.map.memory.mb
and mapreduce.map.memory.mb,mapred
each of the new key are set
as mapred.task.maxvmem / 1024
as new values are in MB
@param vmem Maximum amount of virtual memory in bytes any task of this job
can use.
@see #getMaxVirtualMemoryForTask()
@deprecated
Use {@link #setMemoryForMapTask(long mem)} and
Use {@link #setMemoryForReduceTask(long mem)}]]>
</doc>
</method>
<method name="getMaxPhysicalMemoryForTask" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="this variable is deprecated and nolonger in use.">
<doc>
<![CDATA[@deprecated this variable is deprecated and nolonger in use.]]>
</doc>
</method>
<method name="setMaxPhysicalMemoryForTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mem" type="long"/>
</method>
<field name="MAPRED_TASK_MAXVMEM_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="Use {@link #MAPRED_JOB_MAP_MEMORY_MB_PROPERTY} and
{@link #MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY}">
<doc>
<![CDATA[@deprecated Use {@link #MAPRED_JOB_MAP_MEMORY_MB_PROPERTY} and
{@link #MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY}]]>
</doc>
</field>
<field name="UPPER_LIMIT_ON_TASK_VMEM_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</field>
<field name="MAPRED_TASK_DEFAULT_MAXVMEM_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</field>
<field name="MAPRED_TASK_MAXPMEM_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</field>
<field name="DISABLED_MEMORY_LIMIT" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A value which if set for memory related configuration options,
indicates that the options are turned off.]]>
</doc>
</field>
<field name="MAPRED_LOCAL_DIR_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Property name for the configuration property mapreduce.cluster.local.dir]]>
</doc>
</field>
<field name="DEFAULT_QUEUE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Name of the queue to which jobs will be submitted, if no queue
name is mentioned.]]>
</doc>
</field>
<field name="UNPACK_JAR_PATTERN_DEFAULT" type="java.util.regex.Pattern"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Pattern for the default unpacking behavior for job jars]]>
</doc>
</field>
<field name="MAPRED_TASK_JAVA_OPTS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="Use {@link #MAPRED_MAP_TASK_JAVA_OPTS} or
{@link #MAPRED_REDUCE_TASK_JAVA_OPTS}">
<doc>
<![CDATA[Configuration key to set the java command line options for the child
map and reduce tasks.
Java opts for the task tracker child processes.
The following symbol, if present, will be interpolated: @taskid@.
It is replaced by current TaskID. Any other occurrences of '@' will go
unchanged.
For example, to enable verbose gc logging to a file named for the taskid in
/tmp and to set the heap maximum to be a gigabyte, pass a 'value' of:
-Xmx1024m -verbose:gc -Xloggc:/tmp/@taskid@.gc
The configuration variable {@link #MAPRED_TASK_ULIMIT} can be used to
control the maximum virtual memory of the child processes.
The configuration variable {@link #MAPRED_TASK_ENV} can be used to pass
other environment variables to the child processes.
@deprecated Use {@link #MAPRED_MAP_TASK_JAVA_OPTS} or
{@link #MAPRED_REDUCE_TASK_JAVA_OPTS}]]>
</doc>
</field>
<field name="MAPRED_MAP_TASK_JAVA_OPTS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the java command line options for the map tasks.
Java opts for the task tracker child map processes.
The following symbol, if present, will be interpolated: @taskid@.
It is replaced by current TaskID. Any other occurrences of '@' will go
unchanged.
For example, to enable verbose gc logging to a file named for the taskid in
/tmp and to set the heap maximum to be a gigabyte, pass a 'value' of:
-Xmx1024m -verbose:gc -Xloggc:/tmp/@taskid@.gc
The configuration variable {@link #MAPRED_MAP_TASK_ULIMIT} can be used to
control the maximum virtual memory of the map processes.
The configuration variable {@link #MAPRED_MAP_TASK_ENV} can be used to pass
other environment variables to the map processes.]]>
</doc>
</field>
<field name="MAPRED_REDUCE_TASK_JAVA_OPTS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the java command line options for the reduce tasks.
Java opts for the task tracker child reduce processes.
The following symbol, if present, will be interpolated: @taskid@.
It is replaced by current TaskID. Any other occurrences of '@' will go
unchanged.
For example, to enable verbose gc logging to a file named for the taskid in
/tmp and to set the heap maximum to be a gigabyte, pass a 'value' of:
-Xmx1024m -verbose:gc -Xloggc:/tmp/@taskid@.gc
The configuration variable {@link #MAPRED_REDUCE_TASK_ULIMIT} can be used
to control the maximum virtual memory of the reduce processes.
The configuration variable {@link #MAPRED_REDUCE_TASK_ENV} can be used to
pass process environment variables to the reduce processes.]]>
</doc>
</field>
<field name="DEFAULT_MAPRED_TASK_JAVA_OPTS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAPRED_TASK_ULIMIT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="Use {@link #MAPRED_MAP_TASK_ULIMIT} or
{@link #MAPRED_REDUCE_TASK_ULIMIT}">
<doc>
<![CDATA[Configuration key to set the maximum virutal memory available to the child
map and reduce tasks (in kilo-bytes).
Note: This must be greater than or equal to the -Xmx passed to the JavaVM
via {@link #MAPRED_TASK_JAVA_OPTS}, else the VM might not start.
@deprecated Use {@link #MAPRED_MAP_TASK_ULIMIT} or
{@link #MAPRED_REDUCE_TASK_ULIMIT}]]>
</doc>
</field>
<field name="MAPRED_MAP_TASK_ULIMIT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the maximum virutal memory available to the
map tasks (in kilo-bytes).
Note: This must be greater than or equal to the -Xmx passed to the JavaVM
via {@link #MAPRED_MAP_TASK_JAVA_OPTS}, else the VM might not start.]]>
</doc>
</field>
<field name="MAPRED_REDUCE_TASK_ULIMIT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the maximum virutal memory available to the
reduce tasks (in kilo-bytes).
Note: This must be greater than or equal to the -Xmx passed to the JavaVM
via {@link #MAPRED_REDUCE_TASK_JAVA_OPTS}, else the VM might not start.]]>
</doc>
</field>
<field name="MAPRED_TASK_ENV" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="Use {@link #MAPRED_MAP_TASK_ENV} or
{@link #MAPRED_REDUCE_TASK_ENV}">
<doc>
<![CDATA[Configuration key to set the environment of the child map/reduce tasks.
The format of the value is <code>k1=v1,k2=v2</code>. Further it can
reference existing environment variables via <code>$key</code>.
Example:
<ul>
<li> A=foo - This will set the env variable A to foo. </li>
<li> B=$X:c This is inherit tasktracker's X env variable. </li>
</ul>
@deprecated Use {@link #MAPRED_MAP_TASK_ENV} or
{@link #MAPRED_REDUCE_TASK_ENV}]]>
</doc>
</field>
<field name="MAPRED_MAP_TASK_ENV" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the maximum virutal memory available to the
map tasks.
The format of the value is <code>k1=v1,k2=v2</code>. Further it can
reference existing environment variables via <code>$key</code>.
Example:
<ul>
<li> A=foo - This will set the env variable A to foo. </li>
<li> B=$X:c This is inherit tasktracker's X env variable. </li>
</ul>]]>
</doc>
</field>
<field name="MAPRED_REDUCE_TASK_ENV" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the maximum virutal memory available to the
reduce tasks.
The format of the value is <code>k1=v1,k2=v2</code>. Further it can
reference existing environment variables via <code>$key</code>.
Example:
<ul>
<li> A=foo - This will set the env variable A to foo. </li>
<li> B=$X:c This is inherit tasktracker's X env variable. </li>
</ul>]]>
</doc>
</field>
<field name="MAPRED_MAP_TASK_LOG_LEVEL" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the logging {@link Level} for the map task.
The allowed logging levels are:
OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL.]]>
</doc>
</field>
<field name="MAPRED_REDUCE_TASK_LOG_LEVEL" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set the logging {@link Level} for the reduce task.
The allowed logging levels are:
OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL.]]>
</doc>
</field>
<field name="DEFAULT_LOG_LEVEL" type="org.apache.log4j.Level"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default logging level for map/reduce tasks.]]>
</doc>
</field>
<doc>
<![CDATA[A map/reduce job configuration.
<p><code>JobConf</code> is the primary interface for a user to describe a
map-reduce job to the Hadoop framework for execution. The framework tries to
faithfully execute the job as-is described by <code>JobConf</code>, however:
<ol>
<li>
Some configuration parameters might have been marked as
<a href="{@docRoot}/org/apache/hadoop/conf/Configuration.html#FinalParams">
final</a> by administrators and hence cannot be altered.
</li>
<li>
While some job parameters are straight-forward to set
(e.g. {@link #setNumReduceTasks(int)}), some parameters interact subtly
rest of the framework and/or job-configuration and is relatively more
complex for the user to control finely (e.g. {@link #setNumMapTasks(int)}).
</li>
</ol></p>
<p><code>JobConf</code> typically specifies the {@link Mapper}, combiner
(if any), {@link Partitioner}, {@link Reducer}, {@link InputFormat} and
{@link OutputFormat} implementations to be used etc.
<p>Optionally <code>JobConf</code> is used to specify other advanced facets
of the job such as <code>Comparator</code>s to be used, files to be put in
the {@link DistributedCache}, whether or not intermediate and/or job outputs
are to be compressed (and how), debugability via user-provided scripts
( {@link #setMapDebugScript(String)}/{@link #setReduceDebugScript(String)}),
for doing post-processing on task logs, task's stdout, stderr, syslog.
and etc.</p>
<p>Here is an example on how to configure a job via <code>JobConf</code>:</p>
<p><blockquote><pre>
// Create a new JobConf
JobConf job = new JobConf(new Configuration(), MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
FileInputFormat.setInputPaths(job, new Path("in"));
FileOutputFormat.setOutputPath(job, new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setCombinerClass(MyJob.MyReducer.class);
job.setReducerClass(MyJob.MyReducer.class);
job.setInputFormat(SequenceFileInputFormat.class);
job.setOutputFormat(SequenceFileOutputFormat.class);
</pre></blockquote></p>
@see JobClient
@see ClusterStatus
@see Tool
@see DistributedCache
@deprecated Use {@link Configuration} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobConf -->
<!-- start interface org.apache.hadoop.mapred.JobConfigurable -->
<interface name="JobConfigurable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Initializes a new instance from a {@link JobConf}.
@param job the configuration]]>
</doc>
</method>
<doc>
<![CDATA[That what may be configured.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.JobConfigurable -->
<!-- start interface org.apache.hadoop.mapred.JobContext -->
<interface name="JobContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.JobContext} instead.">
<implements name="org.apache.hadoop.mapreduce.JobContext"/>
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job Configuration
@return JobConf]]>
</doc>
</method>
<method name="getProgressible" return="org.apache.hadoop.util.Progressable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the progress mechanism for reporting progress.
@return progress mechanism]]>
</doc>
</method>
<doc>
<![CDATA[@deprecated Use {@link org.apache.hadoop.mapreduce.JobContext} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.JobContext -->
<!-- start class org.apache.hadoop.mapred.JobID -->
<class name="JobID" extends="org.apache.hadoop.mapreduce.JobID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobID" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a JobID object
@param jtIdentifier jobTracker identifier
@param id job number]]>
</doc>
</constructor>
<constructor name="JobID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="downgrade" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="old" type="org.apache.hadoop.mapreduce.JobID"/>
<doc>
<![CDATA[Downgrade a new JobID to an old one
@param old a new or old JobID
@return either old or a new JobID build to match old]]>
</doc>
</method>
<method name="read" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="forName" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<doc>
<![CDATA[Construct a JobId object from given string
@return constructed JobId object or null if the given String is null
@throws IllegalArgumentException if the given string is malformed]]>
</doc>
</method>
<method name="getJobIDsPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jtIdentifier" type="java.lang.String"/>
<param name="jobId" type="java.lang.Integer"/>
<doc>
<![CDATA[Returns a regex pattern which matches task IDs. Arguments can
be given null, in which case that part of the regex will be generic.
For example to obtain a regex matching <i>any job</i>
run on the jobtracker started at <i>200707121733</i>, we would use :
<pre>
JobID.getTaskIDsPattern("200707121733", null);
</pre>
which will return :
<pre> "job_200707121733_[0-9]*" </pre>
@param jtIdentifier jobTracker identifier, or null
@param jobId job number, or null
@return a regex pattern matching JobIDs]]>
</doc>
</method>
<doc>
<![CDATA[JobID represents the immutable and unique identifier for
the job. JobID consists of two parts. First part
represents the jobtracker identifier, so that jobID to jobtracker map
is defined. For cluster setup this string is the jobtracker
start time, for local setting, it is "local".
Second part of the JobID is the job number. <br>
An example JobID is :
<code>job_200707121733_0003</code> , which represents the third job
running at the jobtracker started at <code>200707121733</code>.
<p>
Applications should never construct or parse JobID strings, but rather
use appropriate constructors or {@link #forName(String)} method.
@see TaskID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobID -->
<!-- start class org.apache.hadoop.mapred.JobPriority -->
<class name="JobPriority" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.JobPriority} instead">
<method name="values" return="org.apache.hadoop.mapred.JobPriority[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobPriority"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Used to describe the priority of the running job.
@deprecated Use {@link org.apache.hadoop.mapreduce.JobPriority} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobPriority -->
<!-- start class org.apache.hadoop.mapred.JobQueueInfo -->
<class name="JobQueueInfo" extends="org.apache.hadoop.mapreduce.QueueInfo"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link QueueInfo} instead">
<constructor name="JobQueueInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for Job Queue Info.]]>
</doc>
</constructor>
<constructor name="JobQueueInfo" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a new JobQueueInfo object using the queue name and the
scheduling information passed.
@param queueName Name of the job queue
@param schedulingInfo Scheduling Information associated with the job
queue]]>
</doc>
</constructor>
<method name="setQueueName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<doc>
<![CDATA[Set the queue name of the JobQueueInfo
@param queueName Name of the job queue.]]>
</doc>
</method>
<method name="setSchedulingInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="schedulingInfo" type="java.lang.String"/>
<doc>
<![CDATA[Set the scheduling information associated to particular job queue
@param schedulingInfo]]>
</doc>
</method>
<method name="setQueueState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="java.lang.String"/>
<doc>
<![CDATA[Set the state of the queue
@param state state of the queue.]]>
</doc>
</method>
<method name="setChildren"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="children" type="java.util.List"/>
</method>
<method name="getChildren" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setProperties"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="props" type="java.util.Properties"/>
</method>
<method name="setJobStatuses"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="stats" type="org.apache.hadoop.mapreduce.JobStatus[]"/>
</method>
<doc>
<![CDATA[Class that contains the information regarding the Job Queues which are
maintained by the Hadoop Map/Reduce framework.
@deprecated Use {@link QueueInfo} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobQueueInfo -->
<!-- start class org.apache.hadoop.mapred.JobStatus -->
<class name="JobStatus" extends="org.apache.hadoop.mapreduce.JobStatus"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.JobStatus} instead">
<constructor name="JobStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, int, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param cleanupProgress The progress made on cleanup
@param runState The current state of the job
@param user userid of the person who submitted the job.
@param jobName user-specified job name.
@param jobFile job configuration file.
@param trackingUrl link to the web-ui for details of the job.]]>
</doc>
</constructor>
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, int, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param runState The current state of the job
@param user userid of the person who submitted the job.
@param jobName user-specified job name.
@param jobFile job configuration file.
@param trackingUrl link to the web-ui for details of the job.]]>
</doc>
</constructor>
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, int, org.apache.hadoop.mapred.JobPriority, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param runState The current state of the job
@param jp Priority of the job.
@param user userid of the person who submitted the job.
@param jobName user-specified job name.
@param jobFile job configuration file.
@param trackingUrl link to the web-ui for details of the job.]]>
</doc>
</constructor>
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, float, int, org.apache.hadoop.mapred.JobPriority, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param setupProgress The progress made on the setup
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param cleanupProgress The progress made on the cleanup
@param runState The current state of the job
@param jp Priority of the job.
@param user userid of the person who submitted the job.
@param jobName user-specified job name.
@param jobFile job configuration file.
@param trackingUrl link to the web-ui for details of the job.]]>
</doc>
</constructor>
<method name="getJobRunState" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="state" type="int"/>
<doc>
<![CDATA[Helper method to get human-readable state of the job.
@param state job state
@return human-readable state of the job]]>
</doc>
</method>
<method name="downgrade" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stat" type="org.apache.hadoop.mapreduce.JobStatus"/>
</method>
<method name="getJobId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="use getJobID instead">
<doc>
<![CDATA[@deprecated use getJobID instead]]>
</doc>
</method>
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return The jobid of the Job]]>
</doc>
</method>
<method name="getJobPriority" return="org.apache.hadoop.mapred.JobPriority"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the priority of the job
@return job priority]]>
</doc>
</method>
<method name="setMapProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the map progress of this job
@param p The value of map progress to set to]]>
</doc>
</method>
<method name="setCleanupProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the cleanup progress of this job
@param p The value of cleanup progress to set to]]>
</doc>
</method>
<method name="setSetupProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the setup progress of this job
@param p The value of setup progress to set to]]>
</doc>
</method>
<method name="setReduceProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the reduce progress of this Job
@param p The value of reduce progress to set to]]>
</doc>
</method>
<method name="setFinishTime"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="finishTime" type="long"/>
<doc>
<![CDATA[Set the finish time of the job
@param finishTime The finishTime of the job]]>
</doc>
</method>
<method name="setHistoryFile"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="historyFile" type="java.lang.String"/>
<doc>
<![CDATA[Set the job history file url for a completed job]]>
</doc>
</method>
<method name="setTrackingUrl"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="trackingUrl" type="java.lang.String"/>
<doc>
<![CDATA[Set the link to the web-ui for details of the job.]]>
</doc>
</method>
<method name="setRetired"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Set the job retire flag to true.]]>
</doc>
</method>
<method name="setRunState"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="int"/>
<doc>
<![CDATA[Change the current run state of the job.]]>
</doc>
</method>
<method name="getRunState" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return running state of the job]]>
</doc>
</method>
<method name="setStartTime"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="startTime" type="long"/>
<doc>
<![CDATA[Set the start time of the job
@param startTime The startTime of the job]]>
</doc>
</method>
<method name="setUsername"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="userName" type="java.lang.String"/>
<doc>
<![CDATA[@param userName The username of the job]]>
</doc>
</method>
<method name="setSchedulingInfo"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="schedulingInfo" type="java.lang.String"/>
<doc>
<![CDATA[Used to set the scheduling information associated to a particular Job.
@param schedulingInfo Scheduling information of the job]]>
</doc>
</method>
<method name="setJobACLs"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="acls" type="java.util.Map"/>
</method>
<method name="setJobPriority"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jp" type="org.apache.hadoop.mapred.JobPriority"/>
<doc>
<![CDATA[Set the priority of the job, defaulting to NORMAL.
@param jp new job priority]]>
</doc>
</method>
<method name="mapProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in maps]]>
</doc>
</method>
<method name="cleanupProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in cleanup]]>
</doc>
</method>
<method name="setupProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in setup]]>
</doc>
</method>
<method name="reduceProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in reduce]]>
</doc>
</method>
<field name="RUNNING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SUCCEEDED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PREP" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="KILLED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Describes the current status of a job. This is
not intended to be a comprehensive piece of data.
For that, look at JobProfile.
@deprecated Use {@link org.apache.hadoop.mapreduce.JobStatus} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobStatus -->
<!-- start class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
<class name="KeyValueLineRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader}
instead">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="KeyValueLineRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="findSeparator" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<param name="sep" type="byte"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class treats a line in the input as a key/value pair separated by a
separator character. The separator can be specified in config file
under the attribute name mapreduce.input.keyvaluelinerecordreader.key.value.separator. The default
separator is the tab character ('\t').
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
<!-- start class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
<class name="KeyValueTextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat}
instead">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="KeyValueTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either linefeed or carriage-return are used to signal end of line. Each line
is divided into key and value parts by a separator byte. If no such a byte
exists, the key will be the entire line and value will be empty.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
<!-- start class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
<class name="LineRecordReader.LineReader" extends="org.apache.hadoop.util.LineReader"
abstract="false"
static="true" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.util.LineReader} instead.">
<constructor name="LineRecordReader.LineReader" type="java.io.InputStream, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<doc>
<![CDATA[A class that provides a line reader from an input stream.
@deprecated Use {@link org.apache.hadoop.util.LineReader} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
<!-- start class org.apache.hadoop.mapred.MapFileOutputFormat -->
<class name="MapFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat} instead">
<constructor name="MapFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReaders" return="org.apache.hadoop.io.MapFile.Reader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the output generated by this format.]]>
</doc>
</method>
<method name="getEntry" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="readers" type="org.apache.hadoop.io.MapFile.Reader[]"/>
<param name="partitioner" type="org.apache.hadoop.mapred.Partitioner"/>
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an entry from output generated by this class.]]>
</doc>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes {@link MapFile}s.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapFileOutputFormat -->
<!-- start interface org.apache.hadoop.mapred.Mapper -->
<interface name="Mapper" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<implements name="org.apache.hadoop.io.Closeable"/>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K1"/>
<param name="value" type="V1"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Maps a single input key/value pair into an intermediate key/value pair.
<p>Output pairs need not be of the same types as input pairs. A given
input pair may map to zero or many output pairs. Output pairs are
collected with calls to
{@link OutputCollector#collect(Object,Object)}.</p>
<p>Applications can use the {@link Reporter} provided to report progress
or just indicate that they are alive. In scenarios where the application
takes an insignificant amount of time to process individual key/value
pairs, this is crucial since the framework might assume that the task has
timed-out and kill that task. The other way of avoiding this is to set
<a href="{@docRoot}/../mapred-default.html#mapreduce.task.timeout">
mapreduce.task.timeout</a> to a high-enough value (or even zero for no
time-outs).</p>
@param key the input key.
@param value the input value.
@param output collects mapped keys and values.
@param reporter facility to report progress.]]>
</doc>
</method>
<doc>
<![CDATA[Maps input key/value pairs to a set of intermediate key/value pairs.
<p>Maps are the individual tasks which transform input records into a
intermediate records. The transformed intermediate records need not be of
the same type as the input records. A given input pair may map to zero or
many output pairs.</p>
<p>The Hadoop Map-Reduce framework spawns one map task for each
{@link InputSplit} generated by the {@link InputFormat} for the job.
<code>Mapper</code> implementations can access the {@link JobConf} for the
job via the {@link JobConfigurable#configure(JobConf)} and initialize
themselves. Similarly they can use the {@link Closeable#close()} method for
de-initialization.</p>
<p>The framework then calls
{@link #map(Object, Object, OutputCollector, Reporter)}
for each key/value pair in the <code>InputSplit</code> for that task.</p>
<p>All intermediate values associated with a given output key are
subsequently grouped by the framework, and passed to a {@link Reducer} to
determine the final output. Users can control the grouping by specifying
a <code>Comparator</code> via
{@link JobConf#setOutputKeyComparatorClass(Class)}.</p>
<p>The grouped <code>Mapper</code> outputs are partitioned per
<code>Reducer</code>. Users can control which keys (and hence records) go to
which <code>Reducer</code> by implementing a custom {@link Partitioner}.
<p>Users can optionally specify a <code>combiner</code>, via
{@link JobConf#setCombinerClass(Class)}, to perform local aggregation of the
intermediate outputs, which helps to cut down the amount of data transferred
from the <code>Mapper</code> to the <code>Reducer</code>.
<p>The intermediate, grouped outputs are always stored in
{@link SequenceFile}s. Applications can specify if and how the intermediate
outputs are to be compressed and which {@link CompressionCodec}s are to be
used via the <code>JobConf</code>.</p>
<p>If the job has
<a href="{@docRoot}/org/apache/hadoop/mapred/JobConf.html#ReducerNone">zero
reduces</a> then the output of the <code>Mapper</code> is directly written
to the {@link FileSystem} without grouping by keys.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyMapper&lt;K extends WritableComparable, V extends Writable&gt;
extends MapReduceBase implements Mapper&lt;K, V, K, V&gt; {
static enum MyCounters { NUM_RECORDS }
private String mapTaskId;
private String inputFile;
private int noRecords = 0;
public void configure(JobConf job) {
mapTaskId = job.get(JobContext.TASK_ATTEMPT_ID);
inputFile = job.get(JobContext.MAP_INPUT_FILE);
}
public void map(K key, V val,
OutputCollector&lt;K, V&gt; output, Reporter reporter)
throws IOException {
// Process the &lt;key, value&gt; pair (assume this takes a while)
// ...
// ...
// Let the framework know that we are alive, and kicking!
// reporter.progress();
// Process some more
// ...
// ...
// Increment the no. of &lt;key, value&gt; pairs processed
++noRecords;
// Increment counters
reporter.incrCounter(NUM_RECORDS, 1);
// Every 100 records update application-level status
if ((noRecords%100) == 0) {
reporter.setStatus(mapTaskId + " processed " + noRecords +
" from input-file: " + inputFile);
}
// Output the result
output.collect(key, val);
}
}
</pre></blockquote></p>
<p>Applications may write a custom {@link MapRunnable} to exert greater
control on map processing e.g. multi-threaded <code>Mapper</code>s etc.</p>
@see JobConf
@see InputFormat
@see Partitioner
@see Reducer
@see MapReduceBase
@see MapRunnable
@see SequenceFile
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Mapper -->
<!-- start class org.apache.hadoop.mapred.MapReduceBase -->
<class name="MapReduceBase" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Closeable"/>
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="MapReduceBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Default implementation that does nothing.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Default implementation that does nothing.]]>
</doc>
</method>
<doc>
<![CDATA[Base class for {@link Mapper} and {@link Reducer} implementations.
<p>Provides default no-op implementations for a few methods, most non-trivial
applications need to override some of them.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapReduceBase -->
<!-- start interface org.apache.hadoop.mapred.MapRunnable -->
<interface name="MapRunnable" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Start mapping input <tt>&lt;key, value&gt;</tt> pairs.
<p>Mapping of input records to output records is complete when this method
returns.</p>
@param input the {@link RecordReader} to read the input records.
@param output the {@link OutputCollector} to collect the outputrecords.
@param reporter {@link Reporter} to report progress, status-updates etc.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Expert: Generic interface for {@link Mapper}s.
<p>Custom implementations of <code>MapRunnable</code> can exert greater
control on map processing e.g. multi-threaded, asynchronous mappers etc.</p>
@see Mapper
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.MapRunnable -->
<!-- start class org.apache.hadoop.mapred.MapRunner -->
<class name="MapRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MapRunnable"/>
<constructor name="MapRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getMapper" return="org.apache.hadoop.mapred.Mapper"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Default {@link MapRunnable} implementation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapRunner -->
<!-- start class org.apache.hadoop.mapred.MultiFileInputFormat -->
<class name="MultiFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapred.lib.CombineFileInputFormat} instead">
<constructor name="MultiFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An abstract {@link InputFormat} that returns {@link MultiFileSplit}'s
in {@link #getSplits(JobConf, int)} method. Splits are constructed from
the files under the input paths. Each split returned contains <i>nearly</i>
equal content length. <br>
Subclasses implement {@link #getRecordReader(InputSplit, JobConf, Reporter)}
to construct <code>RecordReader</code>'s for <code>MultiFileSplit</code>'s.
@see MultiFileSplit
@deprecated Use {@link org.apache.hadoop.mapred.lib.CombineFileInputFormat} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MultiFileInputFormat -->
<!-- start class org.apache.hadoop.mapred.MultiFileSplit -->
<class name="MultiFileSplit" extends="org.apache.hadoop.mapred.lib.CombineFileSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapred.lib.CombineFileSplit} instead">
<constructor name="MultiFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A sub-collection of input files. Unlike {@link FileSplit}, MultiFileSplit
class does not represent a split of a file, but a split of input files
into smaller sets. The atomic unit of split is a file. <br>
MultiFileSplit can be used to implement {@link RecordReader}'s, with
reading one record per file.
@see FileSplit
@see MultiFileInputFormat
@deprecated Use {@link org.apache.hadoop.mapred.lib.CombineFileSplit} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MultiFileSplit -->
<!-- start interface org.apache.hadoop.mapred.OutputCollector -->
<interface name="OutputCollector" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="collect"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Adds a key/value pair to the output.
@param key the key to collect.
@param value to value to collect.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Collects the <code>&lt;key, value&gt;</code> pairs output by {@link Mapper}s
and {@link Reducer}s.
<p><code>OutputCollector</code> is the generalization of the facility
provided by the Map-Reduce framework to collect data output by either the
<code>Mapper</code> or the <code>Reducer</code> i.e. intermediate outputs
or the output of the job.</p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.OutputCollector -->
<!-- start class org.apache.hadoop.mapred.OutputCommitter -->
<class name="OutputCommitter" extends="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.OutputCommitter} instead.">
<constructor name="OutputCommitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setupJob"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For the framework to setup the job output during initialization
@param jobContext Context of the job whose output is being written.
@throws IOException if temporary output could not be created]]>
</doc>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #commitJob(JobContext)} or
{@link #abortJob(JobContext, int)} instead.">
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For cleaning up the job's output after job completion
@param jobContext Context of the job whose output is being written.
@throws IOException
@deprecated Use {@link #commitJob(JobContext)} or
{@link #abortJob(JobContext, int)} instead.]]>
</doc>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For committing job's output after successful job completion. Note that this
is invoked for jobs with final runstate as SUCCESSFUL.
@param jobContext Context of the job whose output is being written.
@throws IOException]]>
</doc>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
<param name="status" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For aborting an unsuccessful job's output. Note that this is invoked for
jobs with final runstate as {@link JobStatus#FAILED} or
{@link JobStatus#KILLED}
@param jobContext Context of the job whose output is being written.
@param status final runstate of the job
@throws IOException]]>
</doc>
</method>
<method name="setupTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sets up output for the task.
@param taskContext Context of the task whose output is being written.
@throws IOException]]>
</doc>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check whether task needs a commit
@param taskContext
@return true/false
@throws IOException]]>
</doc>
</method>
<method name="commitTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[To promote the task's temporary output to final output location
The task's output is moved to the job's output directory.
@param taskContext Context of the task whose output is being written.
@throws IOException if commit is not]]>
</doc>
</method>
<method name="abortTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Discard the task output
@param taskContext
@throws IOException]]>
</doc>
</method>
<method name="setupJob"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="Use {@link #commitJob(org.apache.hadoop.mapreduce.JobContext)}
or {@link #abortJob(org.apache.hadoop.mapreduce.JobContext, org.apache.hadoop.mapreduce.JobStatus.State)}
instead.">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.
@deprecated Use {@link #commitJob(org.apache.hadoop.mapreduce.JobContext)}
or {@link #abortJob(org.apache.hadoop.mapreduce.JobContext, org.apache.hadoop.mapreduce.JobStatus.State)}
instead.]]>
</doc>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="runState" type="org.apache.hadoop.mapreduce.JobStatus.State"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="setupTask"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="commitTask"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<method name="abortTask"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method implements the new interface by calling the old method. Note
that the input types are different between the new and old apis and this
is a bridge between the two.]]>
</doc>
</method>
<doc>
<![CDATA[<code>OutputCommitter</code> describes the commit of task output for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>OutputCommitter</code> of
the job to:<p>
<ol>
<li>
Setup the job during initialization. For example, create the temporary
output directory for the job during the initialization of the job.
</li>
<li>
Cleanup the job after the job completion. For example, remove the
temporary output directory after the job completion.
</li>
<li>
Setup the task temporary output.
</li>
<li>
Check whether a task needs a commit. This is to avoid the commit
procedure if a task does not need commit.
</li>
<li>
Commit of the task output.
</li>
<li>
Discard the task commit.
</li>
</ol>
@see FileOutputCommitter
@see JobContext
@see TaskAttemptContext
@deprecated Use {@link org.apache.hadoop.mapreduce.OutputCommitter} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.OutputCommitter -->
<!-- start interface org.apache.hadoop.mapred.OutputFormat -->
<interface name="OutputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.OutputFormat} instead.">
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the {@link RecordWriter} for the given job.
@param ignored
@param job configuration for the job whose output is being written.
@param name the unique name for this part of the output.
@param progress mechanism for reporting progress while writing to file.
@return a {@link RecordWriter} to write the output for the job.
@throws IOException]]>
</doc>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check for validity of the output-specification for the job.
<p>This is to validate the output specification for the job when it is
a job is submitted. Typically checks that it does not already exist,
throwing an exception when it already exists, so that output is not
overwritten.</p>
@param ignored
@param job job configuration.
@throws IOException when output should not be attempted]]>
</doc>
</method>
<doc>
<![CDATA[<code>OutputFormat</code> describes the output-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>OutputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the output-specification of the job. For e.g. check that the
output directory doesn't already exist.
<li>
Provide the {@link RecordWriter} implementation to be used to write out
the output files of the job. Output files are stored in a
{@link FileSystem}.
</li>
</ol>
@see RecordWriter
@see JobConf
@deprecated Use {@link org.apache.hadoop.mapreduce.OutputFormat} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.OutputFormat -->
<!-- start class org.apache.hadoop.mapred.OutputLogFilter -->
<class name="OutputLogFilter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter}
instead.">
<implements name="org.apache.hadoop.fs.PathFilter"/>
<constructor name="OutputLogFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
</method>
<doc>
<![CDATA[This class filters log files from directory given
It doesnt accept paths having _logs.
This can be used to list paths of output directory as follows:
Path[] fileList = FileUtil.stat2Paths(fs.listStatus(outDir,
new OutputLogFilter()));
@deprecated Use
{@link org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.OutputLogFilter -->
<!-- start interface org.apache.hadoop.mapred.Partitioner -->
<interface name="Partitioner" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Partitioner} instead.">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numPartitions" type="int"/>
<doc>
<![CDATA[Get the paritition number for a given key (hence record) given the total
number of partitions i.e. number of reduce-tasks for the job.
<p>Typically a hash function on a all or a subset of the key.</p>
@param key the key to be paritioned.
@param value the entry value.
@param numPartitions the total number of partitions.
@return the partition number for the <code>key</code>.]]>
</doc>
</method>
<doc>
<![CDATA[Partitions the key space.
<p><code>Partitioner</code> controls the partitioning of the keys of the
intermediate map-outputs. The key (or a subset of the key) is used to derive
the partition, typically by a hash function. The total number of partitions
is the same as the number of reduce tasks for the job. Hence this controls
which of the <code>m</code> reduce tasks the intermediate key (and hence the
record) is sent for reduction.</p>
@see Reducer
@deprecated Use {@link org.apache.hadoop.mapreduce.Partitioner} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Partitioner -->
<!-- start interface org.apache.hadoop.mapred.RecordReader -->
<interface name="RecordReader" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads the next key/value pair from the input for processing.
@param key the key to read data into
@param value the value to read data into
@return true iff a key/value was read, false if at EOF]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an object of the appropriate type to be used as a key.
@return a new key object.]]>
</doc>
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an object of the appropriate type to be used as a value.
@return a new value object.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the current position in the input.
@return the current position in the input.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close this {@link InputSplit} to future operations.
@throws IOException]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[How much of the input has the {@link RecordReader} consumed i.e.
has been processed by?
@return progress from <code>0.0</code> to <code>1.0</code>.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RecordReader</code> reads &lt;key, value&gt; pairs from an
{@link InputSplit}.
<p><code>RecordReader</code>, typically, converts the byte-oriented view of
the input, provided by the <code>InputSplit</code>, and presents a
record-oriented view for the {@link Mapper} & {@link Reducer} tasks for
processing. It thus assumes the responsibility of processing record
boundaries and presenting the tasks with keys and values.</p>
@see InputSplit
@see InputFormat]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RecordReader -->
<!-- start interface org.apache.hadoop.mapred.RecordWriter -->
<interface name="RecordWriter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes a key/value pair.
@param key the key to write.
@param value the value to write.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close this <code>RecordWriter</code> to future operations.
@param reporter facility to report progress.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RecordWriter</code> writes the output &lt;key, value&gt; pairs
to an output file.
<p><code>RecordWriter</code> implementations write the job outputs to the
{@link FileSystem}.
@see OutputFormat]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RecordWriter -->
<!-- start interface org.apache.hadoop.mapred.Reducer -->
<interface name="Reducer" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Reducer} instead.">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<implements name="org.apache.hadoop.io.Closeable"/>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<i>Reduces</i> values for a given key.
<p>The framework calls this method for each
<code>&lt;key, (list of values)></code> pair in the grouped inputs.
Output values must be of the same type as input values. Input keys must
not be altered. The framework will <b>reuse</b> the key and value objects
that are passed into the reduce, therefore the application should clone
the objects they want to keep a copy of. In many cases, all values are
combined into zero or one value.
</p>
<p>Output pairs are collected with calls to
{@link OutputCollector#collect(Object,Object)}.</p>
<p>Applications can use the {@link Reporter} provided to report progress
or just indicate that they are alive. In scenarios where the application
takes an insignificant amount of time to process individual key/value
pairs, this is crucial since the framework might assume that the task has
timed-out and kill that task. The other way of avoiding this is to set
<a href="{@docRoot}/../mapred-default.html#mapreduce.task.timeout">
mapreduce.task.timeout</a> to a high-enough value (or even zero for no
time-outs).</p>
@param key the key.
@param values the list of values to reduce.
@param output to collect keys and combined values.
@param reporter facility to report progress.]]>
</doc>
</method>
<doc>
<![CDATA[Reduces a set of intermediate values which share a key to a smaller set of
values.
<p>The number of <code>Reducer</code>s for the job is set by the user via
{@link JobConf#setNumReduceTasks(int)}. <code>Reducer</code> implementations
can access the {@link JobConf} for the job via the
{@link JobConfigurable#configure(JobConf)} method and initialize themselves.
Similarly they can use the {@link Closeable#close()} method for
de-initialization.</p>
<p><code>Reducer</code> has 3 primary phases:</p>
<ol>
<li>
<h4 id="Shuffle">Shuffle</h4>
<p><code>Reducer</code> is input the grouped output of a {@link Mapper}.
In the phase the framework, for each <code>Reducer</code>, fetches the
relevant partition of the output of all the <code>Mapper</code>s, via HTTP.
</p>
</li>
<li>
<h4 id="Sort">Sort</h4>
<p>The framework groups <code>Reducer</code> inputs by <code>key</code>s
(since different <code>Mapper</code>s may have output the same key) in this
stage.</p>
<p>The shuffle and sort phases occur simultaneously i.e. while outputs are
being fetched they are merged.</p>
<h5 id="SecondarySort">SecondarySort</h5>
<p>If equivalence rules for keys while grouping the intermediates are
different from those for grouping keys before reduction, then one may
specify a <code>Comparator</code> via
{@link JobConf#setOutputValueGroupingComparator(Class)}.Since
{@link JobConf#setOutputKeyComparatorClass(Class)} can be used to
control how intermediate keys are grouped, these can be used in conjunction
to simulate <i>secondary sort on values</i>.</p>
For example, say that you want to find duplicate web pages and tag them
all with the url of the "best" known example. You would set up the job
like:
<ul>
<li>Map Input Key: url</li>
<li>Map Input Value: document</li>
<li>Map Output Key: document checksum, url pagerank</li>
<li>Map Output Value: url</li>
<li>Partitioner: by checksum</li>
<li>OutputKeyComparator: by checksum and then decreasing pagerank</li>
<li>OutputValueGroupingComparator: by checksum</li>
</ul>
</li>
<li>
<h4 id="Reduce">Reduce</h4>
<p>In this phase the
{@link #reduce(Object, Iterator, OutputCollector, Reporter)}
method is called for each <code>&lt;key, (list of values)></code> pair in
the grouped inputs.</p>
<p>The output of the reduce task is typically written to the
{@link FileSystem} via
{@link OutputCollector#collect(Object, Object)}.</p>
</li>
</ol>
<p>The output of the <code>Reducer</code> is <b>not re-sorted</b>.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyReducer&lt;K extends WritableComparable, V extends Writable&gt;
extends MapReduceBase implements Reducer&lt;K, V, K, V&gt; {
static enum MyCounters { NUM_RECORDS }
private String reduceTaskId;
private int noKeys = 0;
public void configure(JobConf job) {
reduceTaskId = job.get(JobContext.TASK_ATTEMPT_ID);
}
public void reduce(K key, Iterator&lt;V&gt; values,
OutputCollector&lt;K, V&gt; output,
Reporter reporter)
throws IOException {
// Process
int noValues = 0;
while (values.hasNext()) {
V value = values.next();
// Increment the no. of values for this key
++noValues;
// Process the &lt;key, value&gt; pair (assume this takes a while)
// ...
// ...
// Let the framework know that we are alive, and kicking!
if ((noValues%10) == 0) {
reporter.progress();
}
// Process some more
// ...
// ...
// Output the &lt;key, value&gt;
output.collect(key, value);
}
// Increment the no. of &lt;key, list of values&gt; pairs processed
++noKeys;
// Increment counters
reporter.incrCounter(NUM_RECORDS, 1);
// Every 100 keys update application-level status
if ((noKeys%100) == 0) {
reporter.setStatus(reduceTaskId + " processed " + noKeys);
}
}
}
</pre></blockquote></p>
@see Mapper
@see Partitioner
@see Reporter
@see MapReduceBase
@deprecated Use {@link org.apache.hadoop.mapreduce.Reducer} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Reducer -->
<!-- start interface org.apache.hadoop.mapred.Reporter -->
<interface name="Reporter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Progressable"/>
<method name="setStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="java.lang.String"/>
<doc>
<![CDATA[Set the status description for the task.
@param status brief description of the current status.]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.Enum"/>
<doc>
<![CDATA[Get the {@link Counter} of the given group with the given name.
@param name counter name
@return the <code>Counter</code> of the given group/name.]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the {@link Counter} of the given group with the given name.
@param group counter group
@param name counter name
@return the <code>Counter</code> of the given group/name.]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the counter identified by the key, which can be of
any {@link Enum} type, by the specified amount.
@param key key to identify the counter to be incremented. The key can be
be any <code>Enum</code>.
@param amount A non-negative amount by which the counter is to
be incremented.]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="counter" type="java.lang.String"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the counter identified by the group and counter name
by the specified amount.
@param group name to identify the group of the counter to be incremented.
@param counter name to identify the counter within the group.
@param amount A non-negative amount by which the counter is to
be incremented.]]>
</doc>
</method>
<method name="getInputSplit" return="org.apache.hadoop.mapred.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="UnsupportedOperationException" type="java.lang.UnsupportedOperationException"/>
<doc>
<![CDATA[Get the {@link InputSplit} object for a map.
@return the <code>InputSplit</code> that the map is reading from.
@throws UnsupportedOperationException if called outside a mapper]]>
</doc>
</method>
<field name="NULL" type="org.apache.hadoop.mapred.Reporter"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A constant of Reporter type that does nothing.]]>
</doc>
</field>
<doc>
<![CDATA[A facility for Map-Reduce applications to report progress and update
counters, status information etc.
<p>{@link Mapper} and {@link Reducer} can use the <code>Reporter</code>
provided to report progress or just indicate that they are alive. In
scenarios where the application takes an insignificant amount of time to
process individual key/value pairs, this is crucial since the framework
might assume that the task has timed-out and kill that task.
<p>Applications can also update {@link Counters} via the provided
<code>Reporter</code> .</p>
@see Progressable
@see Counters]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Reporter -->
<!-- start interface org.apache.hadoop.mapred.RunningJob -->
<interface name="RunningJob" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Job} instead">
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the underlying job configuration
@return the configuration of the job.]]>
</doc>
</method>
<method name="getID" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job identifier.
@return the job identifier.]]>
</doc>
</method>
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="This method is deprecated and will be removed. Applications should
rather use {@link #getID()}.">
<doc>
<![CDATA[@deprecated This method is deprecated and will be removed. Applications should
rather use {@link #getID()}.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the job.
@return the name of the job.]]>
</doc>
</method>
<method name="getJobFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the path of the submitted job configuration.
@return the path of the submitted job configuration.]]>
</doc>
</method>
<method name="getTrackingURL" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the URL where some job progress information will be displayed.
@return the URL where some job progress information will be displayed.]]>
</doc>
</method>
<method name="mapProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's map-tasks, as a float between 0.0
and 1.0. When all map tasks have completed, the function returns 1.0.
@return the progress of the job's map-tasks.
@throws IOException]]>
</doc>
</method>
<method name="reduceProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's reduce-tasks, as a float between 0.0
and 1.0. When all reduce tasks have completed, the function returns 1.0.
@return the progress of the job's reduce-tasks.
@throws IOException]]>
</doc>
</method>
<method name="cleanupProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's cleanup-tasks, as a float between 0.0
and 1.0. When all cleanup tasks have completed, the function returns 1.0.
@return the progress of the job's cleanup-tasks.
@throws IOException]]>
</doc>
</method>
<method name="setupProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's setup-tasks, as a float between 0.0
and 1.0. When all setup tasks have completed, the function returns 1.0.
@return the progress of the job's setup-tasks.
@throws IOException]]>
</doc>
</method>
<method name="isComplete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check if the job is finished or not.
This is a non-blocking call.
@return <code>true</code> if the job is complete, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="isSuccessful" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check if the job completed successfully.
@return <code>true</code> if the job succeeded, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="waitForCompletion"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Blocks until the job is complete.
@throws IOException]]>
</doc>
</method>
<method name="getJobState" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the current state of the Job.
{@link JobStatus}
@throws IOException]]>
</doc>
</method>
<method name="killJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill the running job. Blocks until all job tasks have been
killed as well. If the job is no longer running, it simply returns.
@throws IOException]]>
</doc>
</method>
<method name="setJobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="priority" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the priority of a running job.
@param priority the new priority for the job.
@throws IOException]]>
</doc>
</method>
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="startFrom" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get events indicating completion (success/failure) of component tasks.
@param startFrom index to start fetching events from
@return an array of {@link TaskCompletionEvent}s
@throws IOException]]>
</doc>
</method>
<method name="killTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
<param name="shouldFail" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill indicated task attempt.
@param taskId the id of the task to be terminated.
@param shouldFail if true the task is failed and added to failed tasks
list, otherwise it is just killed, w/o affecting
job failure status.
@throws IOException]]>
</doc>
</method>
<method name="killTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #killTask(TaskAttemptID, boolean)}">
<param name="taskId" type="java.lang.String"/>
<param name="shouldFail" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #killTask(TaskAttemptID, boolean)}]]>
</doc>
</method>
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the counters for this job.
@return the counters for this job.
@throws IOException]]>
</doc>
</method>
<method name="getTaskDiagnostics" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the diagnostic messages for a given task attempt.
@param taskid
@return the list of diagnostic messages for the task
@throws IOException]]>
</doc>
</method>
<method name="getHistoryUrl" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the url where history file is archived. Returns empty string if
history file is not available yet.
@return the url where history file is archived
@throws IOException]]>
</doc>
</method>
<method name="isRetired" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check whether the job has been removed from JobTracker memory and retired.
On retire, the job history file is copied to a location known by
{@link #getHistoryUrl()}
@return <code>true</code> if the job retired, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RunningJob</code> is the user-interface to query for details on a
running Map-Reduce job.
<p>Clients can get hold of <code>RunningJob</code> via the {@link JobClient}
and then query the running-job for details such as name, configuration,
progress etc.</p>
@see JobClient
@deprecated Use {@link org.apache.hadoop.mapreduce.Job} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RunningJob -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
<class name="SequenceFileAsBinaryInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat}
instead">
<constructor name="SequenceFileAsBinaryInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[InputFormat reading keys, values from SequenceFiles in binary (raw)
format.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<class name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getKeyClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the key class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getKeyClassName]]>
</doc>
</method>
<method name="getValueClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the value class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getValueClassName]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.BytesWritable"/>
<param name="val" type="org.apache.hadoop.io.BytesWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read raw bytes from a SequenceFile.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<doc>
<![CDATA[Read records from a SequenceFile as binary (raw) bytes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat -->
<class name="SequenceFileAsBinaryOutputFormat" extends="org.apache.hadoop.mapred.SequenceFileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat}
instead">
<constructor name="SequenceFileAsBinaryOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setSequenceFileOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the {@link SequenceFile}
<p>This allows the user to specify the key class to be different
from the actual class ({@link BytesWritable}) used for writing </p>
@param conf the {@link JobConf} to modify
@param theClass the SequenceFile output key class.]]>
</doc>
</method>
<method name="setSequenceFileOutputValueClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the value class for the {@link SequenceFile}
<p>This allows the user to specify the value class to be different
from the actual class ({@link BytesWritable}) used for writing </p>
@param conf the {@link JobConf} to modify
@param theClass the SequenceFile output key class.]]>
</doc>
</method>
<method name="getSequenceFileOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the key class for the {@link SequenceFile}
@return the key class of the {@link SequenceFile}]]>
</doc>
</method>
<method name="getSequenceFileOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the value class for the {@link SequenceFile}
@return the value class of the {@link SequenceFile}]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes keys, values to
{@link SequenceFile}s in binary(raw) format
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
<class name="SequenceFileAsBinaryOutputFormat.WritableValueBytes" extends="org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<constructor name="SequenceFileAsBinaryOutputFormat.WritableValueBytes"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Inner class used for appendRaw]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
<class name="SequenceFileAsTextInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat}
instead">
<constructor name="SequenceFileAsTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is similar to SequenceFileInputFormat,
except it generates SequenceFileAsTextRecordReader
which converts the input keys and values to their
String forms by calling toString() method.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
<class name="SequenceFileAsTextRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader}
instead">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="SequenceFileAsTextRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class converts the input keys and values to their String forms by calling toString()
method. This class to SequenceFileAsTextInputFormat class is as LineRecordReader
class to TextInputFormat class.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter -->
<class name="SequenceFileInputFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter}
instead">
<constructor name="SequenceFileInputFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a record reader for the given split
@param split file split
@param job job configuration
@param reporter reporter who sends report to task tracker
@return RecordReader]]>
</doc>
</method>
<method name="setFilterClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="filterClass" type="java.lang.Class"/>
<doc>
<![CDATA[set the filter class
@param conf application configuration
@param filterClass filter class]]>
</doc>
</method>
<doc>
<![CDATA[A class that allows a map/red job to work on a sample of sequence files.
The sample is decided by the filter class set by the job.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter -->
<!-- start interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
<interface name="SequenceFileInputFilter.Filter" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.Filter"/>
<doc>
<![CDATA[filter interface]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
<class name="SequenceFileInputFilter.FilterBase" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.SequenceFileInputFilter.Filter"/>
<constructor name="SequenceFileInputFilter.FilterBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[base class for Filters]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
<class name="SequenceFileInputFilter.MD5Filter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.MD5Filter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the filtering frequency in configuration
@param conf configuration
@param frequency filtering frequency]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter according to configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If MD5(key) % frequency==0, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
</doc>
</method>
<field name="MD5_LEN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency <i>f</i>. The filtering criteria is
MD5(key) % f == 0.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
<class name="SequenceFileInputFilter.PercentFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.PercentFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the frequency and stores it in conf
@param conf configuration
@param frequency filtering frequencey]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter by checking the configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If record# % frequency==0, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
</doc>
</method>
<doc>
<![CDATA[This class returns a percentage of records
The percentage is determined by a filtering frequency <i>f</i> using
the criteria record# % f == 0.
For example, if the frequency is 10, one out of 10 records is returned.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
<class name="SequenceFileInputFilter.RegexFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.RegexFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setPattern"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="regex" type="java.lang.String"/>
<exception name="PatternSyntaxException" type="java.util.regex.PatternSyntaxException"/>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the Filter by checking the configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If key matches the regex, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
</doc>
</method>
<doc>
<![CDATA[Records filter by matching key to regex]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFormat -->
<class name="SequenceFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat}
instead.">
<constructor name="SequenceFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for {@link SequenceFile}s.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
<class name="SequenceFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat}
instead.">
<constructor name="SequenceFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReaders" return="org.apache.hadoop.io.SequenceFile.Reader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the output generated by this format.]]>
</doc>
</method>
<method name="getOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link CompressionType} for the output {@link SequenceFile}.
@param conf the {@link JobConf}
@return the {@link CompressionType} for the output {@link SequenceFile},
defaulting to {@link CompressionType#RECORD}]]>
</doc>
</method>
<method name="setOutputCompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<doc>
<![CDATA[Set the {@link CompressionType} for the output {@link SequenceFile}.
@param conf the {@link JobConf} to modify
@param style the {@link CompressionType} for the output
{@link SequenceFile}]]>
</doc>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes {@link SequenceFile}s.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileRecordReader -->
<class name="SequenceFileRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="SequenceFileRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The class of key that must be passed to {@link
#next(Object, Object)}..]]>
</doc>
</method>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The class of value that must be passed to {@link
#next(Object, Object)}..]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCurrentValue"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="conf" type="org.apache.hadoop.conf.Configuration"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link RecordReader} for {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileRecordReader -->
<!-- start class org.apache.hadoop.mapred.SkipBadRecords -->
<class name="SkipBadRecords" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SkipBadRecords"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getAttemptsToStartSkipping" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the number of Task attempts AFTER which skip mode
will be kicked off. When skip mode is kicked off, the
tasks reports the range of records which it will process
next to the TaskTracker. So that on failures, TT knows which
ones are possibly the bad records. On further executions,
those are skipped.
Default value is 2.
@param conf the configuration
@return attemptsToStartSkipping no of task attempts]]>
</doc>
</method>
<method name="setAttemptsToStartSkipping"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="attemptsToStartSkipping" type="int"/>
<doc>
<![CDATA[Set the number of Task attempts AFTER which skip mode
will be kicked off. When skip mode is kicked off, the
tasks reports the range of records which it will process
next to the TaskTracker. So that on failures, TT knows which
ones are possibly the bad records. On further executions,
those are skipped.
Default value is 2.
@param conf the configuration
@param attemptsToStartSkipping no of task attempts]]>
</doc>
</method>
<method name="getAutoIncrMapperProcCount" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the flag which if set to true,
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS} is incremented
by MapRunner after invoking the map function. This value must be set to
false for applications which process the records asynchronously
or buffer the input records. For example streaming.
In such cases applications should increment this counter on their own.
Default value is true.
@param conf the configuration
@return <code>true</code> if auto increment
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS}.
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setAutoIncrMapperProcCount"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="autoIncr" type="boolean"/>
<doc>
<![CDATA[Set the flag which if set to true,
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS} is incremented
by MapRunner after invoking the map function. This value must be set to
false for applications which process the records asynchronously
or buffer the input records. For example streaming.
In such cases applications should increment this counter on their own.
Default value is true.
@param conf the configuration
@param autoIncr whether to auto increment
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS}.]]>
</doc>
</method>
<method name="getAutoIncrReducerProcCount" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the flag which if set to true,
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS} is incremented
by framework after invoking the reduce function. This value must be set to
false for applications which process the records asynchronously
or buffer the input records. For example streaming.
In such cases applications should increment this counter on their own.
Default value is true.
@param conf the configuration
@return <code>true</code> if auto increment
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS}.
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setAutoIncrReducerProcCount"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="autoIncr" type="boolean"/>
<doc>
<![CDATA[Set the flag which if set to true,
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS} is incremented
by framework after invoking the reduce function. This value must be set to
false for applications which process the records asynchronously
or buffer the input records. For example streaming.
In such cases applications should increment this counter on their own.
Default value is true.
@param conf the configuration
@param autoIncr whether to auto increment
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS}.]]>
</doc>
</method>
<method name="getSkipOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the directory to which skipped records are written. By default it is
the sub directory of the output _logs directory.
User can stop writing skipped records by setting the value null.
@param conf the configuration.
@return path skip output directory. Null is returned if this is not set
and output directory is also not set.]]>
</doc>
</method>
<method name="setSkipOutputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the directory to which skipped records are written. By default it is
the sub directory of the output _logs directory.
User can stop writing skipped records by setting the value null.
@param conf the configuration.
@param path skip output directory path]]>
</doc>
</method>
<method name="getMapperMaxSkipRecords" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the number of acceptable skip records surrounding the bad record PER
bad record in mapper. The number includes the bad record as well.
To turn the feature of detection/skipping of bad records off, set the
value to 0.
The framework tries to narrow down the skipped range by retrying
until this threshold is met OR all attempts get exhausted for this task.
Set the value to Long.MAX_VALUE to indicate that framework need not try to
narrow down. Whatever records(depends on application) get skipped are
acceptable.
Default value is 0.
@param conf the configuration
@return maxSkipRecs acceptable skip records.]]>
</doc>
</method>
<method name="setMapperMaxSkipRecords"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="maxSkipRecs" type="long"/>
<doc>
<![CDATA[Set the number of acceptable skip records surrounding the bad record PER
bad record in mapper. The number includes the bad record as well.
To turn the feature of detection/skipping of bad records off, set the
value to 0.
The framework tries to narrow down the skipped range by retrying
until this threshold is met OR all attempts get exhausted for this task.
Set the value to Long.MAX_VALUE to indicate that framework need not try to
narrow down. Whatever records(depends on application) get skipped are
acceptable.
Default value is 0.
@param conf the configuration
@param maxSkipRecs acceptable skip records.]]>
</doc>
</method>
<method name="getReducerMaxSkipGroups" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the number of acceptable skip groups surrounding the bad group PER
bad group in reducer. The number includes the bad group as well.
To turn the feature of detection/skipping of bad groups off, set the
value to 0.
The framework tries to narrow down the skipped range by retrying
until this threshold is met OR all attempts get exhausted for this task.
Set the value to Long.MAX_VALUE to indicate that framework need not try to
narrow down. Whatever groups(depends on application) get skipped are
acceptable.
Default value is 0.
@param conf the configuration
@return maxSkipGrps acceptable skip groups.]]>
</doc>
</method>
<method name="setReducerMaxSkipGroups"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="maxSkipGrps" type="long"/>
<doc>
<![CDATA[Set the number of acceptable skip groups surrounding the bad group PER
bad group in reducer. The number includes the bad group as well.
To turn the feature of detection/skipping of bad groups off, set the
value to 0.
The framework tries to narrow down the skipped range by retrying
until this threshold is met OR all attempts get exhausted for this task.
Set the value to Long.MAX_VALUE to indicate that framework need not try to
narrow down. Whatever groups(depends on application) get skipped are
acceptable.
Default value is 0.
@param conf the configuration
@param maxSkipGrps acceptable skip groups.]]>
</doc>
</method>
<field name="COUNTER_GROUP" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Special counters which are written by the application and are
used by the framework for detecting bad records. For detecting bad records
these counters must be incremented by the application.]]>
</doc>
</field>
<field name="COUNTER_MAP_PROCESSED_RECORDS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of processed map records.
@see SkipBadRecords#getAutoIncrMapperProcCount(Configuration)]]>
</doc>
</field>
<field name="COUNTER_REDUCE_PROCESSED_GROUPS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of processed reduce groups.
@see SkipBadRecords#getAutoIncrReducerProcCount(Configuration)]]>
</doc>
</field>
<doc>
<![CDATA[Utility class for skip bad records functionality. It contains various
settings related to skipping of bad records.
<p>Hadoop provides an optional mode of execution in which the bad records
are detected and skipped in further attempts.
<p>This feature can be used when map/reduce tasks crashes deterministically on
certain input. This happens due to bugs in the map/reduce function. The usual
course would be to fix these bugs. But sometimes this is not possible;
perhaps the bug is in third party libraries for which the source code is
not available. Due to this, the task never reaches to completion even with
multiple attempts and complete data for that task is lost.</p>
<p>With this feature, only a small portion of data is lost surrounding
the bad record, which may be acceptable for some user applications.
see {@link SkipBadRecords#setMapperMaxSkipRecords(Configuration, long)}</p>
<p>The skipping mode gets kicked off after certain no of failures
see {@link SkipBadRecords#setAttemptsToStartSkipping(Configuration, int)}</p>
<p>In the skipping mode, the map/reduce task maintains the record range which
is getting processed at all times. Before giving the input to the
map/reduce function, it sends this record range to the Task tracker.
If task crashes, the Task tracker knows which one was the last reported
range. On further attempts that range get skipped.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SkipBadRecords -->
<!-- start interface org.apache.hadoop.mapred.TaskAttemptContext -->
<interface name="TaskAttemptContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.TaskAttemptContext}
instead.">
<implements name="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<method name="getTaskAttemptID" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProgressible" return="org.apache.hadoop.util.Progressable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[@deprecated Use {@link org.apache.hadoop.mapreduce.TaskAttemptContext}
instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.TaskAttemptContext -->
<!-- start class org.apache.hadoop.mapred.TaskAttemptID -->
<class name="TaskAttemptID" extends="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskAttemptID" type="org.apache.hadoop.mapred.TaskID, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskAttemptID object from given {@link TaskID}.
@param taskId TaskID that this task belongs to
@param id the task attempt number]]>
</doc>
</constructor>
<constructor name="TaskAttemptID" type="java.lang.String, int, boolean, int, int"
static="false" final="false" visibility="public"
deprecated="Use {@link #TaskAttemptID(String, int, TaskType, int, int)}.">
<doc>
<![CDATA[Constructs a TaskId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param isMap whether the tip is a map
@param taskId taskId number
@param id the task attempt number
@deprecated Use {@link #TaskAttemptID(String, int, TaskType, int, int)}.]]>
</doc>
</constructor>
<constructor name="TaskAttemptID" type="java.lang.String, int, org.apache.hadoop.mapreduce.TaskType, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param type the TaskType
@param taskId taskId number
@param id the task attempt number]]>
</doc>
</constructor>
<constructor name="TaskAttemptID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="downgrade" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="old" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<doc>
<![CDATA[Downgrade a new TaskAttemptID to an old one
@param old the new id
@return either old or a new TaskAttemptID constructed to match old]]>
</doc>
</method>
<method name="getTaskID" return="org.apache.hadoop.mapred.TaskID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="read" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="forName" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<doc>
<![CDATA[Construct a TaskAttemptID object from given string
@return constructed TaskAttemptID object or null if the given String is null
@throws IllegalArgumentException if the given string is malformed]]>
</doc>
</method>
<method name="getTaskAttemptIDsPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jtIdentifier" type="java.lang.String"/>
<param name="jobId" type="java.lang.Integer"/>
<param name="isMap" type="java.lang.Boolean"/>
<param name="taskId" type="java.lang.Integer"/>
<param name="attemptId" type="java.lang.Integer"/>
<doc>
<![CDATA[Returns a regex pattern which matches task attempt IDs. Arguments can
be given null, in which case that part of the regex will be generic.
For example to obtain a regex matching <i>all task attempt IDs</i>
of <i>any jobtracker</i>, in <i>any job</i>, of the <i>first
map task</i>, we would use :
<pre>
TaskAttemptID.getTaskAttemptIDsPattern(null, null, true, 1, null);
</pre>
which will return :
<pre> "attempt_[^_]*_[0-9]*_m_000001_[0-9]*" </pre>
@param jtIdentifier jobTracker identifier, or null
@param jobId job number, or null
@param isMap whether the tip is a map, or null
@param taskId taskId number, or null
@param attemptId the task attempt number, or null
@return a regex pattern matching TaskAttemptIDs]]>
</doc>
</method>
<method name="getTaskAttemptIDsPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jtIdentifier" type="java.lang.String"/>
<param name="jobId" type="java.lang.Integer"/>
<param name="type" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskId" type="java.lang.Integer"/>
<param name="attemptId" type="java.lang.Integer"/>
<doc>
<![CDATA[Returns a regex pattern which matches task attempt IDs. Arguments can
be given null, in which case that part of the regex will be generic.
For example to obtain a regex matching <i>all task attempt IDs</i>
of <i>any jobtracker</i>, in <i>any job</i>, of the <i>first
map task</i>, we would use :
<pre>
TaskAttemptID.getTaskAttemptIDsPattern(null, null, TaskType.MAP, 1, null);
</pre>
which will return :
<pre> "attempt_[^_]*_[0-9]*_m_000001_[0-9]*" </pre>
@param jtIdentifier jobTracker identifier, or null
@param jobId job number, or null
@param type the {@link TaskType}
@param taskId taskId number, or null
@param attemptId the task attempt number, or null
@return a regex pattern matching TaskAttemptIDs]]>
</doc>
</method>
<doc>
<![CDATA[TaskAttemptID represents the immutable and unique identifier for
a task attempt. Each task attempt is one particular instance of a Map or
Reduce Task identified by its TaskID.
TaskAttemptID consists of 2 parts. First part is the
{@link TaskID}, that this TaskAttemptID belongs to.
Second part is the task attempt number. <br>
An example TaskAttemptID is :
<code>attempt_200707121733_0003_m_000005_0</code> , which represents the
zeroth task attempt for the fifth map task in the third job
running at the jobtracker started at <code>200707121733</code>.
<p>
Applications should never construct or parse TaskAttemptID strings
, but rather use appropriate constructors or {@link #forName(String)}
method.
@see JobID
@see TaskID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskAttemptID -->
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent -->
<class name="TaskCompletionEvent" extends="org.apache.hadoop.mapreduce.TaskCompletionEvent"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.TaskCompletionEvent} instead">
<constructor name="TaskCompletionEvent"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for Writable.]]>
</doc>
</constructor>
<constructor name="TaskCompletionEvent" type="int, org.apache.hadoop.mapred.TaskAttemptID, int, boolean, org.apache.hadoop.mapred.TaskCompletionEvent.Status, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor. eventId should be created externally and incremented
per event for each job.
@param eventId event id, event id should be unique and assigned in
incrementally, starting from 0.
@param taskId task id
@param status task's status
@param taskTrackerHttp task tracker's host:port for http.]]>
</doc>
</constructor>
<method name="getTaskId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="use {@link #getTaskAttemptId()} instead.">
<doc>
<![CDATA[Returns task id.
@return task id
@deprecated use {@link #getTaskAttemptId()} instead.]]>
</doc>
</method>
<method name="getTaskAttemptId" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task id.
@return task id]]>
</doc>
</method>
<method name="getTaskStatus" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns enum Status.SUCESS or Status.FAILURE.
@return task tracker status]]>
</doc>
</method>
<method name="setTaskId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="use {@link #setTaskAttemptId(TaskAttemptID)} instead.">
<param name="taskId" type="java.lang.String"/>
<doc>
<![CDATA[Sets task id.
@param taskId
@deprecated use {@link #setTaskAttemptId(TaskAttemptID)} instead.]]>
</doc>
</method>
<method name="setTaskAttemptId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
<doc>
<![CDATA[Sets task id.
@param taskId]]>
</doc>
</method>
<method name="setTaskStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="status" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"/>
<doc>
<![CDATA[Set task status.
@param status]]>
</doc>
</method>
<method name="setTaskRunTime"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskCompletionTime" type="int"/>
<doc>
<![CDATA[Set the task completion time
@param taskCompletionTime time (in millisec) the task took to complete]]>
</doc>
</method>
<method name="setEventId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="eventId" type="int"/>
<doc>
<![CDATA[set event Id. should be assigned incrementally starting from 0.
@param eventId]]>
</doc>
</method>
<method name="setTaskTrackerHttp"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskTrackerHttp" type="java.lang.String"/>
<doc>
<![CDATA[Set task tracker http location.
@param taskTrackerHttp]]>
</doc>
</method>
<field name="EMPTY_ARRAY" type="org.apache.hadoop.mapred.TaskCompletionEvent[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This is used to track task completion events on
job tracker.
@deprecated Use
{@link org.apache.hadoop.mapreduce.TaskCompletionEvent} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent -->
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
<class name="TaskCompletionEvent.Status" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
<!-- start class org.apache.hadoop.mapred.TaskID -->
<class name="TaskID" extends="org.apache.hadoop.mapreduce.TaskID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskID" type="org.apache.hadoop.mapreduce.JobID, boolean, int"
static="false" final="false" visibility="public"
deprecated="Use {@link #TaskID(String, int, TaskType, int)}">
<doc>
<![CDATA[Constructs a TaskID object from given {@link JobID}.
@param jobId JobID that this tip belongs to
@param isMap whether the tip is a map
@param id the tip number
@deprecated Use {@link #TaskID(String, int, TaskType, int)}]]>
</doc>
</constructor>
<constructor name="TaskID" type="java.lang.String, int, boolean, int"
static="false" final="false" visibility="public"
deprecated="Use {@link #TaskID(org.apache.hadoop.mapreduce.JobID, TaskType,
int)}">
<doc>
<![CDATA[Constructs a TaskInProgressId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param isMap whether the tip is a map
@param id the tip number
@deprecated Use {@link #TaskID(org.apache.hadoop.mapreduce.JobID, TaskType,
int)}]]>
</doc>
</constructor>
<constructor name="TaskID" type="org.apache.hadoop.mapreduce.JobID, org.apache.hadoop.mapreduce.TaskType, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskID object from given {@link JobID}.
@param jobId JobID that this tip belongs to
@param type the {@link TaskType}
@param id the tip number]]>
</doc>
</constructor>
<constructor name="TaskID" type="java.lang.String, int, org.apache.hadoop.mapreduce.TaskType, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskInProgressId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param type the {@link TaskType}
@param id the tip number]]>
</doc>
</constructor>
<constructor name="TaskID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="downgrade" return="org.apache.hadoop.mapred.TaskID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="old" type="org.apache.hadoop.mapreduce.TaskID"/>
<doc>
<![CDATA[Downgrade a new TaskID to an old one
@param old a new or old TaskID
@return either old or a new TaskID build to match old]]>
</doc>
</method>
<method name="read" return="org.apache.hadoop.mapred.TaskID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskIDsPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="Use {@link TaskID#getTaskIDsPattern(String, Integer, TaskType,
Integer)}">
<param name="jtIdentifier" type="java.lang.String"/>
<param name="jobId" type="java.lang.Integer"/>
<param name="isMap" type="java.lang.Boolean"/>
<param name="taskId" type="java.lang.Integer"/>
<doc>
<![CDATA[Returns a regex pattern which matches task IDs. Arguments can
be given null, in which case that part of the regex will be generic.
For example to obtain a regex matching <i>the first map task</i>
of <i>any jobtracker</i>, of <i>any job</i>, we would use :
<pre>
TaskID.getTaskIDsPattern(null, null, true, 1);
</pre>
which will return :
<pre> "task_[^_]*_[0-9]*_m_000001*" </pre>
@param jtIdentifier jobTracker identifier, or null
@param jobId job number, or null
@param isMap whether the tip is a map, or null
@param taskId taskId number, or null
@return a regex pattern matching TaskIDs
@deprecated Use {@link TaskID#getTaskIDsPattern(String, Integer, TaskType,
Integer)}]]>
</doc>
</method>
<method name="getTaskIDsPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jtIdentifier" type="java.lang.String"/>
<param name="jobId" type="java.lang.Integer"/>
<param name="type" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskId" type="java.lang.Integer"/>
<doc>
<![CDATA[Returns a regex pattern which matches task IDs. Arguments can
be given null, in which case that part of the regex will be generic.
For example to obtain a regex matching <i>the first map task</i>
of <i>any jobtracker</i>, of <i>any job</i>, we would use :
<pre>
TaskID.getTaskIDsPattern(null, null, true, 1);
</pre>
which will return :
<pre> "task_[^_]*_[0-9]*_m_000001*" </pre>
@param jtIdentifier jobTracker identifier, or null
@param jobId job number, or null
@param type the {@link TaskType}, or null
@param taskId taskId number, or null
@return a regex pattern matching TaskIDs]]>
</doc>
</method>
<method name="forName" return="org.apache.hadoop.mapred.TaskID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
</method>
<doc>
<![CDATA[TaskID represents the immutable and unique identifier for
a Map or Reduce Task. Each TaskID encompasses multiple attempts made to
execute the Map or Reduce Task, each of which are uniquely indentified by
their TaskAttemptID.
TaskID consists of 3 parts. First part is the {@link JobID}, that this
TaskInProgress belongs to. Second part of the TaskID is either 'm' or 'r'
representing whether the task is a map task or a reduce task.
And the third part is the task number. <br>
An example TaskID is :
<code>task_200707121733_0003_m_000005</code> , which represents the
fifth map task in the third job running at the jobtracker
started at <code>200707121733</code>.
<p>
Applications should never construct or parse TaskID strings
, but rather use appropriate constructors or {@link #forName(String)}
method.
@see JobID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskID -->
<!-- start class org.apache.hadoop.mapred.TaskLog.Reader -->
<class name="TaskLog.Reader" extends="java.io.InputStream"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLog.Reader" type="org.apache.hadoop.mapred.TaskAttemptID, org.apache.hadoop.mapred.TaskLog.LogName, long, long, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a log file from start to end positions. The offsets may be negative,
in which case they are relative to the end of the file. For example,
Reader(taskid, kind, 0, -1) is the entire file and
Reader(taskid, kind, -4197, -1) is the last 4196 bytes.
@param taskid the id of the task to read the log file for
@param kind the kind of log to read
@param start the offset to read from (negative is relative to tail)
@param end the offset to read upto (negative is relative to tail)
@param isCleanup whether the attempt is cleanup attempt or not
@throws IOException]]>
</doc>
</constructor>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="available" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLog.Reader -->
<!-- start class org.apache.hadoop.mapred.TaskLogAppender -->
<class name="TaskLogAppender" extends="org.apache.log4j.FileAppender"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLogAppender"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="activateOptions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="event" type="org.apache.log4j.spi.LoggingEvent"/>
</method>
<method name="flush"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskId" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Getter/Setter methods for log4j.]]>
</doc>
</method>
<method name="setTaskId"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
</method>
<method name="getTotalLogFileSize" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setTotalLogFileSize"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="logSize" type="long"/>
</method>
<method name="setIsCleanup"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isCleanup" type="boolean"/>
<doc>
<![CDATA[Set whether the task is a cleanup attempt or not.
@param isCleanup
true if the task is cleanup attempt, false otherwise.]]>
</doc>
</method>
<method name="getIsCleanup" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get whether task is cleanup attempt or not.
@return true if the task is cleanup attempt, false otherwise.]]>
</doc>
</method>
<doc>
<![CDATA[A simple log4j-appender for the task child's
map-reduce system logs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLogAppender -->
<!-- start class org.apache.hadoop.mapred.TaskReport -->
<class name="TaskReport" extends="org.apache.hadoop.mapreduce.TaskReport"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.TaskReport} instead">
<constructor name="TaskReport"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getTaskID" return="org.apache.hadoop.mapred.TaskID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The id of the task.]]>
</doc>
</method>
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setSuccessfulAttempt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="org.apache.hadoop.mapred.TaskAttemptID"/>
<doc>
<![CDATA[set successful attempt ID of the task.]]>
</doc>
</method>
<method name="getSuccessfulTaskAttempt" return="org.apache.hadoop.mapred.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the attempt ID that took this task to completion]]>
</doc>
</method>
<method name="setRunningTaskAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="runningAttempts" type="java.util.Collection"/>
<doc>
<![CDATA[set running attempt(s) of the task.]]>
</doc>
</method>
<method name="getRunningTaskAttempts" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the running task attempt IDs for this task]]>
</doc>
</method>
<method name="setFinishTime"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="finishTime" type="long"/>
<doc>
<![CDATA[set finish time of task.
@param finishTime finish time of task.]]>
</doc>
</method>
<method name="setStartTime"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="startTime" type="long"/>
<doc>
<![CDATA[set start time of the task.]]>
</doc>
</method>
<doc>
<![CDATA[A report on the state of a task.
@deprecated Use {@link org.apache.hadoop.mapreduce.TaskReport} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskReport -->
<!-- start class org.apache.hadoop.mapred.TextInputFormat -->
<class name="TextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.TextInputFormat}
instead.">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="TextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either linefeed or carriage-return are used to signal end of line. Keys are
the position in the file, and values are the line of text..
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.TextInputFormat}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TextInputFormat -->
<!-- start class org.apache.hadoop.mapred.TextOutputFormat -->
<class name="TextOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.TextOutputFormat} instead.">
<constructor name="TextOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes plain text files.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.TextOutputFormat} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TextOutputFormat -->
<!-- start class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
<class name="TextOutputFormat.LineRecordWriter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordWriter"/>
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="out" type="java.io.DataOutputStream"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
<!-- start class org.apache.hadoop.mapred.Utils -->
<class name="Utils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Utils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A utility class. It provides
A path filter utility to filter out output/part files in the output dir]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Utils -->
<!-- start class org.apache.hadoop.mapred.Utils.OutputFileUtils -->
<class name="Utils.OutputFileUtils" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Utils.OutputFileUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.mapred.Utils.OutputFileUtils -->
<!-- start class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter -->
<class name="Utils.OutputFileUtils.OutputFilesFilter" extends="org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Utils.OutputFileUtils.OutputFilesFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
</method>
<doc>
<![CDATA[This class filters output(part) files from the given directory
It does not accept files with filenames _logs and _SUCCESS.
This can be used to list paths of output directory as follows:
Path[] fileList = FileUtil.stat2Paths(fs.listStatus(outDir,
new OutputFilesFilter()));]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter -->
<!-- start class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter -->
<class name="Utils.OutputFileUtils.OutputLogFilter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.fs.PathFilter"/>
<constructor name="Utils.OutputFileUtils.OutputLogFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
</method>
<doc>
<![CDATA[This class filters log files from directory given
It doesnt accept paths having _logs.
This can be used to list paths of output directory as follows:
Path[] fileList = FileUtil.stat2Paths(fs.listStatus(outDir,
new OutputLogFilter()));]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter -->
</package>
<package name="org.apache.hadoop.mapred.jobcontrol">
<!-- start class org.apache.hadoop.mapred.jobcontrol.Job -->
<class name="Job" extends="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link ControlledJob} instead.">
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf, java.util.ArrayList"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a job.
@param jobConf a mapred job configuration representing a job to be executed.
@param dependingJobs an array of jobs the current job depends on]]>
</doc>
</constructor>
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getAssignedJobID" return="org.apache.hadoop.mapred.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapred ID of this job as assigned by the
mapred framework.]]>
</doc>
</method>
<method name="setAssignedJobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="setAssignedJobID should not be called.
JOBID is set by the framework.">
<param name="mapredJobID" type="org.apache.hadoop.mapred.JobID"/>
<doc>
<![CDATA[@deprecated setAssignedJobID should not be called.
JOBID is set by the framework.]]>
</doc>
</method>
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapred job conf of this job]]>
</doc>
</method>
<method name="setJobConf"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Set the mapred job conf for this job.
@param jobConf the mapred job conf for this job.]]>
</doc>
</method>
<method name="getState" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the state of this job]]>
</doc>
</method>
<method name="getJobClient" return="org.apache.hadoop.mapred.JobClient"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job client of this job]]>
</doc>
</method>
<method name="getDependingJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the depending jobs of this job]]>
</doc>
</method>
<field name="SUCCESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WAITING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RUNNING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="READY" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEPENDENT_FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[@deprecated Use {@link ControlledJob} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.jobcontrol.Job -->
<!-- start class org.apache.hadoop.mapred.jobcontrol.JobControl -->
<class name="JobControl" extends="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl} instead">
<constructor name="JobControl" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a job control for a group of jobs.
@param groupName a name identifying this group]]>
</doc>
</constructor>
<method name="getWaitingJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the waiting state]]>
</doc>
</method>
<method name="getRunningJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the running state]]>
</doc>
</method>
<method name="getReadyJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the ready state]]>
</doc>
</method>
<method name="getSuccessfulJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the success state]]>
</doc>
</method>
<method name="getFailedJobs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="addJobs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobs" type="java.util.Collection"/>
<doc>
<![CDATA[Add a collection of jobs
@param jobs]]>
</doc>
</method>
<method name="getState" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the thread state]]>
</doc>
</method>
<doc>
<![CDATA[@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.jobcontrol.JobControl -->
</package>
<package name="org.apache.hadoop.mapred.join">
<!-- start class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
<class name="ArrayListBackedIterator" extends="org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator} instead">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
<constructor name="ArrayListBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ArrayListBackedIterator" type="java.util.ArrayList"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. The
implementation uses an {@link java.util.ArrayList} to store elements
added to it, replaying them as requested.
Prefer {@link StreamBackedIterator}.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
<!-- start interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
<interface name="ComposableInputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat} instead">
<implements name="org.apache.hadoop.mapred.InputFormat"/>
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Refinement of InputFormat requiring implementors to provide
ComposableRecordReader instead of RecordReader.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
<!-- start interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
<interface name="ComposableRecordReader" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader} instead">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<implements name="java.lang.Comparable"/>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the position in the collector this class occupies.]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key this RecordReader would supply on a call to next(K,V)]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the head of this RecordReader into the object provided.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the stream is not empty, but provides no guarantee that
a call to next(K,V) will succeed.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[While key-value pairs from this RecordReader match the given key, register
them with the JoinCollector provided.]]>
</doc>
</method>
<doc>
<![CDATA[Additional operations required of a RecordReader to participate in a join.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.CompositeInputFormat -->
<class name="CompositeInputFormat" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat} instead">
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat"/>
<constructor name="CompositeInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Interpret a given string as a composite expression.
{@code
func ::= <ident>([<func>,]*<func>)
func ::= tbl(<class>,"<path>")
class ::= @see java.lang.Class#forName(java.lang.String)
path ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
}
Reads expression from the <tt>mapred.join.expr</tt> property and
user-supplied join types from <tt>mapred.join.define.&lt;ident&gt;</tt>
types. Paths supplied to <tt>tbl</tt> are given as input paths to the
InputFormat class listed.
@see #compose(java.lang.String, java.lang.Class, java.lang.String...)]]>
</doc>
</method>
<method name="addDefaults"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Adds the default set of identifiers to the parser.]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a CompositeInputSplit from the child InputFormats by assigning the
ith split from each child to the ith composite split.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a CompositeRecordReader for the children of this InputFormat
as defined in the init expression.
The outermost join need only be composable, not necessarily a composite.
Mandating TupleWritable isn't strictly correct.]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="java.lang.Class"/>
<param name="path" type="java.lang.String"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given InputFormat class (inf), path (p) return:
{@code tbl(<inf>, <p>) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class"/>
<param name="path" type="java.lang.String[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class"/>
<param name="path" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<doc>
<![CDATA[An InputFormat capable of performing joins over a set of data sources sorted
and partitioned the same way.
@see #setFormat
A user may define new join types by setting the property
<tt>mapred.join.define.&lt;ident&gt;</tt> to a classname. In the expression
<tt>mapred.join.expr</tt>, the identifier will be assumed to be a
ComposableRecordReader.
<tt>mapred.join.keycomparator</tt> can be a classname used to compare keys
in the join.
@see JoinRecordReader
@see MultiFilterRecordReader
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeInputFormat -->
<!-- start class org.apache.hadoop.mapred.join.CompositeInputSplit -->
<class name="CompositeInputSplit" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit} instead">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="CompositeInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CompositeInputSplit" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.mapred.InputSplit"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an InputSplit to this collection.
@throws IOException If capacity was not specified during construction
or if capacity has been reached.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.mapred.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Get ith child InputSplit.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the aggregate length of all child InputSplits currently added.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the length of ith child InputSplit.]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Collect a set of hosts from all child InputSplits.]]>
</doc>
</method>
<method name="getLocation" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[getLocations from ith InputSplit.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write splits in the following format.
{@code
<count><class1><class2>...<classn><split1><split2>...<splitn>
}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}
@throws IOException If the child InputSplit cannot be read, typically
for faliing access checks.]]>
</doc>
</method>
<doc>
<![CDATA[This InputSplit contains a set of child InputSplits. Any InputSplit inserted
into this collection must have a public default constructor.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeInputSplit -->
<!-- start class org.apache.hadoop.mapred.join.CompositeRecordReader -->
<class name="CompositeRecordReader" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader} instead">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="CompositeRecordReader" type="int, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a RecordReader with <tt>capacity</tt> children to position
<tt>id</tt> in the parent reader.
The id of a root CompositeRecordReader is -1 by convention, but relying
on this is not recommended.]]>
</doc>
</constructor>
<method name="combine" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
</method>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the position in the collector this class occupies.]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getRecordReaderQueue" return="java.util.PriorityQueue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return sorted list of RecordReaders for this composite.]]>
</doc>
</method>
<method name="getComparator" return="org.apache.hadoop.io.WritableComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return comparator defining the ordering for RecordReaders in this
composite.]]>
</doc>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rr" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add a RecordReader to this collection.
The id() of a RecordReader determines where in the Tuple its
entry will appear. Adding RecordReaders with the same id has
undefined behavior.]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key for the current join or the value at the top of the
RecordReader heap.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the top of this RR into the given object.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if it is possible that this could emit more values.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Pass skip key to child RRs.]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Obtain an iterator over the child RRs apropos of the value type
ultimately emitted from this join.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[If key provided matches that of this Composite, give JoinCollector
iterator over values it may emit.]]>
</doc>
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For all child RRs offering the key provided, obtain an iterator
at that position in the JoinCollector.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<doc>
<![CDATA[Implement Comparable contract (compare key of join or head of heap
with that of another).]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new key value common to all child RRs.
@throws ClassCastException if key classes differ.]]>
</doc>
</method>
<method name="createInternalValue" return="org.apache.hadoop.mapred.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a value to be used internally for joins.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Unsupported (returns zero in all cases).]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close all child RRs.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Report progress as the minimum of all child RR progress.]]>
</doc>
</method>
<field name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="kids" type="org.apache.hadoop.mapred.join.ComposableRecordReader[]"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
<class name="InnerJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.InnerJoinRecordReader} instead.">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Return true iff the tuple is full (all data sources contain this key).]]>
</doc>
</method>
<doc>
<![CDATA[Full inner join.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.InnerJoinRecordReader} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader -->
<class name="JoinRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.JoinRecordReader} instead">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<constructor name="JoinRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Emit the next set of key, value pairs as defined by the child
RecordReaders and operation associated with this composite RR.]]>
</doc>
</method>
<method name="createValue" return="org.apache.hadoop.mapred.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator wrapping the JoinCollector.]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite joins returning Tuples of arbitrary Writables.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.JoinRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
<class name="JoinRecordReader.JoinDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
<constructor name="JoinRecordReader.JoinDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Since the JoinCollector is effecting our operation, we need only
provide an iterator proxy wrapping its operation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
<class name="MultiFilterRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader} instead">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<constructor name="MultiFilterRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="emit" return="V"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For each tuple emitted, return a value (typically one of the values
in the tuple).
Modifying the Writables in the tuple is permitted and unlikely to affect
join behavior in most cases, but it is not recommended. It's safer to
clone first.]]>
</doc>
</method>
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Default implementation offers {@link #emit} every Tuple from the
collector (the outer join of child RRs).]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator returning a single value from the tuple.
@see MultiFilterDelegationIterator]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite join returning values derived from multiple
sources, but generally not tuples.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<class name="MultiFilterRecordReader.MultiFilterDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
<constructor name="MultiFilterRecordReader.MultiFilterDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Proxy the JoinCollector, but include callback to emit.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<!-- start class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
<class name="OuterJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.OuterJoinRecordReader} instead">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Emit everything from the collector.]]>
</doc>
</method>
<doc>
<![CDATA[Full outer join.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.OuterJoinRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.OverrideRecordReader -->
<class name="OverrideRecordReader" extends="org.apache.hadoop.mapred.join.MultiFilterRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader} instead">
<method name="emit" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Emit the value with the highest position in the tuple.]]>
</doc>
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Instead of filling the JoinCollector with iterators from all
data sources, fill only the rightmost for this key.
This not only saves space by discarding the other sources, but
it also emits the number of key-value pairs in the preferred
RecordReader instead of repeating that stream n times, where
n is the cardinality of the cross product of the discarded
streams for the given key.]]>
</doc>
</method>
<doc>
<![CDATA[Prefer the &quot;rightmost&quot; data source for this key.
For example, <tt>override(S1,S2,S3)</tt> will prefer values
from S3 over S2, and values from S2 over S1 for all keys
emitted from all sources.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.OverrideRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.Parser -->
<class name="Parser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.join.Parser} instead">
<constructor name="Parser"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Very simple shift-reduce parser for join expressions.
This should be sufficient for the user extension permitted now, but ought to
be replaced with a parser generator if more complex grammars are supported.
In particular, this &quot;shift-reduce&quot; parser has no states. Each set
of formals requires a different internal node type, which is responsible for
interpreting the list of tokens it receives. This is sufficient for the
current grammar, but it has several annoying properties that might inhibit
extension. In particular, parenthesis are always function calls; an
algebraic or filter grammar would not only require a node type, but must
also work around the internals of this parser.
For most other cases, adding classes to the hierarchy- particularly by
extending JoinRecordReader and MultiFilterRecordReader- is fairly
straightforward. One need only override the relevant method(s) (usually only
{@link CompositeRecordReader#combine}) and include a property to map its
value to an identifier in the parser.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.join.Parser} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser -->
<!-- start class org.apache.hadoop.mapred.join.Parser.Node -->
<class name="Parser.Node" extends="java.lang.Object"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat"/>
<constructor name="Parser.Node" type="java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="addIdentifier"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="ident" type="java.lang.String"/>
<param name="mcstrSig" type="java.lang.Class[]"/>
<param name="nodetype" type="java.lang.Class"/>
<param name="cl" type="java.lang.Class"/>
<exception name="NoSuchMethodException" type="java.lang.NoSuchMethodException"/>
<doc>
<![CDATA[For a given identifier, add a mapping to the nodetype for the parse
tree and to the ComposableRecordReader to be created, including the
formals required to invoke the constructor.
The nodetype and constructor signature should be filled in from the
child node.]]>
</doc>
</method>
<method name="setID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="id" type="int"/>
</method>
<method name="setKeyComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="cmpcl" type="java.lang.Class"/>
</method>
<field name="rrCstrMap" type="java.util.Map"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="id" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="ident" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="cmpcl" type="java.lang.Class"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.Node -->
<!-- start class org.apache.hadoop.mapred.join.Parser.NodeToken -->
<class name="Parser.NodeToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.NodeToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.NumToken -->
<class name="Parser.NumToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.NumToken" type="double"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.NumToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.StrToken -->
<class name="Parser.StrToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.StrToken" type="org.apache.hadoop.mapred.join.Parser.TType, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.StrToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.Token -->
<class name="Parser.Token" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getType" return="org.apache.hadoop.mapred.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Tagged-union type for tokens from the join expression.
@see Parser.TType]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.Token -->
<!-- start class org.apache.hadoop.mapred.join.Parser.TType -->
<class name="Parser.TType" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.join.Parser.TType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.TType -->
<!-- start interface org.apache.hadoop.mapred.join.ResetableIterator -->
<interface name="ResetableIterator" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.ResetableIterator} instead">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<doc>
<![CDATA[This defines an interface to a stateful Iterator that can replay elements
added to it directly.
Note that this does not extend {@link java.util.Iterator}.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.ResetableIterator} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ResetableIterator -->
<!-- start class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
<class name="ResetableIterator.EMPTY" extends="org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
<constructor name="ResetableIterator.EMPTY"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
<!-- start class org.apache.hadoop.mapred.join.StreamBackedIterator -->
<class name="StreamBackedIterator" extends="org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator} instead">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
<constructor name="StreamBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. This
implementation uses a byte array to store elements added to it.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.StreamBackedIterator -->
<!-- start class org.apache.hadoop.mapred.join.TupleWritable -->
<class name="TupleWritable" extends="org.apache.hadoop.mapreduce.lib.join.TupleWritable"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.TupleWritable} instead">
<constructor name="TupleWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an empty tuple with no allocated storage for writables.]]>
</doc>
</constructor>
<constructor name="TupleWritable" type="org.apache.hadoop.io.Writable[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Initialize tuple with storage; unknown whether any of them contain
&quot;written&quot; values.]]>
</doc>
</constructor>
<doc>
<![CDATA[Writable type storing multiple {@link org.apache.hadoop.io.Writable}s.
This is *not* a general-purpose tuple type. In almost all cases, users are
encouraged to implement their own serializable types, which can perform
better validation and provide more efficient encodings than this class is
capable. TupleWritable relies on the join framework for type safety and
assumes its instances will rarely be persisted, assumptions not only
incompatible with, but contrary to the general case.
@see org.apache.hadoop.io.Writable
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.TupleWritable} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.TupleWritable -->
<!-- start class org.apache.hadoop.mapred.join.WrappedRecordReader -->
<class name="WrappedRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader} instead">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key at the head of this RR.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="qkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the head of this RR into the object supplied.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if the RR- including the k,v pair stored in this object-
is exhausted.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next k,v pair into the head of this object; return true iff
the RR and this are exhausted.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an iterator to the collector at the position occupied by this
RecordReader over the values in this stream paired with the key
provided (ie register a stream of values from this source matching K
with a collector).]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="U"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write key-value pair at the head of this stream to the objects provided;
get next key-value pair from proxied RR.]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request new key from proxied RR.]]>
</doc>
</method>
<method name="createValue" return="U"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request new value from proxied RR.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Request progress from proxied RR.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Request position from proxied RR.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Forward close request to proxied RR.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
<doc>
<![CDATA[Implement Comparable contract (compare key at head of proxied RR
with that of another).]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[Return true iff compareTo(other) retn true.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Proxy class for a RecordReader participating in the join framework.
This class keeps track of the &quot;head&quot; key-value pair for the
provided RecordReader and keeps a store of values matching a key when
this source is participating in a join.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.WrappedRecordReader -->
</package>
<package name="org.apache.hadoop.mapred.lib">
<!-- start class org.apache.hadoop.mapred.lib.BinaryPartitioner -->
<class name="BinaryPartitioner" extends="org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner}
instead.">
<implements name="org.apache.hadoop.mapred.Partitioner"/>
<constructor name="BinaryPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[Partition {@link BinaryComparable} keys using a configurable part of
the bytes array returned by {@link BinaryComparable#getBytes()}.
@see org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.BinaryPartitioner -->
<!-- start class org.apache.hadoop.mapred.lib.ChainMapper -->
<class name="ChainMapper" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.chain.ChainMapper} instead">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="ChainMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor.]]>
</doc>
</constructor>
<method name="addMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="byValue" type="boolean"/>
<param name="mapperConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Adds a Mapper class to the chain job's JobConf.
<p/>
It has to be specified how key and values are passed from one element of
the chain to the next, by value or by reference. If a Mapper leverages the
assumed semantics that the key and values are not modified by the collector
'by value' must be used. If the Mapper does not expect this semantics, as
an optimization to avoid serialization and deserialization 'by reference'
can be used.
<p/>
For the added Mapper the configuration given for it,
<code>mapperConf</code>, have precedence over the job's JobConf. This
precedence is in effect when the task is running.
<p/>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the chain
<p/>
@param job job's JobConf to add the Mapper class.
@param klass the Mapper class to add.
@param inputKeyClass mapper input key class.
@param inputValueClass mapper input value class.
@param outputKeyClass mapper output key class.
@param outputValueClass mapper output value class.
@param byValue indicates if key/values should be passed by value
to the next Mapper in the chain, if any.
@param mapperConf a JobConf with the configuration for the Mapper
class. It is recommended to use a JobConf without default values using the
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Configures the ChainMapper and all the Mappers in the chain.
<p/>
If this method is overriden <code>super.configure(...)</code> should be
invoked at the beginning of the overwriter method.]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="value" type="java.lang.Object"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Chains the <code>map(...)</code> methods of the Mappers in the chain.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Closes the ChainMapper and all the Mappers in the chain.
<p/>
If this method is overriden <code>super.close()</code> should be
invoked at the end of the overwriter method.]]>
</doc>
</method>
<doc>
<![CDATA[The ChainMapper class allows to use multiple Mapper classes within a single
Map task.
<p/>
The Mapper classes are invoked in a chained (or piped) fashion, the output of
the first becomes the input of the second, and so on until the last Mapper,
the output of the last Mapper will be written to the task's output.
<p/>
The key functionality of this feature is that the Mappers in the chain do not
need to be aware that they are executed in a chain. This enables having
reusable specialized Mappers that can be combined to perform composite
operations within a single task.
<p/>
Special care has to be taken when creating chains that the key/values output
by a Mapper are valid for the following Mapper in the chain. It is assumed
all Mappers and the Reduce in the chain use maching output and input key and
value classes as no conversion is done by the chaining code.
<p/>
Using the ChainMapper and the ChainReducer classes is possible to compose
Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
immediate benefit of this pattern is a dramatic reduction in disk IO.
<p/>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the chain.
<p/>
ChainMapper usage pattern:
<p/>
<pre>
...
conf.setJobName("chain");
conf.setInputFormat(TextInputFormat.class);
conf.setOutputFormat(TextOutputFormat.class);
<p/>
JobConf mapAConf = new JobConf(false);
...
ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class,
Text.class, Text.class, true, mapAConf);
<p/>
JobConf mapBConf = new JobConf(false);
...
ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, mapBConf);
<p/>
JobConf reduceConf = new JobConf(false);
...
ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class,
Text.class, Text.class, true, reduceConf);
<p/>
ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, null);
<p/>
ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class,
LongWritable.class, LongWritable.class, true, null);
<p/>
FileInputFormat.setInputPaths(conf, inDir);
FileOutputFormat.setOutputPath(conf, outDir);
...
<p/>
JobClient jc = new JobClient(conf);
RunningJob job = jc.submitJob(conf);
...
</pre>
@deprecated
Use {@link org.apache.hadoop.mapreduce.lib.chain.ChainMapper} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.ChainMapper -->
<!-- start class org.apache.hadoop.mapred.lib.ChainReducer -->
<class name="ChainReducer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.chain.ChainReducer} instead">
<implements name="org.apache.hadoop.mapred.Reducer"/>
<constructor name="ChainReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor.]]>
</doc>
</constructor>
<method name="setReducer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="byValue" type="boolean"/>
<param name="reducerConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Sets the Reducer class to the chain job's JobConf.
<p/>
It has to be specified how key and values are passed from one element of
the chain to the next, by value or by reference. If a Reducer leverages the
assumed semantics that the key and values are not modified by the collector
'by value' must be used. If the Reducer does not expect this semantics, as
an optimization to avoid serialization and deserialization 'by reference'
can be used.
<p/>
For the added Reducer the configuration given for it,
<code>reducerConf</code>, have precedence over the job's JobConf. This
precedence is in effect when the task is running.
<p/>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainReducer, this is done by the setReducer or the addMapper for the last
element in the chain.
@param job job's JobConf to add the Reducer class.
@param klass the Reducer class to add.
@param inputKeyClass reducer input key class.
@param inputValueClass reducer input value class.
@param outputKeyClass reducer output key class.
@param outputValueClass reducer output value class.
@param byValue indicates if key/values should be passed by value
to the next Mapper in the chain, if any.
@param reducerConf a JobConf with the configuration for the Reducer
class. It is recommended to use a JobConf without default values using the
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
</doc>
</method>
<method name="addMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="byValue" type="boolean"/>
<param name="mapperConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Adds a Mapper class to the chain job's JobConf.
<p/>
It has to be specified how key and values are passed from one element of
the chain to the next, by value or by reference. If a Mapper leverages the
assumed semantics that the key and values are not modified by the collector
'by value' must be used. If the Mapper does not expect this semantics, as
an optimization to avoid serialization and deserialization 'by reference'
can be used.
<p/>
For the added Mapper the configuration given for it,
<code>mapperConf</code>, have precedence over the job's JobConf. This
precedence is in effect when the task is running.
<p/>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the chain
.
@param job chain job's JobConf to add the Mapper class.
@param klass the Mapper class to add.
@param inputKeyClass mapper input key class.
@param inputValueClass mapper input value class.
@param outputKeyClass mapper output key class.
@param outputValueClass mapper output value class.
@param byValue indicates if key/values should be passed by value
to the next Mapper in the chain, if any.
@param mapperConf a JobConf with the configuration for the Mapper
class. It is recommended to use a JobConf without default values using the
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Configures the ChainReducer, the Reducer and all the Mappers in the chain.
<p/>
If this method is overriden <code>super.configure(...)</code> should be
invoked at the beginning of the overwriter method.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Chains the <code>reduce(...)</code> method of the Reducer with the
<code>map(...) </code> methods of the Mappers in the chain.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Closes the ChainReducer, the Reducer and all the Mappers in the chain.
<p/>
If this method is overriden <code>super.close()</code> should be
invoked at the end of the overwriter method.]]>
</doc>
</method>
<doc>
<![CDATA[The ChainReducer class allows to chain multiple Mapper classes after a
Reducer within the Reducer task.
<p/>
For each record output by the Reducer, the Mapper classes are invoked in a
chained (or piped) fashion, the output of the first becomes the input of the
second, and so on until the last Mapper, the output of the last Mapper will
be written to the task's output.
<p/>
The key functionality of this feature is that the Mappers in the chain do not
need to be aware that they are executed after the Reducer or in a chain.
This enables having reusable specialized Mappers that can be combined to
perform composite operations within a single task.
<p/>
Special care has to be taken when creating chains that the key/values output
by a Mapper are valid for the following Mapper in the chain. It is assumed
all Mappers and the Reduce in the chain use maching output and input key and
value classes as no conversion is done by the chaining code.
<p/>
Using the ChainMapper and the ChainReducer classes is possible to compose
Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
immediate benefit of this pattern is a dramatic reduction in disk IO.
<p/>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainReducer, this is done by the setReducer or the addMapper for the last
element in the chain.
<p/>
ChainReducer usage pattern:
<p/>
<pre>
...
conf.setJobName("chain");
conf.setInputFormat(TextInputFormat.class);
conf.setOutputFormat(TextOutputFormat.class);
<p/>
JobConf mapAConf = new JobConf(false);
...
ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class,
Text.class, Text.class, true, mapAConf);
<p/>
JobConf mapBConf = new JobConf(false);
...
ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, mapBConf);
<p/>
JobConf reduceConf = new JobConf(false);
...
ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class,
Text.class, Text.class, true, reduceConf);
<p/>
ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, null);
<p/>
ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class,
LongWritable.class, LongWritable.class, true, null);
<p/>
FileInputFormat.setInputPaths(conf, inDir);
FileOutputFormat.setOutputPath(conf, outDir);
...
<p/>
JobClient jc = new JobClient(conf);
RunningJob job = jc.submitJob(conf);
...
</pre>
@deprecated
Use {@link org.apache.hadoop.mapreduce.lib.chain.ChainReducer} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.ChainReducer -->
<!-- start class org.apache.hadoop.mapred.lib.CombineFileInputFormat -->
<class name="CombineFileInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat}">
<implements name="org.apache.hadoop.mapred.InputFormat"/>
<constructor name="CombineFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[default constructor]]>
</doc>
</constructor>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createPool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="Use {@link #createPool(List)}.">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="filters" type="java.util.List"/>
<doc>
<![CDATA[Create a new pool and add the filters to it.
A split cannot have files from different pools.
@deprecated Use {@link #createPool(List)}.]]>
</doc>
</method>
<method name="createPool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="Use {@link #createPool(PathFilter...)}.">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="filters" type="org.apache.hadoop.fs.PathFilter[]"/>
<doc>
<![CDATA[Create a new pool and add the filters to it.
A pathname can satisfy any one of the specified filters.
A split cannot have files from different pools.
@deprecated Use {@link #createPool(PathFilter...)}.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This is not implemented yet.]]>
</doc>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An abstract {@link org.apache.hadoop.mapred.InputFormat} that returns {@link CombineFileSplit}'s
in {@link org.apache.hadoop.mapred.InputFormat#getSplits(JobConf, int)} method.
Splits are constructed from the files under the input paths.
A split cannot have files from different pools.
Each split returned may contain blocks from different files.
If a maxSplitSize is specified, then blocks on the same node are
combined to form a single split. Blocks that are left over are
then combined with other blocks in the same rack.
If maxSplitSize is not specified, then blocks from the same rack
are combined in a single split; no attempt is made to create
node-local splits.
If the maxSplitSize is equal to the block size, then this class
is similar to the default spliting behaviour in Hadoop: each
block is a locally processed split.
Subclasses implement {@link org.apache.hadoop.mapred.InputFormat#getRecordReader(InputSplit, JobConf, Reporter)}
to construct <code>RecordReader</code>'s for <code>CombineFileSplit</code>'s.
@see CombineFileSplit
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.CombineFileInputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.CombineFileRecordReader -->
<class name="CombineFileRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader}">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="CombineFileRecordReader" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.mapred.lib.CombineFileSplit, org.apache.hadoop.mapred.Reporter, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A generic RecordReader that can hand out different recordReaders
for each chunk in the CombineFileSplit.]]>
</doc>
</constructor>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[return the amount of data processed]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[return progress based on the amount of data processed so far.]]>
</doc>
</method>
<method name="initNextRecordReader" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the record reader for the next chunk in this CombineFileSplit.]]>
</doc>
</method>
<field name="split" type="org.apache.hadoop.mapred.lib.CombineFileSplit"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="jc" type="org.apache.hadoop.mapred.JobConf"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="reporter" type="org.apache.hadoop.mapred.Reporter"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rrClass" type="java.lang.Class"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rrConstructor" type="java.lang.reflect.Constructor"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="idx" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="progress" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="curReader" type="org.apache.hadoop.mapred.RecordReader"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A generic RecordReader that can hand out different recordReaders
for each chunk in a {@link CombineFileSplit}.
A CombineFileSplit can combine data chunks from multiple files.
This class allows using different RecordReaders for processing
these data chunks from different files.
@see CombineFileSplit
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.CombineFileRecordReader -->
<!-- start class org.apache.hadoop.mapred.lib.CombineFileSplit -->
<class name="CombineFileSplit" extends="org.apache.hadoop.mapreduce.lib.input.CombineFileSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileSplit}">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="CombineFileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[], long[], java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.lib.CombineFileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy constructor]]>
</doc>
</constructor>
<method name="getJob" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.CombineFileSplit}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.CombineFileSplit -->
<!-- start class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
<class name="FieldSelectionMapReduce" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FieldSelectionMapper} and
{@link FieldSelectionReducer} instead">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<implements name="org.apache.hadoop.mapred.Reducer"/>
<constructor name="FieldSelectionMapReduce"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="val" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The identify function. Input key/value pair is written directly to output.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a mapper/reducer class that can be used to perform
field selections in a manner similar to unix cut. The input data is treated
as fields separated by a user specified separator (the default value is
"\t"). The user can specify a list of fields that form the map output keys,
and a list of fields that form the map output values. If the inputformat is
TextInputFormat, the mapper will ignore the key to the map function. and the
fields are from the value only. Otherwise, the fields are the union of those
from the key and those from the value.
The field separator is under attribute "mapreduce.fieldsel.data.field.separator"
The map output field list spec is under attribute
"mapreduce.fieldsel.map.output.key.value.fields.spec".
The value is expected to be like "keyFieldsSpec:valueFieldsSpec"
key/valueFieldsSpec are comma (,) separated field spec: fieldSpec,fieldSpec,fieldSpec ...
Each field spec can be a simple number (e.g. 5) specifying a specific field, or a range
(like 2-5) to specify a range of fields, or an open range (like 3-) specifying all
the fields starting from field 3. The open range field spec applies value fields only.
They have no effect on the key fields.
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields 4,3,0 and 1 for keys,
and use fields 6,5,1,2,3,7 and above for values.
The reduce output field list spec is under attribute
"mapreduce.fieldsel.reduce.output.key.value.fields.spec".
The reducer extracts output key/value pairs in a similar manner, except that
the key is never ignored.
@deprecated Use {@link FieldSelectionMapper} and
{@link FieldSelectionReducer} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
<!-- start class org.apache.hadoop.mapred.lib.FilterOutputFormat -->
<class name="FilterOutputFormat" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat} instead.">
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
<constructor name="FilterOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FilterOutputFormat" type="org.apache.hadoop.mapred.OutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a FilterOutputFormat based on the supplied output format.
@param out the underlying OutputFormat]]>
</doc>
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="baseOut" type="org.apache.hadoop.mapred.OutputFormat"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[FilterOutputFormat is a convenience class that wraps OutputFormat.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.FilterOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter -->
<class name="FilterOutputFormat.FilterRecordWriter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordWriter"/>
<constructor name="FilterOutputFormat.FilterRecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="FilterOutputFormat.FilterRecordWriter" type="org.apache.hadoop.mapred.RecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="rawWriter" type="org.apache.hadoop.mapred.RecordWriter"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[<code>FilterRecordWriter</code> is a convenience wrapper
class that implements {@link RecordWriter}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter -->
<!-- start class org.apache.hadoop.mapred.lib.HashPartitioner -->
<class name="HashPartitioner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.HashPartitioner} instead.">
<implements name="org.apache.hadoop.mapred.Partitioner"/>
<constructor name="HashPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numReduceTasks" type="int"/>
<doc>
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
</doc>
</method>
<doc>
<![CDATA[Partition keys by their {@link Object#hashCode()}.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.HashPartitioner} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.HashPartitioner -->
<!-- start class org.apache.hadoop.mapred.lib.IdentityMapper -->
<class name="IdentityMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="IdentityMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="val" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The identify function. Input key/value pair is written directly to
output.]]>
</doc>
</method>
<doc>
<![CDATA[Implements the identity function, mapping inputs directly to outputs.
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.IdentityMapper -->
<!-- start class org.apache.hadoop.mapred.lib.IdentityReducer -->
<class name="IdentityReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Reducer} instead.">
<implements name="org.apache.hadoop.mapred.Reducer"/>
<constructor name="IdentityReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes all keys and values directly to output.]]>
</doc>
</method>
<doc>
<![CDATA[Performs no reduction, writing all input values directly to the output.
@deprecated Use {@link org.apache.hadoop.mapreduce.Reducer} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.IdentityReducer -->
<!-- start class org.apache.hadoop.mapred.lib.InputSampler -->
<class name="InputSampler" extends="org.apache.hadoop.mapreduce.lib.partition.InputSampler"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.InputSampler}">
<constructor name="InputSampler" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="writePartitionFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="sampler" type="org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<doc>
<![CDATA[@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.InputSampler}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.InputSampler -->
<!-- start class org.apache.hadoop.mapred.lib.InverseMapper -->
<class name="InverseMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.map.InverseMapper}
instead.">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="InverseMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The inverse function. Input keys and values are swapped.]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Mapper} that swaps keys and values.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.map.InverseMapper}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.InverseMapper -->
<!-- start class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator -->
<class name="KeyFieldBasedComparator" extends="org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator}
instead">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="KeyFieldBasedComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[This comparator implementation provides a subset of the features provided
by the Unix/GNU Sort. In particular, the supported features are:
-n, (Sort numerically)
-r, (Reverse the result of comparison)
-k pos1[,pos2], where pos is of the form f[.c][opts], where f is the number
of the field to use, and c is the number of the first character from the
beginning of the field. Fields and character posns are numbered starting
with 1; a character position of zero in pos2 indicates the field's last
character. If '.c' is omitted from pos1, it defaults to 1 (the beginning
of the field); if omitted from pos2, it defaults to 0 (the end of the
field). opts are ordering options (any of 'nr' as described above).
We assume that the fields in the key are separated by
{@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPERATOR}
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator -->
<!-- start class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
<class name="KeyFieldBasedPartitioner" extends="org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner}
instead">
<implements name="org.apache.hadoop.mapred.Partitioner"/>
<constructor name="KeyFieldBasedPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[Defines a way to partition keys based on certain key fields (also see
{@link KeyFieldBasedComparator}.
The key specification supported is of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field).
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
<!-- start class org.apache.hadoop.mapred.lib.LazyOutputFormat -->
<class name="LazyOutputFormat" extends="org.apache.hadoop.mapred.lib.FilterOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat} instead.">
<constructor name="LazyOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setOutputFormatClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the underlying output format for LazyOutputFormat.
@param job the {@link JobConf} to modify
@param theClass the underlying class]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A Convenience class that creates output lazily.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.LazyOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.LongSumReducer -->
<class name="LongSumReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer}
instead.">
<implements name="org.apache.hadoop.mapred.Reducer"/>
<constructor name="LongSumReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Reducer} that sums long values.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.LongSumReducer -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleInputs -->
<class name="MultipleInputs" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.MultipleInputs} instead">
<constructor name="MultipleInputs"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFormatClass" type="java.lang.Class"/>
<doc>
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} to the list of
inputs for the map-reduce job.
@param conf The configuration of the job
@param path {@link Path} to be added to the list of inputs for the job
@param inputFormatClass {@link InputFormat} class to use for this path]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFormatClass" type="java.lang.Class"/>
<param name="mapperClass" type="java.lang.Class"/>
<doc>
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} and
{@link Mapper} to the list of inputs for the map-reduce job.
@param conf The configuration of the job
@param path {@link Path} to be added to the list of inputs for the job
@param inputFormatClass {@link InputFormat} class to use for this path
@param mapperClass {@link Mapper} class to use for this path]]>
</doc>
</method>
<doc>
<![CDATA[This class supports MapReduce jobs that have multiple input paths with
a different {@link InputFormat} and {@link Mapper} for each path
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.MultipleInputs} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleInputs -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
<class name="MultipleOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead">
<constructor name="MultipleOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a composite record writer that can write key/value data to different
output files
@param fs
the file system to use
@param job
the job conf for the job
@param name
the leaf file name for the output file (such as part-00000")
@param arg3
a progressable for reporting progress.
@return a composite record writer
@throws IOException]]>
</doc>
</method>
<method name="generateLeafFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the leaf name for the output file name. The default behavior does
not change the leaf file name (such as part-00000)
@param name
the leaf file name for the output file
@return the given leaf file name]]>
</doc>
</method>
<method name="generateFileNameForKeyValue" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the file output file name based on the given key and the leaf file
name. The default behavior is that the file name does not depend on the
key.
@param key
the key of the output data
@param name
the leaf file name
@return generated file name]]>
</doc>
</method>
<method name="generateActualKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<doc>
<![CDATA[Generate the actual key from the given key/value. The default behavior is that
the actual key is equal to the given key
@param key
the key of the output data
@param value
the value of the output data
@return the actual key derived from the given key/value]]>
</doc>
</method>
<method name="generateActualValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<doc>
<![CDATA[Generate the actual value from the given key and value. The default behavior is that
the actual value is equal to the given value
@param key
the key of the output data
@param value
the value of the output data
@return the actual value derived from the given key/value]]>
</doc>
</method>
<method name="getInputFileBasedOutputFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the outfile name based on a given anme and the input file name. If
the {@link JobContext#MAP_INPUT_FILE} does not exists (i.e. this is not for a map only job),
the given name is returned unchanged. If the config value for
"num.of.trailing.legs.to.use" is not set, or set 0 or negative, the given
name is returned unchanged. Otherwise, return a file name consisting of the
N trailing legs of the input file name where N is the config value for
"num.of.trailing.legs.to.use".
@param job
the job config
@param name
the output file name
@return the outfile name based on a given anme and the input file name.]]>
</doc>
</method>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param fs
the file system to use
@param job
a job conf object
@param name
the name of the file over which a record writer object will be
constructed
@param arg3
a progressable object
@return A RecordWriter object over the given file
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This abstract class extends the FileOutputFormat, allowing to write the
output data to different output files. There are three basic use cases for
this class.
Case one: This class is used for a map reduce job with at least one reducer.
The reducer wants to write data to different files depending on the actual
keys. It is assumed that a key (or value) encodes the actual key (value)
and the desired location for the actual key (value).
Case two: This class is used for a map only job. The job wants to use an
output file name that is either a part of the input file name of the input
data, or some derivation of it.
Case three: This class is used for a map only job. The job wants to use an
output file name that depends on both the keys and the input file name,
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleOutputs -->
<class name="MultipleOutputs" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead">
<constructor name="MultipleOutputs" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates and initializes multiple named outputs support, it should be
instantiated in the Mapper/Reducer configure method.
@param job the job configuration object]]>
</doc>
</constructor>
<method name="getNamedOutputsList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Returns list of channel names.
@param conf job conf
@return List of channel Names]]>
</doc>
</method>
<method name="isMultiNamedOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<doc>
<![CDATA[Returns if a named output is multiple.
@param conf job conf
@param namedOutput named output
@return <code>true</code> if the name output is multi, <code>false</code>
if it is single. If the name output is not defined it returns
<code>false</code>]]>
</doc>
</method>
<method name="getNamedOutputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<doc>
<![CDATA[Returns the named output OutputFormat.
@param conf job conf
@param namedOutput named output
@return namedOutput OutputFormat]]>
</doc>
</method>
<method name="getNamedOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<doc>
<![CDATA[Returns the key class for a named output.
@param conf job conf
@param namedOutput named output
@return class for the named output key]]>
</doc>
</method>
<method name="getNamedOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<doc>
<![CDATA[Returns the value class for a named output.
@param conf job conf
@param namedOutput named output
@return class of named output value]]>
</doc>
</method>
<method name="addNamedOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<param name="outputFormatClass" type="java.lang.Class"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valueClass" type="java.lang.Class"/>
<doc>
<![CDATA[Adds a named output for the job.
<p/>
@param conf job conf to add the named output
@param namedOutput named output name, it has to be a word, letters
and numbers only, cannot be the word 'part' as
that is reserved for the
default output.
@param outputFormatClass OutputFormat class.
@param keyClass key class
@param valueClass value class]]>
</doc>
</method>
<method name="addMultiNamedOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="namedOutput" type="java.lang.String"/>
<param name="outputFormatClass" type="java.lang.Class"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valueClass" type="java.lang.Class"/>
<doc>
<![CDATA[Adds a multi named output for the job.
<p/>
@param conf job conf to add the named output
@param namedOutput named output name, it has to be a word, letters
and numbers only, cannot be the word 'part' as
that is reserved for the
default output.
@param outputFormatClass OutputFormat class.
@param keyClass key class
@param valueClass value class]]>
</doc>
</method>
<method name="setCountersEnabled"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="enabled" type="boolean"/>
<doc>
<![CDATA[Enables or disables counters for the named outputs.
<p/>
By default these counters are disabled.
<p/>
MultipleOutputs supports counters, by default the are disabled.
The counters group is the {@link MultipleOutputs} class name.
</p>
The names of the counters are the same as the named outputs. For multi
named outputs the name of the counter is the concatenation of the named
output, and underscore '_' and the multiname.
@param conf job conf to enableadd the named output.
@param enabled indicates if the counters will be enabled or not.]]>
</doc>
</method>
<method name="getCountersEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Returns if the counters for the named outputs are enabled or not.
<p/>
By default these counters are disabled.
<p/>
MultipleOutputs supports counters, by default the are disabled.
The counters group is the {@link MultipleOutputs} class name.
</p>
The names of the counters are the same as the named outputs. For multi
named outputs the name of the counter is the concatenation of the named
output, and underscore '_' and the multiname.
@param conf job conf to enableadd the named output.
@return TRUE if the counters are enabled, FALSE if they are disabled.]]>
</doc>
</method>
<method name="getNamedOutputs" return="java.util.Iterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns iterator with the defined name outputs.
@return iterator with the defined named outputs]]>
</doc>
</method>
<method name="getCollector" return="org.apache.hadoop.mapred.OutputCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="namedOutput" type="java.lang.String"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the output collector for a named output.
<p/>
@param namedOutput the named output name
@param reporter the reporter
@return the output collector for the given named output
@throws IOException thrown if output collector could not be created]]>
</doc>
</method>
<method name="getCollector" return="org.apache.hadoop.mapred.OutputCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="namedOutput" type="java.lang.String"/>
<param name="multiName" type="java.lang.String"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the output collector for a multi named output.
<p/>
@param namedOutput the named output name
@param multiName the multi name part
@param reporter the reporter
@return the output collector for the given named output
@throws IOException thrown if output collector could not be created]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Closes all the opened named outputs.
<p/>
If overriden subclasses must invoke <code>super.close()</code> at the
end of their <code>close()</code>
@throws java.io.IOException thrown if any of the MultipleOutput files
could not be closed properly.]]>
</doc>
</method>
<doc>
<![CDATA[The MultipleOutputs class simplifies writting to additional outputs other
than the job default output via the <code>OutputCollector</code> passed to
the <code>map()</code> and <code>reduce()</code> methods of the
<code>Mapper</code> and <code>Reducer</code> implementations.
<p/>
Each additional output, or named output, may be configured with its own
<code>OutputFormat</code>, with its own key class and with its own value
class.
<p/>
A named output can be a single file or a multi file. The later is refered as
a multi named output.
<p/>
A multi named output is an unbound set of files all sharing the same
<code>OutputFormat</code>, key class and value class configuration.
<p/>
When named outputs are used within a <code>Mapper</code> implementation,
key/values written to a name output are not part of the reduce phase, only
key/values written to the job <code>OutputCollector</code> are part of the
reduce phase.
<p/>
MultipleOutputs supports counters, by default the are disabled. The counters
group is the {@link MultipleOutputs} class name.
</p>
The names of the counters are the same as the named outputs. For multi
named outputs the name of the counter is the concatenation of the named
output, and underscore '_' and the multiname.
<p/>
Job configuration usage pattern is:
<pre>
JobConf conf = new JobConf();
conf.setInputPath(inDir);
FileOutputFormat.setOutputPath(conf, outDir);
conf.setMapperClass(MOMap.class);
conf.setReducerClass(MOReduce.class);
...
// Defines additional single text based output 'text' for the job
MultipleOutputs.addNamedOutput(conf, "text", TextOutputFormat.class,
LongWritable.class, Text.class);
// Defines additional multi sequencefile based output 'sequence' for the
// job
MultipleOutputs.addMultiNamedOutput(conf, "seq",
SequenceFileOutputFormat.class,
LongWritable.class, Text.class);
...
JobClient jc = new JobClient();
RunningJob job = jc.submitJob(conf);
...
</pre>
<p/>
Job configuration usage pattern is:
<pre>
public class MOReduce implements
Reducer&lt;WritableComparable, Writable&gt; {
private MultipleOutputs mos;
public void configure(JobConf conf) {
...
mos = new MultipleOutputs(conf);
}
public void reduce(WritableComparable key, Iterator&lt;Writable&gt; values,
OutputCollector output, Reporter reporter)
throws IOException {
...
mos.getCollector("text", reporter).collect(key, new Text("Hello"));
mos.getCollector("seq", "A", reporter).collect(key, new Text("Bye"));
mos.getCollector("seq", "B", reporter).collect(key, new Text("Chau"));
...
}
public void close() throws IOException {
mos.close();
...
}
}
</pre>
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleOutputs -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
<class name="MultipleSequenceFileOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead">
<constructor name="MultipleSequenceFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output data
to different output files in sequence file output format.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
<class name="MultipleTextOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead">
<constructor name="MultipleTextOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output
data to different output files in Text output format.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.MultipleOutputs} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
<class name="MultithreadedMapRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link MultithreadedMapper} instead.">
<implements name="org.apache.hadoop.mapred.MapRunnable"/>
<constructor name="MultithreadedMapRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Multithreaded implementation for @link org.apache.hadoop.mapred.MapRunnable.
<p>
It can be used instead of the default implementation,
@link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU
bound in order to improve throughput.
<p>
Map implementations using this MapRunnable must be thread-safe.
<p>
The Map-Reduce job has to be configured to use this MapRunnable class (using
the JobConf.setMapRunnerClass method) and
the number of thread the thread-pool can use with the
<code>mapred.map.multithreadedrunner.threads</code> property, its default
value is 10 threads.
<p>
@deprecated Use {@link MultithreadedMapper} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
<!-- start class org.apache.hadoop.mapred.lib.NLineInputFormat -->
<class name="NLineInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.input.NLineInputFormat} instead">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="NLineInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Logically splits the set of input files for the job, splits N lines
of the input as one split.
@see org.apache.hadoop.mapred.FileInputFormat#getSplits(JobConf, int)]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[NLineInputFormat which splits N lines of input as one split.
In many "pleasantly" parallel applications, each process/mapper
processes the same input file (s), but with computations are
controlled by different parameters.(Referred to as "parameter sweeps").
One way to achieve this, is to specify a set of parameters
(one set per line) as input in a control file
(which is the input path to the map-reduce application,
where as the input dataset is specified
via a config variable in JobConf.).
The NLineInputFormat can be used in such applications, that splits
the input file such that by default, one line is fed as
a value to one map task, and key is the offset.
i.e. (k,v) is (LongWritable, Text).
The location hints will span the whole mapred cluster.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.input.NLineInputFormat} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.NLineInputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.NullOutputFormat -->
<class name="NullOutputFormat" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat} instead.">
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
<constructor name="NullOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[Consume all outputs and put them in /dev/null.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.NullOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.RegexMapper -->
<class name="RegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.map.RegexMapper}">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="RegexMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.map.RegexMapper}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.RegexMapper -->
<!-- start class org.apache.hadoop.mapred.lib.TokenCountMapper -->
<class name="TokenCountMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper} instead.">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="TokenCountMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that maps text values into <token,freq> pairs. Uses
{@link StringTokenizer} to break text into tokens.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.TokenCountMapper -->
<!-- start class org.apache.hadoop.mapred.lib.TotalOrderPartitioner -->
<class name="TotalOrderPartitioner" extends="org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner}">
<implements name="org.apache.hadoop.mapred.Partitioner"/>
<constructor name="TotalOrderPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[Partitioner effecting a total order by reading split points from
an externally generated source.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.TotalOrderPartitioner -->
</package>
<package name="org.apache.hadoop.mapred.lib.aggregate">
<!-- start class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
<class name="DoubleValueSum" extends="org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="DoubleValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that sums up a sequence of double
values.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
<class name="LongValueMax" extends="org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that maintain the maximum of
a sequence of long values.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
<class name="LongValueMin" extends="org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that maintain the minimum of
a sequence of long values.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
<class name="LongValueSum" extends="org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that sums up
a sequence of long values.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
<class name="StringValueMax" extends="org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that maintain the biggest of
a sequence of strings.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
<class name="StringValueMin" extends="org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that maintain the smallest of
a sequence of strings.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
<class name="UniqValueCount" extends="org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="UniqValueCount"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<constructor name="UniqValueCount" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructor
@param maxNum the limit in the number of unique values to keep.]]>
</doc>
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that dedupes a sequence of objects.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<class name="UserDefinedValueAggregatorDescriptor" extends="org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor}
instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="UserDefinedValueAggregatorDescriptor" type="java.lang.String, org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param className the class name of the user defined descriptor class
@param job a configure object used for decriptor configuration]]>
</doc>
</constructor>
<method name="createInstance" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="className" type="java.lang.String"/>
<doc>
<![CDATA[Create an instance of the given class
@param className the name of the class
@return a dynamically created instance of the given class]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Do nothing.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a wrapper for a user defined value aggregator descriptor.
It servs two functions: One is to create an object of ValueAggregatorDescriptor from the
name of a user defined class that may be dynamically loaded. The other is to
deligate inviokations of generateKeyValPairs function to the created object.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
<interface name="ValueAggregator" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator} instead">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<doc>
<![CDATA[This interface defines the minimal protocol for value aggregators.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
<class name="ValueAggregatorBaseDescriptor" extends="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor}
instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="ValueAggregatorBaseDescriptor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="generateEntry" return="java.util.Map.Entry"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<param name="id" type="java.lang.String"/>
<param name="val" type="org.apache.hadoop.io.Text"/>
<doc>
<![CDATA[@param type the aggregation type
@param id the aggregation id
@param val the val associated with the id to be aggregated
@return an Entry whose key is the aggregation id prefixed with
the aggregation type.]]>
</doc>
</method>
<method name="generateValueAggregator" return="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<doc>
<![CDATA[@param type the aggregation type
@return a value aggregator of the given type.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[get the input file name.
@param job a job configuration object]]>
</doc>
</method>
<field name="UNIQ_VALUE_COUNT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOUBLE_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="VALUE_HISTOGRAM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements the common functionalities of
the subclasses of ValueAggregatorDescriptor class.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
<class name="ValueAggregatorCombiner" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner}
instead">
<constructor name="ValueAggregatorCombiner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Combiner does not need to configure.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Combines values for a given key.
@param key the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values.
@param values the values to combine
@param output to collect combined values]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing.]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="K1"/>
<param name="arg1" type="V1"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic combiner of Aggregate.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
<interface name="ValueAggregatorDescriptor" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor}
instead">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor"/>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Configure the object
@param job
a JobConf object that may contain the information that can be used
to configure the object.]]>
</doc>
</method>
<field name="TYPE_SEPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ONE" type="org.apache.hadoop.io.Text"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This interface defines the contract a value aggregator descriptor must
support. Such a descriptor can be configured with a JobConf object. Its main
function is to generate a list of aggregation-id/value pairs. An aggregation
id encodes an aggregation type which is used to guide the way to aggregate
the value in the reduce/combiner phrase of an Aggregate based job.The mapper in
an Aggregate based map/reduce job may create one or more of
ValueAggregatorDescriptor objects at configuration time. For each input
key/value pair, the mapper will use those objects to create aggregation
id/value pairs.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor}
instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
<class name="ValueAggregatorJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob} instead">
<constructor name="ValueAggregatorJob"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create an Aggregate based map/reduce job.
@param args the arguments used for job creation. Generic hadoop
arguments are accepted.
@return a JobConf object ready for submission.
@throws IOException
@see GenericOptionsParser]]>
</doc>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setAggregatorDescriptors"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="descriptors" type="java.lang.Class[]"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[create and run an Aggregate based map/reduce job.
@param args the arguments used for job creation
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This is the main class for creating a map/reduce job using Aggregate
framework. The Aggregate is a specialization of map/reduce framework,
specilizing for performing various simple aggregations.
Generally speaking, in order to implement an application using Map/Reduce
model, the developer is to implement Map and Reduce functions (and possibly
combine function). However, a lot of applications related to counting and
statistics computing have very similar characteristics. Aggregate abstracts
out the general patterns of these functions and implementing those patterns.
In particular, the package provides generic mapper/redducer/combiner classes,
and a set of built-in value aggregators, and a generic utility class that
helps user create map/reduce jobs using the generic class. The built-in
aggregators include:
sum over numeric values count the number of distinct values compute the
histogram of values compute the minimum, maximum, media,average, standard
deviation of numeric values
The developer using Aggregate will need only to provide a plugin class
conforming to the following interface:
public interface ValueAggregatorDescriptor { public ArrayList<Entry>
generateKeyValPairs(Object key, Object value); public void
configure(JobConfjob); }
The package also provides a base class, ValueAggregatorBaseDescriptor,
implementing the above interface. The user can extend the base class and
implement generateKeyValPairs accordingly.
The primary work of generateKeyValPairs is to emit one or more key/value
pairs based on the input key/value pair. The key in an output key/value pair
encode two pieces of information: aggregation type and aggregation id. The
value will be aggregated onto the aggregation id according the aggregation
type.
This class offers a function to generate a map/reduce job using Aggregate
framework. The function takes the following parameters: input directory spec
input format (text or sequence file) output directory a file specifying the
user plugin class
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
<class name="ValueAggregatorJobBase" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase}
instead">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<implements name="org.apache.hadoop.mapred.Reducer"/>
<constructor name="ValueAggregatorJobBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="logSpec"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="aggregatorDescriptorList" type="java.util.ArrayList"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This abstract class implements some common functionalities of the
the generic mapper, reducer and combiner classes of Aggregate.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
<class name="ValueAggregatorMapper" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper}
instead">
<constructor name="ValueAggregatorMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K1"/>
<param name="value" type="V1"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[the map function. It iterates through the value aggregator descriptor
list to generate aggregation id/value pairs and emit them.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="org.apache.hadoop.io.Text"/>
<param name="arg1" type="java.util.Iterator"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic mapper of Aggregate.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
<class name="ValueAggregatorReducer" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer}
instead">
<constructor name="ValueAggregatorReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param key
the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values. In effect, data
driven computing is achieved. It is assumed that each aggregator's
getReport method emits appropriate output for the aggregator. This
may be further customiized.
@value the values to be aggregated]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="K1"/>
<param name="arg1" type="V1"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic reducer of Aggregate.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer}
instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
<class name="ValueHistogram" extends="org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram} instead">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="ValueHistogram"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This class implements a value aggregator that computes the
histogram of a sequence of strings.
@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
</package>
<package name="org.apache.hadoop.mapred.lib.db">
<!-- start class org.apache.hadoop.mapred.lib.db.DBConfiguration -->
<class name="DBConfiguration" extends="org.apache.hadoop.mapreduce.lib.db.DBConfiguration"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use
{@link org.apache.hadoop.mapreduce.lib.db.DBConfiguration} instead">
<method name="configureDB"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="driverClass" type="java.lang.String"/>
<param name="dbUrl" type="java.lang.String"/>
<param name="userName" type="java.lang.String"/>
<param name="passwd" type="java.lang.String"/>
<doc>
<![CDATA[Sets the DB access related fields in the JobConf.
@param job the job
@param driverClass JDBC Driver class name
@param dbUrl JDBC DB access URL.
@param userName DB access username
@param passwd DB access passwd]]>
</doc>
</method>
<method name="configureDB"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="driverClass" type="java.lang.String"/>
<param name="dbUrl" type="java.lang.String"/>
<doc>
<![CDATA[Sets the DB access related fields in the JobConf.
@param job the job
@param driverClass JDBC Driver class name
@param dbUrl JDBC DB access URL.]]>
</doc>
</method>
<field name="DRIVER_CLASS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The JDBC Driver class name]]>
</doc>
</field>
<field name="URL_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[JDBC Database access URL]]>
</doc>
</field>
<field name="USERNAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[User name to access the database]]>
</doc>
</field>
<field name="PASSWORD_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Password to access the database]]>
</doc>
</field>
<field name="INPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Input table name]]>
</doc>
</field>
<field name="INPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Field names in the Input table]]>
</doc>
</field>
<field name="INPUT_CONDITIONS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[WHERE clause in the input SELECT statement]]>
</doc>
</field>
<field name="INPUT_ORDER_BY_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[ORDER BY clause in the input SELECT statement]]>
</doc>
</field>
<field name="INPUT_QUERY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Whole input query, exluding LIMIT...OFFSET]]>
</doc>
</field>
<field name="INPUT_COUNT_QUERY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Input query to get the count of records]]>
</doc>
</field>
<field name="INPUT_CLASS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Class name implementing DBWritable which will hold input tuples]]>
</doc>
</field>
<field name="OUTPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Output table name]]>
</doc>
</field>
<field name="OUTPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Field names in the Output table]]>
</doc>
</field>
<field name="OUTPUT_FIELD_COUNT_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of fields in the Output table]]>
</doc>
</field>
<doc>
<![CDATA[@deprecated Use
{@link org.apache.hadoop.mapreduce.lib.db.DBConfiguration} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBConfiguration -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat -->
<class name="DBInputFormat" extends="org.apache.hadoop.mapreduce.lib.db.DBInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.db.DBInputFormat} instead.">
<implements name="org.apache.hadoop.mapred.InputFormat"/>
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="DBInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="chunks" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="tableName" type="java.lang.String"/>
<param name="conditions" type="java.lang.String"/>
<param name="orderBy" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<doc>
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
@param job The job
@param inputClass the class object implementing DBWritable, which is the
Java object holding tuple fields.
@param tableName The table to read data from
@param conditions The condition which to select data with, eg. '(updated >
20070101 AND length > 0)'
@param orderBy the fieldNames in the orderBy clause.
@param fieldNames The field names in the table
@see #setInput(JobConf, Class, String, String)]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="inputQuery" type="java.lang.String"/>
<param name="inputCountQuery" type="java.lang.String"/>
<doc>
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
@param job The job
@param inputClass the class object implementing DBWritable, which is the
Java object holding tuple fields.
@param inputQuery the input query to select fields. Example :
"SELECT f1, f2, f3 FROM Mytable ORDER BY f1"
@param inputCountQuery the input query that returns the number of records in
the table.
Example : "SELECT COUNT(f1) FROM Mytable"
@see #setInput(JobConf, Class, String, String, String, String...)]]>
</doc>
</method>
<doc>
<![CDATA[@deprecated
Use {@link org.apache.hadoop.mapreduce.lib.db.DBInputFormat} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit -->
<class name="DBInputFormat.DBInputSplit" extends="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="DBInputFormat.DBInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default Constructor]]>
</doc>
</constructor>
<constructor name="DBInputFormat.DBInputSplit" type="long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convenience Constructor
@param start the index of the first row to select
@param end the index of the last row to select]]>
</doc>
</constructor>
<doc>
<![CDATA[A InputSplit that spans a set of rows]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader -->
<class name="DBInputFormat.DBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DBRecordReader"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader"/>
<constructor name="DBInputFormat.DBRecordReader" type="org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.mapred.JobConf, java.sql.Connection, org.apache.hadoop.mapred.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[@param split The InputSplit to read data for
@throws SQLException]]>
</doc>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.LongWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="createValue" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.LongWritable"/>
<param name="value" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A RecordReader that reads records from a SQL table.
Emits LongWritables containing the record number as
key and DBWritables as value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable -->
<class name="DBInputFormat.NullDBWritable" extends="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.db.DBWritable"/>
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="DBInputFormat.NullDBWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A Class that does nothing, implementing DBWritable]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBOutputFormat -->
<class name="DBOutputFormat" extends="org.apache.hadoop.mapreduce.lib.db.DBOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use org.apache.hadoop.mapreduce.lib.db.DBOutputFormat instead">
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
<constructor name="DBOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filesystem" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filesystem" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="tableName" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<doc>
<![CDATA[Initializes the reduce-part of the job with the appropriate output settings
@param job The job
@param tableName The table to insert data into
@param fieldNames The field names in the table.]]>
</doc>
</method>
<method name="setOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="tableName" type="java.lang.String"/>
<param name="fieldCount" type="int"/>
<doc>
<![CDATA[Initializes the reduce-part of the job with the appropriate output settings
@param job The job
@param tableName The table to insert data into
@param fieldCount the number of fields in the table.]]>
</doc>
</method>
<doc>
<![CDATA[@deprecated Use org.apache.hadoop.mapreduce.lib.db.DBOutputFormat instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter -->
<class name="DBOutputFormat.DBRecordWriter" extends="org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordWriter"/>
<constructor name="DBOutputFormat.DBRecordWriter" type="java.sql.Connection, java.sql.PreparedStatement"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A RecordWriter that writes the reduce output to a SQL table]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter -->
<!-- start interface org.apache.hadoop.mapred.lib.db.DBWritable -->
<interface name="DBWritable" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.db.DBWritable} instead">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBWritable"/>
<doc>
<![CDATA[@deprecated
Use {@link org.apache.hadoop.mapreduce.lib.db.DBWritable} instead]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.lib.db.DBWritable -->
</package>
<package name="org.apache.hadoop.mapred.pipes">
<!-- start class org.apache.hadoop.mapred.pipes.Submitter -->
<class name="Submitter" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="Submitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Submitter" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getExecutable" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the URI of the application's executable.
@param conf
@return the URI where the application's executable is located]]>
</doc>
</method>
<method name="setExecutable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="executable" type="java.lang.String"/>
<doc>
<![CDATA[Set the URI for the application's executable. Normally this is a hdfs:
location.
@param conf
@param executable The URI of the application's executable.]]>
</doc>
</method>
<method name="setIsJavaRecordReader"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the job is using a Java RecordReader.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaRecordReader" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java RecordReader
@param conf the configuration to check
@return is it a Java RecordReader?]]>
</doc>
</method>
<method name="setIsJavaMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the Mapper is written in Java.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaMapper" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java Mapper.
@param conf the configuration to check
@return is it a Java Mapper?]]>
</doc>
</method>
<method name="setIsJavaReducer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the Reducer is written in Java.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaReducer" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java Reducer.
@param conf the configuration to check
@return is it a Java Reducer?]]>
</doc>
</method>
<method name="setIsJavaRecordWriter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the job will use a Java RecordWriter.
@param conf the configuration to modify
@param value the new value to set]]>
</doc>
</method>
<method name="getIsJavaRecordWriter" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Will the reduce use a Java RecordWriter?
@param conf the configuration to check
@return true, if the output of the job will be written by Java]]>
</doc>
</method>
<method name="getKeepCommandFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Does the user want to keep the command file for debugging? If this is
true, pipes will write a copy of the command data to a file in the
task directory named "downlink.data", which may be used to run the C++
program under the debugger. You probably also want to set
JobConf.setKeepFailedTaskFiles(true) to keep the entire directory from
being deleted.
To run using the data file, set the environment variable
"mapreduce.pipes.commandfile" to point to the file.
@param conf the configuration to check
@return will the framework save the command file?]]>
</doc>
</method>
<method name="setKeepCommandFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="keep" type="boolean"/>
<doc>
<![CDATA[Set whether to keep the command file for debugging
@param conf the configuration to modify
@param keep the new value]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="Use {@link Submitter#runJob(JobConf)}">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the map/reduce cluster. All of the necessary modifications
to the job to run under pipes are made to the configuration.
@param conf the job to submit to the cluster (MODIFIED)
@throws IOException
@deprecated Use {@link Submitter#runJob(JobConf)}]]>
</doc>
</method>
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the map/reduce cluster. All of the necessary modifications
to the job to run under pipes are made to the configuration.
@param conf the job to submit to the cluster (MODIFIED)
@throws IOException]]>
</doc>
</method>
<method name="jobSubmit" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the Map-Reduce framework.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param conf the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws IOException]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Submit a pipes job based on the command line arguments.
@param args]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="PRESERVE_COMMANDFILE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="EXECUTABLE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INTERPRETOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IS_JAVA_MAP" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IS_JAVA_RR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IS_JAVA_RW" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IS_JAVA_REDUCE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PARTITIONER" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INPUT_FORMAT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PORT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[The main entry point and job submitter. It may either be used as a command
line-based or API-based method to launch Pipes jobs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.pipes.Submitter -->
</package>
<package name="org.apache.hadoop.mapred.tools">
<!-- start class org.apache.hadoop.mapred.tools.MRAdmin -->
<class name="MRAdmin" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="MRAdmin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="MRAdmin" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="refreshSuperUserGroupsConfiguration" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[refreshSuperUserGroupsConfiguration {@link JobTracker}.
@return exitcode 0 on success, non-zero on failure
@throws IOException]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[Administrative access to Hadoop Map-Reduce.
Currently it only provides the ability to connect to the {@link JobTracker}
and 1) refresh the service-level authorization policy, 2) refresh queue acl
properties.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.tools.MRAdmin -->
</package>
<package name="org.apache.hadoop.mapreduce">
<!-- start class org.apache.hadoop.mapreduce.Cluster -->
<class name="Cluster" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Cluster" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="Cluster" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the <code>Cluster</code>.]]>
</doc>
</method>
<method name="getFileSystem" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the file system where job-specific files are stored
@return object of FileSystem
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapreduce.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get job corresponding to jobid.
@param jobId
@return object of {@link Job}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getQueues" return="org.apache.hadoop.mapreduce.QueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get all the queues in cluster.
@return array of {@link QueueInfo}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getQueue" return="org.apache.hadoop.mapreduce.QueueInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get queue information for the specified name.
@param name queuename
@return object of {@link QueueInfo}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapreduce.ClusterMetrics"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get current cluster status.
@return object of {@link ClusterMetrics}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getActiveTaskTrackers" return="org.apache.hadoop.mapreduce.TaskTrackerInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get all active trackers in the cluster.
@return array of {@link TaskTrackerInfo}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getBlackListedTaskTrackers" return="org.apache.hadoop.mapreduce.TaskTrackerInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get blacklisted trackers.
@return array of {@link TaskTrackerInfo}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapreduce.Job[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get all the jobs in cluster.
@return array of {@link Job}
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getSystemDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Grab the jobtracker system directory path where
job-specific files will be placed.
@return the system directory where job-specific files are to be placed.]]>
</doc>
</method>
<method name="getStagingAreaDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Grab the jobtracker's view of the staging directory path where
job-specific files will be placed.
@return the staging directory where job-specific files are to be placed.]]>
</doc>
</method>
<method name="getJobHistoryUrl" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapreduce.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the job history file path for a given job id. The job history file at
this path may or may not be existing depending on the job completion state.
The file is present only for the completed jobs.
@param jobId the JobID of the job submitted by the current user.
@return the file path of the job history file
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getQueueAclsForCurrentUser" return="org.apache.hadoop.mapreduce.QueueAclsInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Gets the Queue ACLs for current user
@return array of QueueAclsInfo object for current user.
@throws IOException]]>
</doc>
</method>
<method name="getRootQueues" return="org.apache.hadoop.mapreduce.QueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Gets the root level queues.
@return array of JobQueueInfo object.
@throws IOException]]>
</doc>
</method>
<method name="getChildQueues" return="org.apache.hadoop.mapreduce.QueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Returns immediate children of queueName.
@param queueName
@return array of JobQueueInfo which are children of queueName
@throws IOException]]>
</doc>
</method>
<method name="getJobTrackerState" return="org.apache.hadoop.mapreduce.server.jobtracker.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #getJobTrackerStatus()} instead.">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get JobTracker's state
@return {@link State} of the JobTracker
@throws IOException
@throws InterruptedException
@deprecated Use {@link #getJobTrackerStatus()} instead.]]>
</doc>
</method>
<method name="getJobTrackerStatus" return="org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the JobTracker's status.
@return {@link JobTrackerStatus} of the JobTracker
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getTaskTrackerExpiryInterval" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the tasktracker expiry interval for the cluster
@return the expiry interval in msec]]>
</doc>
</method>
<method name="getDelegationToken" return="org.apache.hadoop.security.token.Token"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="renewer" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get a delegation token for the user from the JobTracker.
@param renewer the user who can renew the token
@return the new token
@throws IOException]]>
</doc>
</method>
<method name="renewDelegationToken" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="SecretManager.InvalidToken" type="org.apache.hadoop.security.token.SecretManager.InvalidToken"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Renew a delegation token
@param token the token to renew
@return the new expiration time
@throws InvalidToken
@throws IOException]]>
</doc>
</method>
<method name="cancelDelegationToken"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Cancel a delegation token from the JobTracker
@param token the token to cancel
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Provides a way to access information about the map/reduce cluster.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Cluster -->
<!-- start class org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus -->
<class name="Cluster.JobTrackerStatus" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus -->
<!-- start class org.apache.hadoop.mapreduce.ClusterMetrics -->
<class name="ClusterMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="ClusterMetrics"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ClusterMetrics" type="int, int, int, int, int, int, int, int, int, int, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRunningMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of running map tasks in the cluster.
@return running maps]]>
</doc>
</method>
<method name="getRunningReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of running reduce tasks in the cluster.
@return running reduces]]>
</doc>
</method>
<method name="getOccupiedMapSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get number of occupied map slots in the cluster.
@return occupied map slot count]]>
</doc>
</method>
<method name="getOccupiedReduceSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of occupied reduce slots in the cluster.
@return occupied reduce slot count]]>
</doc>
</method>
<method name="getReservedMapSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get number of reserved map slots in the cluster.
@return reserved map slot count]]>
</doc>
</method>
<method name="getReservedReduceSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of reserved reduce slots in the cluster.
@return reserved reduce slot count]]>
</doc>
</method>
<method name="getMapSlotCapacity" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total number of map slots in the cluster.
@return map slot capacity]]>
</doc>
</method>
<method name="getReduceSlotCapacity" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total number of reduce slots in the cluster.
@return reduce slot capacity]]>
</doc>
</method>
<method name="getTotalJobSubmissions" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total number of job submissions in the cluster.
@return total number of job submissions]]>
</doc>
</method>
<method name="getTaskTrackerCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of active trackers in the cluster.
@return active tracker count.]]>
</doc>
</method>
<method name="getBlackListedTaskTrackerCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of blacklisted trackers in the cluster.
@return blacklisted tracker count]]>
</doc>
</method>
<method name="getDecommissionedTaskTrackerCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of decommissioned trackers in the cluster.
@return decommissioned tracker count]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Status information on the current state of the Map-Reduce cluster.
<p><code>ClusterMetrics</code> provides clients with information such as:
<ol>
<li>
Size of the cluster.
</li>
<li>
Number of blacklisted and decommissioned trackers.
</li>
<li>
Slot capacity of the cluster.
</li>
<li>
The number of currently occupied/reserved map & reduce slots.
</li>
<li>
The number of currently running map & reduce tasks.
</li>
<li>
The number of job submissions.
</li>
</ol></p>
<p>Clients can query for the latest <code>ClusterMetrics</code>, via
{@link Cluster#getClusterStatus()}.</p>
@see Cluster]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.ClusterMetrics -->
<!-- start class org.apache.hadoop.mapreduce.ContextFactory -->
<class name="ContextFactory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ContextFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="cloneContext" return="org.apache.hadoop.mapreduce.JobContext"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="original" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Clone a job or task attempt context with a new configuration.
@param original the original context
@param conf the new configuration
@return a new context object
@throws InterruptedException
@throws IOException]]>
</doc>
</method>
<method name="cloneMapContext" return="org.apache.hadoop.mapreduce.Mapper.Context"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.MapContext"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="reader" type="org.apache.hadoop.mapreduce.RecordReader"/>
<param name="writer" type="org.apache.hadoop.mapreduce.RecordWriter"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Copy a mapper context, optionally replacing the input and output.
@param <K1> input key type
@param <V1> input value type
@param <K2> output key type
@param <V2> output value type
@param context the context to clone
@param conf a new configuration
@param reader Reader to read from. Null means to clone from context.
@param writer Writer to write to. Null means to clone from context.
@return a new context. it will not be the same class as the original.
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<doc>
<![CDATA[A factory to allow applications to deal with inconsistencies between
MapReduce Context Objects API between hadoop-0.20 and later versions.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.ContextFactory -->
<!-- start class org.apache.hadoop.mapreduce.Counter -->
<class name="Counter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="Counter"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<constructor name="Counter" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<constructor name="Counter" type="java.lang.String, java.lang.String, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a counter.
@param name the name within the group's enum.
@param displayName a name to be displayed.
@param value the counter value.]]>
</doc>
</constructor>
<method name="setDisplayName"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="displayName" type="java.lang.String"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the binary representation of the counter]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the binary representation of the counter]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the counter.
@return the user facing name of the counter]]>
</doc>
</method>
<method name="getValue" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[What is the current value of this counter?
@return the current value]]>
</doc>
</method>
<method name="setValue"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="long"/>
<doc>
<![CDATA[Set this counter by the given value
@param value the value to set]]>
</doc>
</method>
<method name="increment"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="incr" type="long"/>
<doc>
<![CDATA[Increment this counter by the given value
@param incr the value to increase this counter by]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericRight" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A named counter that tracks the progress of a map/reduce job.
<p><code>Counters</code> represent global counters, defined either by the
Map-Reduce framework or applications. Each <code>Counter</code> is named by
an {@link Enum} and has a long for the value.</p>
<p><code>Counters</code> are bunched into Groups, each comprising of
counters from a particular <code>Enum</code> class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Counter -->
<!-- start class org.apache.hadoop.mapreduce.CounterGroup -->
<class name="CounterGroup" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<constructor name="CounterGroup" type="java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<constructor name="CounterGroup" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a CounterGroup.
@param name the name of the group's enum.
@param displayName a name to be displayed for the group.]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the internal name of the group
@return the internal name]]>
</doc>
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the display name of the group.
@return the human readable name]]>
</doc>
</method>
<method name="addCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counter" type="org.apache.hadoop.mapreduce.Counter"/>
<doc>
<![CDATA[Add a counter to this group.]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.String"/>
<param name="displayName" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter in a group.
@param counterName the name of the counter
@param displayName the display name of the counter
@return the counter that was found or added]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.String"/>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of counters in this group.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericRight" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="incrAllCounters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rightGroup" type="org.apache.hadoop.mapreduce.CounterGroup"/>
</method>
<doc>
<![CDATA[A group of {@link Counter}s that logically belong together. Typically,
it is an {@link Enum} subclass and the counters are the values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.CounterGroup -->
<!-- start class org.apache.hadoop.mapreduce.Counters -->
<class name="Counters" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<constructor name="Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Counters" type="org.apache.hadoop.mapred.Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Utility method to create a Counters object from the
org.apache.hadoop.mapred counters
@param counters]]>
</doc>
</constructor>
<method name="addGroup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="org.apache.hadoop.mapreduce.CounterGroup"/>
<doc>
<![CDATA[Add a group.]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
</method>
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Find the counter for the given enum. The same enum will always return the
same counter.
@param key the counter key
@return the matching counter object]]>
</doc>
</method>
<method name="getGroupNames" return="java.util.Collection"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the names of all counter classes.
@return Set of counter names.]]>
</doc>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroup" return="org.apache.hadoop.mapreduce.CounterGroup"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the named counter group, or an empty group if there is none
with the specified name.]]>
</doc>
</method>
<method name="countCounters" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of counters, by summing the number of counters
in each group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the set of groups.
The external format is:
#groups (groupName group)*
i.e. the number of groups followed by 0 or more groups, where each
group is of the form:
groupDisplayName #counters (false | true counter)*
where each counter is of the form:
name (false | true displayName) value]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a set of groups.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return textual representation of the counter values.]]>
</doc>
</method>
<method name="incrAllCounters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapreduce.Counters"/>
<doc>
<![CDATA[Increments multiple counters by their amounts in another Counters
instance.
@param other the other Counters instance]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericRight" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.Counters -->
<!-- start class org.apache.hadoop.mapreduce.ID -->
<class name="ID" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="ID" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructs an ID object from the given int]]>
</doc>
</constructor>
<constructor name="ID"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="getId" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[returns the int which represents the identifier]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="that" type="org.apache.hadoop.mapreduce.ID"/>
<doc>
<![CDATA[Compare IDs by associated numbers]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="SEPARATOR" type="char"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="id" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A general identifier, which internally stores the id
as an integer. This is the super class of {@link JobID},
{@link TaskID} and {@link TaskAttemptID}.
@see JobID
@see TaskID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.ID -->
<!-- start class org.apache.hadoop.mapreduce.InputFormat -->
<class name="InputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSplits" return="java.util.List"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Logically split the set of input files for the job.
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
for processing.</p>
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
input files are not physically split into chunks. For e.g. a split could
be <i>&lt;input-file-path, start, offset&gt;</i> tuple. The InputFormat
also creates the {@link RecordReader} to read the {@link InputSplit}.
@param context job configuration.
@return an array of {@link InputSplit}s for the job.]]>
</doc>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Create a record reader for a given split. The framework will call
{@link RecordReader#initialize(InputSplit, TaskAttemptContext)} before
the split is used.
@param split the split to be read
@param context the information about the task
@return a new record reader
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputFormat</code> describes the input-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the input-specification of the job.
<li>
Split-up the input file(s) into logical {@link InputSplit}s, each of
which is then assigned to an individual {@link Mapper}.
</li>
<li>
Provide the {@link RecordReader} implementation to be used to glean
input records from the logical <code>InputSplit</code> for processing by
the {@link Mapper}.
</li>
</ol>
<p>The default behavior of file-based {@link InputFormat}s, typically
sub-classes of {@link FileInputFormat}, is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of the input files. However, the {@link FileSystem} blocksize of
the input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../mapred-default.html#mapreduce.input.fileinputformat.split.minsize">
mapreduce.input.fileinputformat.split.minsize</a>.</p>
<p>Clearly, logical splits based on input-size is insufficient for many
applications since record boundaries are to respected. In such cases, the
application has to also implement a {@link RecordReader} on whom lies the
responsibility to respect record-boundaries and present a record-oriented
view of the logical <code>InputSplit</code> to the individual task.
@see InputSplit
@see RecordReader
@see FileInputFormat]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.InputFormat -->
<!-- start class org.apache.hadoop.mapreduce.InputSplit -->
<class name="InputSplit" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLength" return="long"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the size of the split, so that the input splits can be sorted by size.
@return the number of bytes in the split
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the list of nodes by name where the data for the split would be local.
The locations do not need to be serialized.
@return a new array of the node nodes.
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
individual {@link Mapper}.
<p>Typically, it presents a byte-oriented view on the input and is the
responsibility of {@link RecordReader} of the job to process this and present
a record-oriented view.
@see InputFormat
@see RecordReader]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.InputSplit -->
<!-- start class org.apache.hadoop.mapreduce.Job -->
<class name="Job" extends="org.apache.hadoop.mapreduce.task.JobContextImpl"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.JobContext"/>
<constructor name="Job"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="Job" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="Job" type="org.apache.hadoop.conf.Configuration, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates a new {@link Job} with no particular {@link Cluster} .
A Cluster will be created with a generic {@link Configuration}.
@return the {@link Job} , with no connection to a cluster yet.
@throws IOException]]>
</doc>
</method>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates a new {@link Job} with no particular {@link Cluster} .
A Cluster will be created from the conf parameter only when it's needed.
@param conf the configuration
@return the {@link Job} , with no connection to a cluster yet.
@throws IOException]]>
</doc>
</method>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="jobName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates a new {@link Job} with no particular {@link Cluster} and a given jobName.
A Cluster will be created from the conf parameter only when it's needed.
@param conf the configuration
@return the {@link Job} , with no connection to a cluster yet.
@throws IOException]]>
</doc>
</method>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cluster" type="org.apache.hadoop.mapreduce.Cluster"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cluster" type="org.apache.hadoop.mapreduce.Cluster"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getInstance" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cluster" type="org.apache.hadoop.mapreduce.Cluster"/>
<param name="status" type="org.apache.hadoop.mapreduce.JobStatus"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getStatus" return="org.apache.hadoop.mapreduce.JobStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getJobState" return="org.apache.hadoop.mapreduce.JobStatus.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Returns the current state of the Job.
@return JobStatus#State
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getTrackingURL" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the URL where some job progress information will be displayed.
@return the URL where some job progress information will be displayed.]]>
</doc>
</method>
<method name="getJobFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the path of the submitted job configuration.
@return the path of the submitted job configuration.]]>
</doc>
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get start time of the job.
@return the start time of the job]]>
</doc>
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get finish time of the job.
@return the finish time of the job]]>
</doc>
</method>
<method name="getSchedulingInfo" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get scheduling info of the job.
@return the scheduling info of the job]]>
</doc>
</method>
<method name="getPriority" return="org.apache.hadoop.mapreduce.JobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get scheduling info of the job.
@return the scheduling info of the job]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The user-specified job name.]]>
</doc>
</method>
<method name="getHistoryUrl" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="isRetired" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Dump stats to screen.]]>
</doc>
</method>
<method name="getTaskReports" return="org.apache.hadoop.mapreduce.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="org.apache.hadoop.mapreduce.TaskType"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the information of the current state of the tasks of a job.
@param type Type of the task
@return the list of all of the map tips.
@throws IOException]]>
</doc>
</method>
<method name="mapProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's map-tasks, as a float between 0.0
and 1.0. When all map tasks have completed, the function returns 1.0.
@return the progress of the job's map-tasks.
@throws IOException]]>
</doc>
</method>
<method name="reduceProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's reduce-tasks, as a float between 0.0
and 1.0. When all reduce tasks have completed, the function returns 1.0.
@return the progress of the job's reduce-tasks.
@throws IOException]]>
</doc>
</method>
<method name="cleanupProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's cleanup-tasks, as a float between 0.0
and 1.0. When all cleanup tasks have completed, the function returns 1.0.
@return the progress of the job's cleanup-tasks.
@throws IOException]]>
</doc>
</method>
<method name="setupProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's setup-tasks, as a float between 0.0
and 1.0. When all setup tasks have completed, the function returns 1.0.
@return the progress of the job's setup-tasks.
@throws IOException]]>
</doc>
</method>
<method name="isComplete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Check if the job is finished or not.
This is a non-blocking call.
@return <code>true</code> if the job is complete, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="isSuccessful" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Check if the job completed successfully.
@return <code>true</code> if the job succeeded, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="killJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Kill the running job. Blocks until all job tasks have been
killed as well. If the job is no longer running, it simply returns.
@throws IOException]]>
</doc>
</method>
<method name="setPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="priority" type="org.apache.hadoop.mapreduce.JobPriority"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Set the priority of a running job.
@param priority the new priority for the job.
@throws IOException]]>
</doc>
</method>
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapreduce.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="startFrom" type="int"/>
<param name="numEvents" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get events indicating completion (success/failure) of component tasks.
@param startFrom index to start fetching events from
@param numEvents number of events to fetch
@return an array of {@link TaskCompletionEvent}s
@throws IOException]]>
</doc>
</method>
<method name="killTask" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Kill indicated task attempt.
@param taskId the id of the task to be terminated.
@throws IOException]]>
</doc>
</method>
<method name="failTask" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Fail indicated task attempt.
@param taskId the id of the task to be terminated.
@throws IOException]]>
</doc>
</method>
<method name="getCounters" return="org.apache.hadoop.mapreduce.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Gets the counters for this job. May return null if the job has been
retired and the job is no longer in the completed job store.
@return the counters for this job.
@throws IOException]]>
</doc>
</method>
<method name="getTaskDiagnostics" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Gets the diagnostic messages for a given task attempt.
@param taskid
@return the list of diagnostic messages for the task
@throws IOException]]>
</doc>
</method>
<method name="setNumReduceTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tasks" type="int"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the number of reduce tasks for the job.
@param tasks the number of reduce tasks
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the current working directory for the default file system.
@param dir the new current working directory.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setInputFormatClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the {@link InputFormat} for the job.
@param cls the <code>InputFormat</code> to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setOutputFormatClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the {@link OutputFormat} for the job.
@param cls the <code>OutputFormat</code> to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setMapperClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the {@link Mapper} for the job.
@param cls the <code>Mapper</code> to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setJarByClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<doc>
<![CDATA[Set the Jar by finding where a given class came from.
@param cls the example class]]>
</doc>
</method>
<method name="setJar"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jar" type="java.lang.String"/>
<doc>
<![CDATA[Set the job jar]]>
</doc>
</method>
<method name="setUser"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="user" type="java.lang.String"/>
<doc>
<![CDATA[Set the reported username for this job.
@param user the username for this job.]]>
</doc>
</method>
<method name="setCombinerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the combiner class for the job.
@param cls the combiner to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setReducerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the {@link Reducer} for the job.
@param cls the <code>Reducer</code> to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setPartitionerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the {@link Partitioner} for the job.
@param cls the <code>Partitioner</code> to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setMapOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the key class for the map output data. This allows the user to
specify the map output key class to be different than the final output
value class.
@param theClass the map output key class.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setMapOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the value class for the map output data. This allows the user to
specify the map output value class to be different than the final output
value class.
@param theClass the map output value class.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the key class for the job output data.
@param theClass the key class for the job output data.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the value class for job outputs.
@param theClass the value class for job outputs.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setSortComparatorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Define the comparator that controls how the keys are sorted before they
are passed to the {@link Reducer}.
@param cls the raw comparator
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setGroupingComparatorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Define the comparator that controls which keys are grouped together
for a single call to
{@link Reducer#reduce(Object, Iterable,
org.apache.hadoop.mapreduce.Reducer.Context)}
@param cls the raw comparator to use
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setJobName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
<doc>
<![CDATA[Set the user-specified job name.
@param name the job's new name.
@throws IllegalStateException if the job is submitted]]>
</doc>
</method>
<method name="setSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on, else <code>false</code>.]]>
</doc>
</method>
<method name="setMapSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for map tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for map tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="setReduceSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for reduce tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for reduce tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="setJobSetupCleanupNeeded"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="needed" type="boolean"/>
<doc>
<![CDATA[Specify whether job-setup and job-cleanup is needed for the job
@param needed If <code>true</code>, job-setup and job-cleanup will be
considered from {@link OutputCommitter}
else ignored.]]>
</doc>
</method>
<method name="setCacheArchives"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="archives" type="java.net.URI[]"/>
<doc>
<![CDATA[Set the given set of archives
@param archives The list of archives that need to be localized]]>
</doc>
</method>
<method name="setCacheFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="files" type="java.net.URI[]"/>
<doc>
<![CDATA[Set the given set of files
@param files The list of files that need to be localized]]>
</doc>
</method>
<method name="addCacheArchive"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<doc>
<![CDATA[Add a archives to be localized
@param uri The uri of the cache to be localized]]>
</doc>
</method>
<method name="addCacheFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<doc>
<![CDATA[Add a file to be localized
@param uri The uri of the cache to be localized]]>
</doc>
</method>
<method name="addFileToClassPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an file path to the current set of classpath entries It adds the file
to cache as well.
Files added with this method will not be unpacked while being added to the
classpath.
To add archives to classpath, use the {@link #addArchiveToClassPath(Path)}
method instead.
@param file Path of the file to be added]]>
</doc>
</method>
<method name="addArchiveToClassPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="archive" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an archive path to the current set of classpath entries. It adds the
archive to cache as well.
Archive files will be unpacked and added to the classpath
when being distributed.
@param archive Path of the archive to be added]]>
</doc>
</method>
<method name="createSymlink"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This method allows you to create symlinks in the current working directory
of the task to all the cache files/archives]]>
</doc>
</method>
<method name="setMaxMapAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
map task.
@param n the number of attempts per map task.]]>
</doc>
</method>
<method name="setMaxReduceAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
reduce task.
@param n the number of attempts per reduce task.]]>
</doc>
</method>
<method name="setProfileEnabled"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="boolean"/>
<doc>
<![CDATA[Set whether the system should collect profiler information for some of
the tasks in this job? The information is stored in the user log
directory.
@param newValue true means it should be gathered]]>
</doc>
</method>
<method name="setProfileParams"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="java.lang.String"/>
<doc>
<![CDATA[Set the profiler configuration arguments. If the string contains a '%s' it
will be replaced with the name of the profiling output file when the task
runs.
This value is passed to the task child JVM on the command line.
@param value the configuration string]]>
</doc>
</method>
<method name="setProfileTaskRange"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<param name="newValue" type="java.lang.String"/>
<doc>
<![CDATA[Set the ranges of maps or reduces to profile. setProfileEnabled(true)
must also be called.
@param newValue a set of integer ranges of the map ids]]>
</doc>
</method>
<method name="setCancelDelegationTokenUponJobCompletion"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="boolean"/>
<doc>
<![CDATA[Sets the flag that will allow the JobTracker to cancel the HDFS delegation
tokens upon job completion. Defaults to true.]]>
</doc>
</method>
<method name="submit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Submit the job to the cluster and return immediately.
@throws IOException]]>
</doc>
</method>
<method name="waitForCompletion" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="verbose" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Submit the job to the cluster and wait for it to finish.
@param verbose print the progress to the user
@return true if the job succeeded
@throws IOException thrown if the communication with the
<code>JobTracker</code> is lost]]>
</doc>
</method>
<method name="monitorAndPrintJob" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Monitor a job and print status in real-time as progress is made and tasks
fail.
@return true if the job succeeded
@throws IOException if communication to the JobTracker fails]]>
</doc>
</method>
<method name="getProgressPollInterval" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[The interval at which monitorAndPrintJob() prints status]]>
</doc>
</method>
<method name="getCompletionPollInterval" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[The interval at which waitForCompletion() should check.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapreduce.Job.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the task output filter.
@param conf the configuration.
@return the filter level.]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="newValue" type="org.apache.hadoop.mapreduce.Job.TaskStatusFilter"/>
<doc>
<![CDATA[Modify the Configuration to set the task output filter.
@param conf the Configuration to modify.
@param newValue the value to set.]]>
</doc>
</method>
<field name="OUTPUT_FILTER" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMPLETION_POLL_INTERVAL_KEY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Key in mapred-*.xml that sets completionPollInvervalMillis]]>
</doc>
</field>
<field name="PROGRESS_MONITOR_POLL_INTERVAL_KEY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Key in mapred-*.xml that sets progMonitorPollIntervalMillis]]>
</doc>
</field>
<field name="USED_GENERIC_PARSER" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SUBMIT_REPLICATION" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[The job submitter's view of the Job.
<p>It allows the user to configure the
job, submit it, control its execution, and query the state. The set methods
only work until the job is submitted, afterwards they will throw an
IllegalStateException. </p>
<p>
Normally the user creates the application, describes various facets of the
job via {@link Job} and then submits the job and monitor its progress.</p>
<p>Here is an example on how to submit a job:</p>
<p><blockquote><pre>
// Create a new Job
Job job = new Job(new Configuration());
job.setJarByClass(MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
job.setInputPath(new Path("in"));
job.setOutputPath(new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setReducerClass(MyJob.MyReducer.class);
// Submit the job, then poll for progress until the job is complete
job.waitForCompletion(true);
</pre></blockquote></p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Job -->
<!-- start class org.apache.hadoop.mapreduce.Job.JobState -->
<class name="Job.JobState" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.Job.JobState[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.Job.JobState"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.Job.JobState -->
<!-- start class org.apache.hadoop.mapreduce.Job.TaskStatusFilter -->
<class name="Job.TaskStatusFilter" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.Job.TaskStatusFilter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.Job.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.Job.TaskStatusFilter -->
<!-- start interface org.apache.hadoop.mapreduce.JobContext -->
<interface name="JobContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.MRJobConfig"/>
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the configuration for the job.
@return the shared configuration object]]>
</doc>
</method>
<method name="getCredentials" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get credentials for the job.
@return credentials for the job]]>
</doc>
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the unique ID for the job.
@return the object with the job id]]>
</doc>
</method>
<method name="getNumReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get configured the number of reduce tasks for this job. Defaults to
<code>1</code>.
@return the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the current working directory for the default file system.
@return the directory name.]]>
</doc>
</method>
<method name="getOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the job output data.
@return the key class for the job output data.]]>
</doc>
</method>
<method name="getOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for job outputs.
@return the value class for job outputs.]]>
</doc>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the map output data. If it is not set, use the
(final) output key class. This allows the map output key class to be
different than the final output key class.
@return the map output key class.]]>
</doc>
</method>
<method name="getMapOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for the map output data. If it is not set, use the
(final) output value class This allows the map output value class to be
different than the final output value class.
@return the map output value class.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified job name. This is only used to identify the
job to the user.
@return the job's name, defaulting to "".]]>
</doc>
</method>
<method name="getInputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the {@link InputFormat} class for the job.
@return the {@link InputFormat} class for the job.]]>
</doc>
</method>
<method name="getMapperClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the {@link Mapper} class for the job.
@return the {@link Mapper} class for the job.]]>
</doc>
</method>
<method name="getCombinerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the combiner class for the job.
@return the combiner class for the job.]]>
</doc>
</method>
<method name="getReducerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the {@link Reducer} class for the job.
@return the {@link Reducer} class for the job.]]>
</doc>
</method>
<method name="getOutputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the {@link OutputFormat} class for the job.
@return the {@link OutputFormat} class for the job.]]>
</doc>
</method>
<method name="getPartitionerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Get the {@link Partitioner} class for the job.
@return the {@link Partitioner} class for the job.]]>
</doc>
</method>
<method name="getSortComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link RawComparator} comparator used to compare keys.
@return the {@link RawComparator} comparator used to compare keys.]]>
</doc>
</method>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the pathname of the job's jar.
@return the pathname]]>
</doc>
</method>
<method name="getGroupingComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user defined {@link RawComparator} comparator for
grouping keys of inputs to the reduce.
@return comparator set by the user for grouping values.
@see Job#setGroupingComparatorClass(Class) for details.]]>
</doc>
</method>
<method name="getJobSetupCleanupNeeded" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get whether job-setup and job-cleanup is needed for the job
@return boolean]]>
</doc>
</method>
<method name="getProfileEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get whether the task profiling is enabled.
@return true if some tasks will be profiled]]>
</doc>
</method>
<method name="getProfileParams" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the profiler configuration arguments.
The default value for this property is
"-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s"
@return the parameters to pass to the task child to configure profiling]]>
</doc>
</method>
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<doc>
<![CDATA[Get the range of maps or reduces to profile.
@param isMap is the task a map?
@return the task ranges]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reported username for this job.
@return the username]]>
</doc>
</method>
<method name="getSymlink" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This method checks to see if symlinks are to be create for the
localized cache files in the current working directory
@return true if symlinks are to be created- else return false]]>
</doc>
</method>
<method name="getArchiveClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the archive entries in classpath as an array of Path]]>
</doc>
</method>
<method name="getCacheArchives" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get cache archives set in the Configuration
@return A URI array of the caches set in the Configuration
@throws IOException]]>
</doc>
</method>
<method name="getCacheFiles" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get cache files set in the Configuration
@return A URI array of the files set in the Configuration
@throws IOException]]>
</doc>
</method>
<method name="getLocalCacheArchives" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the path array of the localized caches
@return A path array of localized caches
@throws IOException]]>
</doc>
</method>
<method name="getLocalCacheFiles" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the path array of the localized files
@return A path array of localized files
@throws IOException]]>
</doc>
</method>
<method name="getFileClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the file entries in classpath as an array of Path]]>
</doc>
</method>
<method name="getArchiveTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the timestamps of the archives. Used by internal
DistributedCache and MapReduce code.
@return a string array of timestamps
@throws IOException]]>
</doc>
</method>
<method name="getFileTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the timestamps of the files. Used by internal
DistributedCache and MapReduce code.
@return a string array of timestamps
@throws IOException]]>
</doc>
</method>
<method name="getMaxMapAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
map task, as specified by the <code>mapred.map.max.attempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per map task.]]>
</doc>
</method>
<method name="getMaxReduceAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
reduce task, as specified by the <code>mapred.reduce.max.attempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per reduce task.]]>
</doc>
</method>
<doc>
<![CDATA[A read-only view of the job that is provided to the tasks while they
are running.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.JobContext -->
<!-- start class org.apache.hadoop.mapreduce.JobCounter -->
<class name="JobCounter" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.JobCounter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.JobCounter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.JobCounter -->
<!-- start class org.apache.hadoop.mapreduce.JobID -->
<class name="JobID" extends="org.apache.hadoop.mapred.ID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<constructor name="JobID" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a JobID object
@param jtIdentifier jobTracker identifier
@param id job number]]>
</doc>
</constructor>
<constructor name="JobID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getJtIdentifier" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
<doc>
<![CDATA[Compare JobIds by first jtIdentifiers, then by job numbers]]>
</doc>
</method>
<method name="appendTo" return="java.lang.StringBuilder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="builder" type="java.lang.StringBuilder"/>
<doc>
<![CDATA[Add the stuff after the "job" prefix to the given builder. This is useful,
because the sub-ids use this substring at the start of their string.
@param builder the builder to append to
@return the builder that was passed in]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="forName" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<doc>
<![CDATA[Construct a JobId object from given string
@return constructed JobId object or null if the given String is null
@throws IllegalArgumentException if the given string is malformed]]>
</doc>
</method>
<field name="JOB" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="JOBID_REGEX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="idFormat" type="java.text.NumberFormat"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[JobID represents the immutable and unique identifier for
the job. JobID consists of two parts. First part
represents the jobtracker identifier, so that jobID to jobtracker map
is defined. For cluster setup this string is the jobtracker
start time, for local setting, it is "local".
Second part of the JobID is the job number. <br>
An example JobID is :
<code>job_200707121733_0003</code> , which represents the third job
running at the jobtracker started at <code>200707121733</code>.
<p>
Applications should never construct or parse JobID strings, but rather
use appropriate constructors or {@link #forName(String)} method.
@see TaskID
@see TaskAttemptID
@see org.apache.hadoop.mapred.JobTracker#getNewJobId()
@see org.apache.hadoop.mapred.JobTracker#getStartTime()]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.JobID -->
<!-- start class org.apache.hadoop.mapreduce.JobPriority -->
<class name="JobPriority" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.JobPriority[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.JobPriority"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Used to describe the priority of the running job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.JobPriority -->
<!-- start class org.apache.hadoop.mapreduce.JobStatus -->
<class name="JobStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Cloneable"/>
<constructor name="JobStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="JobStatus" type="org.apache.hadoop.mapreduce.JobID, float, float, float, float, org.apache.hadoop.mapreduce.JobStatus.State, org.apache.hadoop.mapreduce.JobPriority, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param setupProgress The progress made on the setup
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param cleanupProgress The progress made on the cleanup
@param runState The current state of the job
@param jp Priority of the job.
@param user userid of the person who submitted the job.
@param jobName user-specified job name.
@param jobFile job configuration file.
@param trackingUrl link to the web-ui for details of the job.]]>
</doc>
</constructor>
<method name="setMapProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the map progress of this job
@param p The value of map progress to set to]]>
</doc>
</method>
<method name="setCleanupProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the cleanup progress of this job
@param p The value of cleanup progress to set to]]>
</doc>
</method>
<method name="setSetupProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the setup progress of this job
@param p The value of setup progress to set to]]>
</doc>
</method>
<method name="setReduceProgress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="p" type="float"/>
<doc>
<![CDATA[Sets the reduce progress of this Job
@param p The value of reduce progress to set to]]>
</doc>
</method>
<method name="setPriority"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="jp" type="org.apache.hadoop.mapreduce.JobPriority"/>
<doc>
<![CDATA[Set the priority of the job, defaulting to NORMAL.
@param jp new job priority]]>
</doc>
</method>
<method name="setFinishTime"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="finishTime" type="long"/>
<doc>
<![CDATA[Set the finish time of the job
@param finishTime The finishTime of the job]]>
</doc>
</method>
<method name="setHistoryFile"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="historyFile" type="java.lang.String"/>
<doc>
<![CDATA[Set the job history file url for a completed job]]>
</doc>
</method>
<method name="setTrackingUrl"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="trackingUrl" type="java.lang.String"/>
<doc>
<![CDATA[Set the link to the web-ui for details of the job.]]>
</doc>
</method>
<method name="setRetired"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Set the job retire flag to true.]]>
</doc>
</method>
<method name="setState"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="org.apache.hadoop.mapreduce.JobStatus.State"/>
<doc>
<![CDATA[Change the current run state of the job.]]>
</doc>
</method>
<method name="setStartTime"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="startTime" type="long"/>
<doc>
<![CDATA[Set the start time of the job
@param startTime The startTime of the job]]>
</doc>
</method>
<method name="setUsername"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="userName" type="java.lang.String"/>
<doc>
<![CDATA[@param userName The username of the job]]>
</doc>
</method>
<method name="setSchedulingInfo"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="schedulingInfo" type="java.lang.String"/>
<doc>
<![CDATA[Used to set the scheduling information associated to a particular Job.
@param schedulingInfo Scheduling information of the job]]>
</doc>
</method>
<method name="setJobACLs"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="acls" type="java.util.Map"/>
<doc>
<![CDATA[Set the job acls.
@param acls {@link Map} from {@link JobACL} to {@link AccessControlList}]]>
</doc>
</method>
<method name="getMapProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in maps]]>
</doc>
</method>
<method name="getCleanupProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in cleanup]]>
</doc>
</method>
<method name="getSetupProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in setup]]>
</doc>
</method>
<method name="getReduceProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in reduce]]>
</doc>
</method>
<method name="getState" return="org.apache.hadoop.mapreduce.JobStatus.State"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return running state of the job]]>
</doc>
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return start time of the job]]>
</doc>
</method>
<method name="clone" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return The jobid of the Job]]>
</doc>
</method>
<method name="getUsername" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the username of the job]]>
</doc>
</method>
<method name="getSchedulingInfo" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the Scheduling information associated to a particular Job.
@return the scheduling information of the job]]>
</doc>
</method>
<method name="getJobACLs" return="java.util.Map"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job acls.
@return a {@link Map} from {@link JobACL} to {@link AccessControlList}]]>
</doc>
</method>
<method name="getPriority" return="org.apache.hadoop.mapreduce.JobPriority"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the priority of the job
@return job priority]]>
</doc>
</method>
<method name="isJobComplete" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the status is for a completed job.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified job name.]]>
</doc>
</method>
<method name="getJobFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configuration file for the job.]]>
</doc>
</method>
<method name="getTrackingUrl" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the link to the web-ui for details of the job.]]>
</doc>
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the finish time of the job.]]>
</doc>
</method>
<method name="isRetired" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Check whether the job has retired.]]>
</doc>
</method>
<method name="getHistoryFile" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job history file name for a completed job. If job is not
completed or history file not available then return null.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Describes the current status of a job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.JobStatus -->
<!-- start class org.apache.hadoop.mapreduce.JobStatus.State -->
<class name="JobStatus.State" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.JobStatus.State[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.JobStatus.State"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="getValue" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Current state of the job]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.JobStatus.State -->
<!-- start interface org.apache.hadoop.mapreduce.MapContext -->
<interface name="MapContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
<method name="getInputSplit" return="org.apache.hadoop.mapreduce.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the input split for this map.]]>
</doc>
</method>
<doc>
<![CDATA[The context that is given to the {@link Mapper}.
@param <KEYIN> the key input type to the Mapper
@param <VALUEIN> the value input type to the Mapper
@param <KEYOUT> the key output type from the Mapper
@param <VALUEOUT> the value output type from the Mapper]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.MapContext -->
<!-- start class org.apache.hadoop.mapreduce.Mapper -->
<class name="Mapper" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Mapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once at the beginning of the task.]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="KEYIN"/>
<param name="value" type="VALUEIN"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once for each key/value pair in the input split. Most applications
should override this, but the default is the identity function.]]>
</doc>
</method>
<method name="cleanup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once at the end of the task.]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Expert users can override this method for more complete control over the
execution of the Mapper.
@param context
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Maps input key/value pairs to a set of intermediate key/value pairs.
<p>Maps are the individual tasks which transform input records into a
intermediate records. The transformed intermediate records need not be of
the same type as the input records. A given input pair may map to zero or
many output pairs.</p>
<p>The Hadoop Map-Reduce framework spawns one map task for each
{@link InputSplit} generated by the {@link InputFormat} for the job.
<code>Mapper</code> implementations can access the {@link Configuration} for
the job via the {@link JobContext#getConfiguration()}.
<p>The framework first calls
{@link #setup(org.apache.hadoop.mapreduce.Mapper.Context)}, followed by
{@link #map(Object, Object, Context)}
for each key/value pair in the <code>InputSplit</code>. Finally
{@link #cleanup(Context)} is called.</p>
<p>All intermediate values associated with a given output key are
subsequently grouped by the framework, and passed to a {@link Reducer} to
determine the final output. Users can control the sorting and grouping by
specifying two key {@link RawComparator} classes.</p>
<p>The <code>Mapper</code> outputs are partitioned per
<code>Reducer</code>. Users can control which keys (and hence records) go to
which <code>Reducer</code> by implementing a custom {@link Partitioner}.
<p>Users can optionally specify a <code>combiner</code>, via
{@link Job#setCombinerClass(Class)}, to perform local aggregation of the
intermediate outputs, which helps to cut down the amount of data transferred
from the <code>Mapper</code> to the <code>Reducer</code>.
<p>Applications can specify if and how the intermediate
outputs are to be compressed and which {@link CompressionCodec}s are to be
used via the <code>Configuration</code>.</p>
<p>If the job has zero
reduces then the output of the <code>Mapper</code> is directly written
to the {@link OutputFormat} without sorting by keys.</p>
<p>Example:</p>
<p><blockquote><pre>
public class TokenCounterMapper
extends Mapper&lt;Object, Text, Text, IntWritable&gt;{
private final static IntWritable one = new IntWritable(1);
private Text word = new Text();
public void map(Object key, Text value, Context context) throws IOException, InterruptedException {
StringTokenizer itr = new StringTokenizer(value.toString());
while (itr.hasMoreTokens()) {
word.set(itr.nextToken());
context.write(word, one);
}
}
}
</pre></blockquote></p>
<p>Applications may override the {@link #run(Context)} method to exert
greater control on map processing e.g. multi-threaded <code>Mapper</code>s
etc.</p>
@see InputFormat
@see JobContext
@see Partitioner
@see Reducer]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Mapper -->
<!-- start class org.apache.hadoop.mapreduce.Mapper.Context -->
<class name="Mapper.Context" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.MapContext"/>
<constructor name="Mapper.Context"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[The <code>Context</code> passed on to the {@link Mapper} implementations.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Mapper.Context -->
<!-- start class org.apache.hadoop.mapreduce.MarkableIterator -->
<class name="MarkableIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.MarkableIteratorInterface"/>
<constructor name="MarkableIterator" type="java.util.Iterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new iterator layered on the input iterator
@param itr underlying iterator that implements MarkableIteratorInterface]]>
</doc>
</constructor>
<method name="mark"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clearMark"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="VALUE"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[<code>MarkableIterator</code> is a wrapper iterator class that
implements the {@link MarkableIteratorInterface}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.MarkableIterator -->
<!-- start class org.apache.hadoop.mapreduce.OutputCommitter -->
<class name="OutputCommitter" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OutputCommitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setupJob"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For the framework to setup the job output during initialization
@param jobContext Context of the job whose output is being written.
@throws IOException if temporary output could not be created]]>
</doc>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #commitJob(JobContext)} or
{@link #abortJob(JobContext, JobStatus.State)} instead.">
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For cleaning up the job's output after job completion
@param jobContext Context of the job whose output is being written.
@throws IOException
@deprecated Use {@link #commitJob(JobContext)} or
{@link #abortJob(JobContext, JobStatus.State)} instead.]]>
</doc>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For committing job's output after successful job completion. Note that this
is invoked for jobs with final runstate as SUCCESSFUL.
@param jobContext Context of the job whose output is being written.
@throws IOException]]>
</doc>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="state" type="org.apache.hadoop.mapreduce.JobStatus.State"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For aborting an unsuccessful job's output. Note that this is invoked for
jobs with final runstate as {@link JobStatus.State#FAILED} or
{@link JobStatus.State#KILLED}.
@param jobContext Context of the job whose output is being written.
@param state final runstate of the job
@throws IOException]]>
</doc>
</method>
<method name="setupTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sets up output for the task.
@param taskContext Context of the task whose output is being written.
@throws IOException]]>
</doc>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check whether task needs a commit
@param taskContext
@return true/false
@throws IOException]]>
</doc>
</method>
<method name="commitTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[To promote the task's temporary output to final output location
The task's output is moved to the job's output directory.
@param taskContext Context of the task whose output is being written.
@throws IOException if commit is not]]>
</doc>
</method>
<method name="abortTask"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Discard the task output
@param taskContext
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>OutputCommitter</code> describes the commit of task output for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>OutputCommitter</code> of
the job to:<p>
<ol>
<li>
Setup the job during initialization. For example, create the temporary
output directory for the job during the initialization of the job.
</li>
<li>
Cleanup the job after the job completion. For example, remove the
temporary output directory after the job completion.
</li>
<li>
Setup the task temporary output.
</li>
<li>
Check whether a task needs a commit. This is to avoid the commit
procedure if a task does not need commit.
</li>
<li>
Commit of the task output.
</li>
<li>
Discard the task commit.
</li>
</ol>
@see org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
@see JobContext
@see TaskAttemptContext]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.OutputCommitter -->
<!-- start class org.apache.hadoop.mapreduce.OutputFormat -->
<class name="OutputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the {@link RecordWriter} for the given task.
@param context the information about the current task.
@return a {@link RecordWriter} to write the output for the job.
@throws IOException]]>
</doc>
</method>
<method name="checkOutputSpecs"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Check for validity of the output-specification for the job.
<p>This is to validate the output specification for the job when it is
a job is submitted. Typically checks that it does not already exist,
throwing an exception when it already exists, so that output is not
overwritten.</p>
@param context information about the job
@throws IOException when output should not be attempted]]>
</doc>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the output committer for this output format. This is responsible
for ensuring the output is committed correctly.
@param context the task context
@return an output committer
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<doc>
<![CDATA[<code>OutputFormat</code> describes the output-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>OutputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the output-specification of the job. For e.g. check that the
output directory doesn't already exist.
<li>
Provide the {@link RecordWriter} implementation to be used to write out
the output files of the job. Output files are stored in a
{@link FileSystem}.
</li>
</ol>
@see RecordWriter]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.OutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.Partitioner -->
<class name="Partitioner" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Partitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPartition" return="int"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEY"/>
<param name="value" type="VALUE"/>
<param name="numPartitions" type="int"/>
<doc>
<![CDATA[Get the partition number for a given key (hence record) given the total
number of partitions i.e. number of reduce-tasks for the job.
<p>Typically a hash function on a all or a subset of the key.</p>
@param key the key to be partioned.
@param value the entry value.
@param numPartitions the total number of partitions.
@return the partition number for the <code>key</code>.]]>
</doc>
</method>
<doc>
<![CDATA[Partitions the key space.
<p><code>Partitioner</code> controls the partitioning of the keys of the
intermediate map-outputs. The key (or a subset of the key) is used to derive
the partition, typically by a hash function. The total number of partitions
is the same as the number of reduce tasks for the job. Hence this controls
which of the <code>m</code> reduce tasks the intermediate key (and hence the
record) is sent for reduction.</p>
@see Reducer]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Partitioner -->
<!-- start class org.apache.hadoop.mapreduce.QueueAclsInfo -->
<class name="QueueAclsInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="QueueAclsInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for QueueAclsInfo.]]>
</doc>
</constructor>
<constructor name="QueueAclsInfo" type="java.lang.String, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a new QueueAclsInfo object using the queue name and the
queue operations array
@param queueName Name of the job queue
@param operations]]>
</doc>
</constructor>
<method name="getQueueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get queue name.
@return name]]>
</doc>
</method>
<method name="setQueueName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
</method>
<method name="getOperations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get opearations allowed on queue.
@return array of String]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Class to encapsulate Queue ACLs for a particular
user.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.QueueAclsInfo -->
<!-- start class org.apache.hadoop.mapreduce.QueueInfo -->
<class name="QueueInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="QueueInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for QueueInfo.]]>
</doc>
</constructor>
<constructor name="QueueInfo" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a new QueueInfo object using the queue name and the
scheduling information passed.
@param queueName Name of the job queue
@param schedulingInfo Scheduling Information associated with the job
queue]]>
</doc>
</constructor>
<constructor name="QueueInfo" type="java.lang.String, java.lang.String, org.apache.hadoop.mapreduce.QueueState, org.apache.hadoop.mapreduce.JobStatus[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param queueName
@param schedulingInfo
@param state
@param stats]]>
</doc>
</constructor>
<method name="setQueueName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<doc>
<![CDATA[Set the queue name of the JobQueueInfo
@param queueName Name of the job queue.]]>
</doc>
</method>
<method name="getQueueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the queue name from JobQueueInfo
@return queue name]]>
</doc>
</method>
<method name="setSchedulingInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="schedulingInfo" type="java.lang.String"/>
<doc>
<![CDATA[Set the scheduling information associated to particular job queue
@param schedulingInfo]]>
</doc>
</method>
<method name="getSchedulingInfo" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the scheduling information associated to particular job queue.
If nothing is set would return <b>"N/A"</b>
@return Scheduling information associated to particular Job Queue]]>
</doc>
</method>
<method name="setState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="org.apache.hadoop.mapreduce.QueueState"/>
<doc>
<![CDATA[Set the state of the queue
@param state state of the queue.]]>
</doc>
</method>
<method name="getState" return="org.apache.hadoop.mapreduce.QueueState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the queue state
@return the queue state.]]>
</doc>
</method>
<method name="setJobStatuses"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="stats" type="org.apache.hadoop.mapreduce.JobStatus[]"/>
</method>
<method name="getQueueChildren" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get immediate children.
@return list of QueueInfo]]>
</doc>
</method>
<method name="setQueueChildren"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="children" type="java.util.List"/>
</method>
<method name="getProperties" return="java.util.Properties"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get properties.
@return Properties]]>
</doc>
</method>
<method name="setProperties"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="props" type="java.util.Properties"/>
</method>
<method name="getJobStatuses" return="org.apache.hadoop.mapreduce.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the jobs submitted to queue
@return list of JobStatus for the submitted jobs]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Class that contains the information regarding the Job Queues which are
maintained by the Hadoop Map/Reduce framework.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.QueueInfo -->
<!-- start class org.apache.hadoop.mapreduce.QueueState -->
<class name="QueueState" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.QueueState[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.QueueState"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="getStateName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the stateName]]>
</doc>
</method>
<method name="getState" return="org.apache.hadoop.mapreduce.QueueState"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="state" type="java.lang.String"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Enum representing queue state]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.QueueState -->
<!-- start class org.apache.hadoop.mapreduce.RecordReader -->
<class name="RecordReader" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="RecordReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once at initialization.
@param split the split that defines the range of records to read
@param context the information about the task
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="nextKeyValue" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Read the next key, value pair.
@return true if a key/value pair was read
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getCurrentKey" return="KEYIN"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the current key
@return the current key or null if there is no current key
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getCurrentValue" return="VALUEIN"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the current value.
@return the object that was read
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[The current progress of the record reader through its data.
@return a number between 0.0 and 1.0 that is the fraction of the data read
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="close"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the record reader.]]>
</doc>
</method>
<doc>
<![CDATA[The record reader breaks the data into key/value pairs for input to the
{@link Mapper}.
@param <KEYIN>
@param <VALUEIN>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.RecordReader -->
<!-- start class org.apache.hadoop.mapreduce.RecordWriter -->
<class name="RecordWriter" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Writes a key/value pair.
@param key the key to write.
@param value the value to write.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Close this <code>RecordWriter</code> to future operations.
@param context the context of the task
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RecordWriter</code> writes the output &lt;key, value&gt; pairs
to an output file.
<p><code>RecordWriter</code> implementations write the job outputs to the
{@link FileSystem}.
@see OutputFormat]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.RecordWriter -->
<!-- start interface org.apache.hadoop.mapreduce.ReduceContext -->
<interface name="ReduceContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
<method name="nextKey" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Start processing next unique key.]]>
</doc>
</method>
<method name="getValues" return="java.lang.Iterable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Iterate through the values for the current key, reusing the same value
object, which is stored in the context.
@return the series of values associated with the current key. All of the
objects returned directly and indirectly from this method are reused.]]>
</doc>
</method>
<doc>
<![CDATA[The context passed to the {@link Reducer}.
@param <KEYIN> the class of the input keys
@param <VALUEIN> the class of the input values
@param <KEYOUT> the class of the output keys
@param <VALUEOUT> the class of the output values]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.ReduceContext -->
<!-- start interface org.apache.hadoop.mapreduce.ReduceContext.ValueIterator -->
<interface name="ReduceContext.ValueIterator" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.MarkableIteratorInterface"/>
<method name="resetBackupStore"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method is called when the reducer moves from one key to
another.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[{@link Iterator} to iterate over values for a given group of records.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.ReduceContext.ValueIterator -->
<!-- start class org.apache.hadoop.mapreduce.Reducer -->
<class name="Reducer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Reducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once at the start of the task.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="KEYIN"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[This method is called once for each key. Most applications will define
their reduce class by overriding this method. The default implementation
is an identity function.]]>
</doc>
</method>
<method name="cleanup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Called once at the end of the task.]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Advanced application writers can use the
{@link #run(org.apache.hadoop.mapreduce.Reducer.Context)} method to
control how the reduce task works.]]>
</doc>
</method>
<doc>
<![CDATA[Reduces a set of intermediate values which share a key to a smaller set of
values.
<p><code>Reducer</code> implementations
can access the {@link Configuration} for the job via the
{@link JobContext#getConfiguration()} method.</p>
<p><code>Reducer</code> has 3 primary phases:</p>
<ol>
<li>
<h4 id="Shuffle">Shuffle</h4>
<p>The <code>Reducer</code> copies the sorted output from each
{@link Mapper} using HTTP across the network.</p>
</li>
<li>
<h4 id="Sort">Sort</h4>
<p>The framework merge sorts <code>Reducer</code> inputs by
<code>key</code>s
(since different <code>Mapper</code>s may have output the same key).</p>
<p>The shuffle and sort phases occur simultaneously i.e. while outputs are
being fetched they are merged.</p>
<h5 id="SecondarySort">SecondarySort</h5>
<p>To achieve a secondary sort on the values returned by the value
iterator, the application should extend the key with the secondary
key and define a grouping comparator. The keys will be sorted using the
entire key, but will be grouped using the grouping comparator to decide
which keys and values are sent in the same call to reduce.The grouping
comparator is specified via
{@link Job#setGroupingComparatorClass(Class)}. The sort order is
controlled by
{@link Job#setSortComparatorClass(Class)}.</p>
For example, say that you want to find duplicate web pages and tag them
all with the url of the "best" known example. You would set up the job
like:
<ul>
<li>Map Input Key: url</li>
<li>Map Input Value: document</li>
<li>Map Output Key: document checksum, url pagerank</li>
<li>Map Output Value: url</li>
<li>Partitioner: by checksum</li>
<li>OutputKeyComparator: by checksum and then decreasing pagerank</li>
<li>OutputValueGroupingComparator: by checksum</li>
</ul>
</li>
<li>
<h4 id="Reduce">Reduce</h4>
<p>In this phase the
{@link #reduce(Object, Iterable, Context)}
method is called for each <code>&lt;key, (collection of values)&gt;</code> in
the sorted inputs.</p>
<p>The output of the reduce task is typically written to a
{@link RecordWriter} via
{@link Context#write(Object, Object)}.</p>
</li>
</ol>
<p>The output of the <code>Reducer</code> is <b>not re-sorted</b>.</p>
<p>Example:</p>
<p><blockquote><pre>
public class IntSumReducer&lt;Key&gt; extends Reducer&lt;Key,IntWritable,
Key,IntWritable&gt; {
private IntWritable result = new IntWritable();
public void reduce(Key key, Iterable&lt;IntWritable&gt; values,
Context context) throws IOException, InterruptedException {
int sum = 0;
for (IntWritable val : values) {
sum += val.get();
}
result.set(sum);
context.write(key, result);
}
}
</pre></blockquote></p>
@see Mapper
@see Partitioner]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Reducer -->
<!-- start class org.apache.hadoop.mapreduce.Reducer.Context -->
<class name="Reducer.Context" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.ReduceContext"/>
<constructor name="Reducer.Context"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[The <code>Context</code> passed on to the {@link Reducer} implementations.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.Reducer.Context -->
<!-- start interface org.apache.hadoop.mapreduce.TaskAttemptContext -->
<interface name="TaskAttemptContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.JobContext"/>
<implements name="org.apache.hadoop.util.Progressable"/>
<method name="getTaskAttemptID" return="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the unique name for this task attempt.]]>
</doc>
</method>
<method name="setStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="msg" type="java.lang.String"/>
<doc>
<![CDATA[Set the current status of the task to the given string.]]>
</doc>
</method>
<method name="getStatus" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the last set status message.
@return the current status message]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.Enum"/>
<doc>
<![CDATA[Get the {@link Counter} for the given <code>counterName</code>.
@param counterName counter name
@return the <code>Counter</code> for the given <code>counterName</code>]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
<doc>
<![CDATA[Get the {@link Counter} for the given <code>groupName</code> and
<code>counterName</code>.
@param counterName counter name
@return the <code>Counter</code> for the given <code>groupName</code> and
<code>counterName</code>]]>
</doc>
</method>
<doc>
<![CDATA[The context for task attempts.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.TaskAttemptContext -->
<!-- start class org.apache.hadoop.mapreduce.TaskAttemptID -->
<class name="TaskAttemptID" extends="org.apache.hadoop.mapred.ID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskAttemptID" type="org.apache.hadoop.mapreduce.TaskID, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskAttemptID object from given {@link TaskID}.
@param taskId TaskID that this task belongs to
@param id the task attempt number]]>
</doc>
</constructor>
<constructor name="TaskAttemptID" type="java.lang.String, int, org.apache.hadoop.mapreduce.TaskType, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param type the TaskType
@param taskId taskId number
@param id the task attempt number]]>
</doc>
</constructor>
<constructor name="TaskAttemptID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the {@link JobID} object that this task attempt belongs to]]>
</doc>
</method>
<method name="getTaskID" return="org.apache.hadoop.mapreduce.TaskID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the {@link TaskID} object that this task attempt belongs to]]>
</doc>
</method>
<method name="getTaskType" return="org.apache.hadoop.mapreduce.TaskType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the TaskType of the TaskAttemptID]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="appendTo" return="java.lang.StringBuilder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="builder" type="java.lang.StringBuilder"/>
<doc>
<![CDATA[Add the unique string to the StringBuilder
@param builder the builder to append ot
@return the builder that was passed in.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
<doc>
<![CDATA[Compare TaskIds by first tipIds, then by task numbers.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="forName" return="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<doc>
<![CDATA[Construct a TaskAttemptID object from given string
@return constructed TaskAttemptID object or null if the given String is null
@throws IllegalArgumentException if the given string is malformed]]>
</doc>
</method>
<field name="ATTEMPT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[TaskAttemptID represents the immutable and unique identifier for
a task attempt. Each task attempt is one particular instance of a Map or
Reduce Task identified by its TaskID.
TaskAttemptID consists of 2 parts. First part is the
{@link TaskID}, that this TaskAttemptID belongs to.
Second part is the task attempt number. <br>
An example TaskAttemptID is :
<code>attempt_200707121733_0003_m_000005_0</code> , which represents the
zeroth task attempt for the fifth map task in the third job
running at the jobtracker started at <code>200707121733</code>.
<p>
Applications should never construct or parse TaskAttemptID strings
, but rather use appropriate constructors or {@link #forName(String)}
method.
@see JobID
@see TaskID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskAttemptID -->
<!-- start class org.apache.hadoop.mapreduce.TaskCompletionEvent -->
<class name="TaskCompletionEvent" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="TaskCompletionEvent"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for Writable.]]>
</doc>
</constructor>
<constructor name="TaskCompletionEvent" type="int, org.apache.hadoop.mapreduce.TaskAttemptID, int, boolean, org.apache.hadoop.mapreduce.TaskCompletionEvent.Status, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor. eventId should be created externally and incremented
per event for each job.
@param eventId event id, event id should be unique and assigned in
incrementally, starting from 0.
@param taskId task id
@param status task's status
@param taskTrackerHttp task tracker's host:port for http.]]>
</doc>
</constructor>
<method name="getEventId" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns event Id.
@return event id]]>
</doc>
</method>
<method name="getTaskAttemptId" return="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task id.
@return task id]]>
</doc>
</method>
<method name="getStatus" return="org.apache.hadoop.mapreduce.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns enum Status.SUCESS or Status.FAILURE.
@return task tracker status]]>
</doc>
</method>
<method name="getTaskTrackerHttp" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[http location of the tasktracker where this task ran.
@return http location of tasktracker user logs]]>
</doc>
</method>
<method name="getTaskRunTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns time (in millisec) the task took to complete.]]>
</doc>
</method>
<method name="setTaskRunTime"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskCompletionTime" type="int"/>
<doc>
<![CDATA[Set the task completion time
@param taskCompletionTime time (in millisec) the task took to complete]]>
</doc>
</method>
<method name="setEventId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="eventId" type="int"/>
<doc>
<![CDATA[set event Id. should be assigned incrementally starting from 0.
@param eventId]]>
</doc>
</method>
<method name="setTaskAttemptId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<doc>
<![CDATA[Sets task id.
@param taskId]]>
</doc>
</method>
<method name="setTaskStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="status" type="org.apache.hadoop.mapreduce.TaskCompletionEvent.Status"/>
<doc>
<![CDATA[Set task status.
@param status]]>
</doc>
</method>
<method name="setTaskTrackerHttp"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskTrackerHttp" type="java.lang.String"/>
<doc>
<![CDATA[Set task tracker http location.
@param taskTrackerHttp]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isMapTask" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="idWithinJob" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="EMPTY_ARRAY" type="org.apache.hadoop.mapreduce.TaskCompletionEvent[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This is used to track task completion events on
job tracker.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskCompletionEvent -->
<!-- start class org.apache.hadoop.mapreduce.TaskCompletionEvent.Status -->
<class name="TaskCompletionEvent.Status" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.TaskCompletionEvent.Status[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskCompletionEvent.Status -->
<!-- start class org.apache.hadoop.mapreduce.TaskCounter -->
<class name="TaskCounter" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.TaskCounter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.TaskCounter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskCounter -->
<!-- start class org.apache.hadoop.mapreduce.TaskID -->
<class name="TaskID" extends="org.apache.hadoop.mapred.ID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskID" type="org.apache.hadoop.mapreduce.JobID, org.apache.hadoop.mapreduce.TaskType, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskID object from given {@link JobID}.
@param jobId JobID that this tip belongs to
@param type the {@link TaskType} of the task
@param id the tip number]]>
</doc>
</constructor>
<constructor name="TaskID" type="java.lang.String, int, org.apache.hadoop.mapreduce.TaskType, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a TaskInProgressId object from given parts.
@param jtIdentifier jobTracker identifier
@param jobId job number
@param type the TaskType
@param id the tip number]]>
</doc>
</constructor>
<constructor name="TaskID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the {@link JobID} object that this tip belongs to]]>
</doc>
</method>
<method name="getTaskType" return="org.apache.hadoop.mapreduce.TaskType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the type of the task]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
<doc>
<![CDATA[Compare TaskInProgressIds by first jobIds, then by tip numbers. Reduces are
defined as greater then maps.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="appendTo" return="java.lang.StringBuilder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="builder" type="java.lang.StringBuilder"/>
<doc>
<![CDATA[Add the unique string to the given builder.
@param builder the builder to append to
@return the builder that was passed in]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="forName" return="org.apache.hadoop.mapreduce.TaskID"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<doc>
<![CDATA[Construct a TaskID object from given string
@return constructed TaskID object or null if the given String is null
@throws IllegalArgumentException if the given string is malformed]]>
</doc>
</method>
<method name="getRepresentingCharacter" return="char"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="org.apache.hadoop.mapreduce.TaskType"/>
<doc>
<![CDATA[Gets the character representing the {@link TaskType}
@param type the TaskType
@return the character]]>
</doc>
</method>
<method name="getTaskType" return="org.apache.hadoop.mapreduce.TaskType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="char"/>
<doc>
<![CDATA[Gets the {@link TaskType} corresponding to the character
@param c the character
@return the TaskType]]>
</doc>
</method>
<method name="getAllTaskTypes" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="TASK" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="idFormat" type="java.text.NumberFormat"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[TaskID represents the immutable and unique identifier for
a Map or Reduce Task. Each TaskID encompasses multiple attempts made to
execute the Map or Reduce Task, each of which are uniquely indentified by
their TaskAttemptID.
TaskID consists of 3 parts. First part is the {@link JobID}, that this
TaskInProgress belongs to. Second part of the TaskID is either 'm' or 'r'
representing whether the task is a map task or a reduce task.
And the third part is the task number. <br>
An example TaskID is :
<code>task_200707121733_0003_m_000005</code> , which represents the
fifth map task in the third job running at the jobtracker
started at <code>200707121733</code>.
<p>
Applications should never construct or parse TaskID strings
, but rather use appropriate constructors or {@link #forName(String)}
method.
@see JobID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskID -->
<!-- start interface org.apache.hadoop.mapreduce.TaskInputOutputContext -->
<interface name="TaskInputOutputContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Advance to the next key, value pair, returning null if at end.
@return the key object that was read into, or null if no more]]>
</doc>
</method>
<method name="getCurrentKey" return="KEYIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the current key.
@return the current key object or null if there isn't one
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="getCurrentValue" return="VALUEIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the current value.
@return the value object that was read into
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEYOUT"/>
<param name="value" type="VALUEOUT"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Generate an output key/value pair.]]>
</doc>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputCommitter} for the task-attempt.
@return the <code>OutputCommitter</code> for the task-attempt]]>
</doc>
</method>
<doc>
<![CDATA[A context object that allows input and output from the task. It is only
supplied to the {@link Mapper} or {@link Reducer}.
@param <KEYIN> the input key type for the task
@param <VALUEIN> the input value type for the task
@param <KEYOUT> the output key type for the task
@param <VALUEOUT> the output value type for the task]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.TaskInputOutputContext -->
<!-- start class org.apache.hadoop.mapreduce.TaskTrackerInfo -->
<class name="TaskTrackerInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="TaskTrackerInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TaskTrackerInfo" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TaskTrackerInfo" type="java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getTaskTrackerName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the tasktracker's name.
@return tracker's name.]]>
</doc>
</method>
<method name="isBlacklisted" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Whether tracker is blacklisted
@return true if tracker is blacklisted
false otherwise]]>
</doc>
</method>
<method name="getReasonForBlacklist" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the reason for which the tasktracker was blacklisted.
@return reason which tracker was blacklisted]]>
</doc>
</method>
<method name="getBlacklistReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets a descriptive report about why the tasktracker was blacklisted.
@return report describing why the tasktracker was blacklisted.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Information about TaskTracker.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskTrackerInfo -->
<!-- start class org.apache.hadoop.mapreduce.TaskType -->
<class name="TaskType" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.TaskType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.TaskType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.TaskType -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.aggregate">
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum -->
<class name="DoubleValueSum" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="DoubleValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a double value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="double"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a double value.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="getSum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that sums up a sequence of double
values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax -->
<class name="LongValueMax" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newVal" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param newVal
a long value.]]>
</doc>
</method>
<method name="getVal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the maximum of
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin -->
<class name="LongValueMin" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newVal" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param newVal
a long value.]]>
</doc>
</method>
<method name="getVal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the minimum of
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum -->
<class name="LongValueSum" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a long value.]]>
</doc>
</method>
<method name="getSum" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that sums up
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax -->
<class name="StringValueMax" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a string.]]>
</doc>
</method>
<method name="getVal" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the biggest of
a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin -->
<class name="StringValueMin" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a string.]]>
</doc>
</method>
<method name="getVal" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the smallest of
a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount -->
<class name="UniqValueCount" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="UniqValueCount"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<constructor name="UniqValueCount" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructor
@param maxNum the limit in the number of unique values to keep.]]>
</doc>
</constructor>
<method name="setMaxItems" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<doc>
<![CDATA[Set the limit on the number of unique values
@param n the desired limit on the number of unique values
@return the new limit on the number of unique values]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return the number of unique objects aggregated]]>
</doc>
</method>
<method name="getUniqueItems" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the set of the unique objects]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of the unique objects. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<field name="MAX_NUM_UNIQUE_VALUES" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a value aggregator that dedupes a sequence of objects.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<class name="UserDefinedValueAggregatorDescriptor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="UserDefinedValueAggregatorDescriptor" type="java.lang.String, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param className the class name of the user defined descriptor class
@param conf a configure object used for decriptor configuration]]>
</doc>
</constructor>
<method name="createInstance" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="className" type="java.lang.String"/>
<doc>
<![CDATA[Create an instance of the given class
@param className the name of the class
@return a dynamically created instance of the given class]]>
</doc>
</method>
<method name="generateKeyValPairs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate a list of aggregation-id/value pairs for the given
key/value pairs by delegating the invocation to the real object.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of this object.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Do nothing.]]>
</doc>
</method>
<field name="theAggregatorDescriptor" type="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a wrapper for a user defined value
aggregator descriptor.
It serves two functions: One is to create an object of
ValueAggregatorDescriptor from the name of a user defined class
that may be dynamically loaded. The other is to
delegate invocations of generateKeyValPairs function to the created object.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<!-- start interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator -->
<interface name="ValueAggregator" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val the value to be added]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the agregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return an array of values as the outputs of the combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This interface defines the minimal protocol for value aggregators.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor -->
<class name="ValueAggregatorBaseDescriptor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="ValueAggregatorBaseDescriptor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="generateEntry" return="java.util.Map.Entry"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<param name="id" type="java.lang.String"/>
<param name="val" type="org.apache.hadoop.io.Text"/>
<doc>
<![CDATA[@param type the aggregation type
@param id the aggregation id
@param val the val associated with the id to be aggregated
@return an Entry whose key is the aggregation id prefixed with
the aggregation type.]]>
</doc>
</method>
<method name="generateValueAggregator" return="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<param name="uniqCount" type="long"/>
<doc>
<![CDATA[@param type the aggregation type
@param uniqCount the limit in the number of unique values to keep,
if type is UNIQ_VALUE_COUNT
@return a value aggregator of the given type.]]>
</doc>
</method>
<method name="generateKeyValPairs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate 1 or 2 aggregation-id/value pairs for the given key/value pair.
The first id will be of type LONG_VALUE_SUM, with "record_count" as
its aggregation id. If the input is a file split,
the second id of the same type will be generated too, with the file name
as its aggregation id. This achieves the behavior of counting the total
number of records in the input data, and the number of records
in each input file.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[get the input file name.
@param conf a configuration object]]>
</doc>
</method>
<field name="UNIQ_VALUE_COUNT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOUBLE_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="VALUE_HISTOGRAM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="inputFile" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements the common functionalities of
the subclasses of ValueAggregatorDescriptor class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner -->
<class name="ValueAggregatorCombiner" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorCombiner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Combines values for a given key.
@param key the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values.
@param values the values to combine
@param context to collect combined values]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic combiner of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner -->
<!-- start interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor -->
<interface name="ValueAggregatorDescriptor" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="generateKeyValPairs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate a list of aggregation-id/value pairs for
the given key/value pair.
This function is usually called by the mapper of an Aggregate based job.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Configure the object
@param conf
a Configuration object that may contain the information
that can be used to configure the object.]]>
</doc>
</method>
<field name="TYPE_SEPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ONE" type="org.apache.hadoop.io.Text"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This interface defines the contract a value aggregator descriptor must
support. Such a descriptor can be configured with a {@link Configuration}
object. Its main function is to generate a list of aggregation-id/value
pairs. An aggregation id encodes an aggregation type which is used to
guide the way to aggregate the value in the reduce/combiner phrase of an
Aggregate based job.
The mapper in an Aggregate based map/reduce job may create one or more of
ValueAggregatorDescriptor objects at configuration time. For each input
key/value pair, the mapper will use those objects to create aggregation
id/value pairs.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob -->
<class name="ValueAggregatorJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorJob"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create an Aggregate based map/reduce job.
@param conf The configuration for job
@param args the arguments used for job creation. Generic hadoop
arguments are accepted.
@return a Job object ready for submission.
@throws IOException
@see GenericOptionsParser]]>
</doc>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setAggregatorDescriptors" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="descriptors" type="java.lang.Class[]"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[create and run an Aggregate based map/reduce job.
@param args the arguments used for job creation
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This is the main class for creating a map/reduce job using Aggregate
framework. The Aggregate is a specialization of map/reduce framework,
specializing for performing various simple aggregations.
Generally speaking, in order to implement an application using Map/Reduce
model, the developer is to implement Map and Reduce functions (and possibly
combine function). However, a lot of applications related to counting and
statistics computing have very similar characteristics. Aggregate abstracts
out the general patterns of these functions and implementing those patterns.
In particular, the package provides generic mapper/redducer/combiner
classes, and a set of built-in value aggregators, and a generic utility
class that helps user create map/reduce jobs using the generic class.
The built-in aggregators include:
sum over numeric values count the number of distinct values compute the
histogram of values compute the minimum, maximum, media,average, standard
deviation of numeric values
The developer using Aggregate will need only to provide a plugin class
conforming to the following interface:
public interface ValueAggregatorDescriptor { public ArrayList<Entry>
generateKeyValPairs(Object key, Object value); public void
configure(Configuration conf); }
The package also provides a base class, ValueAggregatorBaseDescriptor,
implementing the above interface. The user can extend the base class and
implement generateKeyValPairs accordingly.
The primary work of generateKeyValPairs is to emit one or more key/value
pairs based on the input key/value pair. The key in an output key/value pair
encode two pieces of information: aggregation type and aggregation id. The
value will be aggregated onto the aggregation id according the aggregation
type.
This class offers a function to generate a map/reduce job using Aggregate
framework. The function takes the following parameters: input directory spec
input format (text or sequence file) output directory a file specifying the
user plugin class]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase -->
<class name="ValueAggregatorJobBase" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorJobBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getValueAggregatorDescriptor" return="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="spec" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getAggregatorDescriptors" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="logSpec"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<field name="DESCRIPTOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DESCRIPTOR_NUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="USER_JAR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="aggregatorDescriptorList" type="java.util.ArrayList"
transient="false" volatile="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This abstract class implements some common functionalities of the
the generic mapper, reducer and combiner classes of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper -->
<class name="ValueAggregatorMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K1"/>
<param name="value" type="V1"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[the map function. It iterates through the value aggregator descriptor
list to generate aggregation id/value pairs and emit them.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic mapper of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer -->
<class name="ValueAggregatorReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[@param key
the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values. In effect, data
driven computing is achieved. It is assumed that each aggregator's
getReport method emits appropriate output for the aggregator. This
may be further customized.
@param values the values to be aggregated
@param context]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic reducer of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer -->
<!-- start class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram -->
<class name="ValueHistogram" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator"/>
<constructor name="ValueHistogram"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add the given val to the aggregator.
@param val the value to be added. It is expected to be a string
in the form of xxxx\tnum, meaning xxxx has num occurrences.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of this aggregator.
It includes the following basic statistics of the histogram:
the number of unique values
the minimum value
the media value
the maximum value
the average value
the standard deviation]]>
</doc>
</method>
<method name="getReportDetails" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a string representation of the list of value/frequence pairs of
the histogram]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a list value/frequence pairs.
The return value is expected to be used by the reducer.]]>
</doc>
</method>
<method name="getReportItems" return="java.util.TreeMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a TreeMap representation of the histogram]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that computes the
histogram of a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.chain">
<!-- start class org.apache.hadoop.mapreduce.lib.chain.ChainMapper -->
<class name="ChainMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ChainMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="mapperConf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Adds a {@link Mapper} class to the chain mapper.
<p>
The key and values are passed from one element of the chain to the next, by
value. For the added Mapper the configuration given for it,
<code>mapperConf</code>, have precedence over the job's Configuration. This
precedence is in effect when the task is running.
</p>
<p>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the chain
</p>
@param job
The job.
@param klass
the Mapper class to add.
@param inputKeyClass
mapper input key class.
@param inputValueClass
mapper input value class.
@param outputKeyClass
mapper output key class.
@param outputValueClass
mapper output value class.
@param mapperConf
a configuration for the Mapper class. It is recommended to use a
Configuration without default values using the
<code>Configuration(boolean loadDefaults)</code> constructor with
FALSE.]]>
</doc>
</method>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<doc>
<![CDATA[The ChainMapper class allows to use multiple Mapper classes within a single
Map task.
<p>
The Mapper classes are invoked in a chained (or piped) fashion, the output of
the first becomes the input of the second, and so on until the last Mapper,
the output of the last Mapper will be written to the task's output.
</p>
<p>
The key functionality of this feature is that the Mappers in the chain do not
need to be aware that they are executed in a chain. This enables having
reusable specialized Mappers that can be combined to perform composite
operations within a single task.
</p>
<p>
Special care has to be taken when creating chains that the key/values output
by a Mapper are valid for the following Mapper in the chain. It is assumed
all Mappers and the Reduce in the chain use matching output and input key and
value classes as no conversion is done by the chaining code.
</p>
<p>
Using the ChainMapper and the ChainReducer classes is possible to compose
Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
immediate benefit of this pattern is a dramatic reduction in disk IO.
</p>
<p>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the chain.
</p>
ChainMapper usage pattern:
<p/>
<pre>
...
Job = new Job(conf);
<p/>
Configuration mapAConf = new Configuration(false);
...
ChainMapper.addMapper(job, AMap.class, LongWritable.class, Text.class,
Text.class, Text.class, true, mapAConf);
<p/>
Configuration mapBConf = new Configuration(false);
...
ChainMapper.addMapper(job, BMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, mapBConf);
<p/>
...
<p/>
job.waitForComplettion(true);
...
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.chain.ChainMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.chain.ChainReducer -->
<class name="ChainReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ChainReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setReducer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="reducerConf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Sets the {@link Reducer} class to the chain job.
<p>
The key and values are passed from one element of the chain to the next, by
value. For the added Reducer the configuration given for it,
<code>reducerConf</code>, have precedence over the job's Configuration.
This precedence is in effect when the task is running.
</p>
<p>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainReducer, this is done by the setReducer or the addMapper for the last
element in the chain.
</p>
@param job
the job
@param klass
the Reducer class to add.
@param inputKeyClass
reducer input key class.
@param inputValueClass
reducer input value class.
@param outputKeyClass
reducer output key class.
@param outputValueClass
reducer output value class.
@param reducerConf
a configuration for the Reducer class. It is recommended to use a
Configuration without default values using the
<code>Configuration(boolean loadDefaults)</code> constructor with
FALSE.]]>
</doc>
</method>
<method name="addMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="klass" type="java.lang.Class"/>
<param name="inputKeyClass" type="java.lang.Class"/>
<param name="inputValueClass" type="java.lang.Class"/>
<param name="outputKeyClass" type="java.lang.Class"/>
<param name="outputValueClass" type="java.lang.Class"/>
<param name="mapperConf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Adds a {@link Mapper} class to the chain reducer.
<p>
The key and values are passed from one element of the chain to the next, by
value For the added Mapper the configuration given for it,
<code>mapperConf</code>, have precedence over the job's Configuration. This
precedence is in effect when the task is running.
</p>
<p>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainMapper, this is done by the addMapper for the last mapper in the
chain.
</p>
@param job
The job.
@param klass
the Mapper class to add.
@param inputKeyClass
mapper input key class.
@param inputValueClass
mapper input value class.
@param outputKeyClass
mapper output key class.
@param outputValueClass
mapper output value class.
@param mapperConf
a configuration for the Mapper class. It is recommended to use a
Configuration without default values using the
<code>Configuration(boolean loadDefaults)</code> constructor with
FALSE.]]>
</doc>
</method>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<doc>
<![CDATA[The ChainReducer class allows to chain multiple Mapper classes after a
Reducer within the Reducer task.
<p>
For each record output by the Reducer, the Mapper classes are invoked in a
chained (or piped) fashion. The output of the reducer becomes the input of
the first mapper and output of first becomes the input of the second, and so
on until the last Mapper, the output of the last Mapper will be written to
the task's output.
</p>
<p>
The key functionality of this feature is that the Mappers in the chain do not
need to be aware that they are executed after the Reducer or in a chain. This
enables having reusable specialized Mappers that can be combined to perform
composite operations within a single task.
</p>
<p>
Special care has to be taken when creating chains that the key/values output
by a Mapper are valid for the following Mapper in the chain. It is assumed
all Mappers and the Reduce in the chain use matching output and input key and
value classes as no conversion is done by the chaining code.
</p>
</p> Using the ChainMapper and the ChainReducer classes is possible to
compose Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
immediate benefit of this pattern is a dramatic reduction in disk IO. </p>
<p>
IMPORTANT: There is no need to specify the output key/value classes for the
ChainReducer, this is done by the setReducer or the addMapper for the last
element in the chain.
</p>
ChainReducer usage pattern:
<p/>
<pre>
...
Job = new Job(conf);
....
<p/>
Configuration reduceConf = new Configuration(false);
...
ChainReducer.setReducer(job, XReduce.class, LongWritable.class, Text.class,
Text.class, Text.class, true, reduceConf);
<p/>
ChainReducer.addMapper(job, CMap.class, Text.class, Text.class,
LongWritable.class, Text.class, false, null);
<p/>
ChainReducer.addMapper(job, DMap.class, LongWritable.class, Text.class,
LongWritable.class, LongWritable.class, true, null);
<p/>
...
<p/>
job.waitForCompletion(true);
...
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.chain.ChainReducer -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.db">
<!-- start class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter -->
<class name="BigDecimalSplitter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBSplitter"/>
<constructor name="BigDecimalSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<method name="tryDivide" return="java.math.BigDecimal"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="numerator" type="java.math.BigDecimal"/>
<param name="denominator" type="java.math.BigDecimal"/>
<doc>
<![CDATA[Divide numerator by denominator. If impossible in exact mode, use rounding.]]>
</doc>
</method>
<doc>
<![CDATA[Implement DBSplitter over BigDecimal values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.BooleanSplitter -->
<class name="BooleanSplitter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBSplitter"/>
<constructor name="BooleanSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[Implement DBSplitter over boolean values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.BooleanSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat -->
<class name="DataDrivenDBInputFormat" extends="org.apache.hadoop.mapreduce.lib.db.DBInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="DataDrivenDBInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSplitter" return="org.apache.hadoop.mapreduce.lib.db.DBSplitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="sqlDataType" type="int"/>
<doc>
<![CDATA[@return the DBSplitter implementation to use to divide the table/query into InputSplits.]]>
</doc>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getBoundingValsQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[@return a query which returns the minimum and maximum values for
the order-by column.
The min value should be in the first column, and the
max value should be in the second column of the results.]]>
</doc>
</method>
<method name="setBoundingQuery"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="query" type="java.lang.String"/>
<doc>
<![CDATA[Set the user-defined bounding query to use with a user-defined query.
This *must* include the substring "$CONDITIONS"
(DataDrivenDBInputFormat.SUBSTITUTE_TOKEN) inside the WHERE clause,
so that DataDrivenDBInputFormat knows where to insert split clauses.
e.g., "SELECT foo FROM mytable WHERE $CONDITIONS"
This will be expanded to something like:
SELECT foo FROM mytable WHERE (id &gt; 100) AND (id &lt; 250)
inside each split.]]>
</doc>
</method>
<method name="createDBRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="tableName" type="java.lang.String"/>
<param name="conditions" type="java.lang.String"/>
<param name="splitBy" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<doc>
<![CDATA[Note that the "orderBy" column is called the "splitBy" in this version.
We reuse the same field, but it's not strictly ordering it -- just partitioning
the results.]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="inputQuery" type="java.lang.String"/>
<param name="inputBoundingQuery" type="java.lang.String"/>
<doc>
<![CDATA[setInput() takes a custom query and a separate "bounding query" to use
instead of the custom "count query" used by DBInputFormat.]]>
</doc>
</method>
<field name="SUBSTITUTE_TOKEN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[If users are providing their own query, the following string is expected to
appear in the WHERE clause, which will be substituted with a pair of conditions
on the input to allow input splits to parallelise the import.]]>
</doc>
</field>
<doc>
<![CDATA[A InputFormat that reads input data from an SQL table.
Operates like DBInputFormat, but instead of using LIMIT and OFFSET to demarcate
splits, it tries to generate WHERE clauses which separate the data into roughly
equivalent shards.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit -->
<class name="DataDrivenDBInputFormat.DataDrivenDBInputSplit" extends="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DataDrivenDBInputFormat.DataDrivenDBInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default Constructor]]>
</doc>
</constructor>
<constructor name="DataDrivenDBInputFormat.DataDrivenDBInputSplit" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convenience Constructor
@param lower the string to be put in the WHERE clause to guard on the 'lower' end
@param upper the string to be put in the WHERE clause to guard on the 'upper' end]]>
</doc>
</constructor>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@return The total row count in this split]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="output" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getLowerClause" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getUpperClause" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A InputSplit that spans a set of rows]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader -->
<class name="DataDrivenDBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DBRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DataDrivenDBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[@param split The InputSplit to read data for
@throws SQLException]]>
</doc>
</constructor>
<method name="getSelectQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the query for selecting the records,
subclasses can override this for custom behaviour.]]>
</doc>
</method>
<doc>
<![CDATA[A RecordReader that reads records from a SQL table,
using data-driven WHERE clause splits.
Emits LongWritables containing the record number as
key and DBWritables as value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DateSplitter -->
<class name="DateSplitter" extends="org.apache.hadoop.mapreduce.lib.db.IntegerSplitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DateSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<method name="dateToString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="d" type="java.util.Date"/>
<doc>
<![CDATA[Given a Date 'd', format it as a string for use in a SQL date
comparison operation.
@param d the date to format.
@return the string representing this date in SQL with any appropriate
quotation characters, etc.]]>
</doc>
</method>
<doc>
<![CDATA[Implement DBSplitter over date/time values.
Make use of logic from IntegerSplitter, since date/time are just longs
in Java.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DateSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBConfiguration -->
<class name="DBConfiguration" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DBConfiguration" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configureDB"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="driverClass" type="java.lang.String"/>
<param name="dbUrl" type="java.lang.String"/>
<param name="userName" type="java.lang.String"/>
<param name="passwd" type="java.lang.String"/>
<doc>
<![CDATA[Sets the DB access related fields in the {@link Configuration}.
@param conf the configuration
@param driverClass JDBC Driver class name
@param dbUrl JDBC DB access URL.
@param userName DB access username
@param passwd DB access passwd]]>
</doc>
</method>
<method name="configureDB"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
<param name="driverClass" type="java.lang.String"/>
<param name="dbUrl" type="java.lang.String"/>
<doc>
<![CDATA[Sets the DB access related fields in the JobConf.
@param job the job
@param driverClass JDBC Driver class name
@param dbUrl JDBC DB access URL.]]>
</doc>
</method>
<method name="getConnection" return="java.sql.Connection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[Returns a connection object o the DB
@throws ClassNotFoundException
@throws SQLException]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputTableName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputTableName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tableName" type="java.lang.String"/>
</method>
<method name="getInputFieldNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputFieldNames"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fieldNames" type="java.lang.String[]"/>
</method>
<method name="getInputConditions" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputConditions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conditions" type="java.lang.String"/>
</method>
<method name="getInputOrderBy" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputOrderBy"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="orderby" type="java.lang.String"/>
</method>
<method name="getInputQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputQuery"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
</method>
<method name="getInputCountQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputCountQuery"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
</method>
<method name="setInputBoundingQuery"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
</method>
<method name="getInputBoundingQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setInputClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inputClass" type="java.lang.Class"/>
</method>
<method name="getOutputTableName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setOutputTableName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tableName" type="java.lang.String"/>
</method>
<method name="getOutputFieldNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setOutputFieldNames"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fieldNames" type="java.lang.String[]"/>
</method>
<method name="setOutputFieldCount"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fieldCount" type="int"/>
</method>
<method name="getOutputFieldCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="DRIVER_CLASS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The JDBC Driver class name]]>
</doc>
</field>
<field name="URL_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[JDBC Database access URL]]>
</doc>
</field>
<field name="USERNAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[User name to access the database]]>
</doc>
</field>
<field name="PASSWORD_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Password to access the database]]>
</doc>
</field>
<field name="INPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Input table name]]>
</doc>
</field>
<field name="INPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Field names in the Input table]]>
</doc>
</field>
<field name="INPUT_CONDITIONS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[WHERE clause in the input SELECT statement]]>
</doc>
</field>
<field name="INPUT_ORDER_BY_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[ORDER BY clause in the input SELECT statement]]>
</doc>
</field>
<field name="INPUT_QUERY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Whole input query, exluding LIMIT...OFFSET]]>
</doc>
</field>
<field name="INPUT_COUNT_QUERY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Input query to get the count of records]]>
</doc>
</field>
<field name="INPUT_BOUNDING_QUERY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Input query to get the max and min values of the jdbc.input.query]]>
</doc>
</field>
<field name="INPUT_CLASS_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Class name implementing DBWritable which will hold input tuples]]>
</doc>
</field>
<field name="OUTPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Output table name]]>
</doc>
</field>
<field name="OUTPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Field names in the Output table]]>
</doc>
</field>
<field name="OUTPUT_FIELD_COUNT_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of fields in the Output table]]>
</doc>
</field>
<doc>
<![CDATA[A container for configuration property names for jobs with DB input/output.
The job can be configured using the static methods in this class,
{@link DBInputFormat}, and {@link DBOutputFormat}.
Alternatively, the properties can be set in the configuration with proper
values.
@see DBConfiguration#configureDB(Configuration, String, String, String, String)
@see DBInputFormat#setInput(Job, Class, String, String)
@see DBInputFormat#setInput(Job, Class, String, String, String, String...)
@see DBOutputFormat#setOutput(Job, String, String...)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBConfiguration -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBInputFormat -->
<class name="DBInputFormat" extends="org.apache.hadoop.mapreduce.InputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="DBInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDBConf" return="org.apache.hadoop.mapreduce.lib.db.DBConfiguration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getConnection" return="java.sql.Connection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDBProductName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createDBRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getCountQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the query for getting the total number of rows,
subclasses can override this for custom behaviour.]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="tableName" type="java.lang.String"/>
<param name="conditions" type="java.lang.String"/>
<param name="orderBy" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<doc>
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
@param job The map-reduce job
@param inputClass the class object implementing DBWritable, which is the
Java object holding tuple fields.
@param tableName The table to read data from
@param conditions The condition which to select data with,
eg. '(updated > 20070101 AND length > 0)'
@param orderBy the fieldNames in the orderBy clause.
@param fieldNames The field names in the table
@see #setInput(Job, Class, String, String)]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="inputClass" type="java.lang.Class"/>
<param name="inputQuery" type="java.lang.String"/>
<param name="inputCountQuery" type="java.lang.String"/>
<doc>
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
@param job The map-reduce job
@param inputClass the class object implementing DBWritable, which is the
Java object holding tuple fields.
@param inputQuery the input query to select fields. Example :
"SELECT f1, f2, f3 FROM Mytable ORDER BY f1"
@param inputCountQuery the input query that returns
the number of records in the table.
Example : "SELECT COUNT(f1) FROM Mytable"
@see #setInput(Job, Class, String, String, String, String...)]]>
</doc>
</method>
<method name="closeConnection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A InputFormat that reads input data from an SQL table.
<p>
DBInputFormat emits LongWritables containing the record number as
key and DBWritables as value.
The SQL query, and input class can be using one of the two
setInput methods.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit -->
<class name="DBInputFormat.DBInputSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="DBInputFormat.DBInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default Constructor]]>
</doc>
</constructor>
<constructor name="DBInputFormat.DBInputSplit" type="long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convenience Constructor
@param start the index of the first row to select
@param end the index of the last row to select]]>
</doc>
</constructor>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getStart" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return The index of the first row to select]]>
</doc>
</method>
<method name="getEnd" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return The index of the last row to select]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@return The total row count in this split]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="output" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A InputSplit that spans a set of rows]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable -->
<class name="DBInputFormat.NullDBWritable" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBWritable"/>
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="DBInputFormat.NullDBWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="java.sql.ResultSet"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="java.sql.PreparedStatement"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[A Class that does nothing, implementing DBWritable]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat -->
<class name="DBOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DBOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="constructQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="table" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<doc>
<![CDATA[Constructs the query used as the prepared statement to insert data.
@param table
the table to insert into
@param fieldNames
the fields to insert into. If field names are unknown, supply an
array of nulls.]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="tableName" type="java.lang.String"/>
<param name="fieldNames" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Initializes the reduce-part of the job with
the appropriate output settings
@param job The job
@param tableName The table to insert data into
@param fieldNames The field names in the table.]]>
</doc>
</method>
<method name="setOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="tableName" type="java.lang.String"/>
<param name="fieldCount" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Initializes the reduce-part of the job
with the appropriate output settings
@param job The job
@param tableName The table to insert data into
@param fieldCount the number of fields in the table.]]>
</doc>
</method>
<doc>
<![CDATA[A OutputFormat that sends the reduce output to a SQL table.
<p>
{@link DBOutputFormat} accepts &lt;key,value&gt; pairs, where
key has a type extending DBWritable. Returned {@link RecordWriter}
writes <b>only the key</b> to the database with a batch SQL query.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter -->
<class name="DBOutputFormat.DBRecordWriter" extends="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DBOutputFormat.DBRecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<constructor name="DBOutputFormat.DBRecordWriter" type="java.sql.Connection, java.sql.PreparedStatement"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<method name="getConnection" return="java.sql.Connection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getStatement" return="java.sql.PreparedStatement"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A RecordWriter that writes the reduce output to a SQL table]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.DBRecordReader -->
<class name="DBRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[@param split The InputSplit to read data for
@throws SQLException]]>
</doc>
</constructor>
<method name="executeQuery" return="java.sql.ResultSet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<method name="getSelectQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the query for selecting the records,
subclasses can override this for custom behaviour.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentKey" return="org.apache.hadoop.io.LongWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getCurrentValue" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="createValue" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #nextKeyValue()}">
<param name="key" type="org.apache.hadoop.io.LongWritable"/>
<param name="value" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use {@link #nextKeyValue()}]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getSplit" return="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getFieldNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getTableName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getConditions" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getDBConf" return="org.apache.hadoop.mapreduce.lib.db.DBConfiguration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getConnection" return="java.sql.Connection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getStatement" return="java.sql.PreparedStatement"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="setStatement"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="stmt" type="java.sql.PreparedStatement"/>
</method>
<field name="statement" type="java.sql.PreparedStatement"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A RecordReader that reads records from a SQL table.
Emits LongWritables containing the record number as
key and DBWritables as value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.DBRecordReader -->
<!-- start interface org.apache.hadoop.mapreduce.lib.db.DBSplitter -->
<interface name="DBSplitter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[Given a ResultSet containing one record (and already advanced to that record)
with two columns (a low value, and a high value, both of the same type), determine
a set of splits that span the given values.]]>
</doc>
</method>
<doc>
<![CDATA[DBSplitter will generate DBInputSplits to use with DataDrivenDBInputFormat.
DataDrivenDBInputFormat needs to interpolate between two values that
represent the lowest and highest valued records to import. Depending
on the data-type of the column, this requires different behavior.
DBSplitter implementations should perform this for a data type or family
of data types.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.db.DBSplitter -->
<!-- start interface org.apache.hadoop.mapreduce.lib.db.DBWritable -->
<interface name="DBWritable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="statement" type="java.sql.PreparedStatement"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[Sets the fields of the object in the {@link PreparedStatement}.
@param statement the statement that the fields are put into.
@throws SQLException]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="resultSet" type="java.sql.ResultSet"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[Reads the fields of the object from the {@link ResultSet}.
@param resultSet the {@link ResultSet} to get the fields from.
@throws SQLException]]>
</doc>
</method>
<doc>
<![CDATA[Objects that are read from/written to a database should implement
<code>DBWritable</code>. DBWritable, is similar to {@link Writable}
except that the {@link #write(PreparedStatement)} method takes a
{@link PreparedStatement}, and {@link #readFields(ResultSet)}
takes a {@link ResultSet}.
<p>
Implementations are responsible for writing the fields of the object
to PreparedStatement, and reading the fields of the object from the
ResultSet.
<p>Example:</p>
If we have the following table in the database :
<pre>
CREATE TABLE MyTable (
counter INTEGER NOT NULL,
timestamp BIGINT NOT NULL,
);
</pre>
then we can read/write the tuples from/to the table with :
<p><pre>
public class MyWritable implements Writable, DBWritable {
// Some data
private int counter;
private long timestamp;
//Writable#write() implementation
public void write(DataOutput out) throws IOException {
out.writeInt(counter);
out.writeLong(timestamp);
}
//Writable#readFields() implementation
public void readFields(DataInput in) throws IOException {
counter = in.readInt();
timestamp = in.readLong();
}
public void write(PreparedStatement statement) throws SQLException {
statement.setInt(1, counter);
statement.setLong(2, timestamp);
}
public void readFields(ResultSet resultSet) throws SQLException {
counter = resultSet.getInt(1);
timestamp = resultSet.getLong(2);
}
}
</pre></p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.db.DBWritable -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.FloatSplitter -->
<class name="FloatSplitter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBSplitter"/>
<constructor name="FloatSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[Implement DBSplitter over floating-point values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.FloatSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter -->
<class name="IntegerSplitter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.db.DBSplitter"/>
<constructor name="IntegerSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[Implement DBSplitter over integer values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.MySQLDataDrivenDBRecordReader -->
<class name="MySQLDataDrivenDBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MySQLDataDrivenDBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<method name="executeQuery" return="java.sql.ResultSet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[A RecordReader that reads records from a MySQL table via DataDrivenDBRecordReader]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.MySQLDataDrivenDBRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.MySQLDBRecordReader -->
<class name="MySQLDBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DBRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MySQLDBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<method name="executeQuery" return="java.sql.ResultSet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="query" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
</method>
<doc>
<![CDATA[A RecordReader that reads records from a MySQL table.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.MySQLDBRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat -->
<class name="OracleDataDrivenDBInputFormat" extends="org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="OracleDataDrivenDBInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSplitter" return="org.apache.hadoop.mapreduce.lib.db.DBSplitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="sqlDataType" type="int"/>
<doc>
<![CDATA[@return the DBSplitter implementation to use to divide the table/query into InputSplits.]]>
</doc>
</method>
<method name="createDBRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A InputFormat that reads input data from an SQL table in an Oracle db.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBRecordReader -->
<class name="OracleDataDrivenDBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OracleDataDrivenDBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<doc>
<![CDATA[A RecordReader that reads records from a Oracle table via DataDrivenDBRecordReader]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.OracleDateSplitter -->
<class name="OracleDateSplitter" extends="org.apache.hadoop.mapreduce.lib.db.DateSplitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OracleDateSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="dateToString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="d" type="java.util.Date"/>
</method>
<doc>
<![CDATA[Implement DBSplitter over date/time values returned by an Oracle db.
Make use of logic from DateSplitter, since this just needs to use
some Oracle-specific functions on the formatting end when generating
InputSplits.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.OracleDateSplitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader -->
<class name="OracleDBRecordReader" extends="org.apache.hadoop.mapreduce.lib.db.DBRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OracleDBRecordReader" type="org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.conf.Configuration, java.sql.Connection, org.apache.hadoop.mapreduce.lib.db.DBConfiguration, java.lang.String, java.lang.String[], java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="SQLException" type="java.sql.SQLException"/>
</constructor>
<method name="getSelectQuery" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the query for selecting the records from an Oracle DB.]]>
</doc>
</method>
<method name="setSessionTimeZone"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="conn" type="java.sql.Connection"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[Set session time zone
@param conf The current configuration.
We read the 'oracle.sessionTimeZone' property from here.
@param conn The connection to alter the timezone properties of.]]>
</doc>
</method>
<field name="SESSION_TIMEZONE_KEY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Configuration key to set to a timezone string.]]>
</doc>
</field>
<doc>
<![CDATA[A RecordReader that reads records from an Oracle SQL table.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.db.TextSplitter -->
<class name="TextSplitter" extends="org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TextSplitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="split" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="results" type="java.sql.ResultSet"/>
<param name="colName" type="java.lang.String"/>
<exception name="SQLException" type="java.sql.SQLException"/>
<doc>
<![CDATA[This method needs to determine the splits between two user-provided strings.
In the case where the user's strings are 'A' and 'Z', this is not hard; we
could create two splits from ['A', 'M') and ['M', 'Z'], 26 splits for strings
beginning with each letter, etc.
If a user has provided us with the strings "Ham" and "Haze", however, we need
to create splits that differ in the third letter.
The algorithm used is as follows:
Since there are 2**16 unicode characters, we interpret characters as digits in
base 65536. Given a string 's' containing characters s_0, s_1 .. s_n, we interpret
the string as the number: 0.s_0 s_1 s_2.. s_n in base 65536. Having mapped the
low and high strings into floating-point values, we then use the BigDecimalSplitter
to establish the even split points, then map the resulting floating point values
back into strings.]]>
</doc>
</method>
<doc>
<![CDATA[Implement DBSplitter over text strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.db.TextSplitter -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.fieldsel">
<!-- start class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper -->
<class name="FieldSelectionHelper" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FieldSelectionHelper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FieldSelectionHelper" type="org.apache.hadoop.io.Text, org.apache.hadoop.io.Text"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="parseOutputKeyValueSpec" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="keyValueSpec" type="java.lang.String"/>
<param name="keyFieldList" type="java.util.List"/>
<param name="valueFieldList" type="java.util.List"/>
</method>
<method name="specToString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fieldSeparator" type="java.lang.String"/>
<param name="keyValueSpec" type="java.lang.String"/>
<param name="allValueFieldsFrom" type="int"/>
<param name="keyFieldList" type="java.util.List"/>
<param name="valueFieldList" type="java.util.List"/>
</method>
<method name="getKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="extractOutputKeyValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.String"/>
<param name="val" type="java.lang.String"/>
<param name="fieldSep" type="java.lang.String"/>
<param name="keyFieldList" type="java.util.List"/>
<param name="valFieldList" type="java.util.List"/>
<param name="allValueFieldsFrom" type="int"/>
<param name="ignoreKey" type="boolean"/>
<param name="isMap" type="boolean"/>
</method>
<field name="emptyText" type="org.apache.hadoop.io.Text"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="DATA_FIELD_SEPERATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAP_OUTPUT_KEY_VALUE_SPEC" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="REDUCE_OUTPUT_KEY_VALUE_SPEC" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a mapper/reducer class that can be used to perform
field selections in a manner similar to unix cut. The input data is treated
as fields separated by a user specified separator (the default value is
"\t"). The user can specify a list of fields that form the map output keys,
and a list of fields that form the map output values. If the inputformat is
TextInputFormat, the mapper will ignore the key to the map function. and the
fields are from the value only. Otherwise, the fields are the union of those
from the key and those from the value.
The field separator is under attribute "mapreduce.fieldsel.data.field.separator"
The map output field list spec is under attribute
"mapreduce.fieldsel.map.output.key.value.fields.spec".
The value is expected to be like "keyFieldsSpec:valueFieldsSpec"
key/valueFieldsSpec are comma (,) separated field spec: fieldSpec,fieldSpec,fieldSpec ...
Each field spec can be a simple number (e.g. 5) specifying a specific field, or a range
(like 2-5) to specify a range of fields, or an open range (like 3-) specifying all
the fields starting from field 3. The open range field spec applies value fields only.
They have no effect on the key fields.
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields 4,3,0 and 1 for keys,
and use fields 6,5,1,2,3,7 and above for values.
The reduce output field list spec is under attribute
"mapreduce.fieldsel.reduce.output.key.value.fields.spec".
The reducer extracts output key/value pairs in a similar manner, except that
the key is never ignored.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper -->
<!-- start class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper -->
<class name="FieldSelectionMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FieldSelectionMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="val" type="V"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[The identify function. Input key/value pair is written directly to output.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a mapper class that can be used to perform
field selections in a manner similar to unix cut. The input data is treated
as fields separated by a user specified separator (the default value is
"\t"). The user can specify a list of fields that form the map output keys,
and a list of fields that form the map output values. If the inputformat is
TextInputFormat, the mapper will ignore the key to the map function. and the
fields are from the value only. Otherwise, the fields are the union of those
from the key and those from the value.
The field separator is under attribute "mapreduce.fieldsel.data.field.separator"
The map output field list spec is under attribute
"mapreduce.fieldsel.map.output.key.value.fields.spec".
The value is expected to be like
"keyFieldsSpec:valueFieldsSpec" key/valueFieldsSpec are comma (,) separated
field spec: fieldSpec,fieldSpec,fieldSpec ... Each field spec can be a
simple number (e.g. 5) specifying a specific field, or a range (like 2-5)
to specify a range of fields, or an open range (like 3-) specifying all
the fields starting from field 3. The open range field spec applies value
fields only. They have no effect on the key fields.
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields
4,3,0 and 1 for keys, and use fields 6,5,1,2,3,7 and above for values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer -->
<class name="FieldSelectionReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FieldSelectionReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a reducer class that can be used to perform field
selections in a manner similar to unix cut.
The input data is treated as fields separated by a user specified
separator (the default value is "\t"). The user can specify a list of
fields that form the reduce output keys, and a list of fields that form
the reduce output values. The fields are the union of those from the key
and those from the value.
The field separator is under attribute "mapreduce.fieldsel.data.field.separator"
The reduce output field list spec is under attribute
"mapreduce.fieldsel.reduce.output.key.value.fields.spec".
The value is expected to be like
"keyFieldsSpec:valueFieldsSpec" key/valueFieldsSpec are comma (,)
separated field spec: fieldSpec,fieldSpec,fieldSpec ... Each field spec
can be a simple number (e.g. 5) specifying a specific field, or a range
(like 2-5) to specify a range of fields, or an open range (like 3-)
specifying all the fields starting from field 3. The open range field
spec applies value fields only. They have no effect on the key fields.
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields
4,3,0 and 1 for keys, and use fields 6,5,1,2,3,7 and above for values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.input">
<!-- start class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat -->
<class name="CombineFileInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CombineFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[default constructor]]>
</doc>
</constructor>
<method name="setMaxSplitSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="maxSplitSize" type="long"/>
<doc>
<![CDATA[Specify the maximum size (in bytes) of each split. Each split is
approximately equal to the specified size.]]>
</doc>
</method>
<method name="setMinSplitSizeNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="minSplitSizeNode" type="long"/>
<doc>
<![CDATA[Specify the minimum size (in bytes) of each split per node.
This applies to data that is left over after combining data on a single
node into splits that are of maximum size specified by maxSplitSize.
This leftover data will be combined into its own split if its size
exceeds minSplitSizeNode.]]>
</doc>
</method>
<method name="setMinSplitSizeRack"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="minSplitSizeRack" type="long"/>
<doc>
<![CDATA[Specify the minimum size (in bytes) of each split per rack.
This applies to data that is left over after combining data on a single
rack into splits that are of maximum size specified by maxSplitSize.
This leftover data will be combined into its own split if its size
exceeds minSplitSizeRack.]]>
</doc>
</method>
<method name="createPool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="filters" type="java.util.List"/>
<doc>
<![CDATA[Create a new pool and add the filters to it.
A split cannot have files from different pools.]]>
</doc>
</method>
<method name="createPool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="filters" type="org.apache.hadoop.fs.PathFilter[]"/>
<doc>
<![CDATA[Create a new pool and add the filters to it.
A pathname can satisfy any one of the specified filters.
A split cannot have files from different pools.]]>
</doc>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This is not implemented yet.]]>
</doc>
</method>
<field name="SPLIT_MINSIZE_PERNODE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SPLIT_MINSIZE_PERRACK" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An abstract {@link InputFormat} that returns {@link CombineFileSplit}'s in
{@link InputFormat#getSplits(JobContext)} method.
Splits are constructed from the files under the input paths.
A split cannot have files from different pools.
Each split returned may contain blocks from different files.
If a maxSplitSize is specified, then blocks on the same node are
combined to form a single split. Blocks that are left over are
then combined with other blocks in the same rack.
If maxSplitSize is not specified, then blocks from the same rack
are combined in a single split; no attempt is made to create
node-local splits.
If the maxSplitSize is equal to the block size, then this class
is similar to the default splitting behavior in Hadoop: each
block is a locally processed split.
Subclasses implement
{@link InputFormat#createRecordReader(InputSplit, TaskAttemptContext)}
to construct <code>RecordReader</code>'s for
<code>CombineFileSplit</code>'s.
@see CombineFileSplit]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader -->
<class name="CombineFileRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CombineFileRecordReader" type="org.apache.hadoop.mapreduce.lib.input.CombineFileSplit, org.apache.hadoop.mapreduce.TaskAttemptContext, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A generic RecordReader that can hand out different recordReaders
for each chunk in the CombineFileSplit.]]>
</doc>
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[return progress based on the amount of data processed so far.]]>
</doc>
</method>
<method name="initNextRecordReader" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the record reader for the next chunk in this CombineFileSplit.]]>
</doc>
</method>
<field name="split" type="org.apache.hadoop.mapreduce.lib.input.CombineFileSplit"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rrClass" type="java.lang.Class"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rrConstructor" type="java.lang.reflect.Constructor"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="idx" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="progress" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="curReader" type="org.apache.hadoop.mapreduce.RecordReader"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A generic RecordReader that can hand out different recordReaders
for each chunk in a {@link CombineFileSplit}.
A CombineFileSplit can combine data chunks from multiple files.
This class allows using different RecordReaders for processing
these data chunks from different files.
@see CombineFileSplit]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit -->
<class name="CombineFileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="CombineFileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[default constructor]]>
</doc>
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.fs.Path[], long[], long[], java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.fs.Path[], long[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapreduce.lib.input.CombineFileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy constructor]]>
</doc>
</constructor>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getStartOffsets" return="long[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns an array containing the start offsets of the files in the split]]>
</doc>
</method>
<method name="getLengths" return="long[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns an array containing the lengths of the files in the split]]>
</doc>
</method>
<method name="getOffset" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Returns the start offset of the i<sup>th</sup> Path]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Returns the length of the i<sup>th</sup> Path]]>
</doc>
</method>
<method name="getNumPaths" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of Paths in the split]]>
</doc>
</method>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Returns the i<sup>th</sup> Path]]>
</doc>
</method>
<method name="getPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns all the Paths in the split]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns all the Paths where this input-split resides]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A sub-collection of input files.
Unlike {@link FileSplit}, CombineFileSplit class does not represent
a split of a file, but a split of input files into smaller sets.
A split may contain blocks from different file but all
the blocks in the same split are probably local to some rack <br>
CombineFileSplit can be used to implement {@link RecordReader}'s,
with reading one record per file.
@see FileSplit
@see CombineFileInputFormat]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.FileInputFormat -->
<class name="FileInputFormat" extends="org.apache.hadoop.mapreduce.InputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getFormatMinSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Get the lower bound on split size imposed by the format.
@return the number of bytes of the minimal split for this format]]>
</doc>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="filename" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
<code>FileInputFormat</code> implementations can override this and return
<code>false</code> to ensure that individual input files are never split-up
so that {@link Mapper}s process entire files.
@param context the job context
@param filename the file name to check
@return is this file splitable?]]>
</doc>
</method>
<method name="setInputPathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="filter" type="java.lang.Class"/>
<doc>
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
@param job the job to modify
@param filter the PathFilter class use for filtering the input paths.]]>
</doc>
</method>
<method name="setMinInputSplitSize"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="size" type="long"/>
<doc>
<![CDATA[Set the minimum input split size
@param job the job to modify
@param size the minimum size]]>
</doc>
</method>
<method name="getMinSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the minimum split size
@param job the job
@return the minimum number of bytes that can be in a split]]>
</doc>
</method>
<method name="setMaxInputSplitSize"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="size" type="long"/>
<doc>
<![CDATA[Set the maximum split size
@param job the job to modify
@param size the maximum split size]]>
</doc>
</method>
<method name="getMaxSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the maximum split size.
@param context the job to look at.
@return the maximum number of bytes a split can include]]>
</doc>
</method>
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
@return the PathFilter instance set for the job, NULL if none has been set.]]>
</doc>
</method>
<method name="listStatus" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List input directories.
Subclasses may override to, e.g., select only files matching a regular
expression.
@param job the job to list input paths for
@return array of FileStatus objects
@throws IOException if zero items.]]>
</doc>
</method>
<method name="makeSplit" return="org.apache.hadoop.mapreduce.lib.input.FileSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="length" type="long"/>
<param name="hosts" type="java.lang.String[]"/>
<doc>
<![CDATA[A factory that makes the split for this class. It can be overridden
by sub-classes to make sub-types]]>
</doc>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Generate the list of files and make them into FileSplits.
@param job the job context
@throws IOException]]>
</doc>
</method>
<method name="computeSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blockSize" type="long"/>
<param name="minSize" type="long"/>
<param name="maxSize" type="long"/>
</method>
<method name="getBlockIndex" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sets the given comma separated paths as the list of inputs
for the map-reduce job.
@param job the job
@param commaSeparatedPaths Comma separated paths to be set as
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add the given comma separated paths to the list of inputs for
the map-reduce job.
@param job The job to modify
@param commaSeparatedPaths Comma separated paths to be added to
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the array of {@link Path}s as the list of inputs
for the map-reduce job.
@param job The job to modify
@param inputPaths the {@link Path}s of the input directories/files
for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
@param job The {@link Job} to modify
@param path {@link Path} to be added to the list of inputs for
the map-reduce job.]]>
</doc>
</method>
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
@param context The job
@return the list of input {@link Path}s for the map-reduce job.]]>
</doc>
</method>
<field name="COUNTER_GROUP" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BYTES_READ" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INPUT_DIR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SPLIT_MAXSIZE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SPLIT_MINSIZE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PATHFILTER_CLASS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NUM_INPUT_FILES" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class for file-based {@link InputFormat}s.
<p><code>FileInputFormat</code> is the base class for all file-based
<code>InputFormat</code>s. This provides a generic implementation of
{@link #getSplits(JobContext)}.
Subclasses of <code>FileInputFormat</code> can also override the
{@link #isSplitable(JobContext, Path)} method to ensure input-files are
not split-up and are processed as a whole by {@link Mapper}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.FileInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.FileSplit -->
<class name="FileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a split with host information
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process
@param hosts the list of hosts containing the block, possibly null]]>
</doc>
</constructor>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The file containing this split's data.]]>
</doc>
</method>
<method name="getStart" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The position of the first byte in the file to process.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes in the file to process.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A section of an input file. Returned by {@link
InputFormat#getSplits(JobContext)} and passed to
{@link InputFormat#createRecordReader(InputSplit,TaskAttemptContext)}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.FileSplit -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.InvalidInputException -->
<class name="InvalidInputException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidInputException" type="java.util.List"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create the exception with the given list.
@param probs the list of problems to report. this list is not copied.]]>
</doc>
</constructor>
<method name="getProblems" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the complete list of the problems reported.
@return the list of problems, which must not be modified]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a summary message of the problems found.
@return the concatenated messages from all of the problems.]]>
</doc>
</method>
<doc>
<![CDATA[This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.InvalidInputException -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader -->
<class name="KeyValueLineRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="KeyValueLineRecordReader" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="findSeparator" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<param name="sep" type="byte"/>
</method>
<method name="setKeyValue"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="line" type="byte[]"/>
<param name="lineLen" type="int"/>
<param name="pos" type="int"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getCurrentKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCurrentValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="KEY_VALUE_SEPERATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class treats a line in the input as a key/value pair separated by a
separator character. The separator can be specified in config file
under the attribute name mapreduce.input.keyvaluelinerecordreader.key.value.separator. The default
separator is the tab character ('\t').]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat -->
<class name="KeyValueTextInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="KeyValueTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either line feed or carriage-return are used to signal end of line.
Each line is divided into key and value parts by a separator byte. If no
such a byte exists, the key will be the entire line and value will be empty.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.MultipleInputs -->
<class name="MultipleInputs" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultipleInputs"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFormatClass" type="java.lang.Class"/>
<doc>
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} to the list of
inputs for the map-reduce job.
@param job The {@link Job}
@param path {@link Path} to be added to the list of inputs for the job
@param inputFormatClass {@link InputFormat} class to use for this path]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFormatClass" type="java.lang.Class"/>
<param name="mapperClass" type="java.lang.Class"/>
<doc>
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} and
{@link Mapper} to the list of inputs for the map-reduce job.
@param job The {@link Job}
@param path {@link Path} to be added to the list of inputs for the job
@param inputFormatClass {@link InputFormat} class to use for this path
@param mapperClass {@link Mapper} class to use for this path]]>
</doc>
</method>
<field name="DIR_FORMATS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DIR_MAPPERS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class supports MapReduce jobs that have multiple input paths with
a different {@link InputFormat} and {@link Mapper} for each path]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.MultipleInputs -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat -->
<class name="NLineInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NLineInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Logically splits the set of input files for the job, splits N lines
of the input as one split.
@see FileInputFormat#getSplits(JobContext)]]>
</doc>
</method>
<method name="getSplitsForFile" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="org.apache.hadoop.fs.FileStatus"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="numLinesPerSplit" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setNumLinesPerSplit"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="numLines" type="int"/>
<doc>
<![CDATA[Set the number of lines per split
@param job the job to modify
@param numLines the number of lines per split]]>
</doc>
</method>
<method name="getNumLinesPerSplit" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the number of lines per split
@param job the job
@return the number of lines per split]]>
</doc>
</method>
<field name="LINES_PER_MAP" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[NLineInputFormat which splits N lines of input as one split.
In many "pleasantly" parallel applications, each process/mapper
processes the same input file (s), but with computations are
controlled by different parameters.(Referred to as "parameter sweeps").
One way to achieve this, is to specify a set of parameters
(one set per line) as input in a control file
(which is the input path to the map-reduce application,
where as the input dataset is specified
via a config variable in JobConf.).
The NLineInputFormat can be used in such applications, that splits
the input file such that by default, one line is fed as
a value to one map task, and key is the offset.
i.e. (k,v) is (LongWritable, Text).
The location hints will span the whole mapred cluster.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat -->
<class name="SequenceFileAsBinaryInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsBinaryInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[InputFormat reading keys, values from SequenceFiles in binary (raw)
format.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<class name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentKey" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentValue" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getKeyClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the key class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getKeyClassName]]>
</doc>
</method>
<method name="getValueClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the value class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getValueClassName]]>
</doc>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Read raw bytes from a SequenceFile.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<doc>
<![CDATA[Read records from a SequenceFile as binary (raw) bytes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat -->
<class name="SequenceFileAsTextInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is similar to SequenceFileInputFormat, except it generates
SequenceFileAsTextRecordReader which converts the input keys and values
to their String forms by calling toString() method.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader -->
<class name="SequenceFileAsTextRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsTextRecordReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class converts the input keys and values to their String forms by
calling toString() method. This class to SequenceFileAsTextInputFormat
class is as LineRecordReader class to TextInputFormat class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter -->
<class name="SequenceFileInputFilter" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a record reader for the given split
@param split file split
@param context the task-attempt context
@return RecordReader]]>
</doc>
</method>
<method name="setFilterClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="filterClass" type="java.lang.Class"/>
<doc>
<![CDATA[set the filter class
@param job The job
@param filterClass filter class]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FILTER_CLASS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FILTER_FREQUENCY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FILTER_REGEX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A class that allows a map/red job to work on a sample of sequence files.
The sample is decided by the filter class set by the job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter -->
<!-- start interface org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.Filter -->
<interface name="SequenceFileInputFilter.Filter" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[filter function
Decide if a record should be filtered or not
@param key record key
@return true if a record is accepted; return false otherwise]]>
</doc>
</method>
<doc>
<![CDATA[filter interface]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.Filter -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase -->
<class name="SequenceFileInputFilter.FilterBase" extends="java.lang.Object"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.Filter"/>
<constructor name="SequenceFileInputFilter.FilterBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[base class for Filters]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter -->
<class name="SequenceFileInputFilter.MD5Filter" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.MD5Filter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the filtering frequency in configuration
@param conf configuration
@param frequency filtering frequency]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter according to configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If MD5(key) % frequency==0, return true; otherwise return false
@see Filter#accept(Object)]]>
</doc>
</method>
<field name="MD5_LEN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency <i>f</i>. The filtering criteria is
MD5(key) % f == 0.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter -->
<class name="SequenceFileInputFilter.PercentFilter" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.PercentFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the frequency and stores it in conf
@param conf configuration
@param frequency filtering frequencey]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter by checking the configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If record# % frequency==0, return true; otherwise return false
@see Filter#accept(Object)]]>
</doc>
</method>
<doc>
<![CDATA[This class returns a percentage of records
The percentage is determined by a filtering frequency <i>f</i> using
the criteria record# % f == 0.
For example, if the frequency is 10, one out of 10 records is returned.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter -->
<class name="SequenceFileInputFilter.RegexFilter" extends="org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.RegexFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setPattern"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="regex" type="java.lang.String"/>
<exception name="PatternSyntaxException" type="java.util.regex.PatternSyntaxException"/>
<doc>
<![CDATA[Define the filtering regex and stores it in conf
@param conf where the regex is set
@param regex regex used as a filter]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the Filter by checking the configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[Filtering method
If key matches the regex, return true; otherwise return false
@see Filter#accept(Object)]]>
</doc>
</method>
<doc>
<![CDATA[Records filter by matching key to regex]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat -->
<class name="SequenceFileInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getFormatMinSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="listStatus" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader -->
<class name="SequenceFileRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileRecordReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCurrentValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="conf" type="org.apache.hadoop.conf.Configuration"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link RecordReader} for {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.input.TextInputFormat -->
<class name="TextInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either linefeed or carriage-return are used to signal end of line. Keys are
the position in the file, and values are the line of text..]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.input.TextInputFormat -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.jobcontrol">
<!-- start class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob -->
<class name="ControlledJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ControlledJob" type="org.apache.hadoop.mapreduce.Job, java.util.List"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a job.
@param job a mapreduce job to be executed.
@param dependingJobs an array of jobs the current job depends on]]>
</doc>
</constructor>
<constructor name="ControlledJob" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a job.
@param conf mapred job configuration representing a job to be executed.
@throws IOException]]>
</doc>
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job name of this job]]>
</doc>
</method>
<method name="setJobName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobName" type="java.lang.String"/>
<doc>
<![CDATA[Set the job name for this job.
@param jobName the job name]]>
</doc>
</method>
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job ID of this job assigned by JobControl]]>
</doc>
</method>
<method name="setJobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="id" type="java.lang.String"/>
<doc>
<![CDATA[Set the job ID for this job.
@param id the job ID]]>
</doc>
</method>
<method name="getMapredJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapred ID of this job as assigned by the
mapred framework.]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapreduce.Job"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapreduce job]]>
</doc>
</method>
<method name="setJob"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<doc>
<![CDATA[Set the mapreduce job
@param job the mapreduce job for this job.]]>
</doc>
</method>
<method name="getJobState" return="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the state of this job]]>
</doc>
</method>
<method name="setJobState"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State"/>
<doc>
<![CDATA[Set the state for this job.
@param state the new state for this job.]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the message of this job]]>
</doc>
</method>
<method name="setMessage"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="message" type="java.lang.String"/>
<doc>
<![CDATA[Set the message for this job.
@param message the message for this job.]]>
</doc>
</method>
<method name="getDependentJobs" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the depending jobs of this job]]>
</doc>
</method>
<method name="addDependingJob" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dependingJob" type="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob"/>
<doc>
<![CDATA[Add a job to this jobs' dependency list.
Dependent jobs can only be added while a Job
is waiting to run, not during or afterwards.
@param dependingJob Job that this Job depends on.
@return <tt>true</tt> if the Job was added.]]>
</doc>
</method>
<method name="isCompleted" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return true if this job is in a complete state]]>
</doc>
</method>
<method name="isReady" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return true if this job is in READY state]]>
</doc>
</method>
<method name="killJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="submit"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Submit this job to mapred. The state becomes RUNNING if submission
is successful, FAILED otherwise.]]>
</doc>
</method>
<field name="CREATE_DIR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class encapsulates a MapReduce job and its dependency. It monitors
the states of the depending jobs and updates the state of this job.
A job starts in the WAITING state. If it does not have any depending jobs,
or all of the depending jobs are in SUCCESS state, then the job state
will become READY. If any depending jobs fail, the job will fail too.
When in READY state, the job can be submitted to Hadoop for execution, with
the state changing into RUNNING state. From RUNNING state, the job
can get into SUCCESS or FAILED state, depending
the status of the job execution.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob -->
<!-- start class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State -->
<class name="ControlledJob.State" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State -->
<!-- start class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl -->
<class name="JobControl" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Runnable"/>
<constructor name="JobControl" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a job control for a group of jobs.
@param groupName a name identifying this group]]>
</doc>
</constructor>
<method name="getWaitingJobList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the waiting state]]>
</doc>
</method>
<method name="getRunningJobList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the running state]]>
</doc>
</method>
<method name="getReadyJobsList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the ready state]]>
</doc>
</method>
<method name="getSuccessfulJobList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the success state]]>
</doc>
</method>
<method name="getFailedJobList" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="addJob" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="aJob" type="org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob"/>
<doc>
<![CDATA[Add a new job.
@param aJob the new job]]>
</doc>
</method>
<method name="addJobCollection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobs" type="java.util.Collection"/>
<doc>
<![CDATA[Add a collection of jobs
@param jobs]]>
</doc>
</method>
<method name="getThreadState" return="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the thread state]]>
</doc>
</method>
<method name="stop"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[set the thread state to STOPPING so that the
thread will stop when it wakes up.]]>
</doc>
</method>
<method name="suspend"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[suspend the running thread]]>
</doc>
</method>
<method name="resume"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[resume the suspended thread]]>
</doc>
</method>
<method name="allFinished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The main loop for the thread.
The loop does the following:
Check the states of the running jobs
Update the states of waiting jobs
Submit the jobs in ready state]]>
</doc>
</method>
<doc>
<![CDATA[This class encapsulates a set of MapReduce jobs and its dependency.
It tracks the states of the jobs by placing them into different tables
according to their states.
This class provides APIs for the client app to add a job to the group
and to get the jobs in the group in different states. When a job is
added, an ID unique to the group is assigned to the job.
This class has a thread that submits jobs when they become ready,
monitors the states of the running jobs, and updates the states of jobs
based on the state changes of their depending jobs states. The class
provides APIs for suspending/resuming the thread, and
for stopping the thread.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl -->
<!-- start class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState -->
<class name="JobControl.ThreadState" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.join">
<!-- start class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator -->
<class name="ArrayListBackedIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<constructor name="ArrayListBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ArrayListBackedIterator" type="java.util.ArrayList"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. The
implementation uses an {@link java.util.ArrayList} to store elements
added to it, replaying them as requested.
Prefer {@link StreamBackedIterator}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat -->
<class name="ComposableInputFormat" extends="org.apache.hadoop.mapreduce.InputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ComposableInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<doc>
<![CDATA[Refinement of InputFormat requiring implementors to provide
ComposableRecordReader instead of RecordReader.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader -->
<class name="ComposableRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<constructor name="ComposableRecordReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Additional operations required of a RecordReader to participate in a join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat -->
<class name="CompositeInputFormat" extends="org.apache.hadoop.mapreduce.InputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CompositeInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Interpret a given string as a composite expression.
{@code
func ::= <ident>([<func>,]*<func>)
func ::= tbl(<class>,"<path>")
class ::= @see java.lang.Class#forName(java.lang.String)
path ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
}
Reads expression from the <tt>mapreduce.join.expr</tt> property and
user-supplied join types from <tt>mapreduce.join.define.&lt;ident&gt;</tt>
types. Paths supplied to <tt>tbl</tt> are given as input paths to the
InputFormat class listed.
@see #compose(java.lang.String, java.lang.Class, java.lang.String...)]]>
</doc>
</method>
<method name="addDefaults"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Adds the default set of identifiers to the parser.]]>
</doc>
</method>
<method name="getSplits" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Build a CompositeInputSplit from the child InputFormats by assigning the
ith split from each child to the ith composite split.]]>
</doc>
</method>
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Construct a CompositeRecordReader for the children of this InputFormat
as defined in the init expression.
The outermost join need only be composable, not necessarily a composite.
Mandating TupleWritable isn't strictly correct.]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="java.lang.Class"/>
<param name="path" type="java.lang.String"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given InputFormat class (inf), path (p) return:
{@code tbl(<inf>, <p>) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class"/>
<param name="path" type="java.lang.String[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class"/>
<param name="path" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<field name="JOIN_EXPR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="JOIN_COMPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An InputFormat capable of performing joins over a set of data sources sorted
and partitioned the same way.
@see #setFormat
A user may define new join types by setting the property
<tt>mapreduce.join.define.&lt;ident&gt;</tt> to a classname.
In the expression <tt>mapreduce.join.expr</tt>, the identifier will be
assumed to be a ComposableRecordReader.
<tt>mapreduce.join.keycomparator</tt> can be a classname used to compare
keys in the join.
@see JoinRecordReader
@see MultiFilterRecordReader]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit -->
<class name="CompositeInputSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="CompositeInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CompositeInputSplit" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.mapreduce.InputSplit"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Add an InputSplit to this collection.
@throws IOException If capacity was not specified during construction
or if capacity has been reached.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.mapreduce.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Get ith child InputSplit.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the aggregate length of all child InputSplits currently added.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the length of ith child InputSplit.]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Collect a set of hosts from all child InputSplits.]]>
</doc>
</method>
<method name="getLocation" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[getLocations from ith InputSplit.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write splits in the following format.
{@code
<count><class1><class2>...<classn><split1><split2>...<splitn>
}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}
@throws IOException If the child InputSplit cannot be read, typically
for failing access checks.]]>
</doc>
</method>
<doc>
<![CDATA[This InputSplit contains a set of child InputSplits. Any InputSplit inserted
into this collection must have a public default constructor.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader -->
<class name="CompositeRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="CompositeRecordReader" type="int, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a RecordReader with <tt>capacity</tt> children to position
<tt>id</tt> in the parent reader.
The id of a root CompositeRecordReader is -1 by convention, but relying
on this is not recommended.]]>
</doc>
</constructor>
<method name="combine" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="value" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the position in the collector this class occupies.]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getRecordReaderQueue" return="java.util.PriorityQueue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return sorted list of RecordReaders for this composite.]]>
</doc>
</method>
<method name="getComparator" return="org.apache.hadoop.io.WritableComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return comparator defining the ordering for RecordReaders in this
composite.]]>
</doc>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rr" type="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Add a RecordReader to this collection.
The id() of a RecordReader determines where in the Tuple its
entry will appear. Adding RecordReaders with the same id has
undefined behavior.]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key for the current join or the value at the top of the
RecordReader heap.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the top of this RR into the given object.]]>
</doc>
</method>
<method name="getCurrentKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if it is possible that this could emit more values.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Pass skip key to child RRs.]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Obtain an iterator over the child RRs apropos of the value type
ultimately emitted from this join.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jc" type="org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[If key provided matches that of this Composite, give JoinCollector
iterator over values it may emit.]]>
</doc>
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[For all child RRs offering the key provided, obtain an iterator
at that position in the JoinCollector.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"/>
<doc>
<![CDATA[Implement Comparable contract (compare key of join or head of heap
with that of another).]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new key common to all child RRs.
@throws ClassCastException if key classes differ.]]>
</doc>
</method>
<method name="createTupleWritable" return="org.apache.hadoop.mapreduce.lib.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a value to be used internally for joins.]]>
</doc>
</method>
<method name="getCurrentValue" return="X"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close all child RRs.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Report progress as the minimum of all child RR progress.]]>
</doc>
</method>
<field name="conf" type="org.apache.hadoop.conf.Configuration"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="keyclass" type="java.lang.Class"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="jc" type="org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="kids" type="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader[]"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="key" type="K"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="value" type="X"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector -->
<class name="CompositeRecordReader.JoinCollector" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CompositeRecordReader.JoinCollector" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a collector capable of handling the specified number of
children.]]>
</doc>
</constructor>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="id" type="int"/>
<param name="i" type="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Register a given iterator at position id.]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key associated with this collection.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<doc>
<![CDATA[Codify the contents of the collector to be iterated over.
When this is called, all RecordReaders registered for this
key should have added ResetableIterators.]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Clear all state information.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns false if exhausted or if reset(K) has not been called.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Populate Tuple from iterators.
It should be the case that, given iterators i_1...i_n over values from
sources s_1...s_n sharing key k, repeated calls to next should yield
I x I.]]>
</doc>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Replay the last Tuple emitted.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close all child iterators.]]>
</doc>
</method>
<method name="flush" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the next value into key, value as accepted by the operation
associated with this set of RecordReaders.]]>
</doc>
</method>
<doc>
<![CDATA[Collector for join values.
This accumulates values for a given key from the child RecordReaders. If
one or more child RR contain duplicate keys, this will emit the cross
product of the associated values until exhausted.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.InnerJoinRecordReader -->
<class name="InnerJoinRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.JoinRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<doc>
<![CDATA[Return true iff the tuple is full (all data sources contain this key).]]>
</doc>
</method>
<doc>
<![CDATA[Full inner join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.InnerJoinRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader -->
<class name="JoinRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JoinRecordReader" type="int, org.apache.hadoop.conf.Configuration, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Emit the next set of key, value pairs as defined by the child
RecordReaders and operation associated with this composite RR.]]>
</doc>
</method>
<method name="createValue" return="org.apache.hadoop.mapreduce.lib.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDelegate" return="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator wrapping the JoinCollector.]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite joins returning Tuples of arbitrary Writables.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator -->
<class name="JoinRecordReader.JoinDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<constructor name="JoinRecordReader.JoinDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Since the JoinCollector is effecting our operation, we need only
provide an iterator proxy wrapping its operation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader -->
<class name="MultiFilterRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultiFilterRecordReader" type="int, org.apache.hadoop.conf.Configuration, int, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="emit" return="V"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For each tuple emitted, return a value (typically one of the values
in the tuple).
Modifying the Writables in the tuple is permitted and unlikely to affect
join behavior in most cases, but it is not recommended. It's safer to
clone first.]]>
</doc>
</method>
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<doc>
<![CDATA[Default implementation offers {@link #emit} every Tuple from the
collector (the outer join of child RRs).]]>
</doc>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator returning a single value from the tuple.
@see MultiFilterDelegationIterator]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite join returning values derived from multiple
sources, but generally not tuples.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<class name="MultiFilterRecordReader.MultiFilterDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<constructor name="MultiFilterRecordReader.MultiFilterDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Proxy the JoinCollector, but include callback to emit.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.OuterJoinRecordReader -->
<class name="OuterJoinRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.JoinRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<doc>
<![CDATA[Emit everything from the collector.]]>
</doc>
</method>
<doc>
<![CDATA[Full outer join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.OuterJoinRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader -->
<class name="OverrideRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="emit" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapreduce.lib.join.TupleWritable"/>
<doc>
<![CDATA[Emit the value with the highest position in the tuple.]]>
</doc>
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Instead of filling the JoinCollector with iterators from all
data sources, fill only the rightmost for this key.
This not only saves space by discarding the other sources, but
it also emits the number of key-value pairs in the preferred
RecordReader instead of repeating that stream n times, where
n is the cardinality of the cross product of the discarded
streams for the given key.]]>
</doc>
</method>
<doc>
<![CDATA[Prefer the &quot;rightmost&quot; data source for this key.
For example, <tt>override(S1,S2,S3)</tt> will prefer values
from S3 over S2, and values from S2 over S1 for all keys
emitted from all sources.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser -->
<class name="Parser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Very simple shift-reduce parser for join expressions.
This should be sufficient for the user extension permitted now, but ought to
be replaced with a parser generator if more complex grammars are supported.
In particular, this &quot;shift-reduce&quot; parser has no states. Each set
of formals requires a different internal node type, which is responsible for
interpreting the list of tokens it receives. This is sufficient for the
current grammar, but it has several annoying properties that might inhibit
extension. In particular, parenthesis are always function calls; an
algebraic or filter grammar would not only require a node type, but must
also work around the internals of this parser.
For most other cases, adding classes to the hierarchy- particularly by
extending JoinRecordReader and MultiFilterRecordReader- is fairly
straightforward. One need only override the relevant method(s) (usually only
{@link CompositeRecordReader#combine}) and include a property to map its
value to an identifier in the parser.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.Node -->
<class name="Parser.Node" extends="org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.Node" type="java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="addIdentifier"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="ident" type="java.lang.String"/>
<param name="mcstrSig" type="java.lang.Class[]"/>
<param name="nodetype" type="java.lang.Class"/>
<param name="cl" type="java.lang.Class"/>
<exception name="NoSuchMethodException" type="java.lang.NoSuchMethodException"/>
<doc>
<![CDATA[For a given identifier, add a mapping to the nodetype for the parse
tree and to the ComposableRecordReader to be created, including the
formals required to invoke the constructor.
The nodetype and constructor signature should be filled in from the
child node.]]>
</doc>
</method>
<method name="setID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="id" type="int"/>
</method>
<method name="setKeyComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="cmpcl" type="java.lang.Class"/>
</method>
<field name="rrCstrMap" type="java.util.Map"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="id" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="ident" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="cmpcl" type="java.lang.Class"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.Node -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.NodeToken -->
<class name="Parser.NodeToken" extends="org.apache.hadoop.mapreduce.lib.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getNode" return="org.apache.hadoop.mapreduce.lib.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.NodeToken -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.NumToken -->
<class name="Parser.NumToken" extends="org.apache.hadoop.mapreduce.lib.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.NumToken" type="double"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.NumToken -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.StrToken -->
<class name="Parser.StrToken" extends="org.apache.hadoop.mapreduce.lib.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.StrToken" type="org.apache.hadoop.mapreduce.lib.join.Parser.TType, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.StrToken -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.Token -->
<class name="Parser.Token" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getType" return="org.apache.hadoop.mapreduce.lib.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNode" return="org.apache.hadoop.mapreduce.lib.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Tagged-union type for tokens from the join expression.
@see Parser.TType]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.Token -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.Parser.TType -->
<class name="Parser.TType" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapreduce.lib.join.Parser.TType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.lib.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.Parser.TType -->
<!-- start interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator -->
<interface name="ResetableIterator" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[True if a call to next may return a value. This is permitted false
positives, but not false negatives.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Assign next value to actual.
It is required that elements added to a ResetableIterator be returned in
the same order after a call to {@link #reset} (FIFO).
Note that a call to this may fail for nested joins (i.e. more elements
available, but none satisfying the constraints of the join)]]>
</doc>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Assign last value returned to actual.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Set iterator to return to the start of its range. Must be called after
calling {@link #add} to avoid a ConcurrentModificationException.]]>
</doc>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an element to the collection of elements to iterate over.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close datasources and release resources. Calling methods on the iterator
after calling close has undefined behavior.]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Close datasources, but do not release internal resources. Calling this
method should permit the object to be reused with a different datasource.]]>
</doc>
</method>
<doc>
<![CDATA[This defines an interface to a stateful Iterator that can replay elements
added to it directly.
Note that this does not extend {@link java.util.Iterator}.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY -->
<class name="ResetableIterator.EMPTY" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<constructor name="ResetableIterator.EMPTY"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="U"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="U"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="U"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator -->
<class name="StreamBackedIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.join.ResetableIterator"/>
<constructor name="StreamBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="X"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. This
implementation uses a byte array to store elements added to it.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.TupleWritable -->
<class name="TupleWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<constructor name="TupleWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an empty tuple with no allocated storage for writables.]]>
</doc>
</constructor>
<constructor name="TupleWritable" type="org.apache.hadoop.io.Writable[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Initialize tuple with storage; unknown whether any of them contain
&quot;written&quot; values.]]>
</doc>
</constructor>
<method name="has" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Return true if tuple has an element at the position provided.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Get ith Writable from Tuple.]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of children in this Tuple.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator over the elements in this tuple.
Note that this doesn't flatten the tuple; one may receive tuples
from this iterator.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert Tuple to String as in the following.
<tt>[<child1>,<child2>,...,<childn>]</tt>]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes each Writable to <code>out</code>.
TupleWritable format:
{@code
<count><type1><type2>...<typen><obj1><obj2>...<objn>
}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<field name="written" type="java.util.BitSet"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Writable type storing multiple {@link org.apache.hadoop.io.Writable}s.
This is *not* a general-purpose tuple type. In almost all cases, users are
encouraged to implement their own serializable types, which can perform
better validation and provide more efficient encodings than this class is
capable. TupleWritable relies on the join framework for type safety and
assumes its instances will rarely be persisted, assumptions not only
incompatible with, but contrary to the general case.
@see org.apache.hadoop.io.Writable]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.TupleWritable -->
<!-- start class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader -->
<class name="WrappedRecordReader" extends="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="WrappedRecordReader" type="int"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request new key from proxied RR.]]>
</doc>
</method>
<method name="createValue" return="U"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="key" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key at the head of this RR.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="qkey" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the head of this RR into the object supplied.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if the RR- including the k,v pair stored in this object-
is exhausted.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Add an iterator to the collector at the position occupied by this
RecordReader over the values in this stream paired with the key
provided (ie register a stream of values from this source matching K
with a collector).]]>
</doc>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Read the next k,v pair into the head of this object; return true iff
the RR and this are exhausted.]]>
</doc>
</method>
<method name="getCurrentKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get current key]]>
</doc>
</method>
<method name="getCurrentValue" return="U"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get current value]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Request progress from proxied RR.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Forward close request to proxied RR.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader"/>
<doc>
<![CDATA[Implement Comparable contract (compare key at head of proxied RR
with that of another).]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[Return true iff compareTo(other) retn true.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="empty" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="cmp" type="org.apache.hadoop.io.WritableComparator"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Proxy class for a RecordReader participating in the join framework.
This class keeps track of the &quot;head&quot; key-value pair for the
provided RecordReader and keeps a store of values matching a key when
this source is participating in a join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.map">
<!-- start class org.apache.hadoop.mapreduce.lib.map.InverseMapper -->
<class name="InverseMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InverseMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[The inverse function. Input keys and values are swapped.]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Mapper} that swaps keys and values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.InverseMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper -->
<class name="MultithreadedMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultithreadedMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getNumberOfThreads" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[The number of threads in the thread pool that will run the map function.
@param job the job
@return the number of threads]]>
</doc>
</method>
<method name="setNumberOfThreads"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="threads" type="int"/>
<doc>
<![CDATA[Set the number of threads in the pool for running maps.
@param job the job to modify
@param threads the new number of threads]]>
</doc>
</method>
<method name="getMapperClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the application's mapper class.
@param <K1> the map's input key type
@param <V1> the map's input value type
@param <K2> the map's output key type
@param <V2> the map's output value type
@param job the job
@return the mapper class to run]]>
</doc>
</method>
<method name="setMapperClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="cls" type="java.lang.Class"/>
<doc>
<![CDATA[Set the application's mapper class.
@param <K1> the map input key type
@param <V1> the map input value type
@param <K2> the map output key type
@param <V2> the map output value type
@param job the job to modify
@param cls the class to use as the mapper]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Run the application's maps using a thread pool.]]>
</doc>
</method>
<field name="NUM_THREADS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAP_CLASS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Multithreaded implementation for @link org.apache.hadoop.mapreduce.Mapper.
<p>
It can be used instead of the default implementation,
@link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU
bound in order to improve throughput.
<p>
Mapper implementations using this MapRunnable must be thread-safe.
<p>
The Map-Reduce job has to be configured with the mapper to use via
{@link #setMapperClass(Configuration, Class)} and
the number of thread the thread-pool can use with the
{@link #getNumberOfThreads(Configuration) method. The default
value is 10 threads.
<p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.map.RegexMapper -->
<class name="RegexMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RegexMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="PATTERN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="GROUP" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.RegexMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper -->
<class name="TokenCounterMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TokenCounterMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<doc>
<![CDATA[Tokenize the input values and emit each word with a count of 1.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.map.WrappedMapper -->
<class name="WrappedMapper" extends="org.apache.hadoop.mapreduce.Mapper"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="WrappedMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getMapContext" return="org.apache.hadoop.mapreduce.Mapper.Context"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mapContext" type="org.apache.hadoop.mapreduce.MapContext"/>
<doc>
<![CDATA[Get a wrapped {@link Mapper.Context} for custom implementations.
@param mapContext <code>MapContext</code> to be wrapped
@return a wrapped <code>Mapper.Context</code> for custom implementations]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Mapper} which wraps a given one to allow custom
{@link Mapper.Context} implementations.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.WrappedMapper -->
<!-- start class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context -->
<class name="WrappedMapper.Context" extends="org.apache.hadoop.mapreduce.Mapper.Context"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="WrappedMapper.Context" type="org.apache.hadoop.mapreduce.MapContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getInputSplit" return="org.apache.hadoop.mapreduce.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the input split for this map.]]>
</doc>
</method>
<method name="getCurrentKey" return="KEYIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentValue" return="VALUEIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.Enum"/>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEYOUT"/>
<param name="value" type="VALUEOUT"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getStatus" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskAttemptID" return="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="msg" type="java.lang.String"/>
</method>
<method name="getArchiveClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getArchiveTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCacheArchives" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCacheFiles" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCombinerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroupingComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobSetupCleanupNeeded" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLocalCacheArchives" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocalCacheFiles" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapperClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getMaxMapAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMaxReduceAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPartitionerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getReducerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getSortComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSymlink" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="progress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProfileEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProfileParams" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCredentials" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="mapContext" type="org.apache.hadoop.mapreduce.MapContext"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.output">
<!-- start class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter -->
<class name="FileOutputCommitter" extends="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileOutputCommitter" type="org.apache.hadoop.fs.Path, org.apache.hadoop.mapreduce.TaskAttemptContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a file output committer
@param outputPath the job's output path
@param context the task's context
@throws IOException]]>
</doc>
</constructor>
<method name="setupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the temporary directory that is the root of all of the task
work directories.
@param context the job's context]]>
</doc>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete the temporary directory, including all of the work directories.
Create a _SUCCESS file to make it as successful.
@param context the job's context]]>
</doc>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="state" type="org.apache.hadoop.mapreduce.JobStatus.State"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete the temporary directory, including all of the work directories.
@param context the job's context]]>
</doc>
</method>
<method name="setupTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[No task setup required.]]>
</doc>
</method>
<method name="commitTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Move the files from the work directory to the job output directory
@param context the task context]]>
</doc>
</method>
<method name="abortTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete the work directory
@throws IOException]]>
</doc>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Did this task write any files in the work directory?
@param context the task's context]]>
</doc>
</method>
<method name="getWorkPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the directory that the task should write results into
@return the work directory
@throws IOException]]>
</doc>
</method>
<field name="TEMP_DIR_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Temporary directory name]]>
</doc>
</field>
<field name="SUCCEEDED_FILE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link OutputCommitter} that commits files specified
in job output directory i.e. ${mapreduce.output.fileoutputformat.outputdir}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat -->
<class name="FileOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setCompressOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Set whether the output of the job is compressed.
@param job the job to modify
@param compress should the output of the job be compressed?]]>
</doc>
</method>
<method name="getCompressOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Is the job output compressed?
@param job the Job to look in
@return <code>true</code> if the job output should be compressed,
<code>false</code> otherwise]]>
</doc>
</method>
<method name="setOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="codecClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
@param job the job to modify
@param codecClass the {@link CompressionCodec} to be used to
compress the job outputs]]>
</doc>
</method>
<method name="getOutputCompressorClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="defaultValue" type="java.lang.Class"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
@param job the {@link Job} to look in
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} to be used to compress the
job outputs
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.fs.FileAlreadyExistsException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOutputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
@param job The job to modify
@param outputDir the {@link Path} of the output directory for
the map-reduce job.]]>
</doc>
</method>
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
@return the {@link Path} to the output directory for the map-reduce job.
@see FileOutputFormat#getWorkOutputPath(TaskInputOutputContext)]]>
</doc>
</method>
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get the {@link Path} to the task's temporary output directory
for the map-reduce job
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
<p>Some applications need to create/write-to side-files, which differ from
the actual job-outputs.
<p>In such cases there could be issues with 2 instances of the same TIP
(running simultaneously e.g. speculative tasks) trying to open/write-to the
same file (path) on HDFS. Hence the application-writer will have to pick
unique names per task-attempt (e.g. using the attemptid, say
<tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
<p>To get around this the Map-Reduce framework helps the application-writer
out by maintaining a special
<tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt>
sub-directory for each task-attempt on HDFS where the output of the
task-attempt goes. On successful completion of the task-attempt the files
in the <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> (only)
are <i>promoted</i> to <tt>${mapreduce.output.fileoutputformat.outputdir}</tt>. Of course, the
framework discards the sub-directory of unsuccessful task-attempts. This
is completely transparent to the application.</p>
<p>The application-writer can take advantage of this by creating any
side-files required in a work directory during execution
of his task i.e. via
{@link #getWorkOutputPath(TaskInputOutputContext)}, and
the framework will move them out similarly - thus she doesn't have to pick
unique paths per task-attempt.</p>
<p>The entire discussion holds true for maps of jobs with
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
goes directly to HDFS.</p>
@return the {@link Path} to the task's temporary output directory
for the map-reduce job.]]>
</doc>
</method>
<method name="getPathForWorkFile" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
<param name="name" type="java.lang.String"/>
<param name="extension" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Helper function to generate a {@link Path} for a file that is unique for
the task within the job output directory.
<p>The path can be used to create custom files from within the map and
reduce tasks. The path name will be unique for each task. The path parent
will be the job output directory.</p>ls
<p>This method uses the {@link #getUniqueFile} method to make the file name
unique for the task.</p>
@param context the context for the task.
@param name the name for the file.
@param extension the extension for the file
@return a unique path accross all tasks of the job.]]>
</doc>
</method>
<method name="getUniqueFile" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<param name="name" type="java.lang.String"/>
<param name="extension" type="java.lang.String"/>
<doc>
<![CDATA[Generate a unique filename, based on the task id, name, and extension
@param context the task that is calling this
@param name the base filename
@param extension the filename extension
@return a string like $name-[mrsct]-$id$extension]]>
</doc>
</method>
<method name="getDefaultWorkFile" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<param name="extension" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the default path and filename for the output format.
@param context the task context
@param extension an extension to add to the filename
@return a full path $output/_temporary/$taskid/part-[mr]-$id
@throws IOException]]>
</doc>
</method>
<method name="getOutputName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the base output name for the output file.]]>
</doc>
</method>
<method name="setOutputName"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Set the base output name for output file to be created.]]>
</doc>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="BASE_OUTPUT_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="PART" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="COMPRESS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMPRESS_CODEC" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMPRESS_TYPE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OUTDIR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class for {@link OutputFormat}s that read from {@link FileSystem}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat -->
<class name="FilterOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FilterOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FilterOutputFormat" type="org.apache.hadoop.mapreduce.OutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a FilterOutputFormat based on the underlying output format.
@param baseOut the underlying OutputFormat]]>
</doc>
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="baseOut" type="org.apache.hadoop.mapreduce.OutputFormat"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[FilterOutputFormat is a convenience class that wraps OutputFormat.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter -->
<class name="FilterOutputFormat.FilterRecordWriter" extends="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FilterOutputFormat.FilterRecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FilterOutputFormat.FilterRecordWriter" type="org.apache.hadoop.mapreduce.RecordWriter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="rawWriter" type="org.apache.hadoop.mapreduce.RecordWriter"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[<code>FilterRecordWriter</code> is a convenience wrapper
class that extends the {@link RecordWriter}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat -->
<class name="LazyOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LazyOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setOutputFormatClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the underlying output format for LazyOutputFormat.
@param job the {@link Job} to modify
@param theClass the underlying class]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="OUTPUT_FORMAT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A Convenience class that creates output lazily.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat -->
<class name="MapFileOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReaders" return="org.apache.hadoop.io.MapFile.Reader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the output generated by this format.]]>
</doc>
</method>
<method name="getEntry" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="readers" type="org.apache.hadoop.io.MapFile.Reader[]"/>
<param name="partitioner" type="org.apache.hadoop.mapreduce.Partitioner"/>
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an entry from output generated by this class.]]>
</doc>
</method>
<doc>
<![CDATA[An {@link org.apache.hadoop.mapreduce.OutputFormat} that writes
{@link MapFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs -->
<class name="MultipleOutputs" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultipleOutputs" type="org.apache.hadoop.mapreduce.TaskInputOutputContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates and initializes multiple outputs support,
it should be instantiated in the Mapper/Reducer setup method.
@param context the TaskInputOutputContext object]]>
</doc>
</constructor>
<method name="addNamedOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="namedOutput" type="java.lang.String"/>
<param name="outputFormatClass" type="java.lang.Class"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valueClass" type="java.lang.Class"/>
<doc>
<![CDATA[Adds a named output for the job.
<p/>
@param job job to add the named output
@param namedOutput named output name, it has to be a word, letters
and numbers only, cannot be the word 'part' as
that is reserved for the default output.
@param outputFormatClass OutputFormat class.
@param keyClass key class
@param valueClass value class]]>
</doc>
</method>
<method name="setCountersEnabled"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="enabled" type="boolean"/>
<doc>
<![CDATA[Enables or disables counters for the named outputs.
The counters group is the {@link MultipleOutputs} class name.
The names of the counters are the same as the named outputs. These
counters count the number records written to each output name.
By default these counters are disabled.
@param job job to enable counters
@param enabled indicates if the counters will be enabled or not.]]>
</doc>
</method>
<method name="getCountersEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Returns if the counters for the named outputs are enabled or not.
By default these counters are disabled.
@param job the job
@return TRUE if the counters are enabled, FALSE if they are disabled.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="namedOutput" type="java.lang.String"/>
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Write key and value to the namedOutput.
Output path is a unique file generated for the namedOutput.
For example, {namedOutput}-(m|r)-{part-number}
@param namedOutput the named output name
@param key the key
@param value the value]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="namedOutput" type="java.lang.String"/>
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="baseOutputPath" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Write key and value to baseOutputPath using the namedOutput.
@param namedOutput the named output name
@param key the key
@param value the value
@param baseOutputPath base-output path to write the record to.
Note: Framework will generate unique filename for the baseOutputPath]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEYOUT"/>
<param name="value" type="VALUEOUT"/>
<param name="baseOutputPath" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Write key value to an output file name.
Gets the record writer from job's output format.
Job's output format should be a FileOutputFormat.
@param key the key
@param value the value
@param baseOutputPath base-output path to write the record to.
Note: Framework will generate unique filename for the baseOutputPath]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Closes all the opened outputs.
This should be called from cleanup method of map/reduce task.
If overridden subclasses must invoke <code>super.close()</code> at the
end of their <code>close()</code>]]>
</doc>
</method>
<doc>
<![CDATA[The MultipleOutputs class simplifies writing output data
to multiple outputs
<p>
Case one: writing to additional outputs other than the job default output.
Each additional output, or named output, may be configured with its own
<code>OutputFormat</code>, with its own key class and with its own value
class.
<p>
Case two: to write data to different files provided by user
</p>
<p>
MultipleOutputs supports counters, by default they are disabled. The
counters group is the {@link MultipleOutputs} class name. The names of the
counters are the same as the output name. These count the number records
written to each output name.
</p>
Usage pattern for job submission:
<pre>
Job job = new Job();
FileInputFormat.setInputPath(job, inDir);
FileOutputFormat.setOutputPath(job, outDir);
job.setMapperClass(MOMap.class);
job.setReducerClass(MOReduce.class);
...
// Defines additional single text based output 'text' for the job
MultipleOutputs.addNamedOutput(job, "text", TextOutputFormat.class,
LongWritable.class, Text.class);
// Defines additional sequence-file based output 'sequence' for the job
MultipleOutputs.addNamedOutput(job, "seq",
SequenceFileOutputFormat.class,
LongWritable.class, Text.class);
...
job.waitForCompletion(true);
...
</pre>
<p>
Usage in Reducer:
<pre>
<K, V> String generateFileName(K k, V v) {
return k.toString() + "_" + v.toString();
}
public class MOReduce extends
Reducer&lt;WritableComparable, Writable,WritableComparable, Writable&gt; {
private MultipleOutputs mos;
public void setup(Context context) {
...
mos = new MultipleOutputs(context);
}
public void reduce(WritableComparable key, Iterator&lt;Writable&gt; values,
Context context)
throws IOException {
...
mos.write("text", , key, new Text("Hello"));
mos.write("seq", LongWritable(1), new Text("Bye"), "seq_a");
mos.write("seq", LongWritable(2), key, new Text("Chau"), "seq_b");
mos.write(key, new Text("value"), generateFileName(key, new Text("value")));
...
}
public void cleanup(Context) throws IOException {
mos.close();
...
}
}
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat -->
<class name="NullOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NullOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
</method>
<doc>
<![CDATA[Consume all outputs and put them in /dev/null.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat -->
<class name="SequenceFileAsBinaryOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsBinaryOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setSequenceFileOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the {@link SequenceFile}
<p>This allows the user to specify the key class to be different
from the actual class ({@link BytesWritable}) used for writing </p>
@param job the {@link Job} to modify
@param theClass the SequenceFile output key class.]]>
</doc>
</method>
<method name="setSequenceFileOutputValueClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the value class for the {@link SequenceFile}
<p>This allows the user to specify the value class to be different
from the actual class ({@link BytesWritable}) used for writing </p>
@param job the {@link Job} to modify
@param theClass the SequenceFile output key class.]]>
</doc>
</method>
<method name="getSequenceFileOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the key class for the {@link SequenceFile}
@return the key class of the {@link SequenceFile}]]>
</doc>
</method>
<method name="getSequenceFileOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the value class for the {@link SequenceFile}
@return the value class of the {@link SequenceFile}]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="KEY_CLASS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="VALUE_CLASS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link org.apache.hadoop.mapreduce.OutputFormat} that writes keys,
values to {@link SequenceFile}s in binary(raw) format]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
<class name="SequenceFileAsBinaryOutputFormat.WritableValueBytes" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
<constructor name="SequenceFileAsBinaryOutputFormat.WritableValueBytes"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SequenceFileAsBinaryOutputFormat.WritableValueBytes" type="org.apache.hadoop.io.BytesWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="org.apache.hadoop.io.BytesWritable"/>
</method>
<method name="writeUncompressedBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="outStream" type="java.io.DataOutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeCompressedBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="outStream" type="java.io.DataOutputStream"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Inner class used for appendRaw]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat -->
<class name="SequenceFileOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSequenceWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valueClass" type="java.lang.Class"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the {@link CompressionType} for the output {@link SequenceFile}.
@param job the {@link Job}
@return the {@link CompressionType} for the output {@link SequenceFile},
defaulting to {@link CompressionType#RECORD}]]>
</doc>
</method>
<method name="setOutputCompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<doc>
<![CDATA[Set the {@link CompressionType} for the output {@link SequenceFile}.
@param job the {@link Job} to modify
@param style the {@link CompressionType} for the output
{@link SequenceFile}]]>
</doc>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat -->
<class name="TextOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FileOutputFormat"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TextOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<field name="SEPERATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link OutputFormat} that writes plain text files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat -->
<!-- start class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter -->
<class name="TextOutputFormat.LineRecordWriter" extends="org.apache.hadoop.mapreduce.RecordWriter"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="out" type="java.io.DataOutputStream"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.partition">
<!-- start class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner -->
<class name="BinaryPartitioner" extends="org.apache.hadoop.mapreduce.Partitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="BinaryPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setOffsets"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="left" type="int"/>
<param name="right" type="int"/>
<doc>
<![CDATA[Set the subarray to be used for partitioning to
<code>bytes[left:(right+1)]</code> in Python syntax.
@param conf configuration object
@param left left Python-style offset
@param right right Python-style offset]]>
</doc>
</method>
<method name="setLeftOffset"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="offset" type="int"/>
<doc>
<![CDATA[Set the subarray to be used for partitioning to
<code>bytes[offset:]</code> in Python syntax.
@param conf configuration object
@param offset left Python-style offset]]>
</doc>
</method>
<method name="setRightOffset"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="offset" type="int"/>
<doc>
<![CDATA[Set the subarray to be used for partitioning to
<code>bytes[:(offset+1)]</code> in Python syntax.
@param conf configuration object
@param offset right Python-style offset]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.BinaryComparable"/>
<param name="value" type="V"/>
<param name="numPartitions" type="int"/>
<doc>
<![CDATA[Use (the specified slice of the array returned by)
{@link BinaryComparable#getBytes()} to partition.]]>
</doc>
</method>
<field name="LEFT_OFFSET_PROPERTY_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RIGHT_OFFSET_PROPERTY_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[<p>Partition {@link BinaryComparable} keys using a configurable part of
the bytes array returned by {@link BinaryComparable#getBytes()}.</p>
<p>The subarray to be used for the partitioning can be defined by means
of the following properties:
<ul>
<li>
<i>mapreduce.partition.binarypartitioner.left.offset</i>:
left offset in array (0 by default)
</li>
<li>
<i>mapreduce.partition.binarypartitioner.right.offset</i>:
right offset in array (-1 by default)
</li>
</ul>
Like in Python, both negative and positive offsets are allowed, but
the meaning is slightly different. In case of an array of length 5,
for instance, the possible offsets are:
<pre><code>
+---+---+---+---+---+
| B | B | B | B | B |
+---+---+---+---+---+
0 1 2 3 4
-5 -4 -3 -2 -1
</code></pre>
The first row of numbers gives the position of the offsets 0...5 in
the array; the second row gives the corresponding negative offsets.
Contrary to Python, the specified subarray has byte <code>i</code>
and <code>j</code> as first and last element, repectively, when
<code>i</code> and <code>j</code> are the left and right offset.
<p>For Hadoop programs written in Java, it is advisable to use one of
the following static convenience methods for setting the offsets:
<ul>
<li>{@link #setOffsets}</li>
<li>{@link #setLeftOffset}</li>
<li>{@link #setRightOffset}</li>
</ul></p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner -->
<class name="HashPartitioner" extends="org.apache.hadoop.mapreduce.Partitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="HashPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="numReduceTasks" type="int"/>
<doc>
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
</doc>
</method>
<doc>
<![CDATA[Partition keys by their {@link Object#hashCode()}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.InputSampler -->
<class name="InputSampler" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="InputSampler" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="writePartitionFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="sampler" type="org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Write a partition file for the given job, using the Sampler provided.
Queries the sampler for a sample keyset, sorts by the output key
comparator, selects the keys for each rank, and writes to the destination
returned from {@link TotalOrderPartitioner#getPartitionFile}.]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Driver for InputSampler from the command line.
Configures a JobConf instance and calls {@link #writePartitionFile}.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[Utility for collecting samples and writing a partition file for
{@link TotalOrderPartitioner}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.InputSampler -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler -->
<class name="InputSampler.IntervalSampler" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler"/>
<constructor name="InputSampler.IntervalSampler" type="double"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new IntervalSampler sampling <em>all</em> splits.
@param freq The frequency with which records will be emitted.]]>
</doc>
</constructor>
<constructor name="InputSampler.IntervalSampler" type="double, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new IntervalSampler.
@param freq The frequency with which records will be emitted.
@param maxSplitsSampled The maximum number of splits to examine.
@see #getSample]]>
</doc>
</constructor>
<method name="getSample" return="K[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="org.apache.hadoop.mapreduce.InputFormat"/>
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[For each split sampled, emit when the ratio of the number of records
retained to the total record count is less than the specified
frequency.]]>
</doc>
</method>
<doc>
<![CDATA[Sample from s splits at regular intervals.
Useful for sorted data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler -->
<class name="InputSampler.RandomSampler" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler"/>
<constructor name="InputSampler.RandomSampler" type="double, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new RandomSampler sampling <em>all</em> splits.
This will read every split at the client, which is very expensive.
@param freq Probability with which a key will be chosen.
@param numSamples Total number of samples to obtain from all selected
splits.]]>
</doc>
</constructor>
<constructor name="InputSampler.RandomSampler" type="double, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new RandomSampler.
@param freq Probability with which a key will be chosen.
@param numSamples Total number of samples to obtain from all selected
splits.
@param maxSplitsSampled The maximum number of splits to examine.]]>
</doc>
</constructor>
<method name="getSample" return="K[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="org.apache.hadoop.mapreduce.InputFormat"/>
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Randomize the split order, then take the specified number of keys from
each split sampled, where each key is selected with the specified
probability and possibly replaced by a subsequently selected key when
the quota of keys from that split is satisfied.]]>
</doc>
</method>
<doc>
<![CDATA[Sample from random points in the input.
General-purpose sampler. Takes numSamples / maxSplitsSampled inputs from
each split.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler -->
<!-- start interface org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler -->
<interface name="InputSampler.Sampler" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getSample" return="K[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="org.apache.hadoop.mapreduce.InputFormat"/>
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[For a given job, collect and return a subset of the keys from the
input data.]]>
</doc>
</method>
<doc>
<![CDATA[Interface to sample using an
{@link org.apache.hadoop.mapreduce.InputFormat}.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler -->
<class name="InputSampler.SplitSampler" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler"/>
<constructor name="InputSampler.SplitSampler" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a SplitSampler sampling <em>all</em> splits.
Takes the first numSamples / numSplits records from each split.
@param numSamples Total number of samples to obtain from all selected
splits.]]>
</doc>
</constructor>
<constructor name="InputSampler.SplitSampler" type="int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new SplitSampler.
@param numSamples Total number of samples to obtain from all selected
splits.
@param maxSplitsSampled The maximum number of splits to examine.]]>
</doc>
</constructor>
<method name="getSample" return="K[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="org.apache.hadoop.mapreduce.InputFormat"/>
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[From each split sampled, take the first numSamples / numSplits records.]]>
</doc>
</method>
<doc>
<![CDATA[Samples the first n records from s splits.
Inexpensive way to sample random data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator -->
<class name="KeyFieldBasedComparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="KeyFieldBasedComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<method name="setKeyFieldComparatorOptions"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="keySpec" type="java.lang.String"/>
<doc>
<![CDATA[Set the {@link KeyFieldBasedComparator} options used to compare keys.
@param keySpec the key specification of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field). opts are ordering options. The supported options
are:
-n, (Sort numerically)
-r, (Reverse the result of comparison)]]>
</doc>
</method>
<method name="getKeyFieldComparatorOption" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the {@link KeyFieldBasedComparator} options]]>
</doc>
</method>
<field name="COMPARATOR_OPTIONS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This comparator implementation provides a subset of the features provided
by the Unix/GNU Sort. In particular, the supported features are:
-n, (Sort numerically)
-r, (Reverse the result of comparison)
-k pos1[,pos2], where pos is of the form f[.c][opts], where f is the number
of the field to use, and c is the number of the first character from the
beginning of the field. Fields and character posns are numbered starting
with 1; a character position of zero in pos2 indicates the field's last
character. If '.c' is omitted from pos1, it defaults to 1 (the beginning
of the field); if omitted from pos2, it defaults to 0 (the end of the
field). opts are ordering options (any of 'nr' as described above).
We assume that the fields in the key are separated by
{@link JobContext#MAP_OUTPUT_KEY_FIELD_SEPERATOR}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner -->
<class name="KeyFieldBasedPartitioner" extends="org.apache.hadoop.mapreduce.Partitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="KeyFieldBasedPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numReduceTasks" type="int"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="start" type="int"/>
<param name="end" type="int"/>
<param name="currentHash" type="int"/>
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="hash" type="int"/>
<param name="numReduceTasks" type="int"/>
</method>
<method name="setKeyFieldPartitionerOptions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="keySpec" type="java.lang.String"/>
<doc>
<![CDATA[Set the {@link KeyFieldBasedPartitioner} options used for
{@link Partitioner}
@param keySpec the key specification of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field).]]>
</doc>
</method>
<method name="getKeyFieldPartitionerOption" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
<doc>
<![CDATA[Get the {@link KeyFieldBasedPartitioner} options]]>
</doc>
</method>
<field name="PARTITIONER_OPTIONS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Defines a way to partition keys based on certain key fields (also see
{@link KeyFieldBasedComparator}.
The key specification supported is of the form -k pos1[,pos2], where,
pos is of the form f[.c][opts], where f is the number
of the key field to use, and c is the number of the first character from
the beginning of the field. Fields and character posns are numbered
starting with 1; a character position of zero in pos2 indicates the
field's last character. If '.c' is omitted from pos1, it defaults to 1
(the beginning of the field); if omitted from pos2, it defaults to 0
(the end of the field).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner -->
<!-- start class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner -->
<class name="TotalOrderPartitioner" extends="org.apache.hadoop.mapreduce.Partitioner"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="TotalOrderPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Read in the partition file and build indexing data structures.
If the keytype is {@link org.apache.hadoop.io.BinaryComparable} and
<tt>total.order.partitioner.natural.order</tt> is not false, a trie
of the first <tt>total.order.partitioner.max.trie.depth</tt>(2) + 1 bytes
will be built. Otherwise, keys will be located using a binary search of
the partition keyset using the {@link org.apache.hadoop.io.RawComparator}
defined for this job. The input file must be sorted with the same
comparator and contain {@link Job#getNumReduceTasks()} - 1 keys.]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="numPartitions" type="int"/>
</method>
<method name="setPartitionFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="p" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the path to the SequenceFile storing the sorted partition keyset.
It must be the case that for <tt>R</tt> reduces, there are <tt>R-1</tt>
keys in the SequenceFile.]]>
</doc>
</method>
<method name="getPartitionFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the path to the SequenceFile storing the sorted partition keyset.
@see #setPartitionFile(Configuration, Path)]]>
</doc>
</method>
<field name="DEFAULT_PATH" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PARTITIONER_PATH" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAX_TRIE_DEPTH" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NATURAL_ORDER" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Partitioner effecting a total order by reading split points from
an externally generated source.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner -->
</package>
<package name="org.apache.hadoop.mapreduce.lib.reduce">
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer -->
<class name="IntSumReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IntSumReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="Key"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer -->
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer -->
<class name="LongSumReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LongSumReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEY"/>
<param name="values" type="java.lang.Iterable"/>
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer -->
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer -->
<class name="WrappedReducer" extends="org.apache.hadoop.mapreduce.Reducer"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="WrappedReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getReducerContext" return="org.apache.hadoop.mapreduce.Reducer.Context"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reduceContext" type="org.apache.hadoop.mapreduce.ReduceContext"/>
<doc>
<![CDATA[A a wrapped {@link Reducer.Context} for custom implementations.
@param reduceContext <code>ReduceContext</code> to be wrapped
@return a wrapped <code>Reducer.Context</code> for custom implementations]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Reducer} which wraps a given one to allow for custom
{@link Reducer.Context} implementations.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer -->
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context -->
<class name="WrappedReducer.Context" extends="org.apache.hadoop.mapreduce.Reducer.Context"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="WrappedReducer.Context" type="org.apache.hadoop.mapreduce.ReduceContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getCurrentKey" return="KEYIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCurrentValue" return="VALUEIN"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKeyValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.Enum"/>
</method>
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="KEYOUT"/>
<param name="value" type="VALUEOUT"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getStatus" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskAttemptID" return="org.apache.hadoop.mapreduce.TaskAttemptID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="msg" type="java.lang.String"/>
</method>
<method name="getArchiveClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getArchiveTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCacheArchives" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCacheFiles" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCombinerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroupingComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobSetupCleanupNeeded" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLocalCacheArchives" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocalCacheFiles" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapperClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getMaxMapAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMaxReduceAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputFormatClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPartitionerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getReducerClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
</method>
<method name="getSortComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSymlink" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="progress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getValues" return="java.lang.Iterable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="nextKey" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="getProfileEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProfileParams" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCredentials" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="reduceContext" type="org.apache.hadoop.mapreduce.ReduceContext"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context -->
</package>
<package name="org.apache.hadoop.mapreduce.security">
<!-- start class org.apache.hadoop.mapreduce.security.TokenCache -->
<class name="TokenCache" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TokenCache"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSecretKey" return="byte[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="credentials" type="org.apache.hadoop.security.Credentials"/>
<param name="alias" type="org.apache.hadoop.io.Text"/>
<doc>
<![CDATA[auxiliary method to get user's secret keys..
@param alias
@return secret key from the storage]]>
</doc>
</method>
<method name="obtainTokensForNamenodes"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="credentials" type="org.apache.hadoop.security.Credentials"/>
<param name="ps" type="org.apache.hadoop.fs.Path[]"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Convenience method to obtain delegation tokens from namenodes
corresponding to the paths passed.
@param credentials
@param ps array of paths
@param conf configuration
@throws IOException]]>
</doc>
</method>
<method name="getDelegationToken" return="org.apache.hadoop.security.token.Token"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="credentials" type="org.apache.hadoop.security.Credentials"/>
<param name="namenode" type="java.lang.String"/>
<doc>
<![CDATA[@param namenode
@return delegation token]]>
</doc>
</method>
<method name="loadTokens" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobTokenFile" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[load job token from a file
@param conf
@throws IOException]]>
</doc>
</method>
<method name="setJobToken"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="org.apache.hadoop.security.token.Token"/>
<param name="credentials" type="org.apache.hadoop.security.Credentials"/>
<doc>
<![CDATA[store job token
@param t]]>
</doc>
</method>
<method name="getJobToken" return="org.apache.hadoop.security.token.Token"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="credentials" type="org.apache.hadoop.security.Credentials"/>
<doc>
<![CDATA[@return job token]]>
</doc>
</method>
<field name="JOB_TOKEN_HDFS_FILE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[file name used on HDFS for generated job token]]>
</doc>
</field>
<field name="JOB_TOKENS_FILENAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[conf setting for job tokens cache file name]]>
</doc>
</field>
<doc>
<![CDATA[This class provides user facing APIs for transferring secrets from
the job client to the tasks.
The secrets can be stored just before submission of jobs and read during
the task execution.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.security.TokenCache -->
</package>
<package name="org.apache.hadoop.mapreduce.server.jobtracker">
<!-- start class org.apache.hadoop.mapreduce.server.jobtracker.State -->
<class name="State" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="Use {@link JobTrackerStatus} instead.">
<method name="values" return="org.apache.hadoop.mapreduce.server.jobtracker.State[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapreduce.server.jobtracker.State"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Describes the state of JobTracker
@deprecated Use {@link JobTrackerStatus} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.server.jobtracker.State -->
</package>
<package name="org.apache.hadoop.mapreduce.tools">
<!-- start class org.apache.hadoop.mapreduce.tools.CLI -->
<class name="CLI" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="CLI"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CLI" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="counters" type="org.apache.hadoop.mapreduce.Counters"/>
<param name="counterGroupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getTaskLogURL" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
<param name="baseUrl" type="java.lang.String"/>
</method>
<method name="displayTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
<param name="type" type="java.lang.String"/>
<param name="state" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Display the information about a job's tasks, of a particular type and
in a particular state
@param job the job
@param type the type of the task (map/reduce/setup/cleanup)
@param state the state of the task
(pending/running/completed/failed/killed)]]>
</doc>
</method>
<method name="displayJobList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="jobs" type="org.apache.hadoop.mapreduce.Job[]"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[Interprets the map reduce cli options]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapreduce.tools.CLI -->
</package>
<package name="org.apache.hadoop.fs">
<!-- start class org.apache.hadoop.fs.HarFileSystem -->
<class name="HarFileSystem" extends="org.apache.hadoop.fs.FilterFileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="HarFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[public construction of harfilesystem]]>
</doc>
</constructor>
<constructor name="HarFileSystem" type="org.apache.hadoop.fs.FileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor to create a HarFileSystem with an
underlying filesystem.
@param fs]]>
</doc>
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Initialize a Har filesystem per har archive. The
archive home directory is the top level directory
in the filesystem that contains the HAR archive.
Be careful with this method, you do not want to go
on creating new Filesystem instances per call to
path.getFileSystem().
the uri of Har is
har://underlyingfsscheme-host:port/archivepath.
or
har:///archivepath. This assumes the underlying filesystem
to be used in case not specified.]]>
</doc>
</method>
<method name="getHarVersion" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[return the top level archive.]]>
</doc>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the uri of this filesystem.
The uri is of the form
har://underlyingfsschema-host:port/pathintheunderlyingfs]]>
</doc>
</method>
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.FileStatus"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[get block locations from the underlying fs
@param file the input filestatus to get block locations
@param start the start in the file
@param len the length in the file
@return block locations for this segment of file
@throws IOException]]>
</doc>
</method>
<method name="getHarHash" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[the hash of the path p inside iniside
the filesystem
@param p the path in the harfilesystem
@return the hash code of the path.]]>
</doc>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[return the filestatus of files in har archive.
The permission returned are that of the archive
index files. The permissions are not persisted
while creating a hadoop archive.
@param f the path in har filesystem
@return filestatus.
@throws IOException]]>
</doc>
</method>
<method name="getFileChecksum" return="org.apache.hadoop.fs.FileChecksum"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[@return null since no checksum algorithm is implemented.]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns a har input stream which fakes end of
file. It reads the index files to get the part
file name and the size and start of the file.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="flag" type="java.util.EnumSet"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Not implemented.]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Not implemented.]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[liststatus returns the children of a directory
after looking up the index files.]]>
</doc>
</method>
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[return the top level archive path.]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[not implemented.]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[not implemented.]]>
</doc>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[copies the file in the har filesystem to a local file.]]>
</doc>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[not implemented.]]>
</doc>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[not implemented.]]>
</doc>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[not implemented.]]>
</doc>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="permisssion" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Not implemented.]]>
</doc>
</method>
<field name="VERSION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This is an implementation of the Hadoop Archive
Filesystem. This archive Filesystem has index files
of the form _index* and has contents of the form
part-*. The index files store the indexes of the
real files. The index files are of the form _masterindex
and _index. The master index is a level of indirection
in to the index file to make the look ups faster. the index
file is sorted with hash code of the paths that it contains
and the master index contains pointers to the positions in
index for ranges of hashcodes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.HarFileSystem -->
</package>
<package name="org.apache.hadoop.tools">
<!-- start class org.apache.hadoop.tools.DistCh -->
<class name="DistCh" extends="org.apache.hadoop.tools.DistTool"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[This is the main driver for recursively changing files properties.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[A Map-reduce program to recursively change files properties
such as owner, group and permission.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.DistCh -->
<!-- start class org.apache.hadoop.tools.DistCp -->
<class name="DistCp" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="DistCp" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="copy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="srcPath" type="java.lang.String"/>
<param name="destPath" type="java.lang.String"/>
<param name="logPath" type="org.apache.hadoop.fs.Path"/>
<param name="srcAsList" type="boolean"/>
<param name="ignoreReadFailures" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[This is the main driver for recursively copying directories
across file systems. It takes at least two cmdline parameters. A source
URL and a destination URL. It then essentially does an "ls -lR" on the
source URL, and writes the output in a round-robin manner to all the map
input files. The mapper actually copies the files allotted to it. The
reduce is empty.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="getRandomId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A Map-reduce program to recursively copy directories between
different file-systems.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.DistCp -->
<!-- start class org.apache.hadoop.tools.DistCp.DuplicationException -->
<class name="DistCp.DuplicationException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<field name="ERROR_CODE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Error code for this exception]]>
</doc>
</field>
<doc>
<![CDATA[An exception class for duplicated source files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.DistCp.DuplicationException -->
<!-- start class org.apache.hadoop.tools.HadoopArchives -->
<class name="HadoopArchives" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="HadoopArchives" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[the main driver for creating the archives
it takes at least three command line parameters. The parent path,
The src and the dest. It does an lsr on the source paths.
The mapper created archuves and the reducer creates
the archive index.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[the main functions]]>
</doc>
</method>
<field name="VERSION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[a archive creation utility.
This class provides methods that can be used
to create hadoop archives. For understanding of
Hadoop archives look at {@link HarFileSystem}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.HadoopArchives -->
<!-- start class org.apache.hadoop.tools.Logalyzer -->
<class name="Logalyzer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Logalyzer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doArchive"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="logListURI" type="java.lang.String"/>
<param name="archiveDirectory" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[doArchive: Workhorse function to archive log-files.
@param logListURI : The uri which will serve list of log-files to archive.
@param archiveDirectory : The directory to store archived logfiles.
@throws IOException]]>
</doc>
</method>
<method name="doAnalyze"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inputFilesDirectory" type="java.lang.String"/>
<param name="outputDirectory" type="java.lang.String"/>
<param name="grepPattern" type="java.lang.String"/>
<param name="sortColumns" type="java.lang.String"/>
<param name="columnSeparator" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[doAnalyze:
@param inputFilesDirectory : Directory containing the files to be analyzed.
@param outputDirectory : Directory to store analysis (output).
@param grepPattern : Pattern to *grep* for.
@param sortColumns : Sort specification for output.
@param columnSeparator : Column separator.
@throws IOException]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<field name="SORT_COLUMNS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="COLUMN_SEPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Logalyzer: A utility tool for archiving and analyzing hadoop logs.
<p>
This tool supports archiving and anaylzing (sort/grep) of log-files.
It takes as input
a) Input uri which will serve uris of the logs to be archived.
b) Output directory (not mandatory).
b) Directory on dfs to archive the logs.
c) The sort/grep patterns for analyzing the files and separator for boundaries.
Usage:
Logalyzer -archive -archiveDir <directory to archive logs> -analysis <directory> -logs <log-list uri> -grep <pattern> -sort <col1, col2> -separator <separator>
<p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer -->
<!-- start class org.apache.hadoop.tools.Logalyzer.LogComparator -->
<class name="Logalyzer.LogComparator" extends="org.apache.hadoop.io.Text.Comparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="Logalyzer.LogComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A WritableComparator optimized for UTF8 keys of the logs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer.LogComparator -->
<!-- start class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
<class name="Logalyzer.LogRegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper"/>
<constructor name="Logalyzer.LogRegexMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
</package>
<package name="org.apache.hadoop.tools.rumen">
<!-- start class org.apache.hadoop.tools.rumen.AbstractClusterStory -->
<class name="AbstractClusterStory" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.ClusterStory"/>
<constructor name="AbstractClusterStory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getMachines" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRacks" return="java.util.Set"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRandomMachines" return="org.apache.hadoop.tools.rumen.MachineNode[]"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="expected" type="int"/>
<param name="random" type="java.util.Random"/>
</method>
<method name="buildMachineNodeMap"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getMachineByName" return="org.apache.hadoop.tools.rumen.MachineNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="distance" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.tools.rumen.Node"/>
<param name="b" type="org.apache.hadoop.tools.rumen.Node"/>
</method>
<method name="buildRackNodeMap"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getRackByName" return="org.apache.hadoop.tools.rumen.RackNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="getMaximumDistance" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="parseTopologyTree"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<field name="machineNodes" type="java.util.Set"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rackNodes" type="java.util.Set"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="mNodesFlattened" type="org.apache.hadoop.tools.rumen.MachineNode[]"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="mNodeMap" type="java.util.Map"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="rNodeMap" type="java.util.Map"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="maximumDistance" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[{@link AbstractClusterStory} provides a partial implementation of
{@link ClusterStory} by parsing the topology tree.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.AbstractClusterStory -->
<!-- start class org.apache.hadoop.tools.rumen.CDFPiecewiseLinearRandomGenerator -->
<class name="CDFPiecewiseLinearRandomGenerator" extends="org.apache.hadoop.tools.rumen.CDFRandomGenerator"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CDFPiecewiseLinearRandomGenerator" type="org.apache.hadoop.tools.rumen.LoggedDiscreteCDF"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param cdf
builds a CDFRandomValue engine around this
{@link LoggedDiscreteCDF}, with a defaultly seeded RNG]]>
</doc>
</constructor>
<constructor name="CDFPiecewiseLinearRandomGenerator" type="org.apache.hadoop.tools.rumen.LoggedDiscreteCDF, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param cdf
builds a CDFRandomValue engine around this
{@link LoggedDiscreteCDF}, with an explicitly seeded RNG
@param seed
the random number generator seed]]>
</doc>
</constructor>
<method name="valueAt" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="probability" type="double"/>
<doc>
<![CDATA[TODO This code assumes that the empirical minimum resp. maximum is the
epistomological minimum resp. maximum. This is probably okay for the
minimum, because that likely represents a task where everything went well,
but for the maximum we may want to develop a way of extrapolating past the
maximum.]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.CDFPiecewiseLinearRandomGenerator -->
<!-- start class org.apache.hadoop.tools.rumen.CDFRandomGenerator -->
<class name="CDFRandomGenerator" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="initializeTables"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<param name="cdf" type="org.apache.hadoop.tools.rumen.LoggedDiscreteCDF"/>
</method>
<method name="floorIndex" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="probe" type="double"/>
</method>
<method name="getRankingAt" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="index" type="int"/>
</method>
<method name="getDatumAt" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="index" type="int"/>
</method>
<method name="randomValue" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueAt" return="long"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="probability" type="double"/>
</method>
<doc>
<![CDATA[An instance of this class generates random values that confirm to the
embedded {@link LoggedDiscreteCDF} . The discrete CDF is a pointwise
approximation of the "real" CDF. We therefore have a choice of interpolation
rules.
A concrete subclass of this abstract class will implement valueAt(double)
using a class-dependent interpolation rule.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.CDFRandomGenerator -->
<!-- start interface org.apache.hadoop.tools.rumen.ClusterStory -->
<interface name="ClusterStory" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getMachines" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get all machines of the cluster.
@return A read-only set that contains all machines of the cluster.]]>
</doc>
</method>
<method name="getRacks" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get all racks of the cluster.
@return A read-only set that contains all racks of the cluster.]]>
</doc>
</method>
<method name="getClusterTopology" return="org.apache.hadoop.tools.rumen.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the cluster topology tree.
@return The root node of the cluster topology tree.]]>
</doc>
</method>
<method name="getRandomMachines" return="org.apache.hadoop.tools.rumen.MachineNode[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="expected" type="int"/>
<param name="random" type="java.util.Random"/>
<doc>
<![CDATA[Select a random set of machines.
@param expected The expected sample size.
@param random Random number generator to use.
@return An array of up to expected number of {@link MachineNode}s.]]>
</doc>
</method>
<method name="getMachineByName" return="org.apache.hadoop.tools.rumen.MachineNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get {@link MachineNode} by its host name.
@return The {@link MachineNode} with the same name. Or null if not found.]]>
</doc>
</method>
<method name="getRackByName" return="org.apache.hadoop.tools.rumen.RackNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get {@link RackNode} by its name.
@return The {@link RackNode} with the same name. Or null if not found.]]>
</doc>
</method>
<method name="distance" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.tools.rumen.Node"/>
<param name="b" type="org.apache.hadoop.tools.rumen.Node"/>
<doc>
<![CDATA[Determine the distance between two {@link Node}s. Currently, the distance
is loosely defined as the length of the longer path for either a or b to
reach their common ancestor.
@param a
@param b
@return The distance between {@link Node} a and {@link Node} b.]]>
</doc>
</method>
<method name="getMaximumDistance" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum distance possible between any two nodes.
@return the maximum distance possible between any two nodes.]]>
</doc>
</method>
<doc>
<![CDATA[{@link ClusterStory} represents all configurations of a MapReduce cluster,
including nodes, network topology, and slot configurations.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.ClusterStory -->
<!-- start class org.apache.hadoop.tools.rumen.ClusterTopologyReader -->
<class name="ClusterTopologyReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ClusterTopologyReader" type="org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor.
@param path
Path to the JSON-encoded topology file, possibly compressed.
@param conf
@throws IOException]]>
</doc>
</constructor>
<constructor name="ClusterTopologyReader" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor.
@param input
The input stream for the JSON-encoded topology data.]]>
</doc>
</constructor>
<method name="get" return="org.apache.hadoop.tools.rumen.LoggedNetworkTopology"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link LoggedNetworkTopology} object.
@return The {@link LoggedNetworkTopology} object parsed from the input.]]>
</doc>
</method>
<doc>
<![CDATA[Reading JSON-encoded cluster topology and produce the parsed
{@link LoggedNetworkTopology} object.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.ClusterTopologyReader -->
<!-- start class org.apache.hadoop.tools.rumen.CurrentJHParser -->
<class name="CurrentJHParser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.JobHistoryParser"/>
<constructor name="CurrentJHParser" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="canParse" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Can this parser parse the input?
@param input
@return Whether this parser can parse the input.
@throws IOException]]>
</doc>
</method>
<method name="nextEvent" return="org.apache.hadoop.mapreduce.jobhistory.HistoryEvent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[{@link JobHistoryParser} that parses {@link JobHistory} files produced by
{@link org.apache.hadoop.mapreduce.jobhistory.JobHistory} in the same source
code tree as rumen.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.CurrentJHParser -->
<!-- start interface org.apache.hadoop.tools.rumen.DeepCompare -->
<interface name="DeepCompare" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="myLocation" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
<doc>
<![CDATA[@param other
the other comparand that's being compared to me
@param myLocation
the path that got to me. In the root, myLocation is null. To
process the scalar {@code foo} field of the root we will make a
recursive call with a {@link TreePath} whose {@code fieldName} is
{@code "bar"} and whose {@code index} is -1 and whose {@code
parent} is {@code null}. To process the plural {@code bar} field
of the root we will make a recursive call with a {@link TreePath}
whose fieldName is {@code "foo"} and whose {@code index} is -1 and
whose {@code parent} is also {@code null}.
@throws DeepInequalityException]]>
</doc>
</method>
<doc>
<![CDATA[Classes that implement this interface can deep-compare [for equality only,
not order] with another instance. They do a deep compare. If there is any
semantically significant difference, an implementer throws an Exception to be
thrown with a chain of causes describing the chain of field references and
indices that get you to the miscompared point.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.DeepCompare -->
<!-- start class org.apache.hadoop.tools.rumen.DeepInequalityException -->
<class name="DeepInequalityException" extends="java.lang.Exception"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DeepInequalityException" type="java.lang.String, org.apache.hadoop.tools.rumen.TreePath, java.lang.Throwable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param message
an exception message
@param path
the path that gets from the root to the inequality
This is the constructor that I intend to have used for this
exception.]]>
</doc>
</constructor>
<constructor name="DeepInequalityException" type="java.lang.String, org.apache.hadoop.tools.rumen.TreePath"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param message
an exception message
@param path
the path that gets from the root to the inequality
This is the constructor that I intend to have used for this
exception.]]>
</doc>
</constructor>
<doc>
<![CDATA[We use this exception class in the unit test, and we do a deep comparison
when we run the]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.DeepInequalityException -->
<!-- start class org.apache.hadoop.tools.rumen.DefaultInputDemuxer -->
<class name="DefaultInputDemuxer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.InputDemuxer"/>
<constructor name="DefaultInputDemuxer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="bindTo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getNext" return="org.apache.hadoop.tools.rumen.Pair"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[{@link DefaultInputDemuxer} acts as a pass-through demuxer. It just opens
each file and returns back the input stream. If the input is compressed, it
would return a decompression stream.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.DefaultInputDemuxer -->
<!-- start class org.apache.hadoop.tools.rumen.DefaultOutputter -->
<class name="DefaultOutputter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.Outputter"/>
<constructor name="DefaultOutputter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="output"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="object" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[The default {@link Outputter} that outputs to a plain file. Compression
will be applied if the path has the right suffix.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.DefaultOutputter -->
<!-- start class org.apache.hadoop.tools.rumen.DeskewedJobTraceReader -->
<class name="DeskewedJobTraceReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="DeskewedJobTraceReader" type="org.apache.hadoop.tools.rumen.JobTraceReader, int, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor.
@param reader
the {@link JobTraceReader} that's being protected
@param skewBufferLength
[the number of late jobs that can preced a later out-of-order
earlier job
@throws IOException]]>
</doc>
</constructor>
<constructor name="DeskewedJobTraceReader" type="org.apache.hadoop.tools.rumen.JobTraceReader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.DeskewedJobTraceReader -->
<!-- start class org.apache.hadoop.tools.rumen.Folder -->
<class name="Folder" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="Folder"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[@param args]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Folder -->
<!-- start class org.apache.hadoop.tools.rumen.Hadoop20JHParser -->
<class name="Hadoop20JHParser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.JobHistoryParser"/>
<constructor name="Hadoop20JHParser" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="canParse" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Can this parser parse the input?
@param input
@return Whether this parser can parse the input.
@throws IOException
We will deem a stream to be a good 0.20 job history stream if the
first line is exactly "Meta VERSION=\"1\" ."]]>
</doc>
</method>
<method name="nextEvent" return="org.apache.hadoop.mapreduce.jobhistory.HistoryEvent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[{@link JobHistoryParser} to parse job histories for hadoop 0.20 (META=1).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Hadoop20JHParser -->
<!-- start class org.apache.hadoop.tools.rumen.HadoopLogsAnalyzer -->
<class name="HadoopLogsAnalyzer" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="HadoopLogsAnalyzer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[@param args
Last arg is the input file. That file can be a directory, in which
case you get all the files in sorted order. We will decompress
files whose nmes end in .gz .
switches: -c collect line types.
-d debug mode
-delays print out the delays [interval between job submit time and
launch time]
-runtimes print out the job runtimes
-spreads print out the ratio of 10%ile and 90%ile, of both the
successful map task attempt run times and the the successful
reduce task attempt run times
-tasktimes prints out individual task time distributions
collects all the line types and prints the first example of each
one]]>
</doc>
</method>
<doc>
<![CDATA[This is the main class for rumen log mining functionality.
It reads a directory of job tracker logs, and computes various information
about it. See {@code usage()}, below.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.HadoopLogsAnalyzer -->
<!-- start interface org.apache.hadoop.tools.rumen.InputDemuxer -->
<interface name="InputDemuxer" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<method name="bindTo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Bind the {@link InputDemuxer} to a particular file.
@param path
The path to the file it should bind to.
@param conf
Configuration
@throws IOException
Returns true when the binding succeeds. If the file can be read
but is in the wrong format, returns false. IOException is
reserved for read errors.]]>
</doc>
</method>
<method name="getNext" return="org.apache.hadoop.tools.rumen.Pair"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the next <name, input> pair. The name should preserve the original job
history file or job conf file name. The input object should be closed
before calling getNext() again. The old input object would be invalid after
calling getNext() again.
@return the next <name, input> pair.]]>
</doc>
</method>
<doc>
<![CDATA[{@link InputDemuxer} dem-ultiplexes the input files into individual input
streams.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.InputDemuxer -->
<!-- start class org.apache.hadoop.tools.rumen.Job20LineHistoryEventEmitter -->
<class name="Job20LineHistoryEventEmitter" extends="org.apache.hadoop.tools.rumen.HistoryEventEmitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</class>
<!-- end class org.apache.hadoop.tools.rumen.Job20LineHistoryEventEmitter -->
<!-- start class org.apache.hadoop.tools.rumen.JobBuilder -->
<class name="JobBuilder" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobBuilder" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="process"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="event" type="org.apache.hadoop.mapreduce.jobhistory.HistoryEvent"/>
<doc>
<![CDATA[Process one {@link HistoryEvent}
@param event
The {@link HistoryEvent} to be processed.]]>
</doc>
</method>
<method name="process"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="java.util.Properties"/>
<doc>
<![CDATA[Process a collection of JobConf {@link Properties}. We do not restrict it
to be called once. It is okay to process a conf before, during or after the
events.
@param conf
The job conf properties to be added.]]>
</doc>
</method>
<method name="build" return="org.apache.hadoop.tools.rumen.LoggedJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request the builder to build the final object. Once called, the
{@link JobBuilder} would accept no more events or job-conf properties.
@return Parsed {@link LoggedJob} object.]]>
</doc>
</method>
<doc>
<![CDATA[{@link JobBuilder} builds one job. It processes a sequence of
{@link HistoryEvent}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobBuilder -->
<!-- start class org.apache.hadoop.tools.rumen.JobConfigurationParser -->
<class name="JobConfigurationParser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobConfigurationParser" type="java.util.List"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor
@param interested
properties we should extract from the job configuration xml.]]>
</doc>
</constructor>
<doc>
<![CDATA[{@link JobConfigurationParser} parses the job configuration xml file, and
extracts various framework specific properties. It parses the file using a
stream-parser and thus is more memory efficient. [This optimization may be
postponed for a future release]]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobConfigurationParser -->
<!-- start class org.apache.hadoop.tools.rumen.JobConfPropertyNames -->
<class name="JobConfPropertyNames" extends="java.lang.Enum"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.tools.rumen.JobConfPropertyNames[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.tools.rumen.JobConfPropertyNames"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="getCandidates" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobConfPropertyNames -->
<!-- start interface org.apache.hadoop.tools.rumen.JobHistoryParser -->
<interface name="JobHistoryParser" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<method name="nextEvent" return="org.apache.hadoop.mapreduce.jobhistory.HistoryEvent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the next {@link HistoryEvent}
@return the next {@link HistoryEvent}. If no more events left, return null.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[{@link JobHistoryParser} defines the interface of a Job History file parser.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.JobHistoryParser -->
<!-- start class org.apache.hadoop.tools.rumen.JobHistoryParserFactory -->
<class name="JobHistoryParserFactory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistoryParserFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getParser" return="org.apache.hadoop.tools.rumen.JobHistoryParser"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="ris" type="org.apache.hadoop.tools.rumen.RewindableInputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[{@link JobHistoryParserFactory} is a singleton class that attempts to
determine the version of job history and return a proper parser.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobHistoryParserFactory -->
<!-- start class org.apache.hadoop.tools.rumen.JobHistoryParserFactory.VersionDetector -->
<class name="JobHistoryParserFactory.VersionDetector" extends="java.lang.Enum"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.tools.rumen.JobHistoryParserFactory.VersionDetector[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.tools.rumen.JobHistoryParserFactory.VersionDetector"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobHistoryParserFactory.VersionDetector -->
<!-- start interface org.apache.hadoop.tools.rumen.JobStory -->
<interface name="JobStory" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link JobConf} for the job.
@return the <code>JobConf</code> for the job]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job name.
@return the job name]]>
</doc>
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job ID
@return the job ID]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user who ran the job.
@return the user who ran the job]]>
</doc>
</method>
<method name="getSubmissionTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job submission time.
@return the job submission time]]>
</doc>
</method>
<method name="getNumberMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of maps in the {@link JobStory}.
@return the number of maps in the <code>Job</code>]]>
</doc>
</method>
<method name="getNumberReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of reduce in the {@link JobStory}.
@return the number of reduces in the <code>Job</code>]]>
</doc>
</method>
<method name="getInputSplits" return="org.apache.hadoop.mapreduce.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the input splits for the job.
@return the input splits for the job]]>
</doc>
</method>
<method name="getTaskInfo" return="org.apache.hadoop.tools.rumen.TaskInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskType" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskNumber" type="int"/>
<doc>
<![CDATA[Get {@link TaskInfo} for a given task.
@param taskType {@link TaskType} of the task
@param taskNumber Partition number of the task
@return the <code>TaskInfo</code> for the given task]]>
</doc>
</method>
<method name="getTaskAttemptInfo" return="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskType" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskNumber" type="int"/>
<param name="taskAttemptNumber" type="int"/>
<doc>
<![CDATA[Get {@link TaskAttemptInfo} for a given task-attempt, without regard to
impact of locality (e.g. not needed to make scheduling decisions).
@param taskType {@link TaskType} of the task-attempt
@param taskNumber Partition number of the task-attempt
@param taskAttemptNumber Attempt number of the task
@return the <code>TaskAttemptInfo</code> for the given task-attempt]]>
</doc>
</method>
<method name="getMapTaskAttemptInfoAdjusted" return="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskNumber" type="int"/>
<param name="taskAttemptNumber" type="int"/>
<param name="locality" type="int"/>
<doc>
<![CDATA[Get {@link TaskAttemptInfo} for a given task-attempt, considering impact
of locality.
@param taskNumber Partition number of the task-attempt
@param taskAttemptNumber Attempt number of the task
@param locality Data locality of the task as scheduled in simulation
@return the <code>TaskAttemptInfo</code> for the given task-attempt]]>
</doc>
</method>
<method name="getOutcome" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the outcome of the job execution.
@return The outcome of the job execution.]]>
</doc>
</method>
<method name="getQueueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the queue where the job is submitted.
@return the queue where the job is submitted.]]>
</doc>
</method>
<doc>
<![CDATA[{@link JobStory} represents the runtime information available for a
completed Map-Reduce job.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.JobStory -->
<!-- start interface org.apache.hadoop.tools.rumen.JobStoryProducer -->
<interface name="JobStoryProducer" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<method name="getNextJob" return="org.apache.hadoop.tools.rumen.JobStory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the next job.
@return The next job. Or null if no more job is available.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[{@link JobStoryProducer} produces the sequence of {@link JobStory}'s.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.JobStoryProducer -->
<!-- start class org.apache.hadoop.tools.rumen.JobTraceReader -->
<class name="JobTraceReader" extends="org.apache.hadoop.tools.rumen.JsonObjectMapperParser"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobTraceReader" type="org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor.
@param path
Path to the JSON trace file, possibly compressed.
@param conf
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobTraceReader" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor.
@param input
The input stream for the JSON trace.]]>
</doc>
</constructor>
<doc>
<![CDATA[Reading JSON-encoded job traces and produce {@link LoggedJob} instances.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JobTraceReader -->
<!-- start class org.apache.hadoop.tools.rumen.JsonObjectMapperWriter -->
<class name="JsonObjectMapperWriter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="JsonObjectMapperWriter" type="java.io.OutputStream, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="object" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Simple wrapper around {@link JsonGenerator} to write objects in JSON format.
@param <T> The type of the objects to be written.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.JsonObjectMapperWriter -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedDiscreteCDF -->
<class name="LoggedDiscreteCDF" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<constructor name="LoggedDiscreteCDF"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getMinimum" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRankings" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMaximum" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumberValues" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedDiscreteCDF} is a discrete approximation of a cumulative
distribution function, with this class set up to meet the requirements of the
Jackson JSON parser/generator.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedDiscreteCDF -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedJob -->
<class name="LoggedJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPriority" return="org.apache.hadoop.tools.rumen.LoggedJob.JobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getComputonsPerMapInputByte" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getComputonsPerMapOutputByte" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getComputonsPerReduceInputByte" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getComputonsPerReduceOutputByte" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSubmitTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLaunchTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getHeapMegabytes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTotalMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTotalReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutcome" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobtype" return="org.apache.hadoop.tools.rumen.LoggedJob.JobType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDirectDependantJobs" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapTasks" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getReduceTasks" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOtherTasks" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSuccessfulMapAttemptCDFs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFailedMapAttemptCDFs" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSuccessfulReduceAttemptCDF" return="org.apache.hadoop.tools.rumen.LoggedDiscreteCDF"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFailedReduceAttemptCDF" return="org.apache.hadoop.tools.rumen.LoggedDiscreteCDF"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapperTriesToSucceed" return="double[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFailedMapperFraction" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRelativeTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getQueue" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getClusterMapMB" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getClusterReduceMB" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobMapMB" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobReduceMB" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedDiscreteCDF} is a representation of an hadoop job, with the
details of this class set up to meet the requirements of the Jackson JSON
parser/generator.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedJob -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedJob.JobPriority -->
<class name="LoggedJob.JobPriority" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.tools.rumen.LoggedJob.JobPriority[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.tools.rumen.LoggedJob.JobPriority"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedJob.JobPriority -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedJob.JobType -->
<class name="LoggedJob.JobType" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.tools.rumen.LoggedJob.JobType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.tools.rumen.LoggedJob.JobType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedJob.JobType -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedLocation -->
<class name="LoggedLocation" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<constructor name="LoggedLocation"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLayers" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedLocation} is a representation of a point in an hierarchical
network, represented as a series of membership names, broadest first.
For example, if your network has <i>hosts</i> grouped into <i>racks</i>, then
in onecluster you might have a node {@code node1} on rack {@code rack1}. This
would be represented with a ArrayList of two layers, with two {@link String}
s being {@code "rack1"} and {@code "node1"}.
The details of this class are set up to meet the requirements of the Jackson
JSON parser/generator.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedLocation -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedNetworkTopology -->
<class name="LoggedNetworkTopology" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<constructor name="LoggedNetworkTopology"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getChildren" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedNetworkTopology} represents a tree that in turn represents a
hierarchy of hosts. The current version requires the tree to have all leaves
at the same level.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedNetworkTopology -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedSingleRelativeRanking -->
<class name="LoggedSingleRelativeRanking" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<constructor name="LoggedSingleRelativeRanking"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="getRelativeRanking" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDatum" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedSingleRelativeRanking} represents an X-Y coordinate of a
single point in a discrete CDF.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedSingleRelativeRanking -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedTask -->
<class name="LoggedTask" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="getInputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getAttempts" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPreferredLocations" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskStatus" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskType" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="incorporateCounters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counters" type="org.apache.hadoop.mapreduce.jobhistory.JhCounters"/>
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedTask} represents a [hadoop] task that is part of a hadoop job.
It knows about the [pssibly empty] sequence of attempts, its I/O footprint,
and its runtime.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedTask -->
<!-- start class org.apache.hadoop.tools.rumen.LoggedTaskAttempt -->
<class name="LoggedTaskAttempt" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.DeepCompare"/>
<method name="setUnknownAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="ignored" type="java.lang.Object"/>
</method>
<method name="getShuffleFinished" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSortFinished" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getAttemptID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getResult" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getHostName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getHdfsBytesRead" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getHdfsBytesWritten" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileBytesRead" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileBytesWritten" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapInputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapOutputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapOutputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCombineInputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getReduceInputGroups" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getReduceInputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getReduceShuffleBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getReduceOutputRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSpilledRecords" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLocation" return="org.apache.hadoop.tools.rumen.LoggedLocation"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapInputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="incorporateCounters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counters" type="org.apache.hadoop.mapreduce.jobhistory.JhCounters"/>
</method>
<method name="deepCompare"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comparand" type="org.apache.hadoop.tools.rumen.DeepCompare"/>
<param name="loc" type="org.apache.hadoop.tools.rumen.TreePath"/>
<exception name="DeepInequalityException" type="org.apache.hadoop.tools.rumen.DeepInequalityException"/>
</method>
<doc>
<![CDATA[A {@link LoggedTaskAttempt} represents an attempt to run an hadoop task in a
hadoop job. Note that a task can have several attempts.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.LoggedTaskAttempt -->
<!-- start class org.apache.hadoop.tools.rumen.MachineNode -->
<class name="MachineNode" extends="org.apache.hadoop.tools.rumen.Node"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMemory" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the available physical RAM of the node.
@return The available physical RAM of the node, in KB.]]>
</doc>
</method>
<method name="getMapSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of map slots of the node.
@return The number of map slots of the node.]]>
</doc>
</method>
<method name="getReduceSlots" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of reduce slots of the node.
@return The number of reduce slots fo the node.]]>
</doc>
</method>
<method name="getMemoryPerMapSlot" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the amount of RAM reserved for each map slot.
@return the amount of RAM reserved for each map slot, in KB.]]>
</doc>
</method>
<method name="getMemoryPerReduceSlot" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the amount of RAM reserved for each reduce slot.
@return the amount of RAM reserved for each reduce slot, in KB.]]>
</doc>
</method>
<method name="getNumCores" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of cores of the node.
@return the number of cores of the node.]]>
</doc>
</method>
<method name="getRackNode" return="org.apache.hadoop.tools.rumen.RackNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the rack node that the machine belongs to.
@return The rack node that the machine belongs to. Returns null if the
machine does not belong to any rack.]]>
</doc>
</method>
<method name="addChild" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="child" type="org.apache.hadoop.tools.rumen.Node"/>
</method>
<doc>
<![CDATA[{@link MachineNode} represents the configuration of a cluster node.
{@link MachineNode} should be constructed by {@link MachineNode.Builder}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.MachineNode -->
<!-- start class org.apache.hadoop.tools.rumen.MachineNode.Builder -->
<class name="MachineNode.Builder" extends="java.lang.Object"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<constructor name="MachineNode.Builder" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Start building a new NodeInfo object.
@param name
Unique name of the node. Typically the fully qualified domain
name.]]>
</doc>
</constructor>
<method name="setMemory" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="memory" type="long"/>
<doc>
<![CDATA[Set the physical memory of the node.
@param memory Available RAM in KB.]]>
</doc>
</method>
<method name="setMapSlots" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mapSlots" type="int"/>
<doc>
<![CDATA[Set the number of map slot for the node.
@param mapSlots The number of map slots for the node.]]>
</doc>
</method>
<method name="setReduceSlots" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reduceSlots" type="int"/>
<doc>
<![CDATA[Set the number of reduce slot for the node.
@param reduceSlots The number of reduce slots for the node.]]>
</doc>
</method>
<method name="setMemoryPerMapSlot" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="memoryPerMapSlot" type="long"/>
<doc>
<![CDATA[Set the amount of RAM reserved for each map slot.
@param memoryPerMapSlot The amount of RAM reserved for each map slot, in KB.]]>
</doc>
</method>
<method name="setMemoryPerReduceSlot" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="memoryPerReduceSlot" type="long"/>
<doc>
<![CDATA[Set the amount of RAM reserved for each reduce slot.
@param memoryPerReduceSlot The amount of RAM reserved for each reduce slot, in KB.]]>
</doc>
</method>
<method name="setNumCores" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="numCores" type="int"/>
<doc>
<![CDATA[Set the number of cores for the node.
@param numCores Number of cores for the node.]]>
</doc>
</method>
<method name="cloneFrom" return="org.apache.hadoop.tools.rumen.MachineNode.Builder"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ref" type="org.apache.hadoop.tools.rumen.MachineNode"/>
<doc>
<![CDATA[Clone the settings from a reference {@link MachineNode} object.
@param ref The reference {@link MachineNode} object.]]>
</doc>
</method>
<method name="build" return="org.apache.hadoop.tools.rumen.MachineNode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Build the {@link MachineNode} object.
@return The {@link MachineNode} object being built.]]>
</doc>
</method>
<doc>
<![CDATA[Builder for a NodeInfo object]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.MachineNode.Builder -->
<!-- start class org.apache.hadoop.tools.rumen.MapAttempt20LineHistoryEventEmitter -->
<class name="MapAttempt20LineHistoryEventEmitter" extends="org.apache.hadoop.tools.rumen.TaskAttempt20LineEventEmitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapAttempt20LineHistoryEventEmitter"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.tools.rumen.MapAttempt20LineHistoryEventEmitter -->
<!-- start class org.apache.hadoop.tools.rumen.MapTaskAttemptInfo -->
<class name="MapTaskAttemptInfo" extends="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapTaskAttemptInfo" type="org.apache.hadoop.mapred.TaskStatus.State, org.apache.hadoop.tools.rumen.TaskInfo, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getMapRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the runtime for the <b>map</b> phase of the map-task attempt.
@return the runtime for the <b>map</b> phase of the map-task attempt]]>
</doc>
</method>
<doc>
<![CDATA[{@link MapTaskAttemptInfo} represents the information with regard to a
map task attempt.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.MapTaskAttemptInfo -->
<!-- start class org.apache.hadoop.tools.rumen.Node -->
<class name="Node" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<constructor name="Node" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param name
A unique name to identify a node in the cluster.
@param level
The level of the node in the cluster]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the node.
@return The name of the node.]]>
</doc>
</method>
<method name="getLevel" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the level of the node.
@return The level of the node.]]>
</doc>
</method>
<method name="addChild" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="child" type="org.apache.hadoop.tools.rumen.Node"/>
<doc>
<![CDATA[Add a child node to this node.
@param child The child node to be added. The child node should currently not be belong to another cluster topology.
@return Boolean indicating whether the node is successfully added.]]>
</doc>
</method>
<method name="hasChildren" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Does this node have any children?
@return Boolean indicate whether this node has any children.]]>
</doc>
</method>
<method name="getChildren" return="java.util.Set"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the children of this node.
@return The children of this node. If no child, an empty set will be
returned. The returned set is read-only.]]>
</doc>
</method>
<method name="getParent" return="org.apache.hadoop.tools.rumen.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the parent node.
@return the parent node. If root node, return null.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="org.apache.hadoop.tools.rumen.Node"/>
</method>
<doc>
<![CDATA[{@link Node} represents a node in the cluster topology. A node can be a
{@link MachineNode}, or a {@link RackNode}, etc.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Node -->
<!-- start interface org.apache.hadoop.tools.rumen.Outputter -->
<interface name="Outputter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Initialize the {@link Outputter} to a specific path.
@param path The {@link Path} to the output file.
@param conf Configuration
@throws IOException]]>
</doc>
</method>
<method name="output"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="object" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Output an object.
@param object The objecte.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Interface to output a sequence of objects of type T.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.tools.rumen.Outputter -->
<!-- start class org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants -->
<class name="Pre21JobHistoryConstants" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Pre21JobHistoryConstants"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants -->
<!-- start class org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values -->
<class name="Pre21JobHistoryConstants.Values" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[This enum contains some of the values commonly used by history log events.
since values in history can only be strings - Values.name() is used in
most places in history file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values -->
<!-- start class org.apache.hadoop.tools.rumen.RackNode -->
<class name="RackNode" extends="org.apache.hadoop.tools.rumen.Node"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<constructor name="RackNode" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addChild" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="child" type="org.apache.hadoop.tools.rumen.Node"/>
</method>
<method name="getMachinesInRack" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the machine nodes that belong to the rack.
@return The machine nodes that belong to the rack.]]>
</doc>
</method>
<doc>
<![CDATA[{@link RackNode} represents a rack node in the cluster topology.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.RackNode -->
<!-- start class org.apache.hadoop.tools.rumen.RandomSeedGenerator -->
<class name="RandomSeedGenerator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RandomSeedGenerator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSeed" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="streamId" type="java.lang.String"/>
<param name="masterSeed" type="long"/>
<doc>
<![CDATA[Generates a new random seed.
@param streamId a string identifying the stream of random numbers
@param masterSeed higher level master random seed
@return the random seed. Different (streamId, masterSeed) pairs result in
(vastly) different random seeds.]]>
</doc>
</method>
<doc>
<![CDATA[The purpose of this class is to generate new random seeds from a master
seed. This is needed to make the Random().next*() calls in rumen and mumak
deterministic so that mumak simulations become deterministically replayable.
In these tools we need many independent streams of random numbers, some of
which are created dynamically. We seed these streams with the sub-seeds
returned by RandomSeedGenerator.
For a slightly more complicated approach to generating multiple streams of
random numbers with better theoretical guarantees, see
P. L'Ecuyer, R. Simard, E. J. Chen, and W. D. Kelton,
``An Objected-Oriented Random-Number Package with Many Long Streams and
Substreams'', Operations Research, 50, 6 (2002), 1073--1075
http://www.iro.umontreal.ca/~lecuyer/papers.html
http://www.iro.umontreal.ca/~lecuyer/myftp/streams00/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.RandomSeedGenerator -->
<!-- start class org.apache.hadoop.tools.rumen.ReduceAttempt20LineHistoryEventEmitter -->
<class name="ReduceAttempt20LineHistoryEventEmitter" extends="org.apache.hadoop.tools.rumen.TaskAttempt20LineEventEmitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</class>
<!-- end class org.apache.hadoop.tools.rumen.ReduceAttempt20LineHistoryEventEmitter -->
<!-- start class org.apache.hadoop.tools.rumen.ReduceTaskAttemptInfo -->
<class name="ReduceTaskAttemptInfo" extends="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ReduceTaskAttemptInfo" type="org.apache.hadoop.mapred.TaskStatus.State, org.apache.hadoop.tools.rumen.TaskInfo, long, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getReduceRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the runtime for the <b>reduce</b> phase of the reduce task-attempt.
@return the runtime for the <b>reduce</b> phase of the reduce task-attempt]]>
</doc>
</method>
<method name="getShuffleRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the runtime for the <b>shuffle</b> phase of the reduce task-attempt.
@return the runtime for the <b>shuffle</b> phase of the reduce task-attempt]]>
</doc>
</method>
<method name="getMergeRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the runtime for the <b>merge</b> phase of the reduce task-attempt
@return the runtime for the <b>merge</b> phase of the reduce task-attempt]]>
</doc>
</method>
<method name="getRuntime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[{@link ReduceTaskAttemptInfo} represents the information with regard to a
reduce task attempt.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.ReduceTaskAttemptInfo -->
<!-- start class org.apache.hadoop.tools.rumen.RewindableInputStream -->
<class name="RewindableInputStream" extends="java.io.InputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RewindableInputStream" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor.
@param input]]>
</doc>
</constructor>
<constructor name="RewindableInputStream" type="java.io.InputStream, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor
@param input
input stream.
@param maxBytesToRemember
Maximum number of bytes we need to remember at the beginning of
the stream. If {@link #rewind()} is called after so many bytes are
read from the stream, {@link #rewind()} would fail.]]>
</doc>
</constructor>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rewind" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A simple wrapper class to make any input stream "rewindable". It could be
made more memory efficient by grow the internal buffer adaptively.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.RewindableInputStream -->
<!-- start class org.apache.hadoop.tools.rumen.Task20LineHistoryEventEmitter -->
<class name="Task20LineHistoryEventEmitter" extends="org.apache.hadoop.tools.rumen.HistoryEventEmitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Task20LineHistoryEventEmitter"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.tools.rumen.Task20LineHistoryEventEmitter -->
<!-- start class org.apache.hadoop.tools.rumen.TaskAttempt20LineEventEmitter -->
<class name="TaskAttempt20LineEventEmitter" extends="org.apache.hadoop.tools.rumen.HistoryEventEmitter"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskAttempt20LineEventEmitter"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TaskAttempt20LineEventEmitter -->
<!-- start class org.apache.hadoop.tools.rumen.TaskAttemptInfo -->
<class name="TaskAttemptInfo" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskAttemptInfo" type="org.apache.hadoop.mapred.TaskStatus.State, org.apache.hadoop.tools.rumen.TaskInfo"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="getRunState" return="org.apache.hadoop.mapred.TaskStatus.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the final {@link State} of the task-attempt.
@return the final <code>State</code> of the task-attempt]]>
</doc>
</method>
<method name="getRuntime" return="long"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total runtime for the task-attempt.
@return the total runtime for the task-attempt]]>
</doc>
</method>
<method name="getTaskInfo" return="org.apache.hadoop.tools.rumen.TaskInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link TaskInfo} for the given task-attempt.
@return the <code>TaskInfo</code> for the given task-attempt]]>
</doc>
</method>
<field name="state" type="org.apache.hadoop.mapred.TaskStatus.State"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="taskInfo" type="org.apache.hadoop.tools.rumen.TaskInfo"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[{@link TaskAttemptInfo} is a collection of statistics about a particular
task-attempt gleaned from job-history of the job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TaskAttemptInfo -->
<!-- start class org.apache.hadoop.tools.rumen.TaskInfo -->
<class name="TaskInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskInfo" type="long, int, long, int, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getInputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Raw bytes read from the FileSystem into the task. Note that this
may not always match the input bytes to the task.]]>
</doc>
</method>
<method name="getInputRecords" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Number of records input to this task.]]>
</doc>
</method>
<method name="getOutputBytes" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Raw bytes written to the destination FileSystem. Note that this may
not match output bytes.]]>
</doc>
</method>
<method name="getOutputRecords" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Number of records output from this task.]]>
</doc>
</method>
<method name="getTaskMemory" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Memory used by the task leq the heap size.]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TaskInfo -->
<!-- start class org.apache.hadoop.tools.rumen.TopologyBuilder -->
<class name="TopologyBuilder" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TopologyBuilder"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="process"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="event" type="org.apache.hadoop.mapreduce.jobhistory.HistoryEvent"/>
<doc>
<![CDATA[Process one {@link HistoryEvent}
@param event
The {@link HistoryEvent} to be processed.]]>
</doc>
</method>
<method name="process"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="java.util.Properties"/>
<doc>
<![CDATA[Process a collection of JobConf {@link Properties}. We do not restrict it
to be called once.
@param conf
The job conf properties to be added.]]>
</doc>
</method>
<method name="build" return="org.apache.hadoop.tools.rumen.LoggedNetworkTopology"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request the builder to build the final object. Once called, the
{@link TopologyBuilder} would accept no more events or job-conf properties.
@return Parsed {@link LoggedNetworkTopology} object.]]>
</doc>
</method>
<doc>
<![CDATA[Building the cluster topology.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TopologyBuilder -->
<!-- start class org.apache.hadoop.tools.rumen.TraceBuilder -->
<class name="TraceBuilder" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="TraceBuilder"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[The main driver of the Rumen Parser.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TraceBuilder -->
<!-- start class org.apache.hadoop.tools.rumen.TreePath -->
<class name="TreePath" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TreePath" type="org.apache.hadoop.tools.rumen.TreePath, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TreePath" type="org.apache.hadoop.tools.rumen.TreePath, java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This describes a path from a node to the root. We use it when we compare two
trees during rumen unit tests. If the trees are not identical, this chain
will be converted to a string which describes the path from the root to the
fields that did not compare.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.TreePath -->
<!-- start class org.apache.hadoop.tools.rumen.ZombieCluster -->
<class name="ZombieCluster" extends="org.apache.hadoop.tools.rumen.AbstractClusterStory"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ZombieCluster" type="org.apache.hadoop.tools.rumen.LoggedNetworkTopology, org.apache.hadoop.tools.rumen.MachineNode"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a homogeneous cluster. We assume that the leaves on the topology
are {@link MachineNode}s, and the parents of {@link MachineNode}s are
{@link RackNode}s. We also expect all leaf nodes are on the same level.
@param topology
The network topology.
@param defaultNode
The default node setting.]]>
</doc>
</constructor>
<constructor name="ZombieCluster" type="org.apache.hadoop.fs.Path, org.apache.hadoop.tools.rumen.MachineNode, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a homogeneous cluster. We assume that the leaves on the topology
are {@link MachineNode}s, and the parents of {@link MachineNode}s are
{@link RackNode}s. We also expect all leaf nodes are on the same level.
@param path Path to the JSON-encoded topology file.
@param conf
@param defaultNode
The default node setting.
@throws IOException]]>
</doc>
</constructor>
<constructor name="ZombieCluster" type="java.io.InputStream, org.apache.hadoop.tools.rumen.MachineNode"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a homogeneous cluster. We assume that the leaves on the topology
are {@link MachineNode}s, and the parents of {@link MachineNode}s are
{@link RackNode}s. We also expect all leaf nodes are on the same level.
@param input The input stream for the JSON-encoded topology file.
@param defaultNode
The default node setting.
@throws IOException]]>
</doc>
</constructor>
<method name="getClusterTopology" return="org.apache.hadoop.tools.rumen.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[{@link ZombieCluster} rebuilds the cluster topology using the information
obtained from job history logs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.ZombieCluster -->
<!-- start class org.apache.hadoop.tools.rumen.ZombieJob -->
<class name="ZombieJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.JobStory"/>
<constructor name="ZombieJob" type="org.apache.hadoop.tools.rumen.LoggedJob, org.apache.hadoop.tools.rumen.ClusterStory, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This constructor creates a {@link ZombieJob} with the same semantics as the
{@link LoggedJob} passed in this parameter
@param job
The dead job this ZombieJob instance is based on.
@param cluster
The cluster topology where the dead job ran on. This argument can
be null if we do not have knowledge of the cluster topology.
@param seed
Seed for the random number generator for filling in information
not available from the ZombieJob.]]>
</doc>
</constructor>
<constructor name="ZombieJob" type="org.apache.hadoop.tools.rumen.LoggedJob, org.apache.hadoop.tools.rumen.ClusterStory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This constructor creates a {@link ZombieJob} with the same semantics as the
{@link LoggedJob} passed in this parameter
@param job
The dead job this ZombieJob instance is based on.
@param cluster
The cluster topology where the dead job ran on. This argument can
be null if we do not have knowledge of the cluster topology.]]>
</doc>
</constructor>
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInputSplits" return="org.apache.hadoop.mapreduce.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumberMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumberReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getOutcome" return="org.apache.hadoop.tools.rumen.Pre21JobHistoryConstants.Values"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSubmissionTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getQueueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumLoggedMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Getting the number of map tasks that are actually logged in the trace.
@return The number of map tasks that are actually logged in the trace.]]>
</doc>
</method>
<method name="getNumLoggedReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Getting the number of reduce tasks that are actually logged in the trace.
@return The number of map tasks that are actually logged in the trace.]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLoggedJob" return="org.apache.hadoop.tools.rumen.LoggedJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the underlining {@link LoggedJob} object read directly from the trace.
This is mainly for debugging.
@return the underlining {@link LoggedJob} object]]>
</doc>
</method>
<method name="getTaskAttemptInfo" return="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskType" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskNumber" type="int"/>
<param name="taskAttemptNumber" type="int"/>
<doc>
<![CDATA[Get a {@link TaskAttemptInfo} with a {@link TaskAttemptID} associated with
taskType, taskNumber, and taskAttemptNumber. This function does not care
about locality, and follows the following decision logic: 1. Make up a
{@link TaskAttemptInfo} if the task attempt is missing in trace, 2. Make up
a {@link TaskAttemptInfo} if the task attempt has a KILLED final status in
trace, 3. Otherwise (final state is SUCCEEDED or FAILED), construct the
{@link TaskAttemptInfo} from the trace.]]>
</doc>
</method>
<method name="getTaskInfo" return="org.apache.hadoop.tools.rumen.TaskInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskType" type="org.apache.hadoop.mapreduce.TaskType"/>
<param name="taskNumber" type="int"/>
</method>
<method name="getMapTaskAttemptInfoAdjusted" return="org.apache.hadoop.tools.rumen.TaskAttemptInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskNumber" type="int"/>
<param name="taskAttemptNumber" type="int"/>
<param name="locality" type="int"/>
<doc>
<![CDATA[Get a {@link TaskAttemptInfo} with a {@link TaskAttemptID} associated with
taskType, taskNumber, and taskAttemptNumber. This function considers
locality, and follows the following decision logic: 1. Make up a
{@link TaskAttemptInfo} if the task attempt is missing in trace, 2. Make up
a {@link TaskAttemptInfo} if the task attempt has a KILLED final status in
trace, 3. If final state is FAILED, construct a {@link TaskAttemptInfo}
from the trace, without considering locality. 4. If final state is
SUCCEEDED, construct a {@link TaskAttemptInfo} from the trace, with runtime
scaled according to locality in simulation and locality in trace.]]>
</doc>
</method>
<doc>
<![CDATA[{@link ZombieJob} is a layer above {@link LoggedJob} raw JSON objects.
Each {@link ZombieJob} object represents a job in job history. For everything
that exists in job history, contents are returned unchanged faithfully. To
get input splits of a non-exist task, a non-exist task attempt, or an
ill-formed task attempt, proper objects are made up from statistical
sketches.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.ZombieJob -->
<!-- start class org.apache.hadoop.tools.rumen.ZombieJobProducer -->
<class name="ZombieJobProducer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.tools.rumen.JobStoryProducer"/>
<constructor name="ZombieJobProducer" type="org.apache.hadoop.fs.Path, org.apache.hadoop.tools.rumen.ZombieCluster, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor
@param path
Path to the JSON trace file, possibly compressed.
@param cluster
The topology of the cluster that corresponds to the jobs in the
trace. The argument can be null if we do not have knowledge of the
cluster topology.
@param conf
@throws IOException]]>
</doc>
</constructor>
<constructor name="ZombieJobProducer" type="org.apache.hadoop.fs.Path, org.apache.hadoop.tools.rumen.ZombieCluster, org.apache.hadoop.conf.Configuration, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor
@param path
Path to the JSON trace file, possibly compressed.
@param cluster
The topology of the cluster that corresponds to the jobs in the
trace. The argument can be null if we do not have knowledge of the
cluster topology.
@param conf
@param randomSeed
use a deterministic seed.
@throws IOException]]>
</doc>
</constructor>
<constructor name="ZombieJobProducer" type="java.io.InputStream, org.apache.hadoop.tools.rumen.ZombieCluster"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor
@param input
The input stream for the JSON trace.
@param cluster
The topology of the cluster that corresponds to the jobs in the
trace. The argument can be null if we do not have knowledge of the
cluster topology.
@throws IOException]]>
</doc>
</constructor>
<constructor name="ZombieJobProducer" type="java.io.InputStream, org.apache.hadoop.tools.rumen.ZombieCluster, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructor
@param input
The input stream for the JSON trace.
@param cluster
The topology of the cluster that corresponds to the jobs in the
trace. The argument can be null if we do not have knowledge of the
cluster topology.
@param randomSeed
use a deterministic seed.
@throws IOException]]>
</doc>
</constructor>
<method name="getNextJob" return="org.apache.hadoop.tools.rumen.ZombieJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Producing {@link JobStory}s from job trace.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.rumen.ZombieJobProducer -->
</package>
</api>