HBASE-18700 Document the new changes on mapreduce stuffs
Signed-off-by: Chia-Ping Tsai <chia7712@gmail.com>
This commit is contained in:
parent
73942e37da
commit
811b88a877
|
@ -27,7 +27,7 @@
|
||||||
:icons: font
|
:icons: font
|
||||||
:experimental:
|
:experimental:
|
||||||
|
|
||||||
Apache MapReduce is a software framework used to analyze large amounts of data, and is the framework used most often with link:http://hadoop.apache.org/[Apache Hadoop].
|
Apache MapReduce is a software framework used to analyze large amounts of data. It is provided by link:https://hadoop.apache.org/[Apache Hadoop].
|
||||||
MapReduce itself is out of the scope of this document.
|
MapReduce itself is out of the scope of this document.
|
||||||
A good place to get started with MapReduce is http://hadoop.apache.org/docs/r2.6.0/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html.
|
A good place to get started with MapReduce is http://hadoop.apache.org/docs/r2.6.0/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html.
|
||||||
MapReduce version 2 (MR2)is now part of link:http://hadoop.apache.org/docs/r2.3.0/hadoop-yarn/hadoop-yarn-site/[YARN].
|
MapReduce version 2 (MR2)is now part of link:http://hadoop.apache.org/docs/r2.3.0/hadoop-yarn/hadoop-yarn-site/[YARN].
|
||||||
|
@ -40,44 +40,88 @@ link:http://www.cascading.org/[alternative API] for MapReduce.
|
||||||
.`mapred` and `mapreduce`
|
.`mapred` and `mapreduce`
|
||||||
[NOTE]
|
[NOTE]
|
||||||
====
|
====
|
||||||
There are two mapreduce packages in HBase as in MapReduce itself: _org.apache.hadoop.hbase.mapred_ and _org.apache.hadoop.hbase.mapreduce_.
|
There are two mapreduce packages in HBase as in MapReduce itself: _org.apache.hadoop.hbase.mapred_ and _org.apache.hadoop.hbase.mapreduce_.
|
||||||
The former does old-style API and the latter the new style.
|
The former does old-style API and the latter the new mode.
|
||||||
The latter has more facility though you can usually find an equivalent in the older package.
|
The latter has more facility though you can usually find an equivalent in the older package.
|
||||||
Pick the package that goes with your MapReduce deploy.
|
Pick the package that goes with your MapReduce deploy.
|
||||||
When in doubt or starting over, pick the _org.apache.hadoop.hbase.mapreduce_.
|
When in doubt or starting over, pick _org.apache.hadoop.hbase.mapreduce_.
|
||||||
In the notes below, we refer to o.a.h.h.mapreduce but replace with the o.a.h.h.mapred if that is what you are using.
|
In the notes below, we refer to _o.a.h.h.mapreduce_ but replace with
|
||||||
|
_o.a.h.h.mapred_ if that is what you are using.
|
||||||
====
|
====
|
||||||
|
|
||||||
[[hbase.mapreduce.classpath]]
|
[[hbase.mapreduce.classpath]]
|
||||||
== HBase, MapReduce, and the CLASSPATH
|
== HBase, MapReduce, and the CLASSPATH
|
||||||
|
|
||||||
By default, MapReduce jobs deployed to a MapReduce cluster do not have access to either the HBase configuration under `$HBASE_CONF_DIR` or the HBase classes.
|
By default, MapReduce jobs deployed to a MapReduce cluster do not have access to
|
||||||
|
either the HBase configuration under `$HBASE_CONF_DIR` or the HBase classes.
|
||||||
|
|
||||||
To give the MapReduce jobs the access they need, you could add _hbase-site.xml_ to _$HADOOP_HOME/conf_ and add HBase jars to the _$HADOOP_HOME/lib_ directory.
|
To give the MapReduce jobs the access they need, you could add _hbase-site.xml_to _$HADOOP_HOME/conf_ and add HBase jars to the _$HADOOP_HOME/lib_ directory.
|
||||||
You would then need to copy these changes across your cluster. Or you can edit _$HADOOP_HOME/conf/hadoop-env.sh_ and add them to the `HADOOP_CLASSPATH` variable.
|
You would then need to copy these changes across your cluster. Or you could edit _$HADOOP_HOME/conf/hadoop-env.sh_ and add hbase dependencies to the `HADOOP_CLASSPATH` variable.
|
||||||
However, this approach is not recommended because it will pollute your Hadoop install with HBase references.
|
Neither of these approaches is recommended because it will pollute your Hadoop install with HBase references.
|
||||||
It also requires you to restart the Hadoop cluster before Hadoop can use the HBase data.
|
It also requires you restart the Hadoop cluster before Hadoop can use the HBase data.
|
||||||
|
|
||||||
The recommended approach is to let HBase add its dependency jars itself and use `HADOOP_CLASSPATH` or `-libjars`.
|
The recommended approach is to let HBase add its dependency jars and use `HADOOP_CLASSPATH` or `-libjars`.
|
||||||
|
|
||||||
Since HBase 0.90.x, HBase adds its dependency JARs to the job configuration itself.
|
Since HBase `0.90.x`, HBase adds its dependency JARs to the job configuration itself.
|
||||||
The dependencies only need to be available on the local `CLASSPATH`.
|
The dependencies only need to be available on the local `CLASSPATH` and from here they'll be picked
|
||||||
The following example runs the bundled HBase link:http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/mapreduce/RowCounter.html[RowCounter] MapReduce job against a table named `usertable`.
|
up and bundled into the fat job jar deployed to the MapReduce cluster. A basic trick just passes
|
||||||
If you have not set the environment variables expected in the command (the parts prefixed by a `$` sign and surrounded by curly braces), you can use the actual system paths instead.
|
the full hbase classpath -- all hbase and dependent jars as well as configurations -- to the mapreduce
|
||||||
Be sure to use the correct version of the HBase JAR for your system.
|
job runner letting hbase utility pick out from the full-on classpath what it needs adding them to the
|
||||||
The backticks (``` symbols) cause the shell to execute the sub-commands, setting the output of `hbase classpath` (the command to dump HBase CLASSPATH) to `HADOOP_CLASSPATH`.
|
MapReduce job configuration (See the source at `TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)` for how this is done).
|
||||||
|
|
||||||
|
|
||||||
|
The following example runs the bundled HBase link:https://hbase.apache.org/apidocs/org/apache/hadoop/hbase/mapreduce/RowCounter.html[RowCounter] MapReduce job against a table named `usertable`.
|
||||||
|
It sets into `HADOOP_CLASSPATH` the jars hbase needs to run in an MapReduce context (including configuration files such as hbase-site.xml).
|
||||||
|
Be sure to use the correct version of the HBase JAR for your system; replace the VERSION string in the below command line w/ the version of
|
||||||
|
your local hbase install. The backticks (``` symbols) cause the shell to execute the sub-commands, setting the output of `hbase classpath` into `HADOOP_CLASSPATH`.
|
||||||
This example assumes you use a BASH-compatible shell.
|
This example assumes you use a BASH-compatible shell.
|
||||||
|
|
||||||
[source,bash]
|
[source,bash]
|
||||||
----
|
----
|
||||||
$ HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` ${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/lib/hbase-mapreduce-VERSION.jar rowcounter usertable
|
$ HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` \
|
||||||
|
${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/lib/hbase-mapreduce-VERSION.jar \
|
||||||
|
org.apache.hadoop.hbase.mapreduce.RowCounter usertable
|
||||||
----
|
----
|
||||||
|
|
||||||
When the command runs, internally, the HBase JAR finds the dependencies it needs and adds them to the MapReduce job configuration.
|
The above command will launch a row counting mapreduce job against the hbase cluster that is pointed to by your local configuration on a cluster that the hadoop configs are pointing to.
|
||||||
See the source at `TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)` for how this is done.
|
|
||||||
|
The main for the `hbase-mapreduce.jar` is a Driver that lists a few basic mapreduce tasks that ship with hbase.
|
||||||
|
For example, presuming your install is hbase `2.0.0-SNAPSHOT`:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
|
----
|
||||||
|
$ HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` \
|
||||||
|
${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/lib/hbase-mapreduce-2.0.0-SNAPSHOT.jar
|
||||||
|
An example program must be given as the first argument.
|
||||||
|
Valid program names are:
|
||||||
|
CellCounter: Count cells in HBase table.
|
||||||
|
WALPlayer: Replay WAL files.
|
||||||
|
completebulkload: Complete a bulk data load.
|
||||||
|
copytable: Export a table from local cluster to peer cluster.
|
||||||
|
export: Write table data to HDFS.
|
||||||
|
exportsnapshot: Export the specific snapshot to a given FileSystem.
|
||||||
|
import: Import data written by Export.
|
||||||
|
importtsv: Import data in TSV format.
|
||||||
|
rowcounter: Count rows in HBase table.
|
||||||
|
verifyrep: Compare the data from tables in two different clusters. WARNING: It doesn't work for incrementColumnValues'd cells since the timestamp is changed after being appended to the log.
|
||||||
|
|
||||||
|
----
|
||||||
|
|
||||||
|
You can use the above listed shortnames for mapreduce jobs as in the below re-run of the row counter job (again, presuming your install is hbase `2.0.0-SNAPSHOT`):
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
|
----
|
||||||
|
$ HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` \
|
||||||
|
${HADOOP_HOME}/bin/hadoop jar ${HBASE_HOME}/lib/hbase-mapreduce-2.0.0-SNAPSHOT.jar \
|
||||||
|
rowcounter usertable
|
||||||
|
----
|
||||||
|
|
||||||
|
You might find the more selective `hbase mapredcp` tool output of interest; it lists the minimum set of jars needed
|
||||||
|
to run a basic mapreduce job against an hbase install. It does not include configuration. You'll probably need to add
|
||||||
|
these if you want your MapReduce job to find the target cluster. You'll probably have to also add pointers to extra jars
|
||||||
|
once you start to do anything of substance. Just specify the extras by passing the system propery `-Dtmpjars` when
|
||||||
|
you run `hbase mapredcp`.
|
||||||
|
|
||||||
The command `hbase mapredcp` can also help you dump the CLASSPATH entries required by MapReduce, which are the same jars `TableMapReduceUtil#addDependencyJars` would add.
|
|
||||||
You can add them together with HBase conf directory to `HADOOP_CLASSPATH`.
|
|
||||||
For jobs that do not package their dependencies or call `TableMapReduceUtil#addDependencyJars`, the following command structure is necessary:
|
For jobs that do not package their dependencies or call `TableMapReduceUtil#addDependencyJars`, the following command structure is necessary:
|
||||||
|
|
||||||
[source,bash]
|
[source,bash]
|
||||||
|
|
Loading…
Reference in New Issue