Fixes to site and moved getting started out of javadoc and into book
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1034230 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
433f7a3ff6
commit
72d855549a
|
@ -83,16 +83,25 @@
|
|||
|
||||
<chapter xml:id="getting_started">
|
||||
<title>Getting Started</title>
|
||||
<section >
|
||||
<title>Introduction</title>
|
||||
<para>
|
||||
<link linkend="quickstart">Quick Start</link> will get you up and running
|
||||
on a single-node instance of HBase using the local filesystem.
|
||||
The <link linkend="notsoquick">Not-so-quick Start Guide</link>
|
||||
describes setup of HBase in distributed mode running on top of HDFS.
|
||||
</para>
|
||||
</section>
|
||||
|
||||
<section xml:id="quickstart">
|
||||
<title>Quick Start</title>
|
||||
|
||||
<para><itemizedlist>
|
||||
<para>Here is a quick guide to starting up a standalone HBase
|
||||
instance (an HBase instance that uses the local filesystem rather than
|
||||
Hadoop HDFS), creating a table and inserting rows into a table via the
|
||||
instance that uses the local filesystem. It leads you
|
||||
through creating a table, inserting rows via the
|
||||
<link linkend="shell">HBase Shell</link>, and then cleaning up and shutting
|
||||
down your running instance. The below exercise should take no more than
|
||||
down your instance. The below exercise should take no more than
|
||||
ten minutes (not including download time).
|
||||
</para>
|
||||
|
||||
|
@ -101,7 +110,7 @@
|
|||
|
||||
<para>Choose a download site from this list of <link
|
||||
xlink:href="http://www.apache.org/dyn/closer.cgi/hbase/">Apache
|
||||
Download Mirrors</link>. Click on it. This will take you to a
|
||||
Download Mirrors</link>. Click on suggested top link. This will take you to a
|
||||
mirror of <emphasis>HBase Releases</emphasis>. Click on
|
||||
the folder named <filename>stable</filename> and then download the
|
||||
file that ends in <filename>.tar.gz</filename> to your local filesystem;
|
||||
|
@ -146,7 +155,7 @@ starting master, logging to logs/hbase-user-master-example.org.out</programlisti
|
|||
|
||||
<note>
|
||||
<title>Is <application>java</application> installed?</title>
|
||||
<para>The above presumes a 1.6 version of Oracle
|
||||
<para>The above presumes a 1.6 version of SUN
|
||||
<application>java</application> is installed on your
|
||||
machine and available on your path; i.e. when you type
|
||||
<application>java</application>, you see output that describes the options
|
||||
|
@ -257,6 +266,7 @@ stopping hbase...............</programlisting></para>
|
|||
|
||||
<section xml:id="notsoquick">
|
||||
<title>Not-so-quick Start Guide</title>
|
||||
|
||||
<section xml:id="requirements"><title>Requirements</title>
|
||||
<para>HBase has the following requirements. Please read the
|
||||
section below carefully and ensure that all requirements have been
|
||||
|
@ -271,7 +281,8 @@ Usually you'll want to use the latest version available except the problematic u
|
|||
</section>
|
||||
<section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link></title>
|
||||
<para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>.
|
||||
HBase will lose data unless it is running on an HDFS that has a durable sync.
|
||||
It will not run on hadoop 0.21.x as of this writing.
|
||||
HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>.
|
||||
Currently only the <link xlink:href="http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-append/">branch-0.20-append</link>
|
||||
branch has this attribute. No official releases have been made from this branch as of this writing
|
||||
so you will have to build your own Hadoop from the tip of this branch
|
||||
|
@ -297,6 +308,7 @@ Usually you'll want to use the latest version available except the problematic u
|
|||
</para>
|
||||
</section>
|
||||
|
||||
|
||||
<section xml:id="ulimit">
|
||||
<title><varname>ulimit</varname></title>
|
||||
<para>HBase is a database, it uses a lot of files at the same time.
|
||||
|
@ -330,27 +342,328 @@ Usually you'll want to use the latest version available except the problematic u
|
|||
</programlisting>
|
||||
</para>
|
||||
</section>
|
||||
|
||||
<section xml:id="windows">
|
||||
<title>Windows</title>
|
||||
<para>
|
||||
If you are running HBase on Windows, you must install
|
||||
<link xlink:href="http://cygwin.com/">Cygwin</link>
|
||||
to have a *nix-like environment for the shell scripts. The full details
|
||||
are explained in the <link xlink:href="../cygwin.html">Windows Installation</link>
|
||||
guide.
|
||||
</para>
|
||||
</section>
|
||||
|
||||
</section>
|
||||
|
||||
<section><title>HBase run modes: Standalone, Pseudo-distributed, and Distributed</title>
|
||||
<para>HBase has three different run modes: standalone, this is what is described above in
|
||||
<link linkend="quickstart">Quick Start,</link> pseudo-distributed mode where all
|
||||
daemons run on a single server, and distributed, where each of the daemons runs
|
||||
on different cluster node.</para>
|
||||
<section><title>Standalone HBase</title>
|
||||
<para>TODO</para>
|
||||
</section>
|
||||
<section><title>Pseudo-distributed</title>
|
||||
<para>TODO</para>
|
||||
<section><title>HBase run modes: Standalone and Distributed</title>
|
||||
<para>HBase has two run modes: <link linkend="standalone">standalone</link>
|
||||
and <link linkend="distributed">distributed</link>.</para>
|
||||
|
||||
<para>Whatever your mode, define <code>${HBASE_HOME}</code> to be the location of the root of your HBase installation, e.g.
|
||||
<code>/user/local/hbase</code>. Edit <code>${HBASE_HOME}/conf/hbase-env.sh</code>. In this file you can
|
||||
set the heapsize for HBase, etc. At a minimum, set <code>JAVA_HOME</code> to point at the root of
|
||||
your Java installation.</para>
|
||||
|
||||
<section xml:id="standalone"><title>Standalone HBase</title>
|
||||
<para>This mode is what <link linkend="quickstart">Quick Start</link> covered;
|
||||
all daemons are run in the one JVM and HBase writes the local filesystem.</para>
|
||||
</section>
|
||||
<section><title>Distributed</title>
|
||||
<para>TODO</para>
|
||||
<para>Distributed mode can be subdivided into distributed but all daemons run on a
|
||||
single node AND distibuted with daemons spread across all nodes in the cluster.</para>
|
||||
<para>
|
||||
Distributed modes require an instance of the <emphasis>Hadoop Distributed File System</emphasis> (HDFS).
|
||||
See the Hadoop <link xlink:href="http://hadoop.apache.org/common/docs/current/api/overview-summary.html#overview_description">
|
||||
requirements and instructions</link> for how to set up a HDFS.
|
||||
</para>
|
||||
|
||||
|
||||
|
||||
<section xml:id="pseudo"><title>Pseudo-distributed</title>
|
||||
<para>A pseudo-distributed mode is simply a distributed mode run on a single host.
|
||||
Use this configuration testing and prototyping on hbase. Do not use this configuration
|
||||
for production nor for evaluating HBase performance.
|
||||
</para>
|
||||
<para>Once you have confirmed your HDFS setup, configuring HBase for use on one host requires modification of
|
||||
<filename>./conf/hbase-site.xml</filename>, which needs to be pointed at the running Hadoop HDFS instance.
|
||||
Use <filename>hbase-site.xml</filename> to override the properties defined in
|
||||
<filename>conf/hbase-default.xml</filename> (<filename>hbase-default.xml</filename> itself
|
||||
should never be modified) and for HDFS client configurations.
|
||||
At a minimum, the <varname>hbase.rootdir</varname>,
|
||||
which points HBase at the Hadoop filesystem to use,
|
||||
should be redefined in <filename>hbase-site.xml</filename>. For example,
|
||||
adding the properties below to your <filename>hbase-site.xml</filename> says that HBase
|
||||
should use the <filename>/hbase</filename>
|
||||
directory in the HDFS whose namenode is at port 9000 on your local machine, and that
|
||||
it should run with one replica only (recommended for pseudo-distributed mode):</para>
|
||||
<programlisting>
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.rootdir</name>
|
||||
<value>hdfs://localhost:9000/hbase</value>
|
||||
<description>The directory shared by region servers.
|
||||
</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.replication</name>
|
||||
<value>1</value>
|
||||
<description>The replication count for HLog & HFile storage. Should not be greater than HDFS datanode count.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</programlisting>
|
||||
|
||||
<note>
|
||||
<para>Let HBase create the directory. If you don't, you'll get warning saying HBase
|
||||
needs a migration run because the directory is missing files expected by HBase (it'll
|
||||
create them if you let it).</para>
|
||||
</note>
|
||||
|
||||
<note>
|
||||
<para>Above we bind to localhost. This means that a remote client cannot
|
||||
connect. Amend accordingly, if you want to connect from a remote location.</para>
|
||||
</note>
|
||||
</section>
|
||||
|
||||
<section><title>Distributed across multiple machines</title>
|
||||
|
||||
|
||||
<para>For running a fully-distributed operation on more than one host, the following
|
||||
configurations must be made <emphasis>in addition</emphasis> to those described in the
|
||||
<link linkend="pseudo">pseudo-distributed operation</link> section above.</para>
|
||||
|
||||
<para>In <filename>hbase-site.xml</filename>, set <varname>hbase.cluster.distributed</varname> to <varname>true</varname>.</para>
|
||||
<programlisting>
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.cluster.distributed</name>
|
||||
<value>true</value>
|
||||
<description>The mode the cluster will be in. Possible values are
|
||||
false: standalone and pseudo-distributed setups with managed Zookeeper
|
||||
true: fully-distributed with unmanaged Zookeeper Quorum (see hbase-env.sh)
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</programlisting>
|
||||
|
||||
<para>In fully-distributed mode, you probably want to change your <varname>hbase.rootdir</varname>
|
||||
from localhost to the name of the node running the HDFS NameNode and you should set
|
||||
the dfs.replication to be the number of datanodes you have in your cluster or 3, which
|
||||
ever is the smaller.
|
||||
</para>
|
||||
<para>In addition
|
||||
to <filename>hbase-site.xml</filename> changes, a fully-distributed mode requires that you
|
||||
modify <filename>${HBASE_HOME}/conf/regionservers</filename>.
|
||||
The <filename>regionserver</filename> file lists all hosts running <application>HRegionServer</application>s, one host per line
|
||||
(This file in HBase is like the Hadoop slaves file at <filename>${HADOOP_HOME}/conf/slaves</filename>).</para>
|
||||
|
||||
<para>A distributed HBase depends on a running ZooKeeper cluster. All participating nodes and clients
|
||||
need to be able to get to the running ZooKeeper cluster.
|
||||
HBase by default manages a ZooKeeper cluster for you, or you can manage it on your own and point HBase to it.
|
||||
To toggle HBase management of ZooKeeper, use the <varname>HBASE_MANAGES_ZK</varname> variable in <filename>${HBASE_HOME}/conf/hbase-env.sh</filename>.
|
||||
This variable, which defaults to <varname>true</varname>, tells HBase whether to
|
||||
start/stop the ZooKeeper quorum servers alongside the rest of the servers.</para>
|
||||
|
||||
<para>When HBase manages the ZooKeeper cluster, you can specify ZooKeeper configuration
|
||||
using its canonical <filename>zoo.cfg</filename> file (see below), or
|
||||
just specify ZookKeeper options directly in the <filename>${HBASE_HOME}/conf/hbase-site.xml</filename>
|
||||
(If new to ZooKeeper, go the path of specifying your configuration in HBase's hbase-site.xml).
|
||||
Every ZooKeeper configuration option has a corresponding property in the HBase hbase-site.xml
|
||||
XML configuration file named <varname>hbase.zookeeper.property.OPTION</varname>.
|
||||
For example, the <varname>clientPort</varname> setting in ZooKeeper can be changed by
|
||||
setting the <varname>hbase.zookeeper.property.clientPort</varname> property.
|
||||
For the full list of available properties, see ZooKeeper's <filename>zoo.cfg</filename>.
|
||||
For the default values used by HBase, see <filename>${HBASE_HOME}/conf/hbase-default.xml</filename>.</para>
|
||||
|
||||
<para>At minimum, you should set the list of servers that you want ZooKeeper to run
|
||||
on using the <varname>hbase.zookeeper.quorum</varname> property.
|
||||
This property defaults to <varname>localhost</varname> which is not suitable for a
|
||||
fully distributed HBase (it binds to the local machine only and remote clients
|
||||
will not be able to connect).
|
||||
It is recommended to run a ZooKeeper quorum of 3, 5 or 7 machines, and give each
|
||||
ZooKeeper server around 1GB of RAM, and if possible, its own dedicated disk.
|
||||
For very heavily loaded clusters, run ZooKeeper servers on separate machines from the
|
||||
Region Servers (DataNodes and TaskTrackers).</para>
|
||||
|
||||
|
||||
<para>To point HBase at an existing ZooKeeper cluster, add
|
||||
a suitably configured <filename>zoo.cfg</filename> to the <filename>CLASSPATH</filename>.
|
||||
HBase will see this file and use it to figure out where ZooKeeper is.
|
||||
Additionally set <varname>HBASE_MANAGES_ZK</varname> in <filename>${HBASE_HOME}/conf/hbase-env.sh</filename>
|
||||
to <filename>false</filename> so that HBase doesn't mess with your ZooKeeper setup:</para>
|
||||
<programlisting>
|
||||
...
|
||||
# Tell HBase whether it should manage it's own instance of Zookeeper or not.
|
||||
export HBASE_MANAGES_ZK=false
|
||||
</programlisting>
|
||||
|
||||
<para>As an example, to have HBase manage a ZooKeeper quorum on nodes
|
||||
<empahsis>rs{1,2,3,4,5}.example.com</empahsis>, bound to port 2222 (the default is 2181), use:</para>
|
||||
<programlisting>
|
||||
${HBASE_HOME}/conf/hbase-env.sh:
|
||||
|
||||
...
|
||||
# Tell HBase whether it should manage it's own instance of Zookeeper or not.
|
||||
export HBASE_MANAGES_ZK=true
|
||||
|
||||
${HBASE_HOME}/conf/hbase-site.xml:
|
||||
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.zookeeper.property.clientPort</name>
|
||||
<value>2222</value>
|
||||
<description>Property from ZooKeeper's config zoo.cfg.
|
||||
The port at which the clients will connect.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.zookeeper.quorum</name>
|
||||
<value>rs1.example.com,rs2.example.com,rs3.example.com,rs4.example.com,rs5.example.com</value>
|
||||
<description>Comma separated list of servers in the ZooKeeper Quorum.
|
||||
For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
|
||||
By default this is set to localhost for local and pseudo-distributed modes
|
||||
of operation. For a fully-distributed setup, this should be set to a full
|
||||
list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
|
||||
this is the list of servers which we will start/stop ZooKeeper on.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</programlisting>
|
||||
|
||||
<para>When HBase manages ZooKeeper, it will start/stop the ZooKeeper servers as a part
|
||||
of the regular start/stop scripts. If you would like to run it yourself, you can
|
||||
do:</para>
|
||||
<programlisting>
|
||||
${HBASE_HOME}/bin/hbase-daemons.sh {start,stop} zookeeper
|
||||
</programlisting>
|
||||
|
||||
<para>If you do let HBase manage ZooKeeper for you, make sure you configure
|
||||
where it's data is stored. By default, it will be stored in /tmp which is
|
||||
sometimes cleaned in live systems. Do modify this configuration:</para>
|
||||
<programlisting>
|
||||
<property>
|
||||
<name>hbase.zookeeper.property.dataDir</name>
|
||||
<value>${hbase.tmp.dir}/zookeeper</value>
|
||||
<description>Property from ZooKeeper's config zoo.cfg.
|
||||
The directory where the snapshot is stored.
|
||||
</description>
|
||||
</property>
|
||||
</programlisting>
|
||||
|
||||
<para>Note that you can use HBase in this manner to spin up a ZooKeeper cluster,
|
||||
unrelated to HBase. Just make sure to set <varname>HBASE_MANAGES_ZK</varname> to
|
||||
<varname>false</varname> if you want it to stay up so that when HBase shuts down it
|
||||
doesn't take ZooKeeper with it.</para>
|
||||
|
||||
<para>For more information about setting up a ZooKeeper cluster on your own, see
|
||||
the ZooKeeper <link xlink:href="http://hadoop.apache.org/zookeeper/docs/current/zookeeperStarted.html">Getting Started Guide</link>.
|
||||
HBase currently uses ZooKeeper version 3.3.2, so any cluster setup with a
|
||||
3.x.x version of ZooKeeper should work.</para>
|
||||
|
||||
<para>Of note, if you have made <em>HDFS client configuration</em> on your Hadoop cluster, HBase will not
|
||||
see this configuration unless you do one of the following:</para>
|
||||
<orderedlist>
|
||||
<listitem><para>Add a pointer to your <varname>HADOOP_CONF_DIR</varname> to <varname>CLASSPATH</varname> in <filename>hbase-env.sh</filename>.</para></listitem>
|
||||
<listitem><para>Add a copy of <filename>hdfs-site.xml</filename> (or <filename>hadoop-site.xml</filename>) to <filename>${HBASE_HOME}/conf</filename>, or</para></listitem>
|
||||
<listitem><para>if only a small set of HDFS client configurations, add them to <filename>hbase-site.xml</filename>.</para></listitem>
|
||||
</orderedlist>
|
||||
|
||||
<para>An example of such an HDFS client configuration is <varname>dfs.replication</varname>. If for example,
|
||||
you want to run with a replication factor of 5, hbase will create files with the default of 3 unless
|
||||
you do the above to make the configuration available to HBase.</para>
|
||||
</section>
|
||||
|
||||
<section xml:id="confirm"><title>Running and Confirming Your Installation</title>
|
||||
<para>If you are running in standalone, non-distributed mode, HBase by default uses the local filesystem.</para>
|
||||
|
||||
<para>If you are running a distributed cluster you will need to start the Hadoop DFS daemons and
|
||||
ZooKeeper Quorum before starting HBase and stop the daemons after HBase has shut down.</para>
|
||||
|
||||
<para>Start and stop the Hadoop DFS daemons by running <filename>${HADOOP_HOME}/bin/start-dfs.sh</filename>.
|
||||
You can ensure it started properly by testing the put and get of files into the Hadoop filesystem.
|
||||
HBase does not normally use the mapreduce daemons. These do not need to be started.</para>
|
||||
|
||||
<para>Start up your ZooKeeper cluster.</para>
|
||||
|
||||
<para>Start HBase with the following command:</para>
|
||||
<programlisting>
|
||||
${HBASE_HOME}/bin/start-hbase.sh
|
||||
</programlisting>
|
||||
|
||||
<para>Once HBase has started, enter <filename>${HBASE_HOME}/bin/hbase shell</filename> to obtain a
|
||||
shell against HBase from which you can execute commands.
|
||||
Type 'help' at the shells' prompt to get a list of commands.
|
||||
Test your running install by creating tables, inserting content, viewing content, and then dropping your tables.
|
||||
For example:</para>
|
||||
<programlisting>
|
||||
hbase> # Type "help" to see shell help screen
|
||||
hbase> help
|
||||
hbase> # To create a table named "mylittletable" with a column family of "mylittlecolumnfamily", type
|
||||
hbase> create "mylittletable", "mylittlecolumnfamily"
|
||||
hbase> # To see the schema for you just created "mylittletable" table and its single "mylittlecolumnfamily", type
|
||||
hbase> describe "mylittletable"
|
||||
hbase> # To add a row whose id is "myrow", to the column "mylittlecolumnfamily:x" with a value of 'v', do
|
||||
hbase> put "mylittletable", "myrow", "mylittlecolumnfamily:x", "v"
|
||||
hbase> # To get the cell just added, do
|
||||
hbase> get "mylittletable", "myrow"
|
||||
hbase> # To scan you new table, do
|
||||
hbase> scan "mylittletable"
|
||||
</programlisting>
|
||||
|
||||
<para>To stop HBase, exit the HBase shell and enter:</para>
|
||||
<programlisting>
|
||||
${HBASE_HOME}/bin/stop-hbase.sh
|
||||
</programlisting>
|
||||
|
||||
<para>If you are running a distributed operation, be sure to wait until HBase has shut down completely
|
||||
before stopping the Hadoop daemons.</para>
|
||||
|
||||
<para>The default location for logs is <filename>${HBASE_HOME}/logs</filename>.</para>
|
||||
|
||||
<para>HBase also puts up a UI listing vital attributes. By default its deployed on the master host
|
||||
at port 60010 (HBase RegionServers listen on port 60020 by default and put up an informational
|
||||
http server at 60030).</para>
|
||||
</section>
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
</section>
|
||||
</section>
|
||||
|
||||
|
||||
|
||||
<section><title>Client configuration and dependencies connecting to an HBase cluster</title>
|
||||
<para>TODO</para>
|
||||
</section>
|
||||
|
||||
|
||||
<section xml:id="upgrading">
|
||||
<title>Upgrading your HBase Install</title>
|
||||
<para>This version of 0.90.x HBase can be started on data written by
|
||||
HBase 0.20.x or HBase 0.89.x. There is no need of a migration step.
|
||||
HBase 0.89.x and 0.90.x does write out the name of region directories
|
||||
differently -- it names them with a md5 hash of the region name rather
|
||||
than a jenkins hash -- so this means that once started, there is no
|
||||
going back to HBase 0.20.x.
|
||||
</para>
|
||||
</section>
|
||||
|
||||
|
||||
|
||||
|
||||
<section><title>Example Configurations</title>
|
||||
<para>In this section we provide a few sample configurations.</para>
|
||||
<section><title>Basic Distributed HBase Install</title>
|
||||
|
@ -366,7 +679,7 @@ Below we show what the main configuration files
|
|||
<filename>hbase-env.sh</filename> -- found in the <filename>conf</filename> directory
|
||||
might look like.
|
||||
</para>
|
||||
<section><title><filename>hbase-site.xml</filename></title>
|
||||
<section xml:id="hbase_site"><title><filename>hbase-site.xml</filename></title>
|
||||
<programlisting>
|
||||
<![CDATA[
|
||||
<?xml version="1.0"?>
|
||||
|
@ -404,7 +717,7 @@ might look like.
|
|||
</programlisting>
|
||||
</section>
|
||||
|
||||
<section><title><filename>regionservers</filename></title>
|
||||
<section xml:id="regionservers"><title><filename>regionservers</filename></title>
|
||||
<para>In this file you list the nodes that will run regionservers. In
|
||||
our case we run regionservers on all but the head node example1 which is
|
||||
carrying the HBase master and the HDFS namenode</para>
|
||||
|
@ -420,7 +733,7 @@ might look like.
|
|||
</programlisting>
|
||||
</section>
|
||||
|
||||
<section><title><filename>hbase-env.sh</filename></title>
|
||||
<section xml:id="hbase_env"><title><filename>hbase-env.sh</filename></title>
|
||||
<para>Below we use a <command>diff</command> to show the differences from
|
||||
default in the <filename>hbase-env.sh</filename> file. Here we are setting
|
||||
the HBase heap to be 4G instead of the default 1G.
|
||||
|
@ -487,7 +800,7 @@ index e70ebc6..96f8c27 100644
|
|||
<para></para>
|
||||
</section>
|
||||
|
||||
<section>
|
||||
<section xml:id="log4j">
|
||||
<title><filename>log4j.properties</filename></title>
|
||||
<para></para>
|
||||
</section>
|
||||
|
@ -569,7 +882,7 @@ index e70ebc6..96f8c27 100644
|
|||
</para>
|
||||
</section>
|
||||
|
||||
<section><title>Shell Tricks</title>
|
||||
<section xml:id="shell_tricks"><title>Shell Tricks</title>
|
||||
<section><title><filename>irbrc</filename></title>
|
||||
<para>Create an <filename>.irbrc</filename> file for yourself in your
|
||||
home directory. Add HBase Shell customizations. A useful one is
|
||||
|
@ -639,13 +952,13 @@ index e70ebc6..96f8c27 100644
|
|||
via the table row key -- its primary key.
|
||||
</para>
|
||||
|
||||
<section>
|
||||
<section xml:id="table">
|
||||
<title>Table</title>
|
||||
|
||||
<para></para>
|
||||
</section>
|
||||
|
||||
<section>
|
||||
<section xml:id="row">
|
||||
<title>Row</title>
|
||||
|
||||
<para></para>
|
||||
|
@ -874,7 +1187,7 @@ index e70ebc6..96f8c27 100644
|
|||
|
||||
<subtitle>How HBase is persisted on the Filesystem</subtitle>
|
||||
|
||||
<section>
|
||||
<section xml:id="hfile">
|
||||
<title>HFile</title>
|
||||
|
||||
<section xml:id="hfile_tool">
|
||||
|
@ -1524,8 +1837,8 @@ index e70ebc6..96f8c27 100644
|
|||
</section>
|
||||
</chapter>
|
||||
|
||||
<chapter>
|
||||
<title xml:id="wal">The WAL</title>
|
||||
<chapter xml:id="wal">
|
||||
<title >The WAL</title>
|
||||
|
||||
<subtitle>HBase's<link
|
||||
xlink:href="http://en.wikipedia.org/wiki/Write-ahead_logging"> Write-Ahead
|
||||
|
@ -1540,7 +1853,7 @@ index e70ebc6..96f8c27 100644
|
|||
<para>The HBase WAL is...</para>
|
||||
</section>
|
||||
|
||||
<section>
|
||||
<section xml:id="wal_splitting">
|
||||
<title>WAL splitting</title>
|
||||
|
||||
<subtitle>How edits are recovered from a crashed RegionServer</subtitle>
|
||||
|
@ -1584,8 +1897,8 @@ index e70ebc6..96f8c27 100644
|
|||
|
||||
</chapter>
|
||||
|
||||
<chapter>
|
||||
<title xml:id="blooms">Bloom Filters</title>
|
||||
<chapter xml:id="blooms">
|
||||
<title>Bloom Filters</title>
|
||||
|
||||
<para>Bloom filters were developed over in <link
|
||||
xlink:href="https://issues.apache.org/jira/browse/HBASE-1200">HBase-1200
|
||||
|
@ -1658,7 +1971,7 @@ index e70ebc6..96f8c27 100644
|
|||
</section>
|
||||
</section>
|
||||
|
||||
<section>
|
||||
<section xml:id="bloom_footprint">
|
||||
<title>Bloom StoreFile footprint</title>
|
||||
|
||||
<para>Bloom filters add an entry to the <classname>StoreFile</classname>
|
||||
|
@ -1791,8 +2104,38 @@ index e70ebc6..96f8c27 100644
|
|||
</section>
|
||||
</appendix>
|
||||
|
||||
<appendix xml:id="faq">
|
||||
<title >FAQ</title>
|
||||
<qandaset defaultlabel='faq'>
|
||||
<qandadiv><title>General</title>
|
||||
<qandaentry>
|
||||
<question><para>Are there other HBase FAQs?</para></question>
|
||||
<answer>
|
||||
<para>
|
||||
See the FAQ that is up on the wiki, <link xlink:href="http://wiki.apache.org/hadoop/Hbase/FAQ">HBase Wiki FAQ</link>
|
||||
as well as the <link xlink:href="http://wiki.apache.org/hadoop/Hbase/Troubleshooting">Troubleshooting</link> page and
|
||||
the <link xlink:href="http://wiki.apache.org/hadoop/Hbase/FrequentlySeenErrors">Frequently Seen Errors</link> page.
|
||||
</para>
|
||||
</answer>
|
||||
</qandaentry>
|
||||
</qandadiv>
|
||||
<qandadiv xml:id="ec2"><title>EC2</title>
|
||||
<qandaentry>
|
||||
<question><para>
|
||||
Why doesn't my remote java connection into my ec2 cluster work?
|
||||
</para></question>
|
||||
<answer>
|
||||
<para>
|
||||
See Andrew's answer here, up on the user list: <link xlink:href="http://search-hadoop.com/m/sPdqNFAwyg2">Remote Java client connection into EC2 instance</link>.
|
||||
</para>
|
||||
</answer>
|
||||
</qandaentry>
|
||||
</qandadiv>
|
||||
</qandaset>
|
||||
</appendix>
|
||||
|
||||
<index>
|
||||
|
||||
<index xml:id="book_index">
|
||||
<title>Index</title>
|
||||
</index>
|
||||
</book>
|
||||
|
|
|
@ -1193,7 +1193,7 @@ public class KeyValue implements Writable, HeapSize {
|
|||
* changed to be null). This method does a full copy of the backing byte
|
||||
* array and does not modify the original byte array of this KeyValue.
|
||||
* <p>
|
||||
* This method is used by {@link KeyOnlyFilter} and is an advanced feature of
|
||||
* This method is used by <code>KeyOnlyFilter</code> and is an advanced feature of
|
||||
* KeyValue, proceed with caution.
|
||||
*/
|
||||
public void convertToKeyOnly() {
|
||||
|
|
|
@ -62,7 +62,7 @@ is set to the HBase <code>CLASSPATH</code> via backticking the command
|
|||
etc., dependencies on the passed
|
||||
</code>HADOOP_CLASSPATH</code> and adds the found jars to the mapreduce
|
||||
job configuration. See the source at
|
||||
{@link TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)}
|
||||
<code>TableMapReduceUtil#addDependencyJars(org.apache.hadoop.mapreduce.Job)</code>
|
||||
for how this is done.
|
||||
</p>
|
||||
<p>The above may not work if you are running your HBase from its build directory;
|
||||
|
|
|
@ -2103,6 +2103,7 @@ public class HRegionServer implements HRegionInterface, HBaseRPCErrorHandler,
|
|||
list.add(e.getValue().getRegionInfo());
|
||||
}
|
||||
}
|
||||
Collections.sort(list);
|
||||
return list;
|
||||
}
|
||||
|
||||
|
|
|
@ -40,7 +40,7 @@ import org.apache.hadoop.hbase.util.Bytes;
|
|||
import org.apache.zookeeper.KeeperException;
|
||||
|
||||
/**
|
||||
* Gateway to Replication. Used by {@link HRegionServer}.
|
||||
* Gateway to Replication. Used by {@link org.apache.hadoop.hbase.regionserver.HRegionServer}.
|
||||
*/
|
||||
public class Replication implements WALObserver {
|
||||
private final boolean replication;
|
||||
|
|
|
@ -74,7 +74,7 @@ public abstract class User {
|
|||
/**
|
||||
* Returns the shortened version of the user name -- the portion that maps
|
||||
* to an operating system user name.
|
||||
* @return
|
||||
* @return Short name
|
||||
*/
|
||||
public abstract String getShortName();
|
||||
|
||||
|
|
|
@ -26,352 +26,30 @@
|
|||
|
||||
<h2>Table of Contents</h2>
|
||||
<ul>
|
||||
<li>
|
||||
<a href="#requirements">Requirements</a>
|
||||
<ul>
|
||||
<li><a href="#windows">Windows</a></li>
|
||||
</ul>
|
||||
</li>
|
||||
<li>
|
||||
<a href="#getting_started" >Getting Started</a>
|
||||
<ul>
|
||||
<li><a href="#standalone">Standalone</a></li>
|
||||
<li>
|
||||
<a href="#distributed">Distributed Operation: Pseudo- and Fully-distributed modes</a>
|
||||
<ul>
|
||||
<li><a href="#pseudo-distrib">Pseudo-distributed</a></li>
|
||||
<li><a href="#fully-distrib">Fully-distributed</a></li>
|
||||
</ul>
|
||||
</li>
|
||||
</ul>
|
||||
</li>
|
||||
<li><a href="#runandconfirm">Running and Confirming Your Installation</a></li>
|
||||
<li><a href="#upgrading" >Upgrading</a></li>
|
||||
<li><a href="#client_example">Example API Usage</a></li>
|
||||
<li><a href="#related" >Related Documentation</a></li>
|
||||
</ul>
|
||||
|
||||
|
||||
<h2><a name="getting_started" >Getting Started</a></h2>
|
||||
<p>First review the <a href="http://hbase.apache.org/docs/current/book.html#requirements">requirements</a>
|
||||
section of the HBase Book. A careful reading will save you grief down the road.</p>
|
||||
|
||||
<p>What follows presumes you have obtained a copy of HBase,
|
||||
see <a href="http://hadoop.apache.org/hbase/releases.html">Releases</a>, and are installing
|
||||
for the first time. If upgrading your HBase instance, see <a href="#upgrading">Upgrading</a>.</p>
|
||||
|
||||
<p>Three modes are described: <em>standalone</em>, <em>pseudo-distributed</em> (where all servers are run on
|
||||
a single host), and <em>fully-distributed</em>. If new to HBase start by following the standalone instructions.</p>
|
||||
|
||||
<p>Begin by reading <a href="#requirements">Requirements</a>.</p>
|
||||
|
||||
<p>Whatever your mode, define <code>${HBASE_HOME}</code> to be the location of the root of your HBase installation, e.g.
|
||||
<code>/user/local/hbase</code>. Edit <code>${HBASE_HOME}/conf/hbase-env.sh</code>. In this file you can
|
||||
set the heapsize for HBase, etc. At a minimum, set <code>JAVA_HOME</code> to point at the root of
|
||||
your Java installation.</p>
|
||||
|
||||
<h3><a name="standalone">Standalone mode</a></h3>
|
||||
<p>If you are running a standalone operation, there should be nothing further to configure; proceed to
|
||||
<a href="#runandconfirm">Running and Confirming Your Installation</a>. If you are running a distributed
|
||||
operation, continue reading.</p>
|
||||
|
||||
<h3><a name="distributed">Distributed Operation: Pseudo- and Fully-distributed modes</a></h3>
|
||||
<p>Distributed modes require an instance of the <em>Hadoop Distributed File System</em> (DFS).
|
||||
See the Hadoop <a href="http://hadoop.apache.org/common/docs/r0.20.1/api/overview-summary.html#overview_description">
|
||||
requirements and instructions</a> for how to set up a DFS.</p>
|
||||
|
||||
<h4><a name="pseudo-distrib">Pseudo-distributed mode</a></h4>
|
||||
<p>A pseudo-distributed mode is simply a distributed mode run on a single host.
|
||||
Use this configuration testing and prototyping on hbase. Do not use this configuration
|
||||
for production nor for evaluating HBase performance.
|
||||
<p>See the <a href="../book.html#getting_started">Getting Started</a>
|
||||
section of the <a href="../book.html">HBase Book</a>.
|
||||
</p>
|
||||
<p>Once you have confirmed your DFS setup, configuring HBase for use on one host requires modification of
|
||||
<code>${HBASE_HOME}/conf/hbase-site.xml</code>, which needs to be pointed at the running Hadoop DFS instance.
|
||||
Use <code>hbase-site.xml</code> to override the properties defined in
|
||||
<code>${HBASE_HOME}/conf/hbase-default.xml</code> (<code>hbase-default.xml</code> itself
|
||||
should never be modified) and for HDFS client configurations.
|
||||
At a minimum, the <code>hbase.rootdir</code>,
|
||||
which points HBase at the Hadoop filesystem to use,
|
||||
and the <code>dfs.replication</code>, an hdfs client-side
|
||||
configuration stipulating how many replicas to keep up,
|
||||
should be redefined in <code>hbase-site.xml</code>. For example,
|
||||
adding the properties below to your <code>hbase-site.xml</code> says that HBase
|
||||
should use the <code>/hbase</code>
|
||||
directory in the HDFS whose namenode is at port 9000 on your local machine, and that
|
||||
it should run with one replica only (recommended for pseudo-distributed mode):</p>
|
||||
<blockquote>
|
||||
<pre>
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.rootdir</name>
|
||||
<value>hdfs://localhost:9000/hbase</value>
|
||||
<description>The directory shared by region servers.
|
||||
</description>
|
||||
</property>
|
||||
<property>
|
||||
<name>dfs.replication</name>
|
||||
<value>1</value>
|
||||
<description>The replication count for HLog & HFile storage. Should not be greater than HDFS datanode count.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>Note: Let HBase create the directory. If you don't, you'll get warning saying HBase
|
||||
needs a migration run because the directory is missing files expected by HBase (it'll
|
||||
create them if you let it).</p>
|
||||
<p>Also Note: Above we bind to localhost. This means that a remote client cannot
|
||||
connect. Amend accordingly, if you want to connect from a remote location.</p>
|
||||
|
||||
<h4><a name="fully-distrib">Fully-Distributed Operation</a></h4>
|
||||
<p>For running a fully-distributed operation on more than one host, the following
|
||||
configurations must be made <em>in addition</em> to those described in the
|
||||
<a href="#pseudo-distrib">pseudo-distributed operation</a> section above.</p>
|
||||
|
||||
<p>In <code>hbase-site.xml</code>, set <code>hbase.cluster.distributed</code> to <code>true</code>.</p>
|
||||
<blockquote>
|
||||
<pre>
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.cluster.distributed</name>
|
||||
<value>true</value>
|
||||
<description>The mode the cluster will be in. Possible values are
|
||||
false: standalone and pseudo-distributed setups with managed Zookeeper
|
||||
true: fully-distributed with unmanaged Zookeeper Quorum (see hbase-env.sh)
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>In fully-distributed mode, you probably want to change your <code>hbase.rootdir</code>
|
||||
from localhost to the name of the node running the HDFS NameNode and you should set
|
||||
the dfs.replication to be the number of datanodes you have in your cluster or 3, which
|
||||
ever is the smaller.
|
||||
</p>
|
||||
<p>In addition
|
||||
to <code>hbase-site.xml</code> changes, a fully-distributed mode requires that you
|
||||
modify <code>${HBASE_HOME}/conf/regionservers</code>.
|
||||
The <code>regionserver</code> file lists all hosts running <code>HRegionServer</code>s, one host per line
|
||||
(This file in HBase is like the Hadoop slaves file at <code>${HADOOP_HOME}/conf/slaves</code>).</p>
|
||||
|
||||
<p>A distributed HBase depends on a running ZooKeeper cluster. All participating nodes and clients
|
||||
need to be able to get to the running ZooKeeper cluster.
|
||||
HBase by default manages a ZooKeeper cluster for you, or you can manage it on your own and point HBase to it.
|
||||
To toggle HBase management of ZooKeeper, use the <code>HBASE_MANAGES_ZK</code> variable in <code>${HBASE_HOME}/conf/hbase-env.sh</code>.
|
||||
This variable, which defaults to <code>true</code>, tells HBase whether to
|
||||
start/stop the ZooKeeper quorum servers alongside the rest of the servers.</p>
|
||||
|
||||
<p>When HBase manages the ZooKeeper cluster, you can specify ZooKeeper configuration
|
||||
using its canonical <code>zoo.cfg</code> file (see below), or
|
||||
just specify ZookKeeper options directly in the <code>${HBASE_HOME}/conf/hbase-site.xml</code>
|
||||
(If new to ZooKeeper, go the path of specifying your configuration in HBase's hbase-site.xml).
|
||||
Every ZooKeeper configuration option has a corresponding property in the HBase hbase-site.xml
|
||||
XML configuration file named <code>hbase.zookeeper.property.OPTION</code>.
|
||||
For example, the <code>clientPort</code> setting in ZooKeeper can be changed by
|
||||
setting the <code>hbase.zookeeper.property.clientPort</code> property.
|
||||
For the full list of available properties, see ZooKeeper's <code>zoo.cfg</code>.
|
||||
For the default values used by HBase, see <code>${HBASE_HOME}/conf/hbase-default.xml</code>.</p>
|
||||
|
||||
<p>At minimum, you should set the list of servers that you want ZooKeeper to run
|
||||
on using the <code>hbase.zookeeper.quorum</code> property.
|
||||
This property defaults to <code>localhost</code> which is not suitable for a
|
||||
fully distributed HBase (it binds to the local machine only and remote clients
|
||||
will not be able to connect).
|
||||
It is recommended to run a ZooKeeper quorum of 3, 5 or 7 machines, and give each
|
||||
ZooKeeper server around 1GB of RAM, and if possible, its own dedicated disk.
|
||||
For very heavily loaded clusters, run ZooKeeper servers on separate machines from the
|
||||
Region Servers (DataNodes and TaskTrackers).</p>
|
||||
|
||||
<p>To point HBase at an existing ZooKeeper cluster, add
|
||||
a suitably configured <code>zoo.cfg</code> to the <code>CLASSPATH</code>.
|
||||
HBase will see this file and use it to figure out where ZooKeeper is.
|
||||
Additionally set <code>HBASE_MANAGES_ZK</code> in <code>${HBASE_HOME}/conf/hbase-env.sh</code>
|
||||
to <code>false</code> so that HBase doesn't mess with your ZooKeeper setup:</p>
|
||||
<blockquote>
|
||||
<pre>
|
||||
...
|
||||
# Tell HBase whether it should manage it's own instance of Zookeeper or not.
|
||||
export HBASE_MANAGES_ZK=false
|
||||
</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>As an example, to have HBase manage a ZooKeeper quorum on nodes
|
||||
<em>rs{1,2,3,4,5}.example.com</em>, bound to port 2222 (the default is 2181), use:</p>
|
||||
<blockquote>
|
||||
<pre>
|
||||
${HBASE_HOME}/conf/hbase-env.sh:
|
||||
|
||||
...
|
||||
# Tell HBase whether it should manage it's own instance of Zookeeper or not.
|
||||
export HBASE_MANAGES_ZK=true
|
||||
|
||||
${HBASE_HOME}/conf/hbase-site.xml:
|
||||
|
||||
<configuration>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.zookeeper.property.clientPort</name>
|
||||
<value>2222</value>
|
||||
<description>Property from ZooKeeper's config zoo.cfg.
|
||||
The port at which the clients will connect.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
<property>
|
||||
<name>hbase.zookeeper.quorum</name>
|
||||
<value>rs1.example.com,rs2.example.com,rs3.example.com,rs4.example.com,rs5.example.com</value>
|
||||
<description>Comma separated list of servers in the ZooKeeper Quorum.
|
||||
For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
|
||||
By default this is set to localhost for local and pseudo-distributed modes
|
||||
of operation. For a fully-distributed setup, this should be set to a full
|
||||
list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
|
||||
this is the list of servers which we will start/stop ZooKeeper on.
|
||||
</description>
|
||||
</property>
|
||||
...
|
||||
</configuration>
|
||||
</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>When HBase manages ZooKeeper, it will start/stop the ZooKeeper servers as a part
|
||||
of the regular start/stop scripts. If you would like to run it yourself, you can
|
||||
do:</p>
|
||||
<blockquote>
|
||||
<pre>${HBASE_HOME}/bin/hbase-daemons.sh {start,stop} zookeeper</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>If you do let HBase manage ZooKeeper for you, make sure you configure
|
||||
where it's data is stored. By default, it will be stored in /tmp which is
|
||||
sometimes cleaned in live systems. Do modify this configuration:</p>
|
||||
<pre>
|
||||
<property>
|
||||
<name>hbase.zookeeper.property.dataDir</name>
|
||||
<value>${hbase.tmp.dir}/zookeeper</value>
|
||||
<description>Property from ZooKeeper's config zoo.cfg.
|
||||
The directory where the snapshot is stored.
|
||||
</description>
|
||||
</property>
|
||||
|
||||
</pre>
|
||||
|
||||
<p>Note that you can use HBase in this manner to spin up a ZooKeeper cluster,
|
||||
unrelated to HBase. Just make sure to set <code>HBASE_MANAGES_ZK</code> to
|
||||
<code>false</code> if you want it to stay up so that when HBase shuts down it
|
||||
doesn't take ZooKeeper with it.</p>
|
||||
|
||||
<p>For more information about setting up a ZooKeeper cluster on your own, see
|
||||
the ZooKeeper <a href="http://hadoop.apache.org/zookeeper/docs/current/zookeeperStarted.html">Getting Started Guide</a>.
|
||||
HBase currently uses ZooKeeper version 3.3.1, so any cluster setup with a
|
||||
3.x.x version of ZooKeeper should work.</p>
|
||||
|
||||
<p>Of note, if you have made <em>HDFS client configuration</em> on your Hadoop cluster, HBase will not
|
||||
see this configuration unless you do one of the following:</p>
|
||||
<ul>
|
||||
<li>Add a pointer to your <code>HADOOP_CONF_DIR</code> to <code>CLASSPATH</code> in <code>hbase-env.sh</code>.</li>
|
||||
<li>Add a copy of <code>hdfs-site.xml</code> (or <code>hadoop-site.xml</code>) to <code>${HBASE_HOME}/conf</code>, or</li>
|
||||
<li>if only a small set of HDFS client configurations, add them to <code>hbase-site.xml</code>.</li>
|
||||
</ul>
|
||||
|
||||
<p>An example of such an HDFS client configuration is <code>dfs.replication</code>. If for example,
|
||||
you want to run with a replication factor of 5, hbase will create files with the default of 3 unless
|
||||
you do the above to make the configuration available to HBase.</p>
|
||||
|
||||
|
||||
<h2><a name="runandconfirm">Running and Confirming Your Installation</a></h2>
|
||||
<p>If you are running in standalone, non-distributed mode, HBase by default uses the local filesystem.</p>
|
||||
|
||||
<p>If you are running a distributed cluster you will need to start the Hadoop DFS daemons and
|
||||
ZooKeeper Quorum before starting HBase and stop the daemons after HBase has shut down.</p>
|
||||
|
||||
<p>Start and stop the Hadoop DFS daemons by running <code>${HADOOP_HOME}/bin/start-dfs.sh</code>.
|
||||
You can ensure it started properly by testing the put and get of files into the Hadoop filesystem.
|
||||
HBase does not normally use the mapreduce daemons. These do not need to be started.</p>
|
||||
|
||||
<p>Start up your ZooKeeper cluster.</p>
|
||||
|
||||
<p>Start HBase with the following command:</p>
|
||||
<blockquote>
|
||||
<pre>${HBASE_HOME}/bin/start-hbase.sh</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>Once HBase has started, enter <code>${HBASE_HOME}/bin/hbase shell</code> to obtain a
|
||||
shell against HBase from which you can execute commands.
|
||||
Type 'help' at the shells' prompt to get a list of commands.
|
||||
Test your running install by creating tables, inserting content, viewing content, and then dropping your tables.
|
||||
For example:</p>
|
||||
<blockquote>
|
||||
<pre>
|
||||
hbase> # Type "help" to see shell help screen
|
||||
hbase> help
|
||||
hbase> # To create a table named "mylittletable" with a column family of "mylittlecolumnfamily", type
|
||||
hbase> create "mylittletable", "mylittlecolumnfamily"
|
||||
hbase> # To see the schema for you just created "mylittletable" table and its single "mylittlecolumnfamily", type
|
||||
hbase> describe "mylittletable"
|
||||
hbase> # To add a row whose id is "myrow", to the column "mylittlecolumnfamily:x" with a value of 'v', do
|
||||
hbase> put "mylittletable", "myrow", "mylittlecolumnfamily:x", "v"
|
||||
hbase> # To get the cell just added, do
|
||||
hbase> get "mylittletable", "myrow"
|
||||
hbase> # To scan you new table, do
|
||||
hbase> scan "mylittletable"
|
||||
</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>To stop HBase, exit the HBase shell and enter:</p>
|
||||
<blockquote>
|
||||
<pre>${HBASE_HOME}/bin/stop-hbase.sh</pre>
|
||||
</blockquote>
|
||||
|
||||
<p>If you are running a distributed operation, be sure to wait until HBase has shut down completely
|
||||
before stopping the Hadoop daemons.</p>
|
||||
|
||||
<p>The default location for logs is <code>${HBASE_HOME}/logs</code>.</p>
|
||||
|
||||
<p>HBase also puts up a UI listing vital attributes. By default its deployed on the master host
|
||||
at port 60010 (HBase RegionServers listen on port 60020 by default and put up an informational
|
||||
http server at 60030).</p>
|
||||
|
||||
<h2><a name="upgrading" >Upgrading</a></h2>
|
||||
<p>After installing a new HBase on top of data written by a previous HBase version, before
|
||||
starting your cluster, run the <code>${HBASE_DIR}/bin/hbase migrate</code> migration script.
|
||||
It will make any adjustments to the filesystem data under <code>hbase.rootdir</code> necessary to run
|
||||
the HBase version. It does not change your install unless you explicitly ask it to.</p>
|
||||
|
||||
<h2><a name="client_example">Example API Usage</a></h2>
|
||||
<p>For sample Java code, see <a href="org/apache/hadoop/hbase/client/package-summary.html#package_description">org.apache.hadoop.hbase.client</a> documentation.</p>
|
||||
|
||||
<p>If your client is NOT Java, consider the Thrift or REST libraries.</p>
|
||||
|
||||
<h2><a name="windows">Windows</a></h2>
|
||||
If you are running HBase on Windows, you must install
|
||||
<a href="http://cygwin.com/">Cygwin</a>
|
||||
to have a *nix-like environment for the shell scripts. The full details
|
||||
are explained in
|
||||
the <a href="../cygwin.html">Windows Installation</a>
|
||||
guide.
|
||||
|
||||
<h2><a name="related" >Related Documentation</a></h2>
|
||||
<ul>
|
||||
<li><a href="http://hbase.org">HBase Home Page</a>
|
||||
</li>
|
||||
<li><a href="http://wiki.apache.org/hadoop/Hbase">HBase Wiki</a>
|
||||
</li>
|
||||
<li><a href="http://hadoop.apache.org/">Hadoop Home Page</a>
|
||||
</li>
|
||||
<li><a href="http://wiki.apache.org/hadoop/Hbase/MultipleMasters">Setting up Multiple HBase Masters</a>
|
||||
</li>
|
||||
<li><a href="http://wiki.apache.org/hadoop/Hbase/RollingRestart">Rolling Upgrades</a>
|
||||
</li>
|
||||
<li><a href="org/apache/hadoop/hbase/client/transactional/package-summary.html#package_description">Transactional HBase</a>
|
||||
</li>
|
||||
<li><a href="org/apache/hadoop/hbase/client/tableindexed/package-summary.html">Table Indexed HBase</a>
|
||||
</li>
|
||||
<li><a href="org/apache/hadoop/hbase/stargate/package-summary.html#package_description">Stargate</a> -- a RESTful Web service front end for HBase.
|
||||
<li><a href="http://hbase.org">HBase Home Page</a> </li>
|
||||
<li><a href="http://hbase.org/docs/current/book.html">HBase Book</a> </li>
|
||||
<li><a href="http://wiki.apache.org/hadoop/Hbase">HBase Wiki</a> </li>
|
||||
<li><a href="http://hadoop.apache.org/">Hadoop Home Page</a> </li>
|
||||
</li>
|
||||
</ul>
|
||||
|
||||
|
|
|
@ -1,36 +0,0 @@
|
|||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<!--
|
||||
Copyright 2010 The Apache Software Foundation
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
-->
|
||||
<faqs id="faqs">
|
||||
<part id="General">
|
||||
<faq id="otherfaqs">
|
||||
<question>Are there other HBase FAQs?</question>
|
||||
<answer>
|
||||
<p>See the FAQ that is up on the wiki, <a href="http://wiki.apache.org/hadoop/Hbase/FAQ">HBase Wiki FAQ</a>
|
||||
as well as the <a href="http://wiki.apache.org/hadoop/Hbase/Troubleshooting">Troubleshooting</a> page and
|
||||
the <a href="http://wiki.apache.org/hadoop/Hbase/FrequentlySeenErrors">Frequently Seen Errors</a> page.</p>
|
||||
</answer>
|
||||
</faq>
|
||||
</part>
|
||||
<part id="ec2">
|
||||
<faq id="ec2ips">
|
||||
<question>Why doesn't my remote java connection into my ec2 cluster work?</question>
|
||||
<answer>
|
||||
<p>See Andrew's answer here, up on the user list: <a href="http://search-hadoop.com/m/sPdqNFAwyg2">Remote Java client connection into EC2 instance</a></p>
|
||||
</answer>
|
||||
</faq>
|
||||
</part>
|
||||
</faqs>
|
|
@ -15,7 +15,7 @@
|
|||
<version position="right" />
|
||||
<publishDate position="right" />
|
||||
<body>
|
||||
<menu name="HBase">
|
||||
<menu name="HBase Project">
|
||||
<item name="Overview" href="index.html"/>
|
||||
<item name="License" href="license.html" />
|
||||
<item name="Downloads" href="http://www.apache.org/dyn/closer.cgi/hbase/" />
|
||||
|
@ -23,21 +23,22 @@
|
|||
<item name="Issue Tracking" href="issue-tracking.html" />
|
||||
<item name="Mailing Lists" href="mail-lists.html" />
|
||||
<item name="Source Repository" href="source-repository.html" />
|
||||
<item name="FAQ" href="faq.html" />
|
||||
<item name="Wiki" href="http://wiki.apache.org/hadoop/Hbase" />
|
||||
<item name="Team" href="team-list.html" />
|
||||
</menu>
|
||||
<menu name="Documentation">
|
||||
<item name="Getting Started" href="apidocs/overview-summary.html#overview_description" />
|
||||
<item name="Getting Started: Quick" href="quickstart.html" />
|
||||
<item name="Getting Started: Detailed" href="notsoquick.html" />
|
||||
<item name="API" href="apidocs/index.html" />
|
||||
<item name="X-Ref" href="xref/index.html" />
|
||||
<item name="Book" href="book.html" />
|
||||
<item name="FAQ" href="faq.html" />
|
||||
<item name="Wiki" href="http://wiki.apache.org/hadoop/Hbase" />
|
||||
<item name="ACID Semantics" href="acid-semantics.html" />
|
||||
<item name="Bulk Loads" href="bulk-loads.html" />
|
||||
<item name="Metrics" href="metrics.html" />
|
||||
<item name="HBase on Windows" href="cygwin.html" />
|
||||
<item name="Cluster replication" href="replication.html" />
|
||||
<item name="Pseudo-Distributed HBase" href="pseudo-distributed.html" />
|
||||
<item name="HBase Book" href="book.html" />
|
||||
<item name="Pseudo-Dist. Extras" href="pseudo-distributed.html" />
|
||||
</menu>
|
||||
</body>
|
||||
<skin>
|
||||
|
@ -45,4 +46,3 @@
|
|||
<artifactId>maven-stylus-skin</artifactId>
|
||||
</skin>
|
||||
</project>
|
||||
|
||||
|
|
|
@ -53,9 +53,7 @@ HBase includes:
|
|||
<p>November 15-19th, <a href="http://www.devoxx.com/display/Devoxx2K10/Home">Devoxx</a> features HBase Training and multiple HBase presentations</p>
|
||||
<p>October 12th, HBase-related presentations by core contributors and users at <a href="http://www.cloudera.com/company/press-center/hadoop-world-nyc/">Hadoop World 2010</a></p>
|
||||
<p>October 11th, <a href="http://www.meetup.com/hbaseusergroup/calendar/14606174/">HUG-NYC: HBase User Group NYC Edition</a> (Night before Hadoop World)</p>
|
||||
<p>June 30th, <a href="http://www.meetup.com/hbaseusergroup/calendar/13562846/">HBase Contributor Workshop</a> (Day after Hadoop Summit)</p>
|
||||
<p>May 10th, 2010: HBase graduates from Hadoop sub-project to Apache Top Level Project </p>
|
||||
<p><a href="old_news.html">...</a></p>
|
||||
<p><small><a href="old_news.html">Old News</a></small></p>
|
||||
</section>
|
||||
</body>
|
||||
|
||||
|
|
|
@ -28,6 +28,8 @@
|
|||
</properties>
|
||||
<body>
|
||||
<section name="Old News">
|
||||
<p>June 30th, <a href="http://www.meetup.com/hbaseusergroup/calendar/13562846/">HBase Contributor Workshop</a> (Day after Hadoop Summit)</p>
|
||||
<p>May 10th, 2010: HBase graduates from Hadoop sub-project to Apache Top Level Project </p>
|
||||
<p>Signup for <a href="http://www.meetup.com/hbaseusergroup/calendar/12689490/">HBase User Group Meeting, HUG10</a> hosted by Trend Micro, April 19th, 2010</p>
|
||||
|
||||
<p><a href="http://www.meetup.com/hbaseusergroup/calendar/12689351/">HBase User Group Meeting, HUG9</a> hosted by Mozilla, March 10th, 2010</p>
|
||||
|
|
Loading…
Reference in New Issue