Added more to the book index
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1062517 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
77e26964b8
commit
b24dd7e8a3
|
@ -285,7 +285,7 @@ stopping hbase...............</programlisting></para>
|
||||||
Usually you'll want to use the latest version available except the problematic u18 (u22 is the latest version as of this writing).</para>
|
Usually you'll want to use the latest version available except the problematic u18 (u22 is the latest version as of this writing).</para>
|
||||||
</section>
|
</section>
|
||||||
|
|
||||||
<section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link></title>
|
<section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link><indexterm><primary>Hadoop</primary></indexterm></title>
|
||||||
<para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>.
|
<para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>.
|
||||||
It will not run on hadoop 0.21.x (nor 0.22.x) as of this writing.
|
It will not run on hadoop 0.21.x (nor 0.22.x) as of this writing.
|
||||||
HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>.
|
HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>.
|
||||||
|
@ -343,7 +343,7 @@ be running to use Hadoop's scripts to manage remote Hadoop and HBase daemons.
|
||||||
|
|
||||||
|
|
||||||
<section xml:id="ulimit">
|
<section xml:id="ulimit">
|
||||||
<title><varname>ulimit</varname></title>
|
<title><varname>ulimit</varname><indexterm><primary>ulimit</primary></indexterm></title>
|
||||||
<para>HBase is a database, it uses a lot of files at the same time.
|
<para>HBase is a database, it uses a lot of files at the same time.
|
||||||
The default ulimit -n of 1024 on *nix systems is insufficient.
|
The default ulimit -n of 1024 on *nix systems is insufficient.
|
||||||
Any significant amount of loading will lead you to
|
Any significant amount of loading will lead you to
|
||||||
|
@ -390,7 +390,7 @@ be running to use Hadoop's scripts to manage remote Hadoop and HBase daemons.
|
||||||
</section>
|
</section>
|
||||||
|
|
||||||
<section xml:id="dfs.datanode.max.xcievers">
|
<section xml:id="dfs.datanode.max.xcievers">
|
||||||
<title><varname>dfs.datanode.max.xcievers</varname></title>
|
<title><varname>dfs.datanode.max.xcievers</varname><indexterm><primary>xcievers</primary></indexterm></title>
|
||||||
<para>
|
<para>
|
||||||
An Hadoop HDFS datanode has an upper bound on the number of files
|
An Hadoop HDFS datanode has an upper bound on the number of files
|
||||||
that it will serve at any one time.
|
that it will serve at any one time.
|
||||||
|
@ -1060,13 +1060,14 @@ to ensure well-formedness of your document after an edit session.
|
||||||
HBase ships with a reasonable, conservative configuration that will
|
HBase ships with a reasonable, conservative configuration that will
|
||||||
work on nearly all
|
work on nearly all
|
||||||
machine types that people might want to test with. If you have larger
|
machine types that people might want to test with. If you have larger
|
||||||
machines you might the following configuration options helpful.
|
machines -- HBase has 8G and larger heap -- you might the following configuration options helpful.
|
||||||
|
TODO.
|
||||||
</para>
|
</para>
|
||||||
|
|
||||||
</section>
|
</section>
|
||||||
|
|
||||||
<section xml:id="lzo">
|
<section xml:id="lzo">
|
||||||
<title>LZO compression</title>
|
<title>LZO compression<indexterm><primary>LZO</primary></indexterm></title>
|
||||||
<para>You should consider enabling LZO compression. Its
|
<para>You should consider enabling LZO compression. Its
|
||||||
near-frictionless and in most all cases boosts performance.
|
near-frictionless and in most all cases boosts performance.
|
||||||
</para>
|
</para>
|
||||||
|
@ -1886,10 +1887,14 @@ of all regions.
|
||||||
doing:<programlisting> $ ./<code>bin/hbase org.apache.hadoop.hbase.regionserver.wal.HLog --split hdfs://example.org:9000/hbase/.logs/example.org,60020,1283516293161/</code></programlisting></para>
|
doing:<programlisting> $ ./<code>bin/hbase org.apache.hadoop.hbase.regionserver.wal.HLog --split hdfs://example.org:9000/hbase/.logs/example.org,60020,1283516293161/</code></programlisting></para>
|
||||||
</section>
|
</section>
|
||||||
</section>
|
</section>
|
||||||
|
<section><title>Compression Tool</title>
|
||||||
|
<para>See <link linkend="compression.tool" >Compression Tool</link>.</para>
|
||||||
|
</section>
|
||||||
</appendix>
|
</appendix>
|
||||||
|
|
||||||
<appendix xml:id="compression">
|
<appendix xml:id="compression">
|
||||||
|
|
||||||
<title >Compression In HBase</title>
|
<title >Compression In HBase<indexterm><primary>Compression</primary></indexterm></title>
|
||||||
|
|
||||||
<section id="compression.test">
|
<section id="compression.test">
|
||||||
<title>CompressionTest Tool</title>
|
<title>CompressionTest Tool</title>
|
||||||
|
@ -1947,7 +1952,7 @@ of all regions.
|
||||||
available on the CLASSPATH; in this case it will use native
|
available on the CLASSPATH; in this case it will use native
|
||||||
compressors instead (If the native libs are NOT present,
|
compressors instead (If the native libs are NOT present,
|
||||||
you will see lots of <emphasis>Got brand-new compressor</emphasis>
|
you will see lots of <emphasis>Got brand-new compressor</emphasis>
|
||||||
reports in your logs; TO BE FIXED).
|
reports in your logs; see <link linkend="brand.new.compressor">FAQ</link>).
|
||||||
</para>
|
</para>
|
||||||
</section>
|
</section>
|
||||||
</appendix>
|
</appendix>
|
||||||
|
@ -1966,7 +1971,7 @@ of all regions.
|
||||||
</para>
|
</para>
|
||||||
</answer>
|
</answer>
|
||||||
</qandaentry>
|
</qandaentry>
|
||||||
<qandaentry>
|
<qandaentry xml:id="brand.new.compressor">
|
||||||
<question><para>Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got
|
<question><para>Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got
|
||||||
brand-new compressor' messages?</para></question>
|
brand-new compressor' messages?</para></question>
|
||||||
<answer>
|
<answer>
|
||||||
|
|
Loading…
Reference in New Issue