Added more to the book index

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1062517 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael Stack 2011-01-23 20:28:39 +00:00
parent 77e26964b8
commit b24dd7e8a3
1 changed files with 13 additions and 8 deletions

View File

@ -285,7 +285,7 @@ stopping hbase...............</programlisting></para>
Usually you'll want to use the latest version available except the problematic u18 (u22 is the latest version as of this writing).</para> Usually you'll want to use the latest version available except the problematic u18 (u22 is the latest version as of this writing).</para>
</section> </section>
<section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link></title> <section xml:id="hadoop"><title><link xlink:href="http://hadoop.apache.org">hadoop</link><indexterm><primary>Hadoop</primary></indexterm></title>
<para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>. <para>This version of HBase will only run on <link xlink:href="http://hadoop.apache.org/common/releases.html">Hadoop 0.20.x</link>.
It will not run on hadoop 0.21.x (nor 0.22.x) as of this writing. It will not run on hadoop 0.21.x (nor 0.22.x) as of this writing.
HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>. HBase will lose data unless it is running on an HDFS that has a durable <code>sync</code>.
@ -343,7 +343,7 @@ be running to use Hadoop's scripts to manage remote Hadoop and HBase daemons.
<section xml:id="ulimit"> <section xml:id="ulimit">
<title><varname>ulimit</varname></title> <title><varname>ulimit</varname><indexterm><primary>ulimit</primary></indexterm></title>
<para>HBase is a database, it uses a lot of files at the same time. <para>HBase is a database, it uses a lot of files at the same time.
The default ulimit -n of 1024 on *nix systems is insufficient. The default ulimit -n of 1024 on *nix systems is insufficient.
Any significant amount of loading will lead you to Any significant amount of loading will lead you to
@ -390,7 +390,7 @@ be running to use Hadoop's scripts to manage remote Hadoop and HBase daemons.
</section> </section>
<section xml:id="dfs.datanode.max.xcievers"> <section xml:id="dfs.datanode.max.xcievers">
<title><varname>dfs.datanode.max.xcievers</varname></title> <title><varname>dfs.datanode.max.xcievers</varname><indexterm><primary>xcievers</primary></indexterm></title>
<para> <para>
An Hadoop HDFS datanode has an upper bound on the number of files An Hadoop HDFS datanode has an upper bound on the number of files
that it will serve at any one time. that it will serve at any one time.
@ -1060,13 +1060,14 @@ to ensure well-formedness of your document after an edit session.
HBase ships with a reasonable, conservative configuration that will HBase ships with a reasonable, conservative configuration that will
work on nearly all work on nearly all
machine types that people might want to test with. If you have larger machine types that people might want to test with. If you have larger
machines you might the following configuration options helpful. machines -- HBase has 8G and larger heap -- you might the following configuration options helpful.
TODO.
</para> </para>
</section> </section>
<section xml:id="lzo"> <section xml:id="lzo">
<title>LZO compression</title> <title>LZO compression<indexterm><primary>LZO</primary></indexterm></title>
<para>You should consider enabling LZO compression. Its <para>You should consider enabling LZO compression. Its
near-frictionless and in most all cases boosts performance. near-frictionless and in most all cases boosts performance.
</para> </para>
@ -1886,10 +1887,14 @@ of all regions.
doing:<programlisting> $ ./<code>bin/hbase org.apache.hadoop.hbase.regionserver.wal.HLog --split hdfs://example.org:9000/hbase/.logs/example.org,60020,1283516293161/</code></programlisting></para> doing:<programlisting> $ ./<code>bin/hbase org.apache.hadoop.hbase.regionserver.wal.HLog --split hdfs://example.org:9000/hbase/.logs/example.org,60020,1283516293161/</code></programlisting></para>
</section> </section>
</section> </section>
<section><title>Compression Tool</title>
<para>See <link linkend="compression.tool" >Compression Tool</link>.</para>
</section>
</appendix> </appendix>
<appendix xml:id="compression"> <appendix xml:id="compression">
<title >Compression In HBase</title> <title >Compression In HBase<indexterm><primary>Compression</primary></indexterm></title>
<section id="compression.test"> <section id="compression.test">
<title>CompressionTest Tool</title> <title>CompressionTest Tool</title>
@ -1947,7 +1952,7 @@ of all regions.
available on the CLASSPATH; in this case it will use native available on the CLASSPATH; in this case it will use native
compressors instead (If the native libs are NOT present, compressors instead (If the native libs are NOT present,
you will see lots of <emphasis>Got brand-new compressor</emphasis> you will see lots of <emphasis>Got brand-new compressor</emphasis>
reports in your logs; TO BE FIXED). reports in your logs; see <link linkend="brand.new.compressor">FAQ</link>).
</para> </para>
</section> </section>
</appendix> </appendix>
@ -1966,7 +1971,7 @@ of all regions.
</para> </para>
</answer> </answer>
</qandaentry> </qandaentry>
<qandaentry> <qandaentry xml:id="brand.new.compressor">
<question><para>Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got <question><para>Why are logs flooded with '2011-01-10 12:40:48,407 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new compressor' messages?</para></question> brand-new compressor' messages?</para></question>
<answer> <answer>