HBASE-2076 Many javadoc warnings
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@894511 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
bbefab59d6
commit
42411929ff
|
@ -258,6 +258,7 @@ Release 0.21.0 - Unreleased
|
||||||
HBASE-2062 Metrics documentation outdated (Lars George via JD)
|
HBASE-2062 Metrics documentation outdated (Lars George via JD)
|
||||||
HBASE-2045 Update trunk and branch zk to just-release 3.2.2.
|
HBASE-2045 Update trunk and branch zk to just-release 3.2.2.
|
||||||
HBASE-2074 Improvements to the hadoop-config script (Bassam Tabbara via Stack)
|
HBASE-2074 Improvements to the hadoop-config script (Bassam Tabbara via Stack)
|
||||||
|
HBASE-2076 Many javadoc warnings
|
||||||
|
|
||||||
NEW FEATURES
|
NEW FEATURES
|
||||||
HBASE-1901 "General" partitioner for "hbase-48" bulk (behind the api, write
|
HBASE-1901 "General" partitioner for "hbase-48" bulk (behind the api, write
|
||||||
|
|
|
@ -186,8 +186,8 @@ public class Scan implements Writable {
|
||||||
* @param minStamp minimum timestamp value, inclusive
|
* @param minStamp minimum timestamp value, inclusive
|
||||||
* @param maxStamp maximum timestamp value, exclusive
|
* @param maxStamp maximum timestamp value, exclusive
|
||||||
* @throws IOException if invalid time range
|
* @throws IOException if invalid time range
|
||||||
* @see {@link #setMaxVersions()}
|
* @see #setMaxVersions()
|
||||||
* @see {@link #setMaxVersions(int)}
|
* @see #setMaxVersions(int)
|
||||||
*/
|
*/
|
||||||
public Scan setTimeRange(long minStamp, long maxStamp)
|
public Scan setTimeRange(long minStamp, long maxStamp)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
|
@ -201,8 +201,8 @@ public class Scan implements Writable {
|
||||||
* and you want all versions returned, up the number of versions beyond the
|
* and you want all versions returned, up the number of versions beyond the
|
||||||
* defaut.
|
* defaut.
|
||||||
* @param timestamp version timestamp
|
* @param timestamp version timestamp
|
||||||
* @see {@link #setMaxVersions()}
|
* @see #setMaxVersions()
|
||||||
* @see {@link #setMaxVersions(int)}
|
* @see #setMaxVersions(int)
|
||||||
*/
|
*/
|
||||||
public Scan setTimeStamp(long timestamp) {
|
public Scan setTimeStamp(long timestamp) {
|
||||||
try {
|
try {
|
||||||
|
|
|
@ -54,10 +54,6 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
|
||||||
* loss of data due to region server failure can be tolerated (for example,
|
* loss of data due to region server failure can be tolerated (for example,
|
||||||
* because it is easy to rerun a bulk import).
|
* because it is easy to rerun a bulk import).
|
||||||
* </p>
|
* </p>
|
||||||
*
|
|
||||||
* <p>
|
|
||||||
* See also the {@link IndexBuilder} example.
|
|
||||||
* </p>
|
|
||||||
*/
|
*/
|
||||||
public class MultiTableOutputFormat extends OutputFormat<ImmutableBytesWritable, Writable> {
|
public class MultiTableOutputFormat extends OutputFormat<ImmutableBytesWritable, Writable> {
|
||||||
/** Set this to {@link #WAL_OFF} to turn off write-ahead logging (HLog) */
|
/** Set this to {@link #WAL_OFF} to turn off write-ahead logging (HLog) */
|
||||||
|
|
|
@ -138,12 +138,13 @@ To achieve total ordering, you will likely need to write a Partitioner
|
||||||
that is intimate with your tables key namespace and that knows how
|
that is intimate with your tables key namespace and that knows how
|
||||||
to distribute keys among the reducers so a total order is maintained. If your
|
to distribute keys among the reducers so a total order is maintained. If your
|
||||||
keys are distributed with some regularity across a defined key space -- i.e.
|
keys are distributed with some regularity across a defined key space -- i.e.
|
||||||
you know the start and end keys -- then the {@link SimpleTotalOrderPartitioner}
|
you know the start and end keys -- then the
|
||||||
|
{@link org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner}
|
||||||
may be all you need.
|
may be all you need.
|
||||||
</p>
|
</p>
|
||||||
<p>See org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat for an example
|
<p>See org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat for an example
|
||||||
that puts together {@link org.apache.hadoop.hbase.mapreduce.KeyValueSortReducer},
|
that puts together {@link org.apache.hadoop.hbase.mapreduce.KeyValueSortReducer},
|
||||||
{@link SimpleTotalOrderPartitioner}, and
|
{@link org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner}, and
|
||||||
{@link org.apache.hadoop.hbase.mapreduce.HFileOutputFormat}.</p>
|
{@link org.apache.hadoop.hbase.mapreduce.HFileOutputFormat}.</p>
|
||||||
|
|
||||||
<p>HFileOutputFormat writes HFiles. When your MapReduce file finishes, in your
|
<p>HFileOutputFormat writes HFiles. When your MapReduce file finishes, in your
|
||||||
|
|
|
@ -51,7 +51,7 @@ public interface InternalScanner extends Closeable {
|
||||||
/**
|
/**
|
||||||
* Grab the next row's worth of values with a limit on the number of values
|
* Grab the next row's worth of values with a limit on the number of values
|
||||||
* to return.
|
* to return.
|
||||||
* @param results
|
* @param result
|
||||||
* @param limit
|
* @param limit
|
||||||
* @return true if more rows exist after this one, false if scanner is done
|
* @return true if more rows exist after this one, false if scanner is done
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
|
|
|
@ -98,7 +98,7 @@ import org.apache.hadoop.hbase.util.Threads;
|
||||||
* separate reentrant lock is used.
|
* separate reentrant lock is used.
|
||||||
*
|
*
|
||||||
* <p>To read an HLog, call {@link #getReader(org.apache.hadoop.fs.FileSystem,
|
* <p>To read an HLog, call {@link #getReader(org.apache.hadoop.fs.FileSystem,
|
||||||
* org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration)}.
|
* org.apache.hadoop.fs.Path, org.apache.hadoop.hbase.HBaseConfiguration)}.
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
public class HLog implements HConstants, Syncable {
|
public class HLog implements HConstants, Syncable {
|
||||||
|
@ -352,8 +352,7 @@ public class HLog implements HConstants, Syncable {
|
||||||
* Get a reader for the WAL.
|
* Get a reader for the WAL.
|
||||||
* @param fs
|
* @param fs
|
||||||
* @param path
|
* @param path
|
||||||
* @param keyClass
|
* @param conf
|
||||||
* @param valueClass
|
|
||||||
* @return A WAL reader. Close when done with it.
|
* @return A WAL reader. Close when done with it.
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
|
@ -377,8 +376,7 @@ public class HLog implements HConstants, Syncable {
|
||||||
/**
|
/**
|
||||||
* Get a writer for the WAL.
|
* Get a writer for the WAL.
|
||||||
* @param path
|
* @param path
|
||||||
* @param keyClass
|
* @param conf
|
||||||
* @param valueClass
|
|
||||||
* @return A WAL writer. Close when done with it.
|
* @return A WAL writer. Close when done with it.
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
|
|
Loading…
Reference in New Issue