From 0b8953f29b1f478c8af808b529ebbb75f3a2d417 Mon Sep 17 00:00:00 2001 From: Michael Stack Date: Thu, 23 Apr 2009 21:46:35 +0000 Subject: [PATCH] HBASE-1340 Fix new javadoc warnings git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@768068 13f79535-47bb-0310-9956-ffa450edef68 --- CHANGES.txt | 1 + .../hadoop/hbase/HColumnDescriptor.java | 2 +- .../org/apache/hadoop/hbase/KeyValue.java | 91 ++++++++----------- .../hbase/filter/RowFilterInterface.java | 3 +- .../org/apache/hadoop/hbase/io/RowResult.java | 2 +- .../apache/hadoop/hbase/io/SequenceFile.java | 10 +- .../apache/hadoop/hbase/io/hfile/HFile.java | 1 - .../hbase/mapred/TableInputFormatBase.java | 2 +- .../hbase/regionserver/HAbstractScanner.java | 1 - .../hadoop/hbase/regionserver/LruHashMap.java | 2 +- .../hadoop/hbase/regionserver/Store.java | 3 +- .../rest/filter/RowFilterSetFactory.java | 2 +- .../org/apache/hadoop/hbase/util/Bytes.java | 8 +- 13 files changed, 57 insertions(+), 71 deletions(-) diff --git a/CHANGES.txt b/CHANGES.txt index a6b61fa635a..03812027a7c 100644 --- a/CHANGES.txt +++ b/CHANGES.txt @@ -86,6 +86,7 @@ Release 0.20.0 - Unreleased HBASE-1058 Prevent runaway compactions HBASE-1292 php thrift's getRow() would throw an exception if the row does not exist (Rong-en Fan via Stack) + HBASE-1340 Fix new javadoc warnings (Evgeny Ryabitskiy via Stack) IMPROVEMENTS HBASE-1089 Add count of regions on filesystem to master UI; add percentage diff --git a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 02a850b7910..4e8c39117ec 100644 --- a/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -59,7 +59,7 @@ public class HColumnDescriptor implements ISerializable, WritableComparablerow */ public static KeyValue createFirstOnRow(final byte [] row) { @@ -1192,8 +1179,8 @@ public class KeyValue { } /** - * @param row - * @param ts + * @param row - row key (arbitrary byte array) + * @param ts - timestamp * @return First possible key on passed row and timestamp. */ public static KeyValue createFirstOnRow(final byte [] row, @@ -1202,8 +1189,8 @@ public class KeyValue { } /** - * @param row - * @param ts + * @param row - row key (arbitrary byte array) + * @param ts - timestamp * @return First possible key on passed row, column and timestamp. */ public static KeyValue createFirstOnRow(final byte [] row, final byte [] c, diff --git a/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java b/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java index 7e3a0cc19cd..0db5f455c60 100644 --- a/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java +++ b/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java @@ -115,12 +115,13 @@ public interface RowFilterInterface extends Writable { * but the rest of the row will still get through. * * @param rowKey row key to filter on. - * @param colunmName column name to filter on + * @param columnName column name to filter on * @param columnValue column value to filter on * @return true if row filtered and should not be processed. * @deprecated Use {@link #filterColumn(byte[], int, int, byte[], int, int, byte[], int, int)} * instead. */ + @Deprecated boolean filterColumn(final byte [] rowKey, final byte [] columnName, final byte [] columnValue); diff --git a/src/java/org/apache/hadoop/hbase/io/RowResult.java b/src/java/org/apache/hadoop/hbase/io/RowResult.java index 53b6abde118..e61bd08b5b1 100644 --- a/src/java/org/apache/hadoop/hbase/io/RowResult.java +++ b/src/java/org/apache/hadoop/hbase/io/RowResult.java @@ -268,7 +268,7 @@ public class RowResult implements Writable, SortedMap, } /** - * @param r + * @param l * @return * TODO: This is the glue between old way of doing things and the new. * Herein we are converting our clean KeyValues to old RowResult. diff --git a/src/java/org/apache/hadoop/hbase/io/SequenceFile.java b/src/java/org/apache/hadoop/hbase/io/SequenceFile.java index e270eb30479..101b84aa29e 100644 --- a/src/java/org/apache/hadoop/hbase/io/SequenceFile.java +++ b/src/java/org/apache/hadoop/hbase/io/SequenceFile.java @@ -1787,18 +1787,16 @@ public class SequenceFile { return valClass; } - /** Returns true if values are compressed. - * @return - */ + /** @return true if values are compressed. */ public boolean isCompressed() { return decompress; } - /** Returns true if records are block-compressed. */ + /** @return true if records are block-compressed. */ public boolean isBlockCompressed() { return blockCompressed; } - /** Returns the compression codec of data in this file. */ + /** @return the compression codec of data in this file. */ public CompressionCodec getCompressionCodec() { return codec; } - /** Returns the metadata object of the file */ + /** @return the metadata object of the file */ public Metadata getMetadata() { return this.metadata; } diff --git a/src/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/src/java/org/apache/hadoop/hbase/io/hfile/HFile.java index 5c96420bce5..5414fa39ac8 100644 --- a/src/java/org/apache/hadoop/hbase/io/hfile/HFile.java +++ b/src/java/org/apache/hadoop/hbase/io/hfile/HFile.java @@ -271,7 +271,6 @@ public class HFile { * @param blocksize * @param compress * @param c RawComparator to use. - * @param c * @throws IOException */ public Writer(final FSDataOutputStream ostream, final int blocksize, diff --git a/src/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java b/src/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java index 925080900d4..72dc486a614 100644 --- a/src/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java +++ b/src/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java @@ -46,7 +46,7 @@ import org.apache.hadoop.util.StringUtils; /** * A Base for {@link TableInputFormat}s. Receives a {@link HTable}, a - * {@link Text}[] of input columns and optionally a {@link RowFilterInterface}. + * byte[] of input columns and optionally a {@link RowFilterInterface}. * Subclasses may use other TableRecordReader implementations. *

* An example of a subclass: diff --git a/src/java/org/apache/hadoop/hbase/regionserver/HAbstractScanner.java b/src/java/org/apache/hadoop/hbase/regionserver/HAbstractScanner.java index 1f2440d9bd5..11c64e3a438 100644 --- a/src/java/org/apache/hadoop/hbase/regionserver/HAbstractScanner.java +++ b/src/java/org/apache/hadoop/hbase/regionserver/HAbstractScanner.java @@ -80,7 +80,6 @@ public abstract class HAbstractScanner implements InternalScanner { * Compare the column family and column key using the matchers. The first one * that matches returns true. If no matchers are successful, return false. * - * @param family/store key * @param kv KeyValue to test * @return true if any of the matchers for the column match the column family * and the column key. diff --git a/src/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java b/src/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java index fc5710c4708..0fa23fb17fe 100644 --- a/src/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java +++ b/src/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java @@ -341,7 +341,7 @@ implements HeapSize, Map { /** * Deletes the mapping for the specified key if it exists. * - * @param o the key of the entry to be removed from the map + * @param key the key of the entry to be removed from the map * @return the value associated with the specified key, or null * if no mapping exists. */ diff --git a/src/java/org/apache/hadoop/hbase/regionserver/Store.java b/src/java/org/apache/hadoop/hbase/regionserver/Store.java index c8f95b79c97..1b4d2f71e47 100644 --- a/src/java/org/apache/hadoop/hbase/regionserver/Store.java +++ b/src/java/org/apache/hadoop/hbase/regionserver/Store.java @@ -1008,13 +1008,14 @@ public class Store implements HConstants { * row and timestamp, but not a column name. * * The returned object should map column names to Cells. - * @param origin Where to start searching. Specifies a row and timestamp. + * @param key - Where to start searching. Specifies a row. * Columns are specified in following arguments. * @param columns Can be null which means get all * @param columnPattern Can be null. * @param numVersions * @param versionsCounter Can be null. * @param keyvalues + * @param now - Where to start searching. Specifies a timestamp. * @throws IOException */ public void getFull(KeyValue key, final NavigableSet columns, diff --git a/src/java/org/apache/hadoop/hbase/rest/filter/RowFilterSetFactory.java b/src/java/org/apache/hadoop/hbase/rest/filter/RowFilterSetFactory.java index 581ebe4169e..edcf4b98512 100644 --- a/src/java/org/apache/hadoop/hbase/rest/filter/RowFilterSetFactory.java +++ b/src/java/org/apache/hadoop/hbase/rest/filter/RowFilterSetFactory.java @@ -85,7 +85,7 @@ public class RowFilterSetFactory implements FilterFactory { * } * * @param filter - * @return + * @return RowFilter * @throws org.apache.hadoop.hbase.rest.exception.HBaseRestException */ protected RowFilterInterface getRowFilter(JSONObject filter) diff --git a/src/java/org/apache/hadoop/hbase/util/Bytes.java b/src/java/org/apache/hadoop/hbase/util/Bytes.java index e9f8943002a..a284a9c551e 100644 --- a/src/java/org/apache/hadoop/hbase/util/Bytes.java +++ b/src/java/org/apache/hadoop/hbase/util/Bytes.java @@ -161,9 +161,9 @@ public class Bytes { /** * Write a long value out to the specified byte array position. - * @param bytes the byte array - * @param offset position in the array - * @param b byte to write out + * @param tgtBytes the byte array + * @param tgtOffset position in the array + * @param srcBytes byte to write out * @return incremented offset */ public static int putBytes(byte[] tgtBytes, int tgtOffset, byte[] srcBytes, @@ -687,7 +687,7 @@ public class Bytes { * @param offset the offset in the key you want to find * @param length the length of the key * @param comparator a comparator to compare. - * @return + * @return index of key */ public static int binarySearch(byte [][]arr, byte []key, int offset, int length, RawComparator comparator) {