diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 0f6abf09727..47bafc4e8cf 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -303,7 +303,7 @@ public class HColumnDescriptor implements WritableComparable * The other attributes are defaulted. * * @param familyName Column family name. Must be 'printable' -- digit or - * letter -- and may not contain a : + * letter -- and may not contain a : */ public HColumnDescriptor(final String familyName) { this(Bytes.toBytes(familyName)); @@ -314,7 +314,7 @@ public class HColumnDescriptor implements WritableComparable * The other attributes are defaulted. * * @param familyName Column family name. Must be 'printable' -- digit or - * letter -- and may not contain a : + * letter -- and may not contain a : */ public HColumnDescriptor(final byte [] familyName) { this (familyName == null || familyName.length <= 0? @@ -345,7 +345,7 @@ public class HColumnDescriptor implements WritableComparable /** * Constructor * @param familyName Column family name. Must be 'printable' -- digit or - * letter -- and may not contain a : + * letter -- and may not contain a : * @param maxVersions Maximum number of versions to keep * @param compression Compression type * @param inMemory If true, column data should be kept in an HRegionServer's @@ -376,7 +376,7 @@ public class HColumnDescriptor implements WritableComparable /** * Constructor * @param familyName Column family name. Must be 'printable' -- digit or - * letter -- and may not contain a : + * letter -- and may not contain a : * @param maxVersions Maximum number of versions to keep * @param compression Compression type * @param inMemory If true, column data should be kept in an HRegionServer's @@ -413,7 +413,7 @@ public class HColumnDescriptor implements WritableComparable /** * Constructor * @param familyName Column family name. Must be 'printable' -- digit or - * letter -- and may not contain a : + * letter -- and may not contain a : * @param minVersions Minimum number of versions to keep * @param maxVersions Maximum number of versions to keep * @param keepDeletedCells Whether to retain deleted cells until they expire diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java index 353938d7529..944593038b3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java @@ -249,7 +249,7 @@ public interface Admin extends Abortable, Closeable { * are repeated and if the split key has empty byte array. * * @param desc table descriptor for table - * @throws MasterNotRunningException if master is not running + * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running * @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent * threads, the table may have been created between test-for-existence and attempt-at-creation). * @throws IOException if a remote or network exception occurs diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java index 0378fe3f64d..1a25235d51a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java @@ -109,7 +109,7 @@ public class HConnectionManager extends ConnectionFactory { * {@link HConnectionKey}. * @param conf configuration * @return HConnection object for conf - * @throws ZooKeeperConnectionException + * @throws org.apache.hadoop.hbase.ZooKeeperConnectionException */ @Deprecated public static HConnection getConnection(final Configuration conf) throws IOException { @@ -137,7 +137,7 @@ public class HConnectionManager extends ConnectionFactory { * * @param conf configuration * @return HConnection object for conf - * @throws ZooKeeperConnectionException + * @throws org.apache.hadoop.hbase.ZooKeeperConnectionException */ @Deprecated public static HConnection createConnection(Configuration conf) throws IOException { @@ -163,7 +163,7 @@ public class HConnectionManager extends ConnectionFactory { * @param conf configuration * @param pool the thread pool to use for batch operation in HTables used via this HConnection * @return HConnection object for conf - * @throws ZooKeeperConnectionException + * @throws org.apache.hadoop.hbase.ZooKeeperConnectionException */ @Deprecated public static HConnection createConnection(Configuration conf, ExecutorService pool) @@ -189,7 +189,7 @@ public class HConnectionManager extends ConnectionFactory { * @param conf configuration * @param user the user the connection is for * @return HConnection object for conf - * @throws ZooKeeperConnectionException + * @throws org.apache.hadoop.hbase.ZooKeeperConnectionException */ @Deprecated public static HConnection createConnection(Configuration conf, User user) @@ -216,7 +216,7 @@ public class HConnectionManager extends ConnectionFactory { * @param pool the thread pool to use for batch operation in HTables used via this HConnection * @param user the user the connection is for * @return HConnection object for conf - * @throws ZooKeeperConnectionException + * @throws org.apache.hadoop.hbase.ZooKeeperConnectionException */ @Deprecated public static HConnection createConnection(Configuration conf, ExecutorService pool, User user) diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java index 4b998a6cbd3..ff13c8c5200 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java @@ -170,8 +170,6 @@ public class HTablePool implements Closeable { /** * Get a reference to the specified table from the pool. - *

- *

* * @param tableName * table name @@ -189,7 +187,6 @@ public class HTablePool implements Closeable { /** * Get a reference to the specified table from the pool. - *

* * Create a new one if one is not available. * diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java index 9d67549fdbb..d2959534254 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java @@ -242,7 +242,7 @@ public class Result implements CellScannable, CellScanner { * * WARNING do not use, expensive. This does an arraycopy of the cell[]'s value. * - * Added to ease transition from 0.94 -> 0.96. + * Added to ease transition from 0.94 -> 0.96. * * @deprecated as of 0.96, use {@link #rawCells()} * @return array of KeyValues, empty array if nothing in result. @@ -272,7 +272,7 @@ public class Result implements CellScannable, CellScanner { * * WARNING do not use, expensive. This does an arraycopy of the cell[]'s value. * - * Added to ease transition from 0.94 -> 0.96. + * Added to ease transition from 0.94 -> 0.96. * * @deprecated as of 0.96, use {@link #listCells()} * @return all sorted List of KeyValues; can be null if no cells in the result diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java index 1bb18b2b8e2..19843c4126c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java @@ -353,7 +353,6 @@ public class ReplicationAdmin implements Closeable { * Append the replicable table-cf config of the specified peer * @param id a short that identifies the cluster * @param tableCfs table-cfs config str - * @throws KeeperException */ public void appendPeerTableCFs(String id, String tableCfs) throws ReplicationException { appendPeerTableCFs(id, parseTableCFsFromConfig(tableCfs)); @@ -363,7 +362,6 @@ public class ReplicationAdmin implements Closeable { * Append the replicable table-cf config of the specified peer * @param id a short that identifies the cluster * @param tableCfs A map from tableName to column family names - * @throws KeeperException */ public void appendPeerTableCFs(String id, Map> tableCfs) throws ReplicationException { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java index 774f2681638..0a7a184e581 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java @@ -137,7 +137,7 @@ public abstract class Filter { /** * WARNING: please to not override this method. Instead override {@link #transformCell(Cell)}. - * This is for transition from 0.94 -> 0.96 + * This is for transition from 0.94 -> 0.96 **/ @Deprecated // use Cell transformCell(final Cell) abstract public KeyValue transform(final KeyValue currentKV) throws IOException; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java index c1ffc025b0e..3c6bcabd2b9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java @@ -86,7 +86,7 @@ public abstract class FilterBase extends Filter { /** * WARNING: please to not override this method. Instead override {@link #transformCell(Cell)}. * - * This is for transition from 0.94 -> 0.96 + * This is for transition from 0.94 -> 0.96 */ @Override @Deprecated diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java index 7984d084a8c..95678bbea33 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java @@ -229,7 +229,7 @@ final public class FilterList extends Filter { * * When removing this, its body should be placed in transformCell. * - * This is for transition from 0.94 -> 0.96 + * This is for transition from 0.94 -> 0.96 */ @Deprecated @Override diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java index a48c5158c93..441852740d7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterWrapper.java @@ -130,7 +130,7 @@ final public class FilterWrapper extends Filter { /** * WARNING: please to not override this method. Instead override {@link #transformCell(Cell)}. * - * This is for transition from 0.94 -> 0.96 + * This is for transition from 0.94 -> 0.96 */ @Override @Deprecated diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index eedcfaacdce..dc36bacbd0e 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -280,7 +280,7 @@ public final class ProtobufUtil { * @param bytes Bytes to check. * @param offset offset to start at * @param len length to use - * @return True if passed bytes has {@link ProtobufMagic#PB_MAGIC} for a prefix. + * @return True if passed bytes has {@link #PB_MAGIC} for a prefix. */ public static boolean isPBMagicPrefix(final byte [] bytes, int offset, int len) { if (bytes == null || len < PB_MAGIC.length) return false; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java index 7e6a22b2228..bebbb9460b8 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java @@ -2500,7 +2500,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId, * @return Created KeyValue OR if we find a length of zero, we will return null which * can be useful marking a stream as done. * @throws IOException - * @{@link Deprecated} Use {@link KeyValueUtil#iscreate(InputStream, boolean)} + * @deprecated Use {@link KeyValueUtil#iscreate(InputStream, boolean)} */ @Deprecated public static KeyValue iscreate(final InputStream in) throws IOException { diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java index 38b9470fb0e..bbdece0e962 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java @@ -555,7 +555,8 @@ public class KeyValueUtil { /** * Create a KeyValue reading from the raw InputStream. Named - * iscreate so doesn't clash with {@link #create(DataInput)} + * iscreate so doesn't clash with the create(DataInput) method + * added in 2.0 * * @param in * @param withTags diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeSeeker.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeSeeker.java index 7ea47d9e242..2d0526ca617 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeSeeker.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeSeeker.java @@ -144,19 +144,19 @@ public class PrefixTreeSeeker implements EncodedSeeker { /** * Seek forward only (should be called reseekToKeyInBlock?). - *

- * If the exact key is found look at the seekBefore variable and:
- * - if true: go to the previous key if it's true
+ *

+ * If the exact key is found look at the seekBefore variable and:
+ * - if true: go to the previous key if it's true
* - if false: stay on the exact key - *

+ *

* If the exact key is not found, then go to the previous key *if possible*, but remember to * leave the scanner in a valid state if possible. - *

+ *

* @param keyOnlyBytes KeyValue format of a Cell's key at which to position the seeker * @param offset offset into the keyOnlyBytes array * @param length number of bytes of the keyOnlyBytes array to use * @param forceBeforeOnExactMatch if an exact match is found and seekBefore=true, back up 1 Cell - * @return 0 if the seeker is on the exact key
+ * @return 0 if the seeker is on the exact key
* 1 if the seeker is not on the key for any reason, including seekBefore being true */ @Override diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java index 114933dd5d7..82f00b36c3c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitLogWorkerCoordination.java @@ -34,7 +34,7 @@ import com.google.common.annotations.VisibleForTesting; /** * Coordinated operations for {@link SplitLogWorker} and - * {@link org.apache.hadoop.hbase.regionserver.handler.HLogSplitterHandler} Important + * {@link org.apache.hadoop.hbase.regionserver.handler.WALSplitterHandler} Important * methods for SplitLogWorker:
* {@link #isReady()} called from {@link SplitLogWorker#run()} to check whether the coordination is * ready to supply the tasks
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitTransactionCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitTransactionCoordination.java index 3466d04c7e0..f86017c532d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitTransactionCoordination.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/SplitTransactionCoordination.java @@ -81,8 +81,8 @@ public interface SplitTransactionCoordination { * @param std split transaction details * @param parent * @throws IOException If thrown, transaction failed. Call - * {@link org.apache.hadoop.hbase.regionserver. - * SplitTransaction#rollback(Server, RegionServerServices)} + * {@link org.apache.hadoop.hbase.regionserver.SplitTransaction#rollback( + * Server, RegionServerServices)} */ void completeSplitTransaction(RegionServerServices services, Region first, Region second, SplitTransactionDetails std, Region parent) throws IOException; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java index faa7daf2c09..48ce693b01a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java @@ -55,8 +55,8 @@ import org.apache.hadoop.util.StringUtils; * single-level and multi-level block indexes. * * Examples of how to use the block index writer can be found in - * {@link org.apache.hadoop.hbase.io.hfile.CompoundBloomFilterWriter} and - * {@link HFileWriterWriterV2}. Examples of how to use the reader can be + * {@link org.apache.hadoop.hbase.util.CompoundBloomFilterWriter} and + * {@link HFileWriterV2}. Examples of how to use the reader can be * found in {@link HFileReaderV2} and * {@link org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex}. */ diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 39552117b8c..3e5d9094df2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -1300,7 +1300,8 @@ public class HMaster extends HRegionServer implements MasterServices, Server { * @return true if normalization step was performed successfully, false otherwise * (specifically, if HMaster hasn't been initialized properly or normalization * is globally disabled) - * @throws IOException, CoordinatedStateException + * @throws IOException + * @throws CoordinatedStateException */ public boolean normalizeRegions() throws IOException, CoordinatedStateException { if (!this.initialized) { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java index bf58691796f..75989d545e7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java @@ -36,13 +36,15 @@ import java.util.List; * * Logic in use: * - * - get all regions of a given table - * - get avg size S of each region (by total size of store files reported in RegionLoad) - * - If biggest region is bigger than S * 2, it is kindly requested to split, + *
    + *
  1. get all regions of a given table + *
  2. get avg size S of each region (by total size of store files reported in RegionLoad) + *
  3. If biggest region is bigger than S * 2, it is kindly requested to split, * and normalization stops - * - Otherwise, two smallest region R1 and its smallest neighbor R2 are kindly requested - * to merge, if R1 + R1 < S, and normalization stops - * - Otherwise, no action is performed + *
  4. Otherwise, two smallest region R1 and its smallest neighbor R2 are kindly requested + * to merge, if R1 + R1 < S, and normalization stops + *
  5. Otherwise, no action is performed + *
*/ @InterfaceAudience.Private public class SimpleRegionNormalizer implements RegionNormalizer { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java index 56c37762f66..b8ec4c369af 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java @@ -261,7 +261,7 @@ public class ByteBloomFilter implements BloomFilter, BloomFilterWriter { } /** - * Determines & initializes bloom filter meta data from user config. Call + * Determines & initializes bloom filter meta data from user config. Call * {@link #allocBloom()} to allocate bloom filter data. * * @param maxKeys Maximum expected number of keys that will be stored in this diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java index f773b06f503..0d58cafbb53 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java @@ -40,7 +40,7 @@ import org.apache.zookeeper.KeeperException; import org.apache.zookeeper.KeeperException.NoNodeException; /** - * Tool to migrate zookeeper data of older hbase versions(<0.95.0) to PB. + * Tool to migrate zookeeper data of older hbase versions(<0.95.0) to PB. */ public class ZKDataMigrator extends Configured implements Tool { diff --git a/pom.xml b/pom.xml index f9d3a00858e..fee065aa3d8 100644 --- a/pom.xml +++ b/pom.xml @@ -1158,7 +1158,7 @@ 1.6 2.3.4 1.3.9-1 - 2.9 + 2.10.3 /usr /etc/hbase @@ -2413,6 +2413,19 @@ org.apache.hadoop.hbase.generated.master:org.apache.hadoop.hbase.protobuf.generated 2048m true + + + + org.mockito + mockito-all + ${mockito-all.version} + + + org.hamcrest + hamcrest-core + ${hamcrest.version} + +