diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadSmallValues.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadSmallValues.java index abecd42c18f..86ef389fe3b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadSmallValues.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadSmallValues.java @@ -56,8 +56,8 @@ import org.apache.hadoop.hbase.util.RegionSplitter; import org.apache.hadoop.io.BytesWritable; import org.apache.hadoop.io.MapWritable; import org.apache.hadoop.io.NullWritable; -import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.SequenceFile.CompressionType; +import org.apache.hadoop.io.Writable; import org.apache.hadoop.mapreduce.Counters; import org.apache.hadoop.mapreduce.InputFormat; import org.apache.hadoop.mapreduce.InputSplit; @@ -108,8 +108,8 @@ import org.slf4j.LoggerFactory; *    options outputDir hfile_1 ... hfile_n *

* Where options are:

- *    -d : Width of generated file name for zero padding, default: 5
- *    -n : Total number of blocks to extract, default: unlimited
+ *    -d width: Width of generated file name for zero padding, default: 5
+ *    -n count: Total number of blocks to extract, default: unlimited
*    -r | --random: Shuffle blocks and write them in randomized order * *

@@ -143,7 +143,7 @@ public class IntegrationTestLoadSmallValues extends IntegrationTestBase { REFERENCED, UNREFERENCED, CORRUPT } - protected String args[]; + protected String[] args; protected long numRows = DEFAULT_NUM_ROWS; protected int numMaps = DEFAULT_NUM_MAPS; protected Path outputDir = null; @@ -337,7 +337,7 @@ public class IntegrationTestLoadSmallValues extends IntegrationTestBase { longitude = ((lonRng.nextInt() * 1000) + fracRng.nextInt()) * (r.nextBoolean() ? 1 : -1); // -40 <= +40 C, approximately nine in ten measures are a positive value - temperature = (short) ((r.nextInt() % 40) * (((r.nextInt() % 10) == 1) ? -1 : +1)); + temperature = (short) (r.nextInt(40) * ((r.nextInt(10) == 1) ? -1 : +1)); } } @@ -615,7 +615,7 @@ public class IntegrationTestLoadSmallValues extends IntegrationTestBase { .setColumnFamilies(families) .build(); if (conf.getBoolean(HBaseTestingUtil.PRESPLIT_TEST_TABLE_KEY, DEFAULT_PRESPLIT_TABLE)) { - int numberOfServers = admin.getRegionServers().size(); + int numberOfServers = admin.getRegionServers().size(); if (numberOfServers == 0) { throw new IllegalStateException("No live regionservers"); } diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/util/HFileBlockExtracter.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/util/HFileBlockExtracter.java index 7557049d057..d32c8b08237 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/util/HFileBlockExtracter.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/util/HFileBlockExtracter.java @@ -50,8 +50,8 @@ import org.apache.hbase.thirdparty.org.apache.commons.cli.Options; *

* where options are one or more of: *

- *    -d : Width of generated file name for zero padding, default: 5
- *    -n : Total number of blocks to extract, default: unlimited
+ *    -d width: Width of generated file name for zero padding, default: 5
+ *    -n count: Total number of blocks to extract, default: unlimited
*    -r | --random: Shuffle blocks and write them in randomized order * */ @@ -80,7 +80,8 @@ public class HFileBlockExtracter extends Configured implements Tool { } args = cmd.getArgs(); if (args.length < 2) { - System.out.println("Usage: HFileBlockExtracter ... "); + System.out.println( + "Usage: HFileBlockExtracter ... "); System.out.println("where are:"); System.out.println(opts.toString()); }