diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java index f78dfb199c1..bbafc650b29 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/CatalogReplicaLoadBalanceSimpleSelector.java @@ -103,8 +103,8 @@ class CatalogReplicaLoadBalanceSimpleSelector implements } } - private final ConcurrentMap> - staleCache = new ConcurrentHashMap<>(); + private final ConcurrentMap> staleCache = new ConcurrentHashMap<>(); private volatile int numOfReplicas; private final ChoreService choreService; private final TableName tableName; diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java index 4ecf2d7bf4d..f880d5f2aac 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java @@ -101,7 +101,6 @@ public class ColumnCountGetFilter extends FilterBase { /** * @param pbBytes A pb serialized {@link ColumnCountGetFilter} instance * @return An instance of {@link ColumnCountGetFilter} made from bytes - * @throws org.apache.hadoop.hbase.exceptions.DeserializationException * @see #toByteArray */ public static ColumnCountGetFilter parseFrom(final byte [] pbBytes) diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java index 5151d79023d..39fe4f9b759 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FuzzyRowFilter.java @@ -239,8 +239,8 @@ public class FuzzyRowFilter extends FilterBase { byte[] nextRow() { if (nextRows.isEmpty()) { - throw new IllegalStateException( - "NextRows should not be empty, make sure to call nextRow() after updateTracker() return true"); + throw new IllegalStateException("NextRows should not be empty, " + + "make sure to call nextRow() after updateTracker() return true"); } else { return nextRows.peek().getFirst(); } diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java index b4303cdb76d..f314bede082 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java @@ -399,7 +399,6 @@ public class SingleColumnValueFilter extends FilterBase { /** * @param pbBytes A pb serialized {@link SingleColumnValueFilter} instance * @return An instance of {@link SingleColumnValueFilter} made from bytes - * @throws org.apache.hadoop.hbase.exceptions.DeserializationException * @see #toByteArray */ public static SingleColumnValueFilter parseFrom(final byte [] pbBytes) diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index 25a2f10c3a1..8014525b847 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -2952,9 +2952,7 @@ public final class ProtobufUtil { } /** - * Creates {@link CompactionState} from - * {@link org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState} - * state + * Creates {@link CompactionState} from {@link GetRegionInfoResponse.CompactionState} state * @param state the protobuf CompactionState * @return CompactionState */ @@ -2967,9 +2965,7 @@ public final class ProtobufUtil { } /** - * Creates {@link CompactionState} from - * {@link org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos - * .RegionLoad.CompactionState} state + * Creates {@link CompactionState} from {@link RegionLoad.CompactionState} state * @param state the protobuf CompactionState * @return CompactionState */ @@ -2989,8 +2985,7 @@ public final class ProtobufUtil { } /** - * Creates {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type} - * from {@link SnapshotType} + * Creates {@link SnapshotProtos.SnapshotDescription.Type} from {@link SnapshotType} * @param type the SnapshotDescription type * @return the protobuf SnapshotDescription type */ @@ -3000,8 +2995,8 @@ public final class ProtobufUtil { } /** - * Creates {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription.Type} - * from the type of SnapshotDescription string + * Creates {@link SnapshotProtos.SnapshotDescription.Type} from the type of SnapshotDescription + * string * @param snapshotDesc string representing the snapshot description type * @return the protobuf SnapshotDescription type */ @@ -3011,18 +3006,16 @@ public final class ProtobufUtil { } /** - * Creates {@link SnapshotType} from the type of - * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription} - * @param type the snapshot description type - * @return the protobuf SnapshotDescription type + * Creates {@link SnapshotType} from the {@link SnapshotProtos.SnapshotDescription.Type} + * @param type the snapshot description type + * @return the protobuf SnapshotDescription type */ public static SnapshotType createSnapshotType(SnapshotProtos.SnapshotDescription.Type type) { return SnapshotType.valueOf(type.toString()); } /** - * Convert from {@link SnapshotDescription} to - * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription} + * Convert from {@link SnapshotDescription} to {@link SnapshotProtos.SnapshotDescription} * @param snapshotDesc the POJO SnapshotDescription * @return the protobuf SnapshotDescription */ @@ -3056,9 +3049,7 @@ public final class ProtobufUtil { } /** - * Convert from - * {@link org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription} to - * {@link SnapshotDescription} + * Convert from {@link SnapshotProtos.SnapshotDescription} to {@link SnapshotDescription} * @param snapshotDesc the protobuf SnapshotDescription * @return the POJO SnapshotDescription */ diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java index a6be2378560..7b995b0b8cd 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparatorImpl.java @@ -38,9 +38,9 @@ import org.apache.yetus.audience.InterfaceStability; * {@link BBKVComparator}. Use it when mostly {@link ByteBufferKeyValue}s. *

*/ -@edu.umd.cs.findbugs.annotations.SuppressWarnings( - value="UNKNOWN", - justification="Findbugs doesn't like the way we are negating the result of a compare in below") +@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "UNKNOWN", + justification = "Findbugs doesn't like the way we are negating the result of" + + " a compare in below") @InterfaceAudience.Private @InterfaceStability.Evolving public class CellComparatorImpl implements CellComparator { diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java index 20f8e59ef08..bdb457a0063 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ExtendedCell.java @@ -73,13 +73,14 @@ public interface ExtendedCell extends RawCell, HeapSize { } /** - * @param withTags Whether to write tags. - * @return Bytes count required to serialize this Cell in a {@link KeyValue} format. - *
KeyValue format
+ * KeyValue format + *

* <4 bytes keylength> <4 bytes valuelength> <2 bytes rowlength> * <row> <1 byte columnfamilylength> <columnfamily> <columnqualifier> * <8 bytes timestamp> <1 byte keytype> <value> <2 bytes tagslength> * <tags> + * @param withTags Whether to write tags. + * @return Bytes count required to serialize this Cell in a {@link KeyValue} format. */ // TODO remove the boolean param once HBASE-16706 is done. default int getSerializedSize(boolean withTags) { diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java index efa442bd254..a10e91929f6 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java @@ -2371,21 +2371,23 @@ public class KeyValue implements ExtendedCell, Cloneable { /** * HeapSize implementation - * + *

* We do not count the bytes in the rowCache because it should be empty for a KeyValue in the * MemStore. */ @Override public long heapSize() { - /* - * Deep object overhead for this KV consists of two parts. The first part is the KV object - * itself, while the second part is the backing byte[]. We will only count the array overhead - * from the byte[] only if this is the first KV in there. - */ - return ClassSize.align(FIXED_OVERHEAD) + - (offset == 0 - ? ClassSize.sizeOfByteArray(length) // count both length and object overhead - : length); // only count the number of bytes + // Deep object overhead for this KV consists of two parts. The first part is the KV object + // itself, while the second part is the backing byte[]. We will only count the array overhead + // from the byte[] only if this is the first KV in there. + int fixed = ClassSize.align(FIXED_OVERHEAD); + if (offset == 0) { + // count both length and object overhead + return fixed + ClassSize.sizeOfByteArray(length); + } else { + // only count the number of bytes + return fixed + length; + } } /** diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java index af61256cad0..54b95e3aa53 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java @@ -1589,11 +1589,13 @@ public final class PrivateCellUtil { } private static class FirstOnRowCell extends EmptyCell { + // @formatter:off private static final int FIXED_HEAPSIZE = ClassSize.OBJECT // object + ClassSize.REFERENCE // row array + Bytes.SIZEOF_INT // row offset + Bytes.SIZEOF_SHORT; // row length + // @formatter:on private final byte[] rowArray; private final int roffset; private final short rlength; @@ -1643,11 +1645,13 @@ public final class PrivateCellUtil { } private static class FirstOnRowByteBufferExtendedCell extends EmptyByteBufferExtendedCell { + // @formatter:off private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object + ClassSize.REFERENCE // row buffer + Bytes.SIZEOF_INT // row offset + Bytes.SIZEOF_SHORT; // row length + // @formatter:on private final ByteBuffer rowBuff; private final int roffset; private final short rlength; @@ -1698,11 +1702,12 @@ public final class PrivateCellUtil { } private static class LastOnRowByteBufferExtendedCell extends EmptyByteBufferExtendedCell { - private static final int FIXED_OVERHEAD = - ClassSize.OBJECT // object + // @formatter:off + private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object + ClassSize.REFERENCE // rowBuff + Bytes.SIZEOF_INT // roffset + Bytes.SIZEOF_SHORT; // rlength + // @formatter:on private final ByteBuffer rowBuff; private final int roffset; private final short rlength; @@ -1754,11 +1759,12 @@ public final class PrivateCellUtil { private static class FirstOnRowColByteBufferExtendedCell extends FirstOnRowByteBufferExtendedCell { - private static final int FIXED_OVERHEAD = - FirstOnRowByteBufferExtendedCell.FIXED_OVERHEAD - + ClassSize.REFERENCE * 2 // family buffer and column buffer - + Bytes.SIZEOF_INT * 3 // famOffset, colOffset, colLength - + Bytes.SIZEOF_BYTE; // famLength + // @formatter:off + private static final int FIXED_OVERHEAD = FirstOnRowByteBufferExtendedCell.FIXED_OVERHEAD + + ClassSize.REFERENCE * 2 // family buffer and column buffer + + Bytes.SIZEOF_INT * 3 // famOffset, colOffset, colLength + + Bytes.SIZEOF_BYTE; // famLength + // @formatter:on private final ByteBuffer famBuff; private final int famOffset; private final byte famLength; @@ -1823,11 +1829,12 @@ public final class PrivateCellUtil { } private static class FirstOnRowColCell extends FirstOnRowCell { - private static final long FIXED_HEAPSIZE = - FirstOnRowCell.FIXED_HEAPSIZE + // @formatter:off + private static final long FIXED_HEAPSIZE = FirstOnRowCell.FIXED_HEAPSIZE + Bytes.SIZEOF_BYTE // flength + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength + ClassSize.REFERENCE * 2; // fArray, qArray + // @formatter:on private final byte[] fArray; private final int foffset; private final byte flength; @@ -1886,10 +1893,11 @@ public final class PrivateCellUtil { } private static class FirstOnRowColTSCell extends FirstOnRowColCell { - private static final long FIXED_HEAPSIZE = - FirstOnRowColCell.FIXED_HEAPSIZE - + Bytes.SIZEOF_LONG; // ts + // @formatter:off + private static final long FIXED_HEAPSIZE = FirstOnRowColCell.FIXED_HEAPSIZE + + Bytes.SIZEOF_LONG; // ts private long ts; + // @formatter:on public FirstOnRowColTSCell(byte[] rArray, int roffset, short rlength, byte[] fArray, int foffset, byte flength, byte[] qArray, int qoffset, int qlength, long ts) { @@ -1910,10 +1918,11 @@ public final class PrivateCellUtil { private static class FirstOnRowColTSByteBufferExtendedCell extends FirstOnRowColByteBufferExtendedCell { - private static final int FIXED_OVERHEAD = - FirstOnRowColByteBufferExtendedCell.FIXED_OVERHEAD - + Bytes.SIZEOF_LONG; // ts + // @formatter:off + private static final int FIXED_OVERHEAD = FirstOnRowColByteBufferExtendedCell.FIXED_OVERHEAD + + Bytes.SIZEOF_LONG; // ts private long ts; + // @formatter:on public FirstOnRowColTSByteBufferExtendedCell(ByteBuffer rBuffer, int roffset, short rlength, ByteBuffer fBuffer, int foffset, byte flength, ByteBuffer qBuffer, int qoffset, int qlength, @@ -1934,11 +1943,12 @@ public final class PrivateCellUtil { } private static class LastOnRowCell extends EmptyCell { - private static final int FIXED_OVERHEAD = - ClassSize.OBJECT // object + // @formatter:off + private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object + ClassSize.REFERENCE // row array + Bytes.SIZEOF_INT // row offset + Bytes.SIZEOF_SHORT; // row length + // @formatter:on private final byte[] rowArray; private final int roffset; private final short rlength; @@ -1988,10 +1998,12 @@ public final class PrivateCellUtil { } private static class LastOnRowColCell extends LastOnRowCell { + // @formatter:off private static final long FIXED_OVERHEAD = LastOnRowCell.FIXED_OVERHEAD - + ClassSize.REFERENCE * 2 // fArray and qArray - + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength - + Bytes.SIZEOF_BYTE; // flength + + ClassSize.REFERENCE * 2 // fArray and qArray + + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength + + Bytes.SIZEOF_BYTE; // flength + // @formatter:on private final byte[] fArray; private final int foffset; private final byte flength; @@ -2050,11 +2062,12 @@ public final class PrivateCellUtil { } private static class LastOnRowColByteBufferExtendedCell extends LastOnRowByteBufferExtendedCell { - private static final int FIXED_OVERHEAD = - LastOnRowByteBufferExtendedCell.FIXED_OVERHEAD - + ClassSize.REFERENCE * 2 // fBuffer and qBuffer - + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength - + Bytes.SIZEOF_BYTE; // flength + // @formatter:off + private static final int FIXED_OVERHEAD = LastOnRowByteBufferExtendedCell.FIXED_OVERHEAD + + ClassSize.REFERENCE * 2 // fBuffer and qBuffer + + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength + + Bytes.SIZEOF_BYTE; // flength + // @formatter:on private final ByteBuffer fBuffer; private final int foffset; private final byte flength; @@ -2119,11 +2132,12 @@ public final class PrivateCellUtil { } private static class FirstOnRowDeleteFamilyCell extends EmptyCell { - private static final int FIXED_OVERHEAD = - ClassSize.OBJECT // object + // @formatter:off + private static final int FIXED_OVERHEAD = ClassSize.OBJECT // object + ClassSize.REFERENCE * 2 // fBuffer and qBuffer + Bytes.SIZEOF_INT * 3 // foffset, qoffset, qlength + Bytes.SIZEOF_BYTE; // flength + // @formatter:on private final byte[] row; private final byte[] fam; diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java index f1589ba093b..c8b84574645 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java @@ -164,17 +164,22 @@ public class CoprocessorClassLoader extends ClassLoaderBase { } FileSystem fs = pathPattern.getFileSystem(conf); - Path pathPattern1 = fs.isDirectory(pathPattern) ? - new Path(pathPattern, "*.jar") : pathPattern; // append "*.jar" if a directory is specified - FileStatus[] fileStatuses = fs.globStatus(pathPattern1); // return all files that match the pattern - if (fileStatuses == null || fileStatuses.length == 0) { // if no one matches + // append "*.jar" if a directory is specified + Path pathPattern1 = fs.isDirectory(pathPattern) ? new Path(pathPattern, "*.jar") : pathPattern; + // return all files that match the pattern + FileStatus[] fileStatuses = fs.globStatus(pathPattern1); + if (fileStatuses == null || fileStatuses.length == 0) { + // if no one matches throw new FileNotFoundException(pathPattern1.toString()); } else { boolean validFileEncountered = false; - for (Path path : FileUtil.stat2Paths(fileStatuses)) { // for each file that match the pattern - if (fs.isFile(path)) { // only process files, skip for directories - File dst = new File(parentDirStr, "." + pathPrefix + "." - + path.getName() + "." + System.currentTimeMillis() + ".jar"); + // for each file that match the pattern + for (Path path : FileUtil.stat2Paths(fileStatuses)) { + if (fs.isFile(path)) { + // only process files, skip for directories + File dst = new File(parentDirStr, + "." + pathPrefix + "." + path.getName() + "." + System.currentTimeMillis() + + ".jar"); fs.copyToLocalFile(path, new Path(dst.toString())); dst.deleteOnExit(); @@ -182,7 +187,8 @@ public class CoprocessorClassLoader extends ClassLoaderBase { JarFile jarFile = new JarFile(dst.toString()); try { - Enumeration entries = jarFile.entries(); // get entries inside a jar file + // get entries inside a jar file + Enumeration entries = jarFile.entries(); while (entries.hasMoreElements()) { JarEntry entry = entries.nextElement(); Matcher m = libJarPattern.matcher(entry.getName()); @@ -200,11 +206,12 @@ public class CoprocessorClassLoader extends ClassLoaderBase { } finally { jarFile.close(); } - - validFileEncountered = true; // Set to true when encountering a file + // Set to true when encountering a file + validFileEncountered = true; } } - if (validFileEncountered == false) { // all items returned by globStatus() are directories + if (validFileEncountered == false) { + // all items returned by globStatus() are directories throw new FileNotFoundException("No file found matching " + pathPattern1.toString()); } } diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java index 58075056f38..92c80a9036d 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java @@ -303,8 +303,8 @@ public class DemoClient { m = new Mutation(); m.column = ByteBuffer.wrap(bytes("entry:sqr")); m.isDelete = true; - client.mutateRowTs(demoTable, ByteBuffer.wrap(row), mutations, 1, - dummyAttributes); // shouldn't override latest + // shouldn't override latest + client.mutateRowTs(demoTable, ByteBuffer.wrap(row), mutations, 1, dummyAttributes); printRow(client.getRow(demoTable, ByteBuffer.wrap(row), dummyAttributes)); List versions = client.getVer(demoTable, ByteBuffer.wrap(row), diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java index 990ac8026b8..e7e67beaf45 100644 --- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java +++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.java @@ -481,8 +481,8 @@ public interface MetricsRegionServerSource extends BaseSource, JvmPauseMonitorSo String HEDGED_READ_WINS_DESC = "The number of times we started a hedged read and a hedged read won"; String HEDGED_READ_IN_CUR_THREAD = "hedgedReadOpsInCurThread"; - String HEDGED_READ_IN_CUR_THREAD_DESC = - "The number of times we execute a hedged read in current thread as a fallback for task rejection"; + String HEDGED_READ_IN_CUR_THREAD_DESC = "The number of times we execute a hedged read" + + " in current thread as a fallback for task rejection"; String TOTAL_BYTES_READ = "totalBytesRead"; String TOTAL_BYTES_READ_DESC = "The total number of bytes read from HDFS"; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java index 90086732d59..927bf59ce34 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java @@ -86,21 +86,20 @@ public class IntegrationTestIngestWithACL extends IntegrationTestIngest { tmp.add(sb.toString()); return tmp.toArray(new String[tmp.size()]); } + @Override protected void addOptions() { super.addOptions(); - super.addOptWithArg(OPT_SUPERUSER, - "Super user name used to add the ACL permissions"); + super.addOptWithArg(OPT_SUPERUSER, "Super user name used to add the ACL permissions"); super.addOptWithArg(OPT_USERS, "List of users to be added with the ACLs. Should be comma seperated."); - super - .addOptWithArg( - OPT_AUTHN, - "The name of the properties file that contains kerberos key tab file and principal definitions. " + - "The principal key in the file should be of the form hbase..kerberos.principal." + - " The keytab key in the file should be of the form hbase..keytab.file. Example: " + - "hbase.user1.kerberos.principal=user1/fully.qualified.domain.name@YOUR-REALM.COM, " + - "hbase.user1.keytab.file=."); + super.addOptWithArg(OPT_AUTHN, + "The name of the properties file that contains" + + " kerberos key tab file and principal definitions. The principal key in the file" + + " should be of the form hbase..kerberos.principal. The keytab key in the" + + " file should be of the form hbase..keytab.file. Example:" + + " hbase.user1.kerberos.principal=user1/fully.qualified.domain.name@YOUR-REALM.COM," + + " hbase.user1.keytab.file=."); } @Override diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java index 676b04d3fee..cc8dfcfcad3 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java @@ -49,11 +49,12 @@ public class ServerAndDependenciesKillingMonkeyFactory extends MonkeyFactory { loadProperties(); // Destructive actions to mess things around. Cannot run batch restart. - Action[] actions1 = new Action[]{ + // @formatter:off + Action[] actions1 = new Action[] { new RestartRandomRsExceptMetaAction(60000), new RestartActiveMasterAction(5000), - new RollingBatchRestartRsAction(5000, 1.0f, 2, - true), // only allow 2 servers to be dead. + // only allow 2 servers to be dead. + new RollingBatchRestartRsAction(5000, 1.0f, 2, true), new ForceBalancerAction(), new RestartRandomDataNodeAction(60000), new RestartRandomZKNodeAction(60000), @@ -61,6 +62,7 @@ public class ServerAndDependenciesKillingMonkeyFactory extends MonkeyFactory { new RollingBatchSuspendResumeRsAction(rollingBatchSuspendRSSleepTime, rollingBatchSuspendtRSRatio) }; + // @formatter:on // Action to log more info for debugging Action[] actions2 = new Action[]{ diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java index 6a5efeda845..cfbab8eb955 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java @@ -47,16 +47,18 @@ public class ServerKillingMonkeyFactory extends MonkeyFactory { loadProperties(); // Destructive actions to mess things around. Cannot run batch restart + // @formatter:off Action[] actions1 = new Action[] { - new RestartRandomRsExceptMetaAction(60000), - new RestartActiveMasterAction(5000), - new RollingBatchRestartRsAction(5000, 1.0f, 2, - true), //only allow 2 servers to be dead + new RestartRandomRsExceptMetaAction(60000), + new RestartActiveMasterAction(5000), + // only allow 2 servers to be dead + new RollingBatchRestartRsAction(5000, 1.0f, 2, true), new ForceBalancerAction(), new GracefulRollingRestartRsAction(gracefulRollingRestartTSSLeepTime), new RollingBatchSuspendResumeRsAction(rollingBatchSuspendRSSleepTime, rollingBatchSuspendtRSRatio) }; + // @formatter:on // Action to log more info for debugging Action[] actions2 = new Action[] { diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java index 509972e92aa..cef8e8a8176 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormatBase.java @@ -42,7 +42,7 @@ import org.apache.hadoop.mapred.Reporter; * A Base for {@link TableInputFormat}s. Receives a {@link Table}, a * byte[] of input columns and optionally a {@link Filter}. * Subclasses may use other TableRecordReader implementations. - * + *

* Subclasses MUST ensure initializeTable(Connection, TableName) is called for an instance to * function properly. Each of the entry points to this class used by the MapReduce framework, * {@link #getRecordReader(InputSplit, JobConf, Reporter)} and {@link #getSplits(JobConf, int)}, @@ -98,8 +98,7 @@ implements InputFormat { * Builds a TableRecordReader. If no TableRecordReader was provided, uses * the default. * - * @see org.apache.hadoop.mapred.InputFormat#getRecordReader(InputSplit, - * JobConf, Reporter) + * @see InputFormat#getRecordReader(InputSplit, JobConf, Reporter) */ public RecordReader getRecordReader( InputSplit split, JobConf job, Reporter reporter) @@ -165,7 +164,7 @@ implements InputFormat { /** * Calculates the splits that will serve as input for the map tasks. - * + *

* Splits are created in number equal to the smallest between numSplits and * the number of {@link org.apache.hadoop.hbase.regionserver.HRegion}s in the table. * If the number of splits is smaller than the number of @@ -180,7 +179,7 @@ implements InputFormat { * * @return the input splits * - * @see org.apache.hadoop.mapred.InputFormat#getSplits(org.apache.hadoop.mapred.JobConf, int) + * @see InputFormat#getSplits(org.apache.hadoop.mapred.JobConf, int) */ public InputSplit[] getSplits(JobConf job, int numSplits) throws IOException { if (this.table == null) { @@ -282,7 +281,7 @@ implements InputFormat { * will call {@link #initialize(JobConf)} as a convenient centralized location to handle * retrieving the necessary configuration information and calling * {@link #initializeTable(Connection, TableName)}. - * + *

* Subclasses should implement their initialize call such that it is safe to call multiple times. * The current TableInputFormatBase implementation relies on a non-null table reference to decide * if an initialize call is needed, but this behavior may change in the future. In particular, diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java index d8205c1a0ca..ec789f2bdcd 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java @@ -72,9 +72,7 @@ public abstract class MultiTableInputFormatBase extends * @return The newly created record reader. * @throws IOException When creating the reader fails. * @throws InterruptedException when record reader initialization fails - * @see org.apache.hadoop.mapreduce.InputFormat#createRecordReader( - * org.apache.hadoop.mapreduce.InputSplit, - * org.apache.hadoop.mapreduce.TaskAttemptContext) + * @see InputFormat#createRecordReader(InputSplit, TaskAttemptContext) */ @Override public RecordReader createRecordReader( @@ -152,7 +150,7 @@ public abstract class MultiTableInputFormatBase extends * @param context The current job context. * @return The list of input splits. * @throws IOException When creating the list of splits fails. - * @see org.apache.hadoop.mapreduce.InputFormat#getSplits(org.apache.hadoop.mapreduce.JobContext) + * @see InputFormat#getSplits(org.apache.hadoop.mapreduce.JobContext) */ @Override public List getSplits(JobContext context) throws IOException { diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java index fa712903040..07c0820c1ae 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java @@ -49,8 +49,7 @@ import java.util.Map; * included in each snapshot/scan * pair. * {@link TableMapReduceUtil#initMultiTableSnapshotMapperJob - * (java.util.Map, Class, Class, Class, org.apache.hadoop.mapreduce.Job, boolean, org.apache - * .hadoop.fs.Path)} + * (Map, Class, Class, Class, org.apache.hadoop.mapreduce.Job, boolean, Path)} * can be used to configure the job. *

{@code
  * Job job = new Job(conf);
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
index ef5c161d4cd..76254c3f6ba 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
@@ -78,8 +78,7 @@ public class MultiTableSnapshotInputFormatImpl {
 
   /**
    * Return the list of splits extracted from the scans/snapshots pushed to conf by
-   * {@link
-   * #setInput(org.apache.hadoop.conf.Configuration, java.util.Map, org.apache.hadoop.fs.Path)}
+   * {@link #setInput(Configuration, Map, Path)}
    *
    * @param conf Configuration to determine splits from
    * @return Return the list of splits extracted from the scans/snapshots pushed to conf
@@ -115,7 +114,7 @@ public class MultiTableSnapshotInputFormatImpl {
 
   /**
    * Retrieve the snapshot name -> list<scan> mapping pushed to configuration by
-   * {@link #setSnapshotToScans(org.apache.hadoop.conf.Configuration, java.util.Map)}
+   * {@link #setSnapshotToScans(Configuration, Map)}
    *
    * @param conf Configuration to extract name -> list<scan> mappings from.
    * @return the snapshot name -> list<scan> mapping pushed to configuration
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
index ca82e2a58ee..4ad1935f37a 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
@@ -231,8 +231,9 @@ public class MultithreadedTableMapper extends TableMapper {
     }
   }
 
-  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="REC_CATCH_EXCEPTION",
-      justification="Don't understand why FB is complaining about this one. We do throw exception")
+  @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "REC_CATCH_EXCEPTION",
+      justification = "Don't understand why FB is complaining about this one."
+        + " We do throw exception")
   private class MapRunner implements Runnable {
     private Mapper mapper;
     private Context subcontext;
diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
index 4152182a6d3..3b9a13879e8 100644
--- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
+++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MobRefReporter.java
@@ -65,10 +65,10 @@ import org.slf4j.LoggerFactory;
 
 
 /**
- * Scans a given table + CF for all mob reference cells to get the list of backing mob files.
- * For each referenced file we attempt to verify that said file is on the FileSystem in a place
- * that the MOB system will look when attempting to resolve the actual value.
- *
+ * Scans a given table + CF for all mob reference cells to get the list of backing mob files. For
+ * each referenced file we attempt to verify that said file is on the FileSystem in a place that the
+ * MOB system will look when attempting to resolve the actual value.
+ * 

* The job includes counters that can help provide a rough sketch of the mob data. * *

@@ -94,31 +94,31 @@ import org.slf4j.LoggerFactory;
  *         Number of rows with total size in the 100,000s of bytes=6838
  *         Number of rows with total size in the 1,000,000s of bytes=3162
  * 
- * - * * Map-Reduce Framework:Map input records - the number of rows with mob references - * * Map-Reduce Framework:Reduce output records - the number of unique hfiles referenced - * * MOB:NUM_CELLS - the total number of mob reference cells - * * PROBLEM:Affected rows - the number of rows that reference hfiles with an issue - * * PROBLEM:Problem MOB files - the number of unique hfiles that have an issue - * * CELLS PER ROW: - this counter group gives a histogram of the order of magnitude of the - * number of cells in a given row by grouping by the number of digits used in each count. - * This allows us to see more about the distribution of cells than what we can determine - * with just the cell count and the row count. In this particular example we can see that - * all of our rows have somewhere between 1 - 9 cells. - * * ROWS WITH PROBLEMS PER FILE: - this counter group gives a histogram of the order of - * magnitude of the number of rows in each of the hfiles with a problem. e.g. in the - * example there are 2 hfiles and they each have the same order of magnitude number of rows, - * specifically between 100 and 999. - * * SIZES OF CELLS: - this counter group gives a histogram of the order of magnitude of - * the size of mob values according to our reference cells. e.g. in the example above we - * have cell sizes that are all between 10,000 bytes and 9,999,999 bytes. From this - * histogram we can also see that _most_ cells are 100,000 - 999,000 bytes and the smaller - * and bigger ones are outliers making up less than 2% of mob cells. - * * SIZES OF ROWS: - this counter group gives a histogram of the order of magnitude of the - * size of mob values across each row according to our reference cells. In the example above - * we have rows that are are between 100,000 bytes and 9,999,999 bytes. We can also see that - * about 2/3rd of our rows are 100,000 - 999,999 bytes. - * + *
    + *
  1. Map-Reduce Framework:Map input records - the number of rows with mob references
  2. + *
  3. Map-Reduce Framework:Reduce output records - the number of unique hfiles referenced
  4. + *
  5. MOB:NUM_CELLS - the total number of mob reference cells
  6. + *
  7. PROBLEM:Affected rows - the number of rows that reference hfiles with an issue
  8. + *
  9. PROBLEM:Problem MOB files - the number of unique hfiles that have an issue
  10. + *
  11. CELLS PER ROW: - this counter group gives a histogram of the order of magnitude of the number + * of cells in a given row by grouping by the number of digits used in each count. This allows us to + * see more about the distribution of cells than what we can determine with just the cell count and + * the row count. In this particular example we can see that all of our rows have somewhere between + * 1 - 9 cells.
  12. + *
  13. ROWS WITH PROBLEMS PER FILE: - this counter group gives a histogram of the order of magnitude + * of the number of rows in each of the hfiles with a problem. e.g. in the example there are 2 + * hfiles and they each have the same order of magnitude number of rows, specifically between 100 + * and 999.
  14. + *
  15. SIZES OF CELLS: - this counter group gives a histogram of the order of magnitude of the size + * of mob values according to our reference cells. e.g. in the example above we have cell sizes that + * are all between 10,000 bytes and 9,999,999 bytes. From this histogram we can also see that _most_ + * cells are 100,000 - 999,000 bytes and the smaller and bigger ones are outliers making up less + * than 2% of mob cells.
  16. + *
  17. SIZES OF ROWS: - this counter group gives a histogram of the order of magnitude of the size + * of mob values across each row according to our reference cells. In the example above we have rows + * that are are between 100,000 bytes and 9,999,999 bytes. We can also see that about 2/3rd of our + * rows are 100,000 - 999,999 bytes.
  18. + *
* Generates a report that gives one file status per line, with tabs dividing fields. * *
@@ -133,32 +133,31 @@ import org.slf4j.LoggerFactory;
  * 
* * Possible results are listed; the first three indicate things are working properly. - * * MOB DIR - the reference is in the normal MOB area for the given table and CF - * * HLINK TO ARCHIVE FOR SAME TABLE - the reference is present in the archive area for this - * table and CF - * * HLINK TO ARCHIVE FOR OTHER TABLE - the reference is present in a different table and CF, - * either in the MOB or archive areas (e.g. from a snapshot restore or clone) - * * ARCHIVE WITH HLINK BUT NOT FROM OUR TABLE - the reference is currently present in the archive - * area for this table and CF, but it is kept there because a _different_ table has a - * reference to it (e.g. from a snapshot clone). If these other tables are removed then - * the file will likely be deleted unless there is a snapshot also referencing it. - * * ARCHIVE BUT NO HLINKS - the reference is currently present in the archive for this table and - * CF, but there are no references present to prevent its removal. Unless it is newer than - * the general TTL (default 5 minutes) or referenced in a snapshot it will be subject to - * cleaning. - * * ARCHIVE BUT FAILURE WHILE CHECKING HLINKS - Check the job logs to see why things failed while - * looking for why this file is being kept around. - * * MISSING FILE - We couldn't find the reference on the FileSystem. Either there is dataloss due - * to a bug in the MOB storage system or the MOB storage is damaged but in an edge case that - * allows it to work for now. You can verify which by doing a raw reference scan to get the - * referenced hfile and check the underlying filesystem. See the ref guide section on mob - * for details. - * * HLINK BUT POINT TO MISSING FILE - There is a pointer in our mob area for this table and CF - * to a file elsewhere on the FileSystem, however the file it points to no longer exists. - * * MISSING FILE BUT FAILURE WHILE CHECKING HLINKS - We could not find the referenced file, - * however you should check the job logs to see why we couldn't check to see if there is a - * pointer to the referenced file in our archive or another table's archive or mob area. - * + *
    + *
  1. MOB DIR - the reference is in the normal MOB area for the given table and CF
  2. + *
  3. HLINK TO ARCHIVE FOR SAME TABLE - the reference is present in the archive area for this table + * and CF
  4. + *
  5. HLINK TO ARCHIVE FOR OTHER TABLE - the reference is present in a different table and CF, + * either in the MOB or archive areas (e.g. from a snapshot restore or clone)
  6. + *
  7. ARCHIVE WITH HLINK BUT NOT FROM OUR TABLE - the reference is currently present in the archive + * area for this table and CF, but it is kept there because a _different_ table has a reference to + * it (e.g. from a snapshot clone). If these other tables are removed then the file will likely be + * deleted unless there is a snapshot also referencing it.
  8. + *
  9. ARCHIVE BUT NO HLINKS - the reference is currently present in the archive for this table and + * CF, but there are no references present to prevent its removal. Unless it is newer than the + * general TTL (default 5 minutes) or referenced in a snapshot it will be subject to cleaning.
  10. + *
  11. ARCHIVE BUT FAILURE WHILE CHECKING HLINKS - Check the job logs to see why things failed while + * looking for why this file is being kept around.
  12. + *
  13. MISSING FILE - We couldn't find the reference on the FileSystem. Either there is dataloss due + * to a bug in the MOB storage system or the MOB storage is damaged but in an edge case that allows + * it to work for now. You can verify which by doing a raw reference scan to get the referenced + * hfile and check the underlying filesystem. See the ref guide section on mob for details.
  14. + *
  15. HLINK BUT POINT TO MISSING FILE - There is a pointer in our mob area for this table and CF to + * a file elsewhere on the FileSystem, however the file it points to no longer exists.
  16. + *
  17. MISSING FILE BUT FAILURE WHILE CHECKING HLINKS - We could not find the referenced file, + * however you should check the job logs to see why we couldn't check to see if there is a pointer + * to the referenced file in our archive or another table's archive or mob area.
  18. + *
*/ @InterfaceAudience.Private public class MobRefReporter extends Configured implements Tool { diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java index aaa49e607e7..f9f16b5e350 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java @@ -297,15 +297,16 @@ public class TestImportExport { IMPORT_TABLE, FQ_OUTPUT_DIR }; assertTrue(runImport(args)); - /* exportedTableIn94Format contains 5 rows - ROW COLUMN+CELL - r1 column=f1:c1, timestamp=1383766761171, value=val1 - r2 column=f1:c1, timestamp=1383766771642, value=val2 - r3 column=f1:c1, timestamp=1383766777615, value=val3 - r4 column=f1:c1, timestamp=1383766785146, value=val4 - r5 column=f1:c1, timestamp=1383766791506, value=val5 - */ - assertEquals(5, UTIL.countRows(t)); + // @formatter:off + // exportedTableIn94Format contains 5 rows + // ROW COLUMN+CELL + // r1 column=f1:c1, timestamp=1383766761171, value=val1 + // r2 column=f1:c1, timestamp=1383766771642, value=val2 + // r3 column=f1:c1, timestamp=1383766777615, value=val3 + // r4 column=f1:c1, timestamp=1383766785146, value=val4 + // r5 column=f1:c1, timestamp=1383766791506, value=val5 + // @formatter:on + assertEquals(5, UTIL.countRows(t)); } } @@ -330,12 +331,9 @@ public class TestImportExport { p.addColumn(FAMILYA, QUAL, now + 3, QUAL); p.addColumn(FAMILYA, QUAL, now + 4, QUAL); t.put(p); - - String[] args = new String[] { - "-D" + ExportUtils.EXPORT_BATCHING + "=" + EXPORT_BATCH_SIZE, // added scanner batching arg. - name.getMethodName(), - FQ_OUTPUT_DIR - }; + // added scanner batching arg. + String[] args = new String[] { "-D" + ExportUtils.EXPORT_BATCHING + "=" + EXPORT_BATCH_SIZE, + name.getMethodName(), FQ_OUTPUT_DIR }; assertTrue(runExport(args)); FileSystem fs = FileSystem.get(UTIL.getConfiguration()); diff --git a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java index 99f8e3df447..808b77bc9d6 100644 --- a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java +++ b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/model/TestRowModel.java @@ -48,8 +48,9 @@ public class TestRowModel extends TestModelBase { public TestRowModel() throws Exception { super(RowModel.class); AS_XML = - "" + - "dGVzdHZhbHVlMQ=="; + "" + "" + + "dGVzdHZhbHVlMQ==" + + ""; AS_JSON = "{\"key\":\"dGVzdHJvdzE=\",\"Cell\":[{\"column\":\"dGVzdGNvbHVtbjE=\"," + diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java index 85abc722044..142827be70c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/errorhandling/ForeignException.java @@ -157,7 +157,6 @@ public class ForeignException extends IOException { * @param bytes * @return the ForeignExcpetion instance * @throws InvalidProtocolBufferException if there was deserialization problem this is thrown. - * @throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException */ public static ForeignException deserialize(byte[] bytes) throws IOException { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java index ed0e84deace..d106cf72382 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java @@ -533,13 +533,13 @@ public final class HFile { /** * @param fs filesystem * @param path Path to file to read - * @param cacheConf This must not be null. @see - * {@link org.apache.hadoop.hbase.io.hfile.CacheConfig#CacheConfig(Configuration)} + * @param cacheConf This must not be null. * @param primaryReplicaReader true if this is a reader for primary replica * @param conf Configuration * @return an active Reader instance * @throws IOException Will throw a CorruptHFileException (DoNotRetryIOException subtype) if hfile * is corrupt/invalid. + * @see CacheConfig#CacheConfig(Configuration) */ public static Reader createReader(FileSystem fs, Path path, CacheConfig cacheConf, boolean primaryReplicaReader, Configuration conf) throws IOException { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java index cd347f43822..0108c145f75 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java @@ -151,11 +151,10 @@ public class HMasterCommandLine extends ServerCommandLine { if (shutDownCluster) { return stopMaster(); } - System.err.println( - "To shutdown the master run " + - "hbase-daemon.sh stop master or send a kill signal to " + - "the HMaster pid, " + - "and to stop HBase Cluster run \"stop-hbase.sh\" or \"hbase master stop --shutDownCluster\""); + System.err.println("To shutdown the master run " + + "hbase-daemon.sh stop master or send a kill signal to the HMaster pid, " + + "and to stop HBase Cluster run \"stop-hbase.sh\" or \"hbase master " + + "stop --shutDownCluster\""); return 1; } else if ("clear".equals(command)) { return (ZNodeClearer.clear(getConf()) ? 0 : 1); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java index 25b661c69ce..02863b843ed 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java @@ -297,9 +297,9 @@ public class MasterWalManager { splitLog(serverNames, META_FILTER); } - @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="UL_UNRELEASED_LOCK", justification= - "We only release this lock when we set it. Updates to code that uses it should verify use " + - "of the guard boolean.") + @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "UL_UNRELEASED_LOCK", + justification = "We only release this lock when we set it. Updates to code " + + "that uses it should verify use of the guard boolean.") List getLogDirs(final Set serverNames) throws IOException { List logDirs = new ArrayList<>(); boolean needReleaseLock = false; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java index 186a8ff11bb..16185cb7949 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java @@ -16,6 +16,7 @@ * limitations under the License. */ package org.apache.hadoop.hbase.master; + import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.CHECK; import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.FORCE; import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.DELETED; @@ -155,10 +156,10 @@ public class SplitLogManager { /** * Get a list of paths that need to be split given a set of server-specific directories and * optionally a filter. - * + *

* See {@link AbstractFSWALProvider#getServerNameFromWALDirectoryName} for more info on directory * layout. - * + *

* Should be package-private, but is needed by * {@link org.apache.hadoop.hbase.wal.WALSplitter#split(Path, Path, Path, FileSystem, * Configuration, org.apache.hadoop.hbase.wal.WALFactory)} for tests. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java index 291c419bf9f..73c1ff20371 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java @@ -268,11 +268,11 @@ class SimpleRegionNormalizer implements RegionNormalizer, ConfigurationObserver } /** - * @param tableRegions regions of table to normalize + * Also make sure tableRegions contains regions of the same table + * @param tableRegions regions of table to normalize * @param tableDescriptor the TableDescriptor * @return average region size depending on - * @see org.apache.hadoop.hbase.client.TableDescriptor#getNormalizerTargetRegionCount() - * Also make sure tableRegions contains regions of the same table + * @see TableDescriptor#getNormalizerTargetRegionCount() */ private double getAverageRegionSizeMb(final List tableRegions, final TableDescriptor tableDescriptor) { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java index 8157af99ba4..8d583bdb295 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java @@ -61,8 +61,9 @@ import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloneSnapshotState; +import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloneSnapshotStateData; +import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreParentToChildRegionsPair; import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription; @InterfaceAudience.Private @@ -267,11 +268,10 @@ public class CloneSnapshotProcedure throws IOException { super.serializeStateData(serializer); - MasterProcedureProtos.CloneSnapshotStateData.Builder cloneSnapshotMsg = - MasterProcedureProtos.CloneSnapshotStateData.newBuilder() - .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser())) - .setSnapshot(this.snapshot) - .setTableSchema(ProtobufUtil.toTableSchema(tableDescriptor)); + CloneSnapshotStateData.Builder cloneSnapshotMsg = CloneSnapshotStateData.newBuilder() + .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser())) + .setSnapshot(this.snapshot) + .setTableSchema(ProtobufUtil.toTableSchema(tableDescriptor)); cloneSnapshotMsg.setRestoreAcl(restoreAcl); if (newRegions != null) { @@ -285,11 +285,11 @@ public class CloneSnapshotProcedure while (it.hasNext()) { final Map.Entry> entry = it.next(); - MasterProcedureProtos.RestoreParentToChildRegionsPair.Builder parentToChildrenPair = - MasterProcedureProtos.RestoreParentToChildRegionsPair.newBuilder() - .setParentRegionName(entry.getKey()) - .setChild1RegionName(entry.getValue().getFirst()) - .setChild2RegionName(entry.getValue().getSecond()); + RestoreParentToChildRegionsPair.Builder parentToChildrenPair = + RestoreParentToChildRegionsPair.newBuilder() + .setParentRegionName(entry.getKey()) + .setChild1RegionName(entry.getValue().getFirst()) + .setChild2RegionName(entry.getValue().getSecond()); cloneSnapshotMsg.addParentToChildRegionsPairList(parentToChildrenPair); } } @@ -301,8 +301,7 @@ public class CloneSnapshotProcedure throws IOException { super.deserializeStateData(serializer); - MasterProcedureProtos.CloneSnapshotStateData cloneSnapshotMsg = - serializer.deserialize(MasterProcedureProtos.CloneSnapshotStateData.class); + CloneSnapshotStateData cloneSnapshotMsg = serializer.deserialize(CloneSnapshotStateData.class); setUser(MasterProcedureUtil.toUserInfo(cloneSnapshotMsg.getUserInfo())); snapshot = cloneSnapshotMsg.getSnapshot(); tableDescriptor = ProtobufUtil.toTableDescriptor(cloneSnapshotMsg.getTableSchema()); @@ -319,8 +318,8 @@ public class CloneSnapshotProcedure } if (cloneSnapshotMsg.getParentToChildRegionsPairListCount() > 0) { parentsToChildrenPairMap = new HashMap<>(); - for (MasterProcedureProtos.RestoreParentToChildRegionsPair parentToChildrenPair: - cloneSnapshotMsg.getParentToChildRegionsPairListList()) { + for (RestoreParentToChildRegionsPair parentToChildrenPair : cloneSnapshotMsg + .getParentToChildRegionsPairListList()) { parentsToChildrenPairMap.put( parentToChildrenPair.getParentRegionName(), new Pair<>( diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java index e9440621e44..8f2a9c86afb 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java @@ -57,8 +57,9 @@ import org.slf4j.LoggerFactory; import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; +import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreParentToChildRegionsPair; import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreSnapshotState; +import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RestoreSnapshotStateData; import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription; @InterfaceAudience.Private @@ -237,11 +238,10 @@ public class RestoreSnapshotProcedure throws IOException { super.serializeStateData(serializer); - MasterProcedureProtos.RestoreSnapshotStateData.Builder restoreSnapshotMsg = - MasterProcedureProtos.RestoreSnapshotStateData.newBuilder() - .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser())) - .setSnapshot(this.snapshot) - .setModifiedTableSchema(ProtobufUtil.toTableSchema(modifiedTableDescriptor)); + RestoreSnapshotStateData.Builder restoreSnapshotMsg = RestoreSnapshotStateData.newBuilder() + .setUserInfo(MasterProcedureUtil.toProtoUserInfo(getUser())) + .setSnapshot(this.snapshot) + .setModifiedTableSchema(ProtobufUtil.toTableSchema(modifiedTableDescriptor)); if (regionsToRestore != null) { for (RegionInfo hri: regionsToRestore) { @@ -264,11 +264,11 @@ public class RestoreSnapshotProcedure while (it.hasNext()) { final Map.Entry> entry = it.next(); - MasterProcedureProtos.RestoreParentToChildRegionsPair.Builder parentToChildrenPair = - MasterProcedureProtos.RestoreParentToChildRegionsPair.newBuilder() - .setParentRegionName(entry.getKey()) - .setChild1RegionName(entry.getValue().getFirst()) - .setChild2RegionName(entry.getValue().getSecond()); + RestoreParentToChildRegionsPair.Builder parentToChildrenPair = + RestoreParentToChildRegionsPair.newBuilder() + .setParentRegionName(entry.getKey()) + .setChild1RegionName(entry.getValue().getFirst()) + .setChild2RegionName(entry.getValue().getSecond()); restoreSnapshotMsg.addParentToChildRegionsPairList (parentToChildrenPair); } } @@ -281,8 +281,8 @@ public class RestoreSnapshotProcedure throws IOException { super.deserializeStateData(serializer); - MasterProcedureProtos.RestoreSnapshotStateData restoreSnapshotMsg = - serializer.deserialize(MasterProcedureProtos.RestoreSnapshotStateData.class); + RestoreSnapshotStateData restoreSnapshotMsg = + serializer.deserialize(RestoreSnapshotStateData.class); setUser(MasterProcedureUtil.toUserInfo(restoreSnapshotMsg.getUserInfo())); snapshot = restoreSnapshotMsg.getSnapshot(); modifiedTableDescriptor = @@ -313,8 +313,8 @@ public class RestoreSnapshotProcedure } } if (restoreSnapshotMsg.getParentToChildRegionsPairListCount() > 0) { - for (MasterProcedureProtos.RestoreParentToChildRegionsPair parentToChildrenPair: - restoreSnapshotMsg.getParentToChildRegionsPairListList()) { + for (RestoreParentToChildRegionsPair parentToChildrenPair : restoreSnapshotMsg + .getParentToChildRegionsPairListList()) { parentsToChildrenPairMap.put( parentToChildrenPair.getParentRegionName(), new Pair<>( diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java index 6da708381af..9e922ee4393 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/namequeues/impl/BalancerRejectionQueueService.java @@ -79,8 +79,8 @@ public class BalancerRejectionQueueService implements NamedQueueService { return; } if (!(namedQueuePayload instanceof BalancerRejectionDetails)) { - LOG.warn( - "BalancerRejectionQueueService: NamedQueuePayload is not of type BalancerRejectionDetails."); + LOG.warn("BalancerRejectionQueueService: NamedQueuePayload is not of type" + + " BalancerRejectionDetails."); return; } BalancerRejectionDetails balancerRejectionDetails = (BalancerRejectionDetails) namedQueuePayload; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java index b285e049bdf..e17f8996431 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java @@ -34,9 +34,10 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; */ @InterfaceAudience.Private @InterfaceStability.Evolving -@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", - justification="FindBugs seems confused; says bypassGlobals, namepaceLimiters, and " + - "tableLimiters are mostly synchronized...but to me it looks like they are totally synchronized") +@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "IS2_INCONSISTENT_SYNC", + justification = "FindBugs seems confused; says bypassGlobals, namepaceLimiters, and " + + "tableLimiters are mostly synchronized..." + + "but to me it looks like they are totally synchronized") public class UserQuotaState extends QuotaState { private Map namespaceLimiters = null; private Map tableLimiters = null; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java index 56dab21baf2..4ffb600dd09 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java @@ -55,9 +55,11 @@ public abstract class AbstractMemStore implements MemStore { protected RegionServicesForStores regionServices; + // @formatter:off public final static long FIXED_OVERHEAD = (long) ClassSize.OBJECT - + (5 * ClassSize.REFERENCE) - + (2 * Bytes.SIZEOF_LONG); // snapshotId, timeOfOldestEdit + + (5 * ClassSize.REFERENCE) + + (2 * Bytes.SIZEOF_LONG); // snapshotId, timeOfOldestEdit + // @formatter:on public final static long DEEP_OVERHEAD = FIXED_OVERHEAD; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java index 63eba8b1a30..01904c685d6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java @@ -2057,8 +2057,9 @@ public class RSRpcServices implements HBaseRPCErrorHandler, throw new ServiceException(ie); } // We are assigning meta, wait a little for regionserver to finish initialization. - int timeout = regionServer.getConfiguration().getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, - HConstants.DEFAULT_HBASE_RPC_TIMEOUT) >> 2; // Quarter of RPC timeout + // Default to quarter of RPC timeout + int timeout = regionServer.getConfiguration() + .getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, HConstants.DEFAULT_HBASE_RPC_TIMEOUT) >> 2; long endTime = System.currentTimeMillis() + timeout; synchronized (regionServer.online) { try { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java index a1d082e9701..5ac86970c9d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java @@ -200,10 +200,10 @@ public class ReplicationSource implements ReplicationSourceInterface { this.waitOnEndpointSeconds = this.conf.getInt(WAIT_ON_ENDPOINT_SECONDS, DEFAULT_WAIT_ON_ENDPOINT_SECONDS); decorateConf(); - this.sleepForRetries = - this.conf.getLong("replication.source.sleepforretries", 1000); // 1 second - this.maxRetriesMultiplier = - this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per + // 1 second + this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000); + // 5 minutes @ 1 sec per + this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300); this.queueSizePerGroup = this.conf.getInt("hbase.regionserver.maxlogs", 32); this.logQueue = new ReplicationSourceLogQueue(conf, metrics, this); this.queueStorage = queueStorage; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java index f188e7ba50d..a5c7b16f529 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceShipper.java @@ -80,14 +80,15 @@ public class ReplicationSourceShipper extends Thread { this.walGroupId = walGroupId; this.logQueue = logQueue; this.source = source; - this.sleepForRetries = - this.conf.getLong("replication.source.sleepforretries", 1000); // 1 second - this.maxRetriesMultiplier = - this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per + // 1 second + this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000); + // 5 minutes @ 1 sec per + this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300); + // 20 seconds this.getEntriesTimeout = - this.conf.getInt("replication.source.getEntries.timeout", DEFAULT_TIMEOUT); // 20 seconds + this.conf.getInt("replication.source.getEntries.timeout", DEFAULT_TIMEOUT); this.shipEditsTimeout = this.conf.getInt(HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT, - HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT_DFAULT); + HConstants.REPLICATION_SOURCE_SHIPEDITS_TIMEOUT_DFAULT); } @Override diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java index 11090448c7c..c61494e12c6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.java @@ -106,10 +106,10 @@ class ReplicationSourceWALReader extends Thread { int batchCount = conf.getInt("replication.source.nb.batches", 1); this.totalBufferUsed = source.getSourceManager().getTotalBufferUsed(); this.totalBufferQuota = source.getSourceManager().getTotalBufferLimit(); - this.sleepForRetries = - this.conf.getLong("replication.source.sleepforretries", 1000); // 1 second - this.maxRetriesMultiplier = - this.conf.getInt("replication.source.maxretriesmultiplier", 300); // 5 minutes @ 1 sec per + // 1 second + this.sleepForRetries = this.conf.getLong("replication.source.sleepforretries", 1000); + // 5 minutes @ 1 sec per + this.maxRetriesMultiplier = this.conf.getInt("replication.source.maxretriesmultiplier", 300); this.eofAutoRecovery = conf.getBoolean("replication.source.eof.autorecovery", false); this.entryBatchQueue = new LinkedBlockingQueue<>(batchCount); this.walGroupId = walGroupId; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java index f540b536ee2..3f47a3cba5f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java @@ -72,7 +72,7 @@ public class TokenUtil { /** - * See {@link ClientTokenUtil#toToken(org.apache.hadoop.security.token.Token)}. + * See {@link ClientTokenUtil#toToken(Token)}. * @deprecated External users should not use this method. Please post on * the HBase dev mailing list if you need this method. Internal * HBase code should use {@link ClientTokenUtil} instead. @@ -83,8 +83,7 @@ public class TokenUtil { } /** - * See {@link ClientTokenUtil#obtainToken(org.apache.hadoop.hbase.client.Connection, - * org.apache.hadoop.hbase.security.User)}. + * See {@link ClientTokenUtil#obtainToken(Connection, User)}. * @deprecated External users should not use this method. Please post on * the HBase dev mailing list if you need this method. Internal * HBase code should use {@link ClientTokenUtil} instead. @@ -96,8 +95,7 @@ public class TokenUtil { } /** - * See {@link ClientTokenUtil#obtainAndCacheToken(org.apache.hadoop.hbase.client.Connection, - * org.apache.hadoop.hbase.security.User)}. + * See {@link ClientTokenUtil#obtainAndCacheToken(Connection, User)}. */ public static void obtainAndCacheToken(final Connection conn, User user) @@ -106,7 +104,7 @@ public class TokenUtil { } /** - * See {@link ClientTokenUtil#toToken(org.apache.hadoop.security.token.Token)}. + * See {@link ClientTokenUtil#toToken(org.apache.hadoop.hbase.shaded.protobuf.generated.AuthenticationProtos.Token)}. * @deprecated External users should not use this method. Please post on * the HBase dev mailing list if you need this method. Internal * HBase code should use {@link ClientTokenUtil} instead. diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java index 908d3e6db77..4c1e5be695c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/EntryBuffers.java @@ -167,8 +167,8 @@ public class EntryBuffers { internify(entry); entries.add(entry); // TODO linkedlist entry - long incrHeap = entry.getEdit().heapSize() + - ClassSize.align(2 * ClassSize.REFERENCE); // WALKey pointers + // entry size plus WALKey pointers + long incrHeap = entry.getEdit().heapSize() + ClassSize.align(2 * ClassSize.REFERENCE); heapInBuffer += incrHeap; return incrHeap; } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index 7c11124711c..ff386cfa232 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -2920,9 +2920,8 @@ public class HBaseTestingUtility extends HBaseZKTestingUtility { if (jobConf == null) { jobConf = mrCluster.createJobConf(); } - - jobConf.set("mapreduce.cluster.local.dir", - conf.get("mapreduce.cluster.local.dir")); //Hadoop MiniMR overwrites this while it should not + // Hadoop MiniMR overwrites this while it should not + jobConf.set("mapreduce.cluster.local.dir", conf.get("mapreduce.cluster.local.dir")); LOG.info("Mini mapreduce cluster started"); // In hadoop2, YARN/MR2 starts a mini cluster with its own conf instance and updates settings. diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java index 9ab021aeb9d..53a84366d1d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java @@ -223,8 +223,9 @@ public class TestRegionObserverScannerOpenHook { Get get = new Get(ROW); Result r = region.get(get); assertNull( - "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: " - + r, r.listCells()); + "Got an unexpected number of rows - " + + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r, + r.listCells()); HBaseTestingUtility.closeRegionAndWAL(region); } @@ -250,8 +251,9 @@ public class TestRegionObserverScannerOpenHook { Get get = new Get(ROW); Result r = region.get(get); assertNull( - "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: " - + r, r.listCells()); + "Got an unexpected number of rows - " + + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r, + r.listCells()); HBaseTestingUtility.closeRegionAndWAL(region); } @@ -269,15 +271,19 @@ public class TestRegionObserverScannerOpenHook { } public CountDownLatch getCompactionStateChangeLatch() { - if (compactionStateChangeLatch == null) compactionStateChangeLatch = new CountDownLatch(1); + if (compactionStateChangeLatch == null) { + compactionStateChangeLatch = new CountDownLatch(1); + } return compactionStateChangeLatch; } @Override public boolean compact(CompactionContext compaction, HStore store, - ThroughputController throughputController) throws IOException { + ThroughputController throughputController) throws IOException { boolean ret = super.compact(compaction, store, throughputController); - if (ret) compactionStateChangeLatch.countDown(); + if (ret) { + compactionStateChangeLatch.countDown(); + } return ret; } @@ -341,14 +347,16 @@ public class TestRegionObserverScannerOpenHook { Get get = new Get(ROW); Result r = table.get(get); assertNull( - "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor. Found: " - + r, r.listCells()); + "Got an unexpected number of rows - " + + "no data should be returned with the NoDataFromScan coprocessor. Found: " + r, + r.listCells()); get = new Get(Bytes.toBytes("anotherrow")); r = table.get(get); assertNull( - "Got an unexpected number of rows - no data should be returned with the NoDataFromScan coprocessor Found: " - + r, r.listCells()); + "Got an unexpected number of rows - " + + "no data should be returned with the NoDataFromScan coprocessor Found: " + r, + r.listCells()); table.close(); UTIL.shutdownMiniCluster(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java index a42410a2aab..289ed670841 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/favored/TestFavoredNodeAssignmentHelper.java @@ -190,8 +190,9 @@ public class TestFavoredNodeAssignmentHelper { // the primary can be assigned but the secondary/tertiary would be null Map rackToServerCount = new HashMap<>(); rackToServerCount.put("rack1", 1); - Triple, FavoredNodeAssignmentHelper, List> - primaryRSMapAndHelper = secondaryAndTertiaryRSPlacementHelper(1, rackToServerCount); + Triple, FavoredNodeAssignmentHelper, + List> primaryRSMapAndHelper = + secondaryAndTertiaryRSPlacementHelper(1, rackToServerCount); FavoredNodeAssignmentHelper helper = primaryRSMapAndHelper.getSecond(); Map primaryRSMap = primaryRSMapAndHelper.getFirst(); List regions = primaryRSMapAndHelper.getThird(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java index a5215bfe1d9..9f9e2540819 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestFixedFileTrailer.java @@ -206,9 +206,10 @@ public class TestFixedFileTrailer { String msg = ex.getMessage(); String cleanMsg = msg.replaceAll( "^(java(\\.[a-zA-Z]+)+:\\s+)?|\\s+\\(.*\\)\\s*$", ""); - assertEquals("Actual exception message is \"" + msg + "\".\n" + - "Cleaned-up message", // will be followed by " expected: ..." - "Invalid HFile version: " + invalidVersion, cleanMsg); + // will be followed by " expected: ..." + assertEquals("Actual exception message is \"" + msg + "\".\nCleaned-up message", + "Invalid HFile version: " + invalidVersion, + cleanMsg); LOG.info("Got an expected exception: " + msg); } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java index 50c8bc24071..d1e1f5cbeeb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java @@ -308,28 +308,29 @@ public class TestHFileBlock { @Test public void testGzipCompression() throws IOException { - final String correctTestBlockStr = - "DATABLK*\\x00\\x00\\x00>\\x00\\x00\\x0F\\xA0\\xFF\\xFF\\xFF\\xFF" - + "\\xFF\\xFF\\xFF\\xFF" - + "\\x0" + ChecksumType.getDefaultChecksumType().getCode() - + "\\x00\\x00@\\x00\\x00\\x00\\x00[" - // gzip-compressed block: http://www.gzip.org/zlib/rfc-gzip.html - + "\\x1F\\x8B" // gzip magic signature - + "\\x08" // Compression method: 8 = "deflate" - + "\\x00" // Flags - + "\\x00\\x00\\x00\\x00" // mtime - + "\\x00" // XFL (extra flags) - // OS (0 = FAT filesystems, 3 = Unix). However, this field - // sometimes gets set to 0 on Linux and Mac, so we reset it to 3. - // This appears to be a difference caused by the availability - // (and use) of the native GZ codec. - + "\\x03" - + "\\xED\\xC3\\xC1\\x11\\x00 \\x08\\xC00DD\\xDD\\x7Fa" - + "\\xD6\\xE8\\xA3\\xB9K\\x84`\\x96Q\\xD3\\xA8\\xDB\\xA8e\\xD4c" - + "\\xD46\\xEA5\\xEA3\\xEA7\\xE7\\x00LI\\x5Cs\\xA0\\x0F\\x00\\x00" - + "\\x00\\x00\\x00\\x00"; // 4 byte checksum (ignored) - final int correctGzipBlockLength = 95; - final String testBlockStr = createTestBlockStr(GZ, correctGzipBlockLength, false); + // @formatter:off + String correctTestBlockStr = "DATABLK*\\x00\\x00\\x00>\\x00\\x00\\x0F\\xA0\\xFF\\xFF\\xFF\\xFF" + + "\\xFF\\xFF\\xFF\\xFF" + + "\\x0" + ChecksumType.getDefaultChecksumType().getCode() + + "\\x00\\x00@\\x00\\x00\\x00\\x00[" + // gzip-compressed block: http://www.gzip.org/zlib/rfc-gzip.html + + "\\x1F\\x8B" // gzip magic signature + + "\\x08" // Compression method: 8 = "deflate" + + "\\x00" // Flags + + "\\x00\\x00\\x00\\x00" // mtime + + "\\x00" // XFL (extra flags) + // OS (0 = FAT filesystems, 3 = Unix). However, this field + // sometimes gets set to 0 on Linux and Mac, so we reset it to 3. + // This appears to be a difference caused by the availability + // (and use) of the native GZ codec. + + "\\x03" + + "\\xED\\xC3\\xC1\\x11\\x00 \\x08\\xC00DD\\xDD\\x7Fa" + + "\\xD6\\xE8\\xA3\\xB9K\\x84`\\x96Q\\xD3\\xA8\\xDB\\xA8e\\xD4c" + + "\\xD46\\xEA5\\xEA3\\xEA7\\xE7\\x00LI\\x5Cs\\xA0\\x0F\\x00\\x00" + + "\\x00\\x00\\x00\\x00"; // 4 byte checksum (ignored) + // @formatter:on + int correctGzipBlockLength = 95; + String testBlockStr = createTestBlockStr(GZ, correctGzipBlockLength, false); // We ignore the block checksum because createTestBlockStr can change the // gzip header after the block is produced assertEquals(correctTestBlockStr.substring(0, correctGzipBlockLength - 4), diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java index b5fbc96d9f6..46c544ecde5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java @@ -75,34 +75,24 @@ public class TestMetricsRegion { // test region with replica id > 0 mr = new MetricsRegion(new MetricsRegionWrapperStub(1), new Configuration()); agg = mr.getSource().getAggregateSource(); - HELPER.assertGauge( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_storeCount", - 101, agg); - HELPER.assertGauge( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_storeFileCount", - 102, agg); - HELPER.assertGauge( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_memstoreSize", - 103, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_" + - "filteredReadRequestCount", - 107, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_replicaid", - 1, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_compactionsQueuedCount", - 4, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_flushesQueuedCount", - 6, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_maxCompactionQueueSize", - 4, agg); - HELPER.assertCounter( - "namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001_metric_maxFlushQueueSize", - 6, agg); + HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_storeCount", 101, agg); + HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_storeFileCount", 102, agg); + HELPER.assertGauge("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_memstoreSize", 103, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_filteredReadRequestCount", 107, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_replicaid", 1, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_compactionsQueuedCount", 4, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_flushesQueuedCount", 6, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_maxCompactionQueueSize", 4, agg); + HELPER.assertCounter("namespace_TestNS_table_MetricsRegionWrapperStub_region_DEADBEEF001" + + "_metric_maxFlushQueueSize", 6, agg); mr.close(); } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java index fd1862f01c0..e1d15178a88 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionMergeTransactionOnCluster.java @@ -392,17 +392,30 @@ public class TestRegionMergeTransactionOnCluster { for (Pair p : currentRegionToServers) { currentRegions.add(p.getFirst()); } - assertTrue(initialRegions.contains(mergedRegions.getFirst())); //this is the first region - assertTrue(initialRegions.contains(RegionReplicaUtil - .getRegionInfoForReplica(mergedRegions.getFirst(), 1))); //this is the replica of the first region - assertTrue(initialRegions.contains(mergedRegions.getSecond())); //this is the second region - assertTrue(initialRegions.contains(RegionReplicaUtil - .getRegionInfoForReplica(mergedRegions.getSecond(), 1))); //this is the replica of the second region - assertTrue(!initialRegions.contains(currentRegions.get(0))); //this is the new region - assertTrue(!initialRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); //replica of the new region - assertTrue(currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); //replica of the new region - assertTrue(!currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1))); //replica of the merged region - assertTrue(!currentRegions.contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1))); //replica of the merged region + // this is the first region + assertTrue(initialRegions.contains(mergedRegions.getFirst())); + // this is the replica of the first region + assertTrue(initialRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1))); + // this is the second region + assertTrue(initialRegions.contains(mergedRegions.getSecond())); + // this is the replica of the second region + assertTrue(initialRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1))); + // this is the new region + assertTrue(!initialRegions.contains(currentRegions.get(0))); + // replica of the new region + assertTrue(!initialRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); + // replica of the new region + assertTrue(currentRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(currentRegions.get(0), 1))); + // replica of the merged region + assertTrue(!currentRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getFirst(), 1))); + // replica of the merged region + assertTrue(!currentRegions + .contains(RegionReplicaUtil.getRegionInfoForReplica(mergedRegions.getSecond(), 1))); table.close(); } finally { TEST_UTIL.deleteTable(tableName); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java index 21a2bfbb993..7768eaa1319 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java @@ -707,22 +707,30 @@ public class TestStoreScanner { @Test public void testWildCardScannerUnderDeletes() throws IOException { - KeyValue [] kvs = new KeyValue [] { - create("R1", "cf", "a", 2, KeyValue.Type.Put, "dont-care"), // inc - // orphaned delete column. - create("R1", "cf", "a", 1, KeyValue.Type.DeleteColumn, "dont-care"), - // column b - create("R1", "cf", "b", 2, KeyValue.Type.Put, "dont-care"), // inc - create("R1", "cf", "b", 1, KeyValue.Type.Put, "dont-care"), // inc - // column c - create("R1", "cf", "c", 10, KeyValue.Type.Delete, "dont-care"), - create("R1", "cf", "c", 10, KeyValue.Type.Put, "dont-care"), // no - create("R1", "cf", "c", 9, KeyValue.Type.Put, "dont-care"), // inc - // column d - create("R1", "cf", "d", 11, KeyValue.Type.Put, "dont-care"), // inc - create("R1", "cf", "d", 10, KeyValue.Type.DeleteColumn, "dont-care"), - create("R1", "cf", "d", 9, KeyValue.Type.Put, "dont-care"), // no - create("R1", "cf", "d", 8, KeyValue.Type.Put, "dont-care"), // no + KeyValue[] kvs = new KeyValue[] { + // inc + create("R1", "cf", "a", 2, KeyValue.Type.Put, "dont-care"), + // orphaned delete column. + create("R1", "cf", "a", 1, KeyValue.Type.DeleteColumn, "dont-care"), + // column b + // inc + create("R1", "cf", "b", 2, KeyValue.Type.Put, "dont-care"), + // inc + create("R1", "cf", "b", 1, KeyValue.Type.Put, "dont-care"), + // column c + create("R1", "cf", "c", 10, KeyValue.Type.Delete, "dont-care"), + // no + create("R1", "cf", "c", 10, KeyValue.Type.Put, "dont-care"), + // inc + create("R1", "cf", "c", 9, KeyValue.Type.Put, "dont-care"), + // column d + // inc + create("R1", "cf", "d", 11, KeyValue.Type.Put, "dont-care"), + create("R1", "cf", "d", 10, KeyValue.Type.DeleteColumn, "dont-care"), + // no + create("R1", "cf", "d", 9, KeyValue.Type.Put, "dont-care"), + // no + create("R1", "cf", "d", 8, KeyValue.Type.Put, "dont-care"), }; List scanners = scanFixture(kvs); @@ -925,6 +933,7 @@ public class TestStoreScanner { return now; } }); + // @formatter:off KeyValue[] kvs = new KeyValue[]{ /*0*/ new KeyValue(Bytes.toBytes("R1"), Bytes.toBytes("cf"), null, now - 100, KeyValue.Type.DeleteFamily), // live @@ -943,22 +952,23 @@ public class TestStoreScanner { /*7*/ create("R1", "cf", "a", now - 100, KeyValue.Type.DeleteColumn, "dont-care"), // max-version /*8*/ create("R1", "cf", "b", now - 600, - KeyValue.Type.DeleteColumn, "dont-care"), //expired + KeyValue.Type.DeleteColumn, "dont-care"), // expired /*9*/ create("R1", "cf", "b", now - 70, - KeyValue.Type.Put, "v2"), //live + KeyValue.Type.Put, "v2"), // live /*10*/ create("R1", "cf", "b", now - 750, - KeyValue.Type.Put, "v1"), //expired + KeyValue.Type.Put, "v1"), // expired /*11*/ create("R1", "cf", "c", now - 500, - KeyValue.Type.Delete, "dontcare"), //expired + KeyValue.Type.Delete, "dontcare"), // expired /*12*/ create("R1", "cf", "c", now - 600, - KeyValue.Type.Put, "v1"), //expired + KeyValue.Type.Put, "v1"), // expired /*13*/ create("R1", "cf", "c", now - 1000, - KeyValue.Type.Delete, "dontcare"), //expired + KeyValue.Type.Delete, "dontcare"), // expired /*14*/ create("R1", "cf", "d", now - 60, - KeyValue.Type.Put, "expired put"), //live + KeyValue.Type.Put, "expired put"), // live /*15*/ create("R1", "cf", "d", now - 100, - KeyValue.Type.Delete, "not-expired delete"), //live + KeyValue.Type.Delete, "not-expired delete"), // live }; + // @formatter:on List scanners = scanFixture(kvs); ScanInfo scanInfo = new ScanInfo(CONF, Bytes.toBytes("cf"), 0 /* minVersions */, diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java index 8f98d7b43d0..64fa54c44a3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestCompactionWithThroughputController.java @@ -17,6 +17,9 @@ */ package org.apache.hadoop.hbase.regionserver.throttle; +import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND; +import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND; +import static org.apache.hadoop.hbase.regionserver.throttle.PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertTrue; @@ -110,14 +113,8 @@ public class TestCompactionWithThroughputController { conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MIN_KEY, 100); conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MAX_KEY, 200); conf.setInt(HStore.BLOCKING_STOREFILES_KEY, 10000); - conf.setLong( - PressureAwareCompactionThroughputController - .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, - throughputLimit); - conf.setLong( - PressureAwareCompactionThroughputController - .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, - throughputLimit); + conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, throughputLimit); + conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, throughputLimit); conf.set(CompactionThroughputControllerFactory.HBASE_THROUGHPUT_CONTROLLER_KEY, PressureAwareCompactionThroughputController.class.getName()); TEST_UTIL.startMiniCluster(1); @@ -182,21 +179,13 @@ public class TestCompactionWithThroughputController { public void testThroughputTuning() throws Exception { Configuration conf = TEST_UTIL.getConfiguration(); conf.set(StoreEngine.STORE_ENGINE_CLASS_KEY, DefaultStoreEngine.class.getName()); - conf.setLong( - PressureAwareCompactionThroughputController - .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, - 20L * 1024 * 1024); - conf.setLong( - PressureAwareCompactionThroughputController - .HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, - 10L * 1024 * 1024); + conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_HIGHER_BOUND, 20L * 1024 * 1024); + conf.setLong(HBASE_HSTORE_COMPACTION_MAX_THROUGHPUT_LOWER_BOUND, 10L * 1024 * 1024); conf.setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MIN_KEY, 4); conf.setInt(HStore.BLOCKING_STOREFILES_KEY, 6); conf.set(CompactionThroughputControllerFactory.HBASE_THROUGHPUT_CONTROLLER_KEY, PressureAwareCompactionThroughputController.class.getName()); - conf.setInt( - PressureAwareCompactionThroughputController.HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD, - 1000); + conf.setInt(HBASE_HSTORE_COMPACTION_THROUGHPUT_TUNE_PERIOD, 1000); TEST_UTIL.startMiniCluster(1); Connection conn = ConnectionFactory.createConnection(conf); try { diff --git a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java index 602eff87acd..b0eb375611e 100644 --- a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java +++ b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKMulti.java @@ -288,11 +288,15 @@ public class TestZKMulti { // test that, even with operations that fail, the ones that would pass will pass // with runSequentialOnMultiFailure ops = new LinkedList<>(); - ops.add(ZKUtilOp.setData(path1, Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1)))); // pass - ops.add(ZKUtilOp.deleteNodeFailSilent(path2)); // pass - ops.add(ZKUtilOp.deleteNodeFailSilent(path3)); // fail -- node doesn't exist - ops.add(ZKUtilOp.createAndFailSilent(path4, - Bytes.add(Bytes.toBytes(path4), Bytes.toBytes(path4)))); // pass + // pass + ops.add(ZKUtilOp.setData(path1, Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1)))); + // pass + ops.add(ZKUtilOp.deleteNodeFailSilent(path2)); + // fail -- node doesn't exist + ops.add(ZKUtilOp.deleteNodeFailSilent(path3)); + // pass + ops.add( + ZKUtilOp.createAndFailSilent(path4, Bytes.add(Bytes.toBytes(path4), Bytes.toBytes(path4)))); ZKUtil.multiOrSequential(zkw, ops, true); assertTrue(Bytes.equals(ZKUtil.getData(zkw, path1), Bytes.add(Bytes.toBytes(path1), Bytes.toBytes(path1))));