diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/OzoneConfiguration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/OzoneConfiguration.java index bd7ecc3110d..ed118bfb7dd 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/OzoneConfiguration.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/OzoneConfiguration.java @@ -148,8 +148,8 @@ public int hashCode(){ @Override public boolean equals(Object obj) { - return (obj instanceof Property) && (((Property) obj).getName()).equals - (this.getName()); + return (obj instanceof Property) && (((Property) obj).getName()) + .equals(this.getName()); } } } \ No newline at end of file diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/jmx/ServiceRuntimeInfo.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/jmx/ServiceRuntimeInfo.java index 8250bed2b93..085872b8d60 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/jmx/ServiceRuntimeInfo.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/jmx/ServiceRuntimeInfo.java @@ -41,14 +41,14 @@ public interface ServiceRuntimeInfo { String getVersion(); /** - * Get the version of software running on the Namenode + * Get the version of software running on the Namenode. * * @return a string representing the version */ String getSoftwareVersion(); /** - * Get the compilation information which contains date, user and branch + * Get the compilation information which contains date, user and branch. * * @return the compilation information, as a JSON string. */ diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/ozone/client/io/ChunkGroupOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/ozone/client/io/ChunkGroupOutputStream.java index d0975a8ffc4..d9f271da2fe 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/ozone/client/io/ChunkGroupOutputStream.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/ozone/client/io/ChunkGroupOutputStream.java @@ -332,8 +332,8 @@ public Builder setRequestID(String id) { return this; } - public Builder setType(ReplicationType type) { - this.type = type; + public Builder setType(ReplicationType replicationType) { + this.type = replicationType; return this; } diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/container/common/helpers/ContainerInfo.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/container/common/helpers/ContainerInfo.java index f99fdf85d19..ff8bcdb1643 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/container/common/helpers/ContainerInfo.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/container/common/helpers/ContainerInfo.java @@ -249,8 +249,8 @@ public Builder setState(OzoneProtos.LifeCycleState lifeCycleState) { return this; } - public Builder setPipeline(Pipeline pipeline) { - this.pipeline = pipeline; + public Builder setPipeline(Pipeline containerPipeline) { + this.pipeline = containerPipeline; return this; } @@ -269,18 +269,18 @@ public Builder setNumberOfKeys(long keyCount) { return this; } - public Builder setStateEnterTime(long stateEnterTime) { - this.stateEnterTime = stateEnterTime; + public Builder setStateEnterTime(long time) { + this.stateEnterTime = time; return this; } - public Builder setOwner(OzoneProtos.Owner owner) { - this.owner = owner; + public Builder setOwner(OzoneProtos.Owner containerOwner) { + this.owner = containerOwner; return this; } - public Builder setContainerName(String containerName) { - this.containerName = containerName; + public Builder setContainerName(String container) { + this.containerName = container; return this; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-balancer.sh b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-balancer.sh index df044fe48ba..bde4b3a175b 100755 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-balancer.sh +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-balancer.sh @@ -17,6 +17,10 @@ MYNAME="${BASH_SOURCE-$0}" +## @description usage info +## @audience private +## @stability evolving +## @replaceable no function hadoop_usage { hadoop_add_option "--buildpaths" "attempt to add class files from build tree" diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-ozone.sh b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-ozone.sh index ca6b6b82554..097490e6206 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-ozone.sh +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/start-ozone.sh @@ -17,7 +17,10 @@ # Start hadoop hdfs and ozone daemons. # Run this on master node. - +## @description usage info +## @audience private +## @stability evolving +## @replaceable no function hadoop_usage { echo "Usage: start-ozone.sh" diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-balancer.sh b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-balancer.sh index ec94080ece5..487aee8925e 100755 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-balancer.sh +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-balancer.sh @@ -17,6 +17,10 @@ MYNAME="${BASH_SOURCE-$0}" +## @description usage info +## @audience private +## @stability evolving +## @replaceable no function hadoop_usage { hadoop_add_option "--buildpaths" "attempt to add class files from build tree" diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-ozone.sh b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-ozone.sh index 15255c02993..bf7309d6595 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-ozone.sh +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/stop-ozone.sh @@ -17,7 +17,10 @@ # Stop hdfs and ozone daemons. # Run this on master node. - +## @description usage info +## @audience private +## @stability evolving +## @replaceable no function hadoop_usage { echo "Usage: stop-ozone.sh" diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/cblock/jscsiHelper/cache/impl/CBlockLocalCache.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/cblock/jscsiHelper/cache/impl/CBlockLocalCache.java index 1c93b40febb..1149164f946 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/cblock/jscsiHelper/cache/impl/CBlockLocalCache.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/cblock/jscsiHelper/cache/impl/CBlockLocalCache.java @@ -434,88 +434,88 @@ private static long computeCacheSize(Configuration configuration, /** * Sets the Config to be used by this cache. * - * @param configuration - Config + * @param conf - Config * @return Builder */ - public Builder setConfiguration(Configuration configuration) { - this.configuration = configuration; + public Builder setConfiguration(Configuration conf) { + this.configuration = conf; return this; } /** * Sets the user name who is the owner of this volume. * - * @param userName - name of the owner, please note this is not the current + * @param user - name of the owner, please note this is not the current * user name. * @return - Builder */ - public Builder setUserName(String userName) { - this.userName = userName; + public Builder setUserName(String user) { + this.userName = user; return this; } /** * Sets the VolumeName. * - * @param volumeName - Name of the volume + * @param volume - Name of the volume * @return Builder */ - public Builder setVolumeName(String volumeName) { - this.volumeName = volumeName; + public Builder setVolumeName(String volume) { + this.volumeName = volume; return this; } /** * Sets the Pipelines that form this volume. * - * @param pipelines - list of pipelines + * @param pipelineList - list of pipelines * @return Builder */ - public Builder setPipelines(List pipelines) { - this.pipelines = pipelines; + public Builder setPipelines(List pipelineList) { + this.pipelines = pipelineList; return this; } /** * Sets the Client Manager that manages the communication with containers. * - * @param clientManager - clientManager. + * @param xceiverClientManager - clientManager. * @return - Builder */ - public Builder setClientManager(XceiverClientManager clientManager) { - this.clientManager = clientManager; + public Builder setClientManager(XceiverClientManager xceiverClientManager) { + this.clientManager = xceiverClientManager; return this; } /** * Sets the block size -- Typical sizes are 4KB, 8KB etc. * - * @param blockSize - BlockSize. + * @param size - BlockSize. * @return - Builder */ - public Builder setBlockSize(int blockSize) { - this.blockSize = blockSize; + public Builder setBlockSize(int size) { + this.blockSize = size; return this; } /** * Sets the volumeSize. * - * @param volumeSize - VolumeSize + * @param size - VolumeSize * @return - Builder */ - public Builder setVolumeSize(long volumeSize) { - this.volumeSize = volumeSize; + public Builder setVolumeSize(long size) { + this.volumeSize = size; return this; } /** * Set flusher. - * @param flusher - cache Flusher + * @param containerCacheFlusher - cache Flusher * @return Builder. */ - public Builder setFlusher(ContainerCacheFlusher flusher) { - this.flusher = flusher; + public Builder setFlusher(ContainerCacheFlusher containerCacheFlusher) { + this.flusher = containerCacheFlusher; return this; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java index 349ed3f8cf4..a31b77292c1 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java @@ -321,9 +321,9 @@ public boolean isDaemonStopped() { /** * Create a command handler thread. * - * @param conf + * @param config */ - private void initCommandHandlerThread(Configuration conf) { + private void initCommandHandlerThread(Configuration config) { /** * Task that periodically checks if we have any outstanding commands. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/CommandDispatcher.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/CommandDispatcher.java index c682000eeb9..6670ea9d83d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/CommandDispatcher.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/CommandDispatcher.java @@ -125,37 +125,37 @@ public Builder addHandler(CommandHandler handler) { /** * Add the OzoneContainer. * - * @param container - ozone container. + * @param ozoneContainer - ozone container. * @return Builder */ - public Builder setContainer(OzoneContainer container) { - Preconditions.checkNotNull(container); - this.container = container; + public Builder setContainer(OzoneContainer ozoneContainer) { + Preconditions.checkNotNull(ozoneContainer); + this.container = ozoneContainer; return this; } /** * Set the Connection Manager. * - * @param connectionManager + * @param scmConnectionManager * @return this */ public Builder setConnectionManager(SCMConnectionManager - connectionManager) { - Preconditions.checkNotNull(connectionManager); - this.connectionManager = connectionManager; + scmConnectionManager) { + Preconditions.checkNotNull(scmConnectionManager); + this.connectionManager = scmConnectionManager; return this; } /** * Sets the Context. * - * @param context - StateContext + * @param stateContext - StateContext * @return this */ - public Builder setContext(StateContext context) { - Preconditions.checkNotNull(context); - this.context = context; + public Builder setContext(StateContext stateContext) { + Preconditions.checkNotNull(stateContext); + this.context = stateContext; return this; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/LongMetric.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/LongMetric.java index 5fc22b10eae..df3f4626016 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/LongMetric.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/LongMetric.java @@ -92,32 +92,32 @@ public Long get() { /** * Sets the value of this metric. * - * @param value - value of the metric. + * @param setValue - value of the metric. */ @Override - public void set(Long value) { - this.value = value; + public void set(Long setValue) { + this.value = setValue; } /** * Adds a value of to the base. * - * @param value - value + * @param addValue - value */ @Override - public void add(Long value) { - this.value += value; + public void add(Long addValue) { + this.value += addValue; } /** * subtract a value. * - * @param value value + * @param subValue value */ @Override - public void subtract(Long value) { - this.value -= value; + public void subtract(Long subValue) { + this.value -= subValue; } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/SCMNodeStat.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/SCMNodeStat.java index bb189748717..64f76a6b0a5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/SCMNodeStat.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/scm/container/placement/metrics/SCMNodeStat.java @@ -73,26 +73,26 @@ public LongMetric getRemaining() { /** * Set the capacity, used and remaining space on a datanode. * - * @param capacity in bytes - * @param used in bytes - * @param remaining in bytes + * @param newCapacity in bytes + * @param newUsed in bytes + * @param newRemaining in bytes */ @VisibleForTesting - public void set(long capacity, long used, long remaining) { - Preconditions.checkNotNull(capacity, "Capacity cannot be null"); - Preconditions.checkNotNull(used, "used cannot be null"); - Preconditions.checkNotNull(remaining, "remaining cannot be null"); + public void set(long newCapacity, long newUsed, long newRemaining) { + Preconditions.checkNotNull(newCapacity, "Capacity cannot be null"); + Preconditions.checkNotNull(newUsed, "used cannot be null"); + Preconditions.checkNotNull(newRemaining, "remaining cannot be null"); - Preconditions.checkArgument(capacity >= 0, "Capacity cannot be " + + Preconditions.checkArgument(newCapacity >= 0, "Capacity cannot be " + "negative."); - Preconditions.checkArgument(used >= 0, "used space cannot be " + + Preconditions.checkArgument(newUsed >= 0, "used space cannot be " + "negative."); - Preconditions.checkArgument(remaining >= 0, "remaining cannot be " + + Preconditions.checkArgument(newRemaining >= 0, "remaining cannot be " + "negative"); - this.capacity = new LongMetric(capacity); - this.scmUsed = new LongMetric(used); - this.remaining = new LongMetric(remaining); + this.capacity = new LongMetric(newCapacity); + this.scmUsed = new LongMetric(newUsed); + this.remaining = new LongMetric(newRemaining); } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/interfaces/Keys.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/interfaces/Keys.java index 853a9f1b41b..f9255f2d54a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/interfaces/Keys.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/interfaces/Keys.java @@ -65,9 +65,9 @@ public interface Keys { */ @PUT @Consumes(MediaType.WILDCARD) - @ApiOperation(value = "Adds a key to an existing bucket.",notes = "If the " - + "object already exists this call will overwrite or add with new version " - + "number if the bucket versioning is turned on.") + @ApiOperation(value = "Adds a key to an existing bucket.", notes = "If the " + + "object already exists this call will overwrite or add with new version" + + " number if the bucket versioning is turned on.") @ApiImplicitParams({ @ApiImplicitParam(name = "x-ozone-version", example = "v1", required = true, paramType = "header"), diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/proto/StorageContainerDatanodeProtocol.proto b/hadoop-hdfs-project/hadoop-hdfs/src/main/proto/StorageContainerDatanodeProtocol.proto index 88518e3fe62..2cbe37e5536 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/proto/StorageContainerDatanodeProtocol.proto +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/proto/StorageContainerDatanodeProtocol.proto @@ -268,9 +268,9 @@ message ContainerBlocksDeletionACKProto { * Here is a simple state diagram that shows how a datanode would boot up and * communicate with SCM. * - * ----------------------- + * ----------------------- * | Start | - * ---------- ------------ + * ---------- ------------ * | * | * | @@ -278,19 +278,19 @@ message ContainerBlocksDeletionACKProto { * | * | * | - * ----------v------------- - * | Searching for SCM ------------ + * ----------v------------- + * | Searching for SCM ------------ * ---------- ------------- | * | | * | | - * | ----------v------------- + * | ----------v------------- * | | Register if needed | - * | ----------- ------------ + * | ----------- ------------ * | | * v | * ----------- ---------------- | - * --------- Heartbeat state <-------- - * | --------^------------------- + * --------- Heartbeat state <-------- + * | --------^------------------- * | | * | | * | | @@ -298,7 +298,7 @@ message ContainerBlocksDeletionACKProto { * | | * | | * | | - * ------------------ + * ------------------ * * * diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestBufferManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestBufferManager.java index 3b5f362fe09..4211255e38e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestBufferManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestBufferManager.java @@ -83,7 +83,7 @@ public static void shutdown() throws InterruptedException { if (cluster != null) { cluster.shutdown(); } - IOUtils.cleanup(null, storageContainerLocationClient, cluster); + IOUtils.cleanupWithLogger(null, storageContainerLocationClient, cluster); } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestCBlockReadWrite.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestCBlockReadWrite.java index 2d1fc5a726e..e0389de62f3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestCBlockReadWrite.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestCBlockReadWrite.java @@ -90,7 +90,7 @@ public static void shutdown() throws InterruptedException { if (cluster != null) { cluster.shutdown(); } - IOUtils.cleanup(null, storageContainerLocationClient, cluster); + IOUtils.cleanupWithLogger(null, storageContainerLocationClient, cluster); } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestLocalBlockCache.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestLocalBlockCache.java index f60c4b7dd19..ea7c8dd090b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestLocalBlockCache.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/cblock/TestLocalBlockCache.java @@ -96,7 +96,7 @@ public static void shutdown() throws InterruptedException { if (cluster != null) { cluster.shutdown(); } - IOUtils.cleanup(null, storageContainerLocationClient, cluster); + IOUtils.cleanupWithLogger(null, storageContainerLocationClient, cluster); } /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java index dcaf68440a2..c11323ab682 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java @@ -169,10 +169,10 @@ private void sleepIfNeeded() { @Override public StorageContainerDatanodeProtocolProtos.SCMHeartbeatResponseProto sendHeartbeat(DatanodeID datanodeID, SCMNodeReport nodeReport, - ReportState reportState) throws IOException { + ReportState scmReportState) throws IOException { rpcCount.incrementAndGet(); heartbeatCount.incrementAndGet(); - this.reportState = reportState; + this.reportState = scmReportState; sleepIfNeeded(); List cmdResponses = new LinkedList<>(); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestBucketInfo.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestBucketInfo.java index 0678b630531..2e699225947 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestBucketInfo.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestBucketInfo.java @@ -29,6 +29,9 @@ import java.util.LinkedList; import java.util.List; +/** + * Test Ozone Bucket Info operation. + */ public class TestBucketInfo { @Test public void testBucketInfoJson() throws IOException { diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestErrorCode.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestErrorCode.java index 73b624367d2..abb61bb8c38 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestErrorCode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestErrorCode.java @@ -25,6 +25,9 @@ import static junit.framework.TestCase.assertEquals; import static org.apache.hadoop.ozone.web.utils.OzoneUtils.getRequestID; +/** + * Test Ozone Error Codes. + */ public class TestErrorCode { /** * Test Error Generator functions. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestOzoneWebAccess.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestOzoneWebAccess.java index 381ba37a26e..d75dce278e5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestOzoneWebAccess.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestOzoneWebAccess.java @@ -46,6 +46,9 @@ import static org.apache.hadoop.ozone.web.utils.OzoneUtils.getRequestID; import static org.junit.Assert.assertEquals; +/** + * Test Ozone Access through REST protocol. + */ public class TestOzoneWebAccess { /** * Set the timeout for every test. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestQuota.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestQuota.java index 9b5a578dad9..d777d0cd897 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestQuota.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestQuota.java @@ -29,7 +29,9 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; - +/** + * Test Ozone Volume Quota. + */ public class TestQuota { @Test public void testParseQuota() { diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestUtils.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestUtils.java index 4891e4cb6f0..d3f8f5e659c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestUtils.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestUtils.java @@ -29,6 +29,9 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +/** + * Test Ozone Utility operations like verifying resource name. + */ public class TestUtils { /** diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestVolumeStructs.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestVolumeStructs.java index ef5a3468006..b433be6fb51 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestVolumeStructs.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/TestVolumeStructs.java @@ -28,6 +28,9 @@ import static org.junit.Assert.assertEquals; +/** + * Test Ozone Volume info structure. + */ public class TestVolumeStructs { @Test diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestBuckets.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestBuckets.java index 90e587145cb..c3c9a4ab1bf 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestBuckets.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestBuckets.java @@ -44,7 +44,9 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; - +/** + * Test Ozone Bucket Lifecycle. + */ public class TestBuckets { /** * Set the timeout for every test. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestKeys.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestKeys.java index c3b2dcc8a06..b44d782e1d2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestKeys.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestKeys.java @@ -74,6 +74,9 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +/** + * Test Ozone Key Lifecycle. + */ public class TestKeys { /** * Set the timeout for every test. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestVolume.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestVolume.java index 3f124a3f1ba..322216a79fe 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestVolume.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/ozone/web/client/TestVolume.java @@ -58,6 +58,9 @@ import static org.mockito.Mockito.times; import static org.mockito.Mockito.verify; +/** + * Test Ozone Volumes Lifecycle. + */ public class TestVolume { private static MiniOzoneCluster cluster = null; private static OzoneRestClient ozoneRestClient = null; diff --git a/hadoop-tools/hadoop-ozone/src/main/java/org/apache/hadoop/fs/ozone/Constants.java b/hadoop-tools/hadoop-ozone/src/main/java/org/apache/hadoop/fs/ozone/Constants.java index 1f9b05519ef..992d43a077f 100644 --- a/hadoop-tools/hadoop-ozone/src/main/java/org/apache/hadoop/fs/ozone/Constants.java +++ b/hadoop-tools/hadoop-ozone/src/main/java/org/apache/hadoop/fs/ozone/Constants.java @@ -21,7 +21,7 @@ /** * Constants for Ozone FileSystem implementation. */ -public class Constants { +public final class Constants { public static final String OZONE_URI_SCHEME = "o3"; diff --git a/hadoop-tools/hadoop-ozone/src/test/java/org/apache/hadoop/fs/ozone/contract/OzoneContract.java b/hadoop-tools/hadoop-ozone/src/test/java/org/apache/hadoop/fs/ozone/contract/OzoneContract.java index 21db6975dc8..04e6c4f5bb2 100644 --- a/hadoop-tools/hadoop-ozone/src/test/java/org/apache/hadoop/fs/ozone/contract/OzoneContract.java +++ b/hadoop-tools/hadoop-ozone/src/test/java/org/apache/hadoop/fs/ozone/contract/OzoneContract.java @@ -48,7 +48,7 @@ class OzoneContract extends AbstractFSContract { private static StorageHandler storageHandler; private static final String CONTRACT_XML = "contract/ozone.xml"; - public OzoneContract(Configuration conf) { + OzoneContract(Configuration conf) { super(conf); //insert the base features addConfResource(CONTRACT_XML);