HDDS-843. [JDK11] Fix Javadoc errors in hadoop-hdds-server-scm module. Contributed by Dinesh Chitlangia.
This commit is contained in:
parent
57866b366f
commit
d8ec017a8d
|
@ -1,18 +1,19 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with this
|
||||
* work for additional information regarding copyright ownership. The ASF
|
||||
* licenses this file to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
* <p/>
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
* <p/>
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
* License for the specific language governing permissions and limitations under
|
||||
* the License.
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdds.scm.exceptions;
|
||||
// Exceptions thrown by SCM.
|
|
@ -23,7 +23,6 @@ import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
|
|||
import org.apache.hadoop.hdds.scm.pipeline.PipelineID;
|
||||
import org.apache.hadoop.hdds.scm.container.placement.metrics.SCMNodeMetric;
|
||||
import org.apache.hadoop.hdds.scm.container.placement.metrics.SCMNodeStat;
|
||||
import org.apache.hadoop.hdds.scm.exceptions.SCMException;
|
||||
import org.apache.hadoop.hdds.scm.node.states.NodeNotFoundException;
|
||||
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
|
||||
import org.apache.hadoop.hdds.protocol.proto.HddsProtos.NodeState;
|
||||
|
@ -40,17 +39,17 @@ import java.util.UUID;
|
|||
|
||||
/**
|
||||
* A node manager supports a simple interface for managing a datanode.
|
||||
* <p/>
|
||||
* <p>
|
||||
* 1. A datanode registers with the NodeManager.
|
||||
* <p/>
|
||||
* <p>
|
||||
* 2. If the node is allowed to register, we add that to the nodes that we need
|
||||
* to keep track of.
|
||||
* <p/>
|
||||
* <p>
|
||||
* 3. A heartbeat is made by the node at a fixed frequency.
|
||||
* <p/>
|
||||
* <p>
|
||||
* 4. A node can be in any of these 4 states: {HEALTHY, STALE, DEAD,
|
||||
* DECOMMISSIONED}
|
||||
* <p/>
|
||||
* <p>
|
||||
* HEALTHY - It is a datanode that is regularly heartbeating us.
|
||||
*
|
||||
* STALE - A datanode for which we have missed few heart beats.
|
||||
|
@ -135,8 +134,8 @@ public interface NodeManager extends StorageContainerNodeProtocol,
|
|||
* Remaps datanode to containers mapping to the new set of containers.
|
||||
* @param datanodeDetails - DatanodeDetails
|
||||
* @param containerIds - Set of containerIDs
|
||||
* @throws SCMException - if datanode is not known. For new datanode use
|
||||
* addDatanodeInContainerMap call.
|
||||
* @throws NodeNotFoundException - if datanode is not known. For new datanode
|
||||
* use addDatanodeInContainerMap call.
|
||||
*/
|
||||
void setContainers(DatanodeDetails datanodeDetails,
|
||||
Set<ContainerID> containerIds) throws NodeNotFoundException;
|
||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.hadoop.hdds.protocol.proto
|
|||
import org.apache.hadoop.hdds.scm.container.ContainerID;
|
||||
import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
|
||||
import org.apache.hadoop.hdds.scm.pipeline.PipelineID;
|
||||
import org.apache.hadoop.hdds.scm.exceptions.SCMException;
|
||||
import org.apache.hadoop.hdds.scm.node.states.NodeAlreadyExistsException;
|
||||
import org.apache.hadoop.hdds.scm.node.states.NodeNotFoundException;
|
||||
import org.apache.hadoop.hdds.scm.server.StorageContainerManager;
|
||||
|
@ -281,7 +280,6 @@ public class SCMNodeManager
|
|||
*
|
||||
* @param datanodeDetails - DatanodeDetailsProto.
|
||||
* @return SCMheartbeat response.
|
||||
* @throws IOException
|
||||
*/
|
||||
@Override
|
||||
public List<SCMCommand> processHeartbeat(DatanodeDetails datanodeDetails) {
|
||||
|
@ -396,8 +394,8 @@ public class SCMNodeManager
|
|||
* Update set of containers available on a datanode.
|
||||
* @param datanodeDetails - DatanodeID
|
||||
* @param containerIds - Set of containerIDs
|
||||
* @throws SCMException - if datanode is not known. For new datanode use
|
||||
* addDatanodeInContainerMap call.
|
||||
* @throws NodeNotFoundException - if datanode is not known. For new datanode
|
||||
* use addDatanodeInContainerMap call.
|
||||
*/
|
||||
@Override
|
||||
public void setContainers(DatanodeDetails datanodeDetails,
|
||||
|
|
|
@ -69,7 +69,7 @@ public final class HddsTestUtils {
|
|||
* Creates list of ContainerInfo.
|
||||
*
|
||||
* @param numContainers number of ContainerInfo to be included in list.
|
||||
* @return List<ContainerInfo>
|
||||
* @return {@literal List<ContainerInfo>}
|
||||
*/
|
||||
public static List<ContainerInfo> getContainerInfo(int numContainers) {
|
||||
List<ContainerInfo> containerInfoList = new ArrayList<>();
|
||||
|
|
Loading…
Reference in New Issue