HDFS-11358. DiskBalancer: Report command supports reading nodes from host file. Contributed by Yiqun Lin.

This commit is contained in:
Yiqun Lin 2017-03-21 17:44:17 +08:00
parent 1a98a279f4
commit 3b908f71c5
6 changed files with 95 additions and 18 deletions

View File

@ -41,6 +41,7 @@ public class DiskBalancerException extends IOException {
UNKNOWN_KEY, UNKNOWN_KEY,
INVALID_NODE, INVALID_NODE,
DATANODE_STATUS_NOT_REGULAR, DATANODE_STATUS_NOT_REGULAR,
INVALID_HOST_FILE_PATH,
} }
private final Result result; private final Result result;

View File

@ -49,17 +49,17 @@ import org.apache.hadoop.hdfs.server.diskbalancer.datamodel.DiskBalancerVolumeSe
import org.apache.hadoop.hdfs.tools.DiskBalancerCLI; import org.apache.hadoop.hdfs.tools.DiskBalancerCLI;
import org.apache.hadoop.net.NetUtils; import org.apache.hadoop.net.NetUtils;
import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.HostsFileReader;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import java.io.Closeable; import java.io.Closeable;
import java.io.FileNotFoundException;
import java.io.IOException; import java.io.IOException;
import java.io.PrintStream; import java.io.PrintStream;
import java.net.InetSocketAddress; import java.net.InetSocketAddress;
import java.net.URI; import java.net.URI;
import java.net.URL; import java.net.URL;
import java.nio.charset.Charset;
import java.nio.file.Files;
import java.nio.file.Paths; import java.nio.file.Paths;
import java.text.SimpleDateFormat; import java.text.SimpleDateFormat;
import java.util.Collections; import java.util.Collections;
@ -268,16 +268,33 @@ public abstract class Command extends Configured implements Closeable {
if ((listArg == null) || listArg.isEmpty()) { if ((listArg == null) || listArg.isEmpty()) {
return resultSet; return resultSet;
} }
if (listArg.startsWith("file://")) { if (listArg.startsWith("file://")) {
listURL = new URL(listArg); listURL = new URL(listArg);
byte[] data = Files.readAllBytes(Paths.get(listURL.getPath())); try {
nodeData = new String(data, Charset.forName("UTF-8")); HostsFileReader.readFileToSet("include",
Paths.get(listURL.getPath()).toString(), resultSet);
} catch (FileNotFoundException e) {
String warnMsg = String
.format("The input host file path '%s' is not a valid path. "
+ "Please make sure the host file exists.", listArg);
throw new DiskBalancerException(warnMsg,
DiskBalancerException.Result.INVALID_HOST_FILE_PATH);
}
} else { } else {
nodeData = listArg; nodeData = listArg;
String[] nodes = nodeData.split(",");
if (nodes.length == 0) {
String warnMsg = "The number of input nodes is 0. "
+ "Please input the valid nodes.";
throw new DiskBalancerException(warnMsg,
DiskBalancerException.Result.INVALID_NODE);
} }
String[] nodes = nodeData.split(",");
Collections.addAll(resultSet, nodes); Collections.addAll(resultSet, nodes);
}
return resultSet; return resultSet;
} }

View File

@ -148,7 +148,9 @@ public class ReportCommand extends Command {
* Reporting volume information for specific DataNode(s) * Reporting volume information for specific DataNode(s)
*/ */
outputLine = String.format( outputLine = String.format(
"Reporting volume information for DataNode(s) '%s'.", nodeVal); "Reporting volume information for DataNode(s). "
+ "These DataNode(s) are parsed from '%s'.", nodeVal);
recordOutput(result, outputLine); recordOutput(result, outputLine);
List<DiskBalancerDataNode> dbdns = Lists.newArrayList(); List<DiskBalancerDataNode> dbdns = Lists.newArrayList();
@ -224,7 +226,7 @@ public class ReportCommand extends Command {
+ "hdfs diskbalancer -report\n" + "hdfs diskbalancer -report\n"
+ "hdfs diskbalancer -report -top 5\n" + "hdfs diskbalancer -report -top 5\n"
+ "hdfs diskbalancer -report " + "hdfs diskbalancer -report "
+ "-node [<DataNodeID|IP|Hostname>,...]"; + "-node <file://> | [<DataNodeID|IP|Hostname>,...]";
HelpFormatter helpFormatter = new HelpFormatter(); HelpFormatter helpFormatter = new HelpFormatter();
helpFormatter.printHelp("hdfs diskbalancer -fs http://namenode.uri " + helpFormatter.printHelp("hdfs diskbalancer -fs http://namenode.uri " +

View File

@ -418,7 +418,7 @@ Usage:
[-query <datanode>] [-query <datanode>]
[-cancel <planfile>] [-cancel <planfile>]
[-cancel <planID> -node <datanode>] [-cancel <planID> -node <datanode>]
[-report -node [<DataNodeID|IP|Hostname>,...]] [-report -node <file://> | [<DataNodeID|IP|Hostname>,...]]
[-report -node -top <topnum>] [-report -node -top <topnum>]
| COMMAND\_OPTION | Description | | COMMAND\_OPTION | Description |

View File

@ -102,9 +102,9 @@ or
Plan ID can be read from datanode using query command. Plan ID can be read from datanode using query command.
### Report ### Report
Report command provides detailed report of specified node(s) or top nodes that will benefit from running disk balancer. Report command provides detailed report of specified node(s) or top nodes that will benefit from running disk balancer. The node(s) can be specified by a host file or comma-separated list of nodes.
`hdfs diskbalancer -fs http://namenode.uri -report -node [<DataNodeID|IP|Hostname>,...]` `hdfs diskbalancer -fs http://namenode.uri -report -node <file://> | [<DataNodeID|IP|Hostname>,...]`
or or

View File

@ -34,6 +34,8 @@ import static org.junit.Assert.assertThat;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import java.io.ByteArrayOutputStream; import java.io.ByteArrayOutputStream;
import java.io.File;
import java.io.FileWriter;
import java.io.PrintStream; import java.io.PrintStream;
import java.net.URI; import java.net.URI;
import java.util.List; import java.util.List;
@ -675,14 +677,18 @@ public class TestDiskBalancerCommand {
REPORT, NODE, dataNodeUuid1, dataNodeUuid2); REPORT, NODE, dataNodeUuid1, dataNodeUuid2);
final String cmdLine = String.format("hdfs diskbalancer %s", planArg); final String cmdLine = String.format("hdfs diskbalancer %s", planArg);
List<String> outputs = runCommand(cmdLine, cluster); List<String> outputs = runCommand(cmdLine, cluster);
verifyOutputsOfReportCommand(outputs, dataNodeUuid1, dataNodeUuid2, true);
}
assertThat( private void verifyOutputsOfReportCommand(List<String> outputs,
outputs.get(0), String dataNodeUuid1, String dataNodeUuid2, boolean inputNodesStr) {
containsString("Processing report command")); assertThat(outputs.get(0), containsString("Processing report command"));
assertThat( if (inputNodesStr) {
outputs.get(1), assertThat(outputs.get(1),
is(allOf(containsString("Reporting volume information for DataNode"), is(allOf(containsString("Reporting volume information for DataNode"),
containsString(dataNodeUuid1), containsString(dataNodeUuid2)))); containsString(dataNodeUuid1), containsString(dataNodeUuid2))));
}
// Since the order of input nodes will be disrupted when parse // Since the order of input nodes will be disrupted when parse
// the node string, we should compare UUID with both output lines. // the node string, we should compare UUID with both output lines.
assertTrue(outputs.get(2).contains(dataNodeUuid1) assertTrue(outputs.get(2).contains(dataNodeUuid1)
@ -714,4 +720,55 @@ public class TestDiskBalancerCommand {
, invalidNode, invalidNode); , invalidNode, invalidNode);
assertTrue(outputs.get(2).contains(invalidNodeInfo)); assertTrue(outputs.get(2).contains(invalidNodeInfo));
} }
@Test(timeout = 60000)
public void testReportCommandWithNullNodes() throws Exception {
// don't input nodes
final String planArg = String.format("-%s -%s ,", REPORT, NODE);
final String cmdLine = String.format("hdfs diskbalancer %s", planArg);
List<String> outputs = runCommand(cmdLine, cluster);
String invalidNodeInfo = "The number of input nodes is 0. "
+ "Please input the valid nodes.";
assertTrue(outputs.get(2).contains(invalidNodeInfo));
}
@Test(timeout = 60000)
public void testReportCommandWithReadingHostFile() throws Exception {
final String testDir = GenericTestUtils.getTestDir().getAbsolutePath();
File includeFile = new File(testDir, "diskbalancer.include");
String filePath = testDir + "/diskbalancer.include";
String dataNodeUuid1 = cluster.getDataNodes().get(0).getDatanodeUuid();
String dataNodeUuid2 = cluster.getDataNodes().get(1).getDatanodeUuid();
FileWriter fw = new FileWriter(filePath);
fw.write("#This-is-comment\n");
fw.write(dataNodeUuid1 + "\n");
fw.write(dataNodeUuid2 + "\n");
fw.close();
final String planArg = String.format("-%s -%s file://%s",
REPORT, NODE, filePath);
final String cmdLine = String.format("hdfs diskbalancer %s", planArg);
List<String> outputs = runCommand(cmdLine, cluster);
verifyOutputsOfReportCommand(outputs, dataNodeUuid1, dataNodeUuid2, false);
includeFile.delete();
}
@Test(timeout = 60000)
public void testReportCommandWithInvalidHostFilePath() throws Exception {
final String testDir = GenericTestUtils.getTestDir().getAbsolutePath();
String invalidFilePath = testDir + "/diskbalancer-invalid.include";
final String planArg = String.format("-%s -%s file://%s",
REPORT, NODE, invalidFilePath);
final String cmdLine = String.format("hdfs diskbalancer %s", planArg);
List<String> outputs = runCommand(cmdLine, cluster);
String invalidNodeInfo = String.format(
"The input host file path 'file://%s' is not a valid path.", invalidFilePath);
assertTrue(outputs.get(2).contains(invalidNodeInfo));
}
} }