HDFS-4996. Merging change r1504679 from trunk to branch-2.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1504682 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
9b1bc20045
commit
bddf1659f8
|
@ -381,6 +381,7 @@ Administration Commands
|
|||
*-----------------+-----------------------------------------------------------+
|
||||
| -metasave filename | Save Namenode's primary data structures to <filename> in
|
||||
| the directory specified by hadoop.log.dir property.
|
||||
| <filename> is overwritten if it exists.
|
||||
| <filename> will contain one line for each of the following\
|
||||
| 1. Datanodes heart beating with Namenode\
|
||||
| 2. Blocks waiting to be replicated\
|
||||
|
|
|
@ -229,6 +229,9 @@ Release 2.1.0-beta - 2013-07-02
|
|||
HDFS-4992. Make balancer's mover thread count and dispatcher thread count
|
||||
configurable. (Max Lapan via szetszwo)
|
||||
|
||||
HDFS-4996. ClientProtocol#metaSave can be made idempotent by overwriting the
|
||||
output file instead of appending to it. (cnauroth)
|
||||
|
||||
OPTIMIZATIONS
|
||||
|
||||
HDFS-4465. Optimize datanode ReplicasMap and ReplicaInfo. (atm)
|
||||
|
|
|
@ -1182,7 +1182,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
|
|||
checkOperation(OperationCategory.UNCHECKED);
|
||||
File file = new File(System.getProperty("hadoop.log.dir"), filename);
|
||||
PrintWriter out = new PrintWriter(new BufferedWriter(
|
||||
new OutputStreamWriter(new FileOutputStream(file, true), Charsets.UTF_8)));
|
||||
new OutputStreamWriter(new FileOutputStream(file), Charsets.UTF_8)));
|
||||
metaSave(out);
|
||||
out.flush();
|
||||
out.close();
|
||||
|
|
|
@ -628,6 +628,7 @@ public class DFSAdmin extends FsShell {
|
|||
|
||||
String metaSave = "-metasave <filename>: \tSave Namenode's primary data structures\n" +
|
||||
"\t\tto <filename> in the directory specified by hadoop.log.dir property.\n" +
|
||||
"\t\t<filename> is overwritten if it exists.\n" +
|
||||
"\t\t<filename> will contain one line for each of the following\n" +
|
||||
"\t\t\t1. Datanodes heart beating with Namenode\n" +
|
||||
"\t\t\t2. Blocks waiting to be replicated\n" +
|
||||
|
|
|
@ -18,9 +18,11 @@
|
|||
package org.apache.hadoop.hdfs.server.namenode;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.DataInputStream;
|
||||
import java.io.File;
|
||||
import java.io.FileInputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStreamReader;
|
||||
|
@ -31,6 +33,7 @@ import org.apache.hadoop.fs.CommonConfigurationKeys;
|
|||
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
|
@ -47,6 +50,7 @@ public class TestMetaSave {
|
|||
static final int blockSize = 8192;
|
||||
private static MiniDFSCluster cluster = null;
|
||||
private static FileSystem fileSys = null;
|
||||
private static FSNamesystem namesystem = null;
|
||||
|
||||
private void createFile(FileSystem fileSys, Path name) throws IOException {
|
||||
FSDataOutputStream stm = fileSys.create(name, true, fileSys.getConf()
|
||||
|
@ -72,6 +76,7 @@ public class TestMetaSave {
|
|||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(NUM_DATA_NODES).build();
|
||||
cluster.waitActive();
|
||||
fileSys = cluster.getFileSystem();
|
||||
namesystem = cluster.getNamesystem();
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -79,9 +84,6 @@ public class TestMetaSave {
|
|||
*/
|
||||
@Test
|
||||
public void testMetaSave() throws IOException, InterruptedException {
|
||||
|
||||
final FSNamesystem namesystem = cluster.getNamesystem();
|
||||
|
||||
for (int i = 0; i < 2; i++) {
|
||||
Path file = new Path("/filestatus" + i);
|
||||
createFile(fileSys, file);
|
||||
|
@ -95,9 +97,8 @@ public class TestMetaSave {
|
|||
namesystem.metaSave("metasave.out.txt");
|
||||
|
||||
// Verification
|
||||
String logFile = System.getProperty("hadoop.log.dir") + "/"
|
||||
+ "metasave.out.txt";
|
||||
FileInputStream fstream = new FileInputStream(logFile);
|
||||
FileInputStream fstream = new FileInputStream(getLogFile(
|
||||
"metasave.out.txt"));
|
||||
DataInputStream in = new DataInputStream(fstream);
|
||||
BufferedReader reader = null;
|
||||
try {
|
||||
|
@ -124,9 +125,6 @@ public class TestMetaSave {
|
|||
@Test
|
||||
public void testMetasaveAfterDelete()
|
||||
throws IOException, InterruptedException {
|
||||
|
||||
final FSNamesystem namesystem = cluster.getNamesystem();
|
||||
|
||||
for (int i = 0; i < 2; i++) {
|
||||
Path file = new Path("/filestatus" + i);
|
||||
createFile(fileSys, file);
|
||||
|
@ -142,11 +140,10 @@ public class TestMetaSave {
|
|||
namesystem.metaSave("metasaveAfterDelete.out.txt");
|
||||
|
||||
// Verification
|
||||
String logFile = System.getProperty("hadoop.log.dir") + "/"
|
||||
+ "metasaveAfterDelete.out.txt";
|
||||
BufferedReader reader = null;
|
||||
try {
|
||||
FileInputStream fstream = new FileInputStream(logFile);
|
||||
FileInputStream fstream = new FileInputStream(getLogFile(
|
||||
"metasaveAfterDelete.out.txt"));
|
||||
DataInputStream in = new DataInputStream(fstream);
|
||||
reader = new BufferedReader(new InputStreamReader(in));
|
||||
reader.readLine();
|
||||
|
@ -166,6 +163,42 @@ public class TestMetaSave {
|
|||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Tests that metasave overwrites the output file (not append).
|
||||
*/
|
||||
@Test
|
||||
public void testMetaSaveOverwrite() throws Exception {
|
||||
// metaSave twice.
|
||||
namesystem.metaSave("metaSaveOverwrite.out.txt");
|
||||
namesystem.metaSave("metaSaveOverwrite.out.txt");
|
||||
|
||||
// Read output file.
|
||||
FileInputStream fis = null;
|
||||
InputStreamReader isr = null;
|
||||
BufferedReader rdr = null;
|
||||
try {
|
||||
fis = new FileInputStream(getLogFile("metaSaveOverwrite.out.txt"));
|
||||
isr = new InputStreamReader(fis);
|
||||
rdr = new BufferedReader(isr);
|
||||
|
||||
// Validate that file was overwritten (not appended) by checking for
|
||||
// presence of only one "Live Datanodes" line.
|
||||
boolean foundLiveDatanodesLine = false;
|
||||
String line = rdr.readLine();
|
||||
while (line != null) {
|
||||
if (line.startsWith("Live Datanodes")) {
|
||||
if (foundLiveDatanodesLine) {
|
||||
fail("multiple Live Datanodes lines, output file not overwritten");
|
||||
}
|
||||
foundLiveDatanodesLine = true;
|
||||
}
|
||||
line = rdr.readLine();
|
||||
}
|
||||
} finally {
|
||||
IOUtils.cleanup(null, rdr, isr, fis);
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDown() throws IOException {
|
||||
if (fileSys != null)
|
||||
|
@ -173,4 +206,14 @@ public class TestMetaSave {
|
|||
if (cluster != null)
|
||||
cluster.shutdown();
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns a File for the given name inside the log directory.
|
||||
*
|
||||
* @param name String file name
|
||||
* @return File for given name inside log directory
|
||||
*/
|
||||
private static File getLogFile(String name) {
|
||||
return new File(System.getProperty("hadoop.log.dir"), name);
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue