HDFS-10383. Safely close resources in DFSTestUtil. Contributed by Mingliang Liu.
(cherry picked from commit ccb1cade5b
)
This commit is contained in:
parent
a166286b9e
commit
1470af8c95
|
@ -18,6 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_FILE_BUFFER_SIZE_KEY;
|
||||
import static org.apache.hadoop.fs.CreateFlag.CREATE;
|
||||
import static org.apache.hadoop.fs.CreateFlag.LAZY_PERSIST;
|
||||
import static org.apache.hadoop.fs.CreateFlag.OVERWRITE;
|
||||
|
@ -336,17 +337,10 @@ public class DFSTestUtil {
|
|||
|
||||
public static byte[] readFileBuffer(FileSystem fs, Path fileName)
|
||||
throws IOException {
|
||||
ByteArrayOutputStream os = new ByteArrayOutputStream();
|
||||
try {
|
||||
FSDataInputStream in = fs.open(fileName);
|
||||
try {
|
||||
IOUtils.copyBytes(in, os, 1024, true);
|
||||
return os.toByteArray();
|
||||
} finally {
|
||||
in.close();
|
||||
}
|
||||
} finally {
|
||||
os.close();
|
||||
try (ByteArrayOutputStream os = new ByteArrayOutputStream();
|
||||
FSDataInputStream in = fs.open(fileName)) {
|
||||
IOUtils.copyBytes(in, os, 1024, true);
|
||||
return os.toByteArray();
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -356,9 +350,7 @@ public class DFSTestUtil {
|
|||
throw new IOException("Mkdirs failed to create " +
|
||||
fileName.getParent().toString());
|
||||
}
|
||||
FSDataOutputStream out = null;
|
||||
try {
|
||||
out = fs.create(fileName, replFactor);
|
||||
try (FSDataOutputStream out = fs.create(fileName, replFactor)) {
|
||||
byte[] toWrite = new byte[1024];
|
||||
Random rb = new Random(seed);
|
||||
long bytesToWrite = fileLen;
|
||||
|
@ -369,10 +361,6 @@ public class DFSTestUtil {
|
|||
out.write(toWrite, 0, bytesToWriteNext);
|
||||
bytesToWrite -= bytesToWriteNext;
|
||||
}
|
||||
out.close();
|
||||
out = null;
|
||||
} finally {
|
||||
IOUtils.closeStream(out);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -394,51 +382,40 @@ public class DFSTestUtil {
|
|||
boolean isLazyPersist, int bufferLen, long fileLen, long blockSize,
|
||||
short replFactor, long seed, boolean flush,
|
||||
InetSocketAddress[] favoredNodes) throws IOException {
|
||||
assert bufferLen > 0;
|
||||
if (!fs.mkdirs(fileName.getParent())) {
|
||||
assert bufferLen > 0;
|
||||
if (!fs.mkdirs(fileName.getParent())) {
|
||||
throw new IOException("Mkdirs failed to create " +
|
||||
fileName.getParent().toString());
|
||||
}
|
||||
FSDataOutputStream out = null;
|
||||
EnumSet<CreateFlag> createFlags = EnumSet.of(CREATE);
|
||||
createFlags.add(OVERWRITE);
|
||||
if (isLazyPersist) {
|
||||
createFlags.add(LAZY_PERSIST);
|
||||
}
|
||||
try {
|
||||
if (favoredNodes == null) {
|
||||
out = fs.create(
|
||||
fileName,
|
||||
FsPermission.getFileDefault(),
|
||||
createFlags,
|
||||
fs.getConf().getInt(
|
||||
CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096),
|
||||
replFactor, blockSize, null);
|
||||
} else {
|
||||
out = ((DistributedFileSystem) fs).create(fileName,
|
||||
FsPermission.getDefault(), true, bufferLen, replFactor, blockSize,
|
||||
null, favoredNodes);
|
||||
fileName.getParent().toString());
|
||||
}
|
||||
EnumSet<CreateFlag> createFlags = EnumSet.of(CREATE);
|
||||
createFlags.add(OVERWRITE);
|
||||
if (isLazyPersist) {
|
||||
createFlags.add(LAZY_PERSIST);
|
||||
}
|
||||
try (FSDataOutputStream out = (favoredNodes == null) ?
|
||||
fs.create(fileName, FsPermission.getFileDefault(), createFlags,
|
||||
fs.getConf().getInt(IO_FILE_BUFFER_SIZE_KEY, 4096), replFactor,
|
||||
blockSize, null)
|
||||
:
|
||||
((DistributedFileSystem) fs).create(fileName, FsPermission.getDefault(),
|
||||
true, bufferLen, replFactor, blockSize, null, favoredNodes)
|
||||
) {
|
||||
if (fileLen > 0) {
|
||||
byte[] toWrite = new byte[bufferLen];
|
||||
Random rb = new Random(seed);
|
||||
long bytesToWrite = fileLen;
|
||||
while (bytesToWrite>0) {
|
||||
while (bytesToWrite > 0) {
|
||||
rb.nextBytes(toWrite);
|
||||
int bytesToWriteNext = (bufferLen < bytesToWrite) ? bufferLen
|
||||
: (int) bytesToWrite;
|
||||
: (int) bytesToWrite;
|
||||
|
||||
out.write(toWrite, 0, bytesToWriteNext);
|
||||
bytesToWrite -= bytesToWriteNext;
|
||||
out.write(toWrite, 0, bytesToWriteNext);
|
||||
bytesToWrite -= bytesToWriteNext;
|
||||
}
|
||||
if (flush) {
|
||||
out.hsync();
|
||||
}
|
||||
}
|
||||
} finally {
|
||||
if (out != null) {
|
||||
out.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -455,20 +432,18 @@ public class DFSTestUtil {
|
|||
|
||||
for (int idx = 0; idx < nFiles; idx++) {
|
||||
Path fPath = new Path(root, files[idx].getName());
|
||||
FSDataInputStream in = fs.open(fPath);
|
||||
byte[] toRead = new byte[files[idx].getSize()];
|
||||
byte[] toCompare = new byte[files[idx].getSize()];
|
||||
Random rb = new Random(files[idx].getSeed());
|
||||
rb.nextBytes(toCompare);
|
||||
in.readFully(0, toRead);
|
||||
in.close();
|
||||
for (int i = 0; i < toRead.length; i++) {
|
||||
if (toRead[i] != toCompare[i]) {
|
||||
return false;
|
||||
try (FSDataInputStream in = fs.open(fPath)) {
|
||||
byte[] toRead = new byte[files[idx].getSize()];
|
||||
byte[] toCompare = new byte[files[idx].getSize()];
|
||||
Random rb = new Random(files[idx].getSeed());
|
||||
rb.nextBytes(toCompare);
|
||||
in.readFully(0, toRead);
|
||||
for (int i = 0; i < toRead.length; i++) {
|
||||
if (toRead[i] != toCompare[i]) {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
}
|
||||
toRead = null;
|
||||
toCompare = null;
|
||||
}
|
||||
|
||||
return true;
|
||||
|
@ -502,16 +477,13 @@ public class DFSTestUtil {
|
|||
*/
|
||||
public static boolean allBlockReplicasCorrupt(MiniDFSCluster cluster,
|
||||
Path file, int blockNo) throws IOException {
|
||||
DFSClient client = new DFSClient(new InetSocketAddress("localhost",
|
||||
cluster.getNameNodePort()), cluster.getConfiguration(0));
|
||||
LocatedBlocks blocks;
|
||||
try {
|
||||
blocks = client.getNamenode().getBlockLocations(
|
||||
try (DFSClient client = new DFSClient(new InetSocketAddress("localhost",
|
||||
cluster.getNameNodePort()), cluster.getConfiguration(0))) {
|
||||
LocatedBlocks blocks;
|
||||
blocks = client.getNamenode().getBlockLocations(
|
||||
file.toString(), 0, Long.MAX_VALUE);
|
||||
} finally {
|
||||
client.close();
|
||||
return blocks.get(blockNo).isCorrupt();
|
||||
}
|
||||
return blocks.get(blockNo).isCorrupt();
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -791,12 +763,9 @@ public class DFSTestUtil {
|
|||
}
|
||||
|
||||
public static ExtendedBlock getFirstBlock(FileSystem fs, Path path) throws IOException {
|
||||
HdfsDataInputStream in = (HdfsDataInputStream) fs.open(path);
|
||||
try {
|
||||
try (HdfsDataInputStream in = (HdfsDataInputStream) fs.open(path)) {
|
||||
in.readByte();
|
||||
return in.getCurrentBlock();
|
||||
} finally {
|
||||
in.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -807,8 +776,9 @@ public class DFSTestUtil {
|
|||
|
||||
public static List<LocatedBlock> getAllBlocks(FileSystem fs, Path path)
|
||||
throws IOException {
|
||||
HdfsDataInputStream in = (HdfsDataInputStream) fs.open(path);
|
||||
return in.getAllBlocks();
|
||||
try (HdfsDataInputStream in = (HdfsDataInputStream) fs.open(path)) {
|
||||
return in.getAllBlocks();
|
||||
}
|
||||
}
|
||||
|
||||
public static Token<BlockTokenIdentifier> getBlockToken(
|
||||
|
@ -817,11 +787,14 @@ public class DFSTestUtil {
|
|||
}
|
||||
|
||||
public static String readFile(File f) throws IOException {
|
||||
StringBuilder b = new StringBuilder();
|
||||
BufferedReader in = new BufferedReader(new FileReader(f));
|
||||
for(int c; (c = in.read()) != -1; b.append((char)c));
|
||||
in.close();
|
||||
return b.toString();
|
||||
try (BufferedReader in = new BufferedReader(new FileReader(f))) {
|
||||
StringBuilder b = new StringBuilder();
|
||||
int c;
|
||||
while ((c = in.read()) != -1) {
|
||||
b.append((char) c);
|
||||
}
|
||||
return b.toString();
|
||||
}
|
||||
}
|
||||
|
||||
/* Write the given string to the given file */
|
||||
|
@ -830,18 +803,20 @@ public class DFSTestUtil {
|
|||
if (fs.exists(p)) {
|
||||
fs.delete(p, true);
|
||||
}
|
||||
InputStream is = new ByteArrayInputStream(s.getBytes());
|
||||
FSDataOutputStream os = fs.create(p);
|
||||
IOUtils.copyBytes(is, os, s.length(), true);
|
||||
try (InputStream is = new ByteArrayInputStream(s.getBytes());
|
||||
FSDataOutputStream os = fs.create(p)) {
|
||||
IOUtils.copyBytes(is, os, s.length());
|
||||
}
|
||||
}
|
||||
|
||||
/* Append the given string to the given file */
|
||||
public static void appendFile(FileSystem fs, Path p, String s)
|
||||
throws IOException {
|
||||
assert fs.exists(p);
|
||||
InputStream is = new ByteArrayInputStream(s.getBytes());
|
||||
FSDataOutputStream os = fs.append(p);
|
||||
IOUtils.copyBytes(is, os, s.length(), true);
|
||||
try (InputStream is = new ByteArrayInputStream(s.getBytes());
|
||||
FSDataOutputStream os = fs.append(p)) {
|
||||
IOUtils.copyBytes(is, os, s.length());
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -858,9 +833,9 @@ public class DFSTestUtil {
|
|||
byte[] toAppend = new byte[length];
|
||||
Random random = new Random();
|
||||
random.nextBytes(toAppend);
|
||||
FSDataOutputStream out = fs.append(p);
|
||||
out.write(toAppend);
|
||||
out.close();
|
||||
try (FSDataOutputStream out = fs.append(p)) {
|
||||
out.write(toAppend);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -978,35 +953,32 @@ public class DFSTestUtil {
|
|||
*/
|
||||
public static byte[] loadFile(String filename) throws IOException {
|
||||
File file = new File(filename);
|
||||
DataInputStream in = new DataInputStream(new FileInputStream(file));
|
||||
byte[] content = new byte[(int)file.length()];
|
||||
try {
|
||||
try (DataInputStream in = new DataInputStream(new FileInputStream(file))) {
|
||||
byte[] content = new byte[(int) file.length()];
|
||||
in.readFully(content);
|
||||
} finally {
|
||||
IOUtils.cleanup(LOG, in);
|
||||
return content;
|
||||
}
|
||||
return content;
|
||||
}
|
||||
|
||||
/** For {@link TestTransferRbw} */
|
||||
public static BlockOpResponseProto transferRbw(final ExtendedBlock b,
|
||||
final DFSClient dfsClient, final DatanodeInfo... datanodes) throws IOException {
|
||||
assertEquals(2, datanodes.length);
|
||||
final Socket s = DataStreamer.createSocketForPipeline(datanodes[0],
|
||||
datanodes.length, dfsClient);
|
||||
final long writeTimeout = dfsClient.getDatanodeWriteTimeout(datanodes.length);
|
||||
final DataOutputStream out = new DataOutputStream(new BufferedOutputStream(
|
||||
NetUtils.getOutputStream(s, writeTimeout),
|
||||
DFSUtilClient.getSmallBufferSize(dfsClient.getConfiguration())));
|
||||
final DataInputStream in = new DataInputStream(NetUtils.getInputStream(s));
|
||||
try (Socket s = DataStreamer.createSocketForPipeline(datanodes[0],
|
||||
datanodes.length, dfsClient);
|
||||
DataOutputStream out = new DataOutputStream(new BufferedOutputStream(
|
||||
NetUtils.getOutputStream(s, writeTimeout),
|
||||
DFSUtilClient.getSmallBufferSize(dfsClient.getConfiguration())));
|
||||
DataInputStream in = new DataInputStream(NetUtils.getInputStream(s))) {
|
||||
// send the request
|
||||
new Sender(out).transferBlock(b, new Token<BlockTokenIdentifier>(),
|
||||
dfsClient.clientName, new DatanodeInfo[]{datanodes[1]},
|
||||
new StorageType[]{StorageType.DEFAULT});
|
||||
out.flush();
|
||||
|
||||
// send the request
|
||||
new Sender(out).transferBlock(b, new Token<BlockTokenIdentifier>(),
|
||||
dfsClient.clientName, new DatanodeInfo[]{datanodes[1]},
|
||||
new StorageType[]{StorageType.DEFAULT});
|
||||
out.flush();
|
||||
|
||||
return BlockOpResponseProto.parseDelimitedFrom(in);
|
||||
return BlockOpResponseProto.parseDelimitedFrom(in);
|
||||
}
|
||||
}
|
||||
|
||||
public static void setFederatedConfiguration(MiniDFSCluster cluster,
|
||||
|
@ -1554,13 +1526,12 @@ public class DFSTestUtil {
|
|||
*/
|
||||
public static void verifyFilesEqual(FileSystem fs, Path p1, Path p2, int len)
|
||||
throws IOException {
|
||||
final FSDataInputStream in1 = fs.open(p1);
|
||||
final FSDataInputStream in2 = fs.open(p2);
|
||||
for (int i = 0; i < len; i++) {
|
||||
assertEquals("Mismatch at byte " + i, in1.read(), in2.read());
|
||||
try (FSDataInputStream in1 = fs.open(p1);
|
||||
FSDataInputStream in2 = fs.open(p2)) {
|
||||
for (int i = 0; i < len; i++) {
|
||||
assertEquals("Mismatch at byte " + i, in1.read(), in2.read());
|
||||
}
|
||||
}
|
||||
in1.close();
|
||||
in2.close();
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -1573,20 +1544,15 @@ public class DFSTestUtil {
|
|||
* @throws IOException
|
||||
*/
|
||||
public static void verifyFilesNotEqual(FileSystem fs, Path p1, Path p2,
|
||||
int len)
|
||||
throws IOException {
|
||||
final FSDataInputStream in1 = fs.open(p1);
|
||||
final FSDataInputStream in2 = fs.open(p2);
|
||||
try {
|
||||
int len) throws IOException {
|
||||
try (FSDataInputStream in1 = fs.open(p1);
|
||||
FSDataInputStream in2 = fs.open(p2)) {
|
||||
for (int i = 0; i < len; i++) {
|
||||
if (in1.read() != in2.read()) {
|
||||
return;
|
||||
}
|
||||
}
|
||||
fail("files are equal, but should not be");
|
||||
} finally {
|
||||
in1.close();
|
||||
in2.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1697,13 +1663,13 @@ public class DFSTestUtil {
|
|||
int ret = 0;
|
||||
try {
|
||||
ByteArrayOutputStream bs = new ByteArrayOutputStream(1024);
|
||||
PrintStream out = new PrintStream(bs);
|
||||
System.setOut(out);
|
||||
System.setErr(out);
|
||||
ret = tool.run(cmds);
|
||||
System.out.flush();
|
||||
System.err.flush();
|
||||
out.close();
|
||||
try (PrintStream out = new PrintStream(bs)) {
|
||||
System.setOut(out);
|
||||
System.setErr(out);
|
||||
ret = tool.run(cmds);
|
||||
System.out.flush();
|
||||
System.err.flush();
|
||||
}
|
||||
output = bs.toString();
|
||||
} finally {
|
||||
System.setOut(origOut);
|
||||
|
@ -1805,9 +1771,9 @@ public class DFSTestUtil {
|
|||
ExtendedBlock blk, int dnIndex, int lenDelta) throws IOException {
|
||||
File blockFile = cluster.getBlockFile(dnIndex, blk);
|
||||
if (blockFile != null && blockFile.exists()) {
|
||||
RandomAccessFile raFile = new RandomAccessFile(blockFile, "rw");
|
||||
raFile.setLength(raFile.length()+lenDelta);
|
||||
raFile.close();
|
||||
try (RandomAccessFile raFile = new RandomAccessFile(blockFile, "rw")) {
|
||||
raFile.setLength(raFile.length() + lenDelta);
|
||||
}
|
||||
return true;
|
||||
}
|
||||
LOG.info("failed to change length of block " + blk);
|
||||
|
|
Loading…
Reference in New Issue