HDFS-6294. Use INode IDs to avoid conflicts when a file open for write is renamed (cmccabe)

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1593638 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Colin McCabe 2014-05-09 22:55:17 +00:00
parent 29a124d09d
commit fea7cd3c12
25 changed files with 326 additions and 157 deletions

View File

@ -89,6 +89,9 @@ Release 2.5.0 - UNRELEASED
HDFS-6295. Add "decommissioning" state and node state filtering to HDFS-6295. Add "decommissioning" state and node state filtering to
dfsadmin. (wang) dfsadmin. (wang)
HDFS-6294. Use INode IDs to avoid conflicts when a file open for write is
renamed. (cmccabe)
OPTIMIZATIONS OPTIMIZATIONS
HDFS-6214. Webhdfs has poor throughput for files >2GB (daryn) HDFS-6214. Webhdfs has poor throughput for files >2GB (daryn)

View File

@ -504,8 +504,8 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
* that are currently being written by this client. * that are currently being written by this client.
* Note that a file can only be written by a single client. * Note that a file can only be written by a single client.
*/ */
private final Map<String, DFSOutputStream> filesBeingWritten private final Map<Long, DFSOutputStream> filesBeingWritten
= new HashMap<String, DFSOutputStream>(); = new HashMap<Long, DFSOutputStream>();
/** /**
* Same as this(NameNode.getAddress(conf), conf); * Same as this(NameNode.getAddress(conf), conf);
@ -734,14 +734,14 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
} }
/** Get a lease and start automatic renewal */ /** Get a lease and start automatic renewal */
private void beginFileLease(final String src, final DFSOutputStream out) private void beginFileLease(final long inodeId, final DFSOutputStream out)
throws IOException { throws IOException {
getLeaseRenewer().put(src, out, this); getLeaseRenewer().put(inodeId, out, this);
} }
/** Stop renewal of lease for the file. */ /** Stop renewal of lease for the file. */
void endFileLease(final String src) throws IOException { void endFileLease(final long inodeId) throws IOException {
getLeaseRenewer().closeFile(src, this); getLeaseRenewer().closeFile(inodeId, this);
} }
@ -749,9 +749,9 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
* enforced to consistently update its local dfsclients array and * enforced to consistently update its local dfsclients array and
* client's filesBeingWritten map. * client's filesBeingWritten map.
*/ */
void putFileBeingWritten(final String src, final DFSOutputStream out) { void putFileBeingWritten(final long inodeId, final DFSOutputStream out) {
synchronized(filesBeingWritten) { synchronized(filesBeingWritten) {
filesBeingWritten.put(src, out); filesBeingWritten.put(inodeId, out);
// update the last lease renewal time only when there was no // update the last lease renewal time only when there was no
// writes. once there is one write stream open, the lease renewer // writes. once there is one write stream open, the lease renewer
// thread keeps it updated well with in anyone's expiration time. // thread keeps it updated well with in anyone's expiration time.
@ -762,9 +762,9 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
} }
/** Remove a file. Only called from LeaseRenewer. */ /** Remove a file. Only called from LeaseRenewer. */
void removeFileBeingWritten(final String src) { void removeFileBeingWritten(final long inodeId) {
synchronized(filesBeingWritten) { synchronized(filesBeingWritten) {
filesBeingWritten.remove(src); filesBeingWritten.remove(inodeId);
if (filesBeingWritten.isEmpty()) { if (filesBeingWritten.isEmpty()) {
lastLeaseRenewal = 0; lastLeaseRenewal = 0;
} }
@ -849,14 +849,14 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
/** Close/abort all files being written. */ /** Close/abort all files being written. */
private void closeAllFilesBeingWritten(final boolean abort) { private void closeAllFilesBeingWritten(final boolean abort) {
for(;;) { for(;;) {
final String src; final long inodeId;
final DFSOutputStream out; final DFSOutputStream out;
synchronized(filesBeingWritten) { synchronized(filesBeingWritten) {
if (filesBeingWritten.isEmpty()) { if (filesBeingWritten.isEmpty()) {
return; return;
} }
src = filesBeingWritten.keySet().iterator().next(); inodeId = filesBeingWritten.keySet().iterator().next();
out = filesBeingWritten.remove(src); out = filesBeingWritten.remove(inodeId);
} }
if (out != null) { if (out != null) {
try { try {
@ -866,8 +866,8 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
out.close(); out.close();
} }
} catch(IOException ie) { } catch(IOException ie) {
LOG.error("Failed to " + (abort? "abort": "close") + " file " + src, LOG.error("Failed to " + (abort? "abort": "close") +
ie); " inode " + inodeId, ie);
} }
} }
} }
@ -1481,7 +1481,7 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
final DFSOutputStream result = DFSOutputStream.newStreamForCreate(this, final DFSOutputStream result = DFSOutputStream.newStreamForCreate(this,
src, masked, flag, createParent, replication, blockSize, progress, src, masked, flag, createParent, replication, blockSize, progress,
buffersize, dfsClientConf.createChecksum(checksumOpt), favoredNodeStrs); buffersize, dfsClientConf.createChecksum(checksumOpt), favoredNodeStrs);
beginFileLease(src, result); beginFileLease(result.getFileId(), result);
return result; return result;
} }
@ -1529,7 +1529,7 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
flag, createParent, replication, blockSize, progress, buffersize, flag, createParent, replication, blockSize, progress, buffersize,
checksum); checksum);
} }
beginFileLease(src, result); beginFileLease(result.getFileId(), result);
return result; return result;
} }
@ -1617,7 +1617,7 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
+ src + " on client " + clientName); + src + " on client " + clientName);
} }
final DFSOutputStream result = callAppend(stat, src, buffersize, progress); final DFSOutputStream result = callAppend(stat, src, buffersize, progress);
beginFileLease(src, result); beginFileLease(result.getFileId(), result);
return result; return result;
} }
@ -2454,8 +2454,8 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory {
} }
@VisibleForTesting @VisibleForTesting
ExtendedBlock getPreviousBlock(String file) { ExtendedBlock getPreviousBlock(long fileId) {
return filesBeingWritten.get(file).getBlock(); return filesBeingWritten.get(fileId).getBlock();
} }
/** /**

View File

@ -1017,7 +1017,7 @@ public class DFSOutputStream extends FSOutputSummer
//get a new datanode //get a new datanode
final DatanodeInfo[] original = nodes; final DatanodeInfo[] original = nodes;
final LocatedBlock lb = dfsClient.namenode.getAdditionalDatanode( final LocatedBlock lb = dfsClient.namenode.getAdditionalDatanode(
src, block, nodes, storageIDs, src, fileId, block, nodes, storageIDs,
failed.toArray(new DatanodeInfo[failed.size()]), failed.toArray(new DatanodeInfo[failed.size()]),
1, dfsClient.clientName); 1, dfsClient.clientName);
setPipeline(lb); setPipeline(lb);
@ -1273,7 +1273,8 @@ public class DFSOutputStream extends FSOutputSummer
if (!success) { if (!success) {
DFSClient.LOG.info("Abandoning " + block); DFSClient.LOG.info("Abandoning " + block);
dfsClient.namenode.abandonBlock(block, src, dfsClient.clientName); dfsClient.namenode.abandonBlock(block, fileId, src,
dfsClient.clientName);
block = null; block = null;
DFSClient.LOG.info("Excluding datanode " + nodes[errorIndex]); DFSClient.LOG.info("Excluding datanode " + nodes[errorIndex]);
excludedNodes.put(nodes[errorIndex], nodes[errorIndex]); excludedNodes.put(nodes[errorIndex], nodes[errorIndex]);
@ -1923,7 +1924,8 @@ public class DFSOutputStream extends FSOutputSummer
// namenode. // namenode.
if (persistBlocks.getAndSet(false) || updateLength) { if (persistBlocks.getAndSet(false) || updateLength) {
try { try {
dfsClient.namenode.fsync(src, dfsClient.clientName, lastBlockLength); dfsClient.namenode.fsync(src, fileId,
dfsClient.clientName, lastBlockLength);
} catch (IOException ioe) { } catch (IOException ioe) {
DFSClient.LOG.warn("Unable to persist blocks in hflush for " + src, ioe); DFSClient.LOG.warn("Unable to persist blocks in hflush for " + src, ioe);
// If we got an error here, it might be because some other thread called // If we got an error here, it might be because some other thread called
@ -2044,7 +2046,7 @@ public class DFSOutputStream extends FSOutputSummer
streamer.setLastException(new IOException("Lease timeout of " streamer.setLastException(new IOException("Lease timeout of "
+ (dfsClient.getHdfsTimeout()/1000) + " seconds expired.")); + (dfsClient.getHdfsTimeout()/1000) + " seconds expired."));
closeThreads(true); closeThreads(true);
dfsClient.endFileLease(src); dfsClient.endFileLease(fileId);
} }
// shutdown datastreamer and responseprocessor threads. // shutdown datastreamer and responseprocessor threads.
@ -2098,7 +2100,7 @@ public class DFSOutputStream extends FSOutputSummer
ExtendedBlock lastBlock = streamer.getBlock(); ExtendedBlock lastBlock = streamer.getBlock();
closeThreads(false); closeThreads(false);
completeFile(lastBlock); completeFile(lastBlock);
dfsClient.endFileLease(src); dfsClient.endFileLease(fileId);
} catch (ClosedChannelException e) { } catch (ClosedChannelException e) {
} finally { } finally {
closed = true; closed = true;
@ -2191,7 +2193,7 @@ public class DFSOutputStream extends FSOutputSummer
} }
@VisibleForTesting @VisibleForTesting
long getFileId() { public long getFileId() {
return fileId; return fileId;
} }
} }

View File

@ -281,7 +281,7 @@ class LeaseRenewer {
&& Time.now() - emptyTime > gracePeriod; && Time.now() - emptyTime > gracePeriod;
} }
synchronized void put(final String src, final DFSOutputStream out, synchronized void put(final long inodeId, final DFSOutputStream out,
final DFSClient dfsc) { final DFSClient dfsc) {
if (dfsc.isClientRunning()) { if (dfsc.isClientRunning()) {
if (!isRunning() || isRenewerExpired()) { if (!isRunning() || isRenewerExpired()) {
@ -319,7 +319,7 @@ class LeaseRenewer {
}); });
daemon.start(); daemon.start();
} }
dfsc.putFileBeingWritten(src, out); dfsc.putFileBeingWritten(inodeId, out);
emptyTime = Long.MAX_VALUE; emptyTime = Long.MAX_VALUE;
} }
} }
@ -330,8 +330,8 @@ class LeaseRenewer {
} }
/** Close a file. */ /** Close a file. */
void closeFile(final String src, final DFSClient dfsc) { void closeFile(final long inodeId, final DFSClient dfsc) {
dfsc.removeFileBeingWritten(src); dfsc.removeFileBeingWritten(inodeId);
synchronized(this) { synchronized(this) {
if (dfsc.isFilesBeingWrittenEmpty()) { if (dfsc.isFilesBeingWrittenEmpty()) {

View File

@ -290,13 +290,20 @@ public interface ClientProtocol {
* file. * file.
* Any partial writes to the block will be discarded. * Any partial writes to the block will be discarded.
* *
* @param b Block to abandon
* @param fileId The id of the file where the block resides. Older clients
* will pass GRANDFATHER_INODE_ID here.
* @param src The path of the file where the block resides.
* @param holder Lease holder.
*
* @throws AccessControlException If access is denied * @throws AccessControlException If access is denied
* @throws FileNotFoundException file <code>src</code> is not found * @throws FileNotFoundException file <code>src</code> is not found
* @throws UnresolvedLinkException If <code>src</code> contains a symlink * @throws UnresolvedLinkException If <code>src</code> contains a symlink
* @throws IOException If an I/O error occurred * @throws IOException If an I/O error occurred
*/ */
@Idempotent @Idempotent
public void abandonBlock(ExtendedBlock b, String src, String holder) public void abandonBlock(ExtendedBlock b, long fileId,
String src, String holder)
throws AccessControlException, FileNotFoundException, throws AccessControlException, FileNotFoundException,
UnresolvedLinkException, IOException; UnresolvedLinkException, IOException;
@ -344,6 +351,7 @@ public interface ClientProtocol {
* Get a datanode for an existing pipeline. * Get a datanode for an existing pipeline.
* *
* @param src the file being written * @param src the file being written
* @param fileId the ID of the file being written
* @param blk the block being written * @param blk the block being written
* @param existings the existing nodes in the pipeline * @param existings the existing nodes in the pipeline
* @param excludes the excluded nodes * @param excludes the excluded nodes
@ -359,8 +367,10 @@ public interface ClientProtocol {
* @throws IOException If an I/O error occurred * @throws IOException If an I/O error occurred
*/ */
@Idempotent @Idempotent
public LocatedBlock getAdditionalDatanode(final String src, final ExtendedBlock blk, public LocatedBlock getAdditionalDatanode(final String src,
final DatanodeInfo[] existings, final String[] existingStorageIDs, final long fileId, final ExtendedBlock blk,
final DatanodeInfo[] existings,
final String[] existingStorageIDs,
final DatanodeInfo[] excludes, final DatanodeInfo[] excludes,
final int numAdditionalNodes, final String clientName final int numAdditionalNodes, final String clientName
) throws AccessControlException, FileNotFoundException, ) throws AccessControlException, FileNotFoundException,
@ -896,6 +906,8 @@ public interface ClientProtocol {
* Write all metadata for this file into persistent storage. * Write all metadata for this file into persistent storage.
* The file must be currently open for writing. * The file must be currently open for writing.
* @param src The string representation of the path * @param src The string representation of the path
* @param inodeId The inode ID, or GRANDFATHER_INODE_ID if the client is
* too old to support fsync with inode IDs.
* @param client The string representation of the client * @param client The string representation of the client
* @param lastBlockLength The length of the last block (under construction) * @param lastBlockLength The length of the last block (under construction)
* to be reported to NameNode * to be reported to NameNode
@ -905,7 +917,8 @@ public interface ClientProtocol {
* @throws IOException If an I/O error occurred * @throws IOException If an I/O error occurred
*/ */
@Idempotent @Idempotent
public void fsync(String src, String client, long lastBlockLength) public void fsync(String src, long inodeId, String client,
long lastBlockLength)
throws AccessControlException, FileNotFoundException, throws AccessControlException, FileNotFoundException,
UnresolvedLinkException, IOException; UnresolvedLinkException, IOException;

View File

@ -422,8 +422,8 @@ public class ClientNamenodeProtocolServerSideTranslatorPB implements
public AbandonBlockResponseProto abandonBlock(RpcController controller, public AbandonBlockResponseProto abandonBlock(RpcController controller,
AbandonBlockRequestProto req) throws ServiceException { AbandonBlockRequestProto req) throws ServiceException {
try { try {
server.abandonBlock(PBHelper.convert(req.getB()), req.getSrc(), server.abandonBlock(PBHelper.convert(req.getB()), req.getFileId(),
req.getHolder()); req.getSrc(), req.getHolder());
} catch (IOException e) { } catch (IOException e) {
throw new ServiceException(e); throw new ServiceException(e);
} }
@ -461,7 +461,7 @@ public class ClientNamenodeProtocolServerSideTranslatorPB implements
List<String> existingStorageIDsList = req.getExistingStorageUuidsList(); List<String> existingStorageIDsList = req.getExistingStorageUuidsList();
List<DatanodeInfoProto> excludesList = req.getExcludesList(); List<DatanodeInfoProto> excludesList = req.getExcludesList();
LocatedBlock result = server.getAdditionalDatanode(req.getSrc(), LocatedBlock result = server.getAdditionalDatanode(req.getSrc(),
PBHelper.convert(req.getBlk()), req.getFileId(), PBHelper.convert(req.getBlk()),
PBHelper.convert(existingList.toArray( PBHelper.convert(existingList.toArray(
new DatanodeInfoProto[existingList.size()])), new DatanodeInfoProto[existingList.size()])),
existingStorageIDsList.toArray( existingStorageIDsList.toArray(
@ -819,7 +819,8 @@ public class ClientNamenodeProtocolServerSideTranslatorPB implements
public FsyncResponseProto fsync(RpcController controller, public FsyncResponseProto fsync(RpcController controller,
FsyncRequestProto req) throws ServiceException { FsyncRequestProto req) throws ServiceException {
try { try {
server.fsync(req.getSrc(), req.getClient(), req.getLastBlockLength()); server.fsync(req.getSrc(), req.getFileId(),
req.getClient(), req.getLastBlockLength());
return VOID_FSYNC_RESPONSE; return VOID_FSYNC_RESPONSE;
} catch (IOException e) { } catch (IOException e) {
throw new ServiceException(e); throw new ServiceException(e);

View File

@ -329,11 +329,12 @@ public class ClientNamenodeProtocolTranslatorPB implements
} }
@Override @Override
public void abandonBlock(ExtendedBlock b, String src, String holder) public void abandonBlock(ExtendedBlock b, long fileId, String src,
throws AccessControlException, FileNotFoundException, String holder) throws AccessControlException, FileNotFoundException,
UnresolvedLinkException, IOException { UnresolvedLinkException, IOException {
AbandonBlockRequestProto req = AbandonBlockRequestProto.newBuilder() AbandonBlockRequestProto req = AbandonBlockRequestProto.newBuilder()
.setB(PBHelper.convert(b)).setSrc(src).setHolder(holder).build(); .setB(PBHelper.convert(b)).setSrc(src).setHolder(holder)
.setFileId(fileId).build();
try { try {
rpcProxy.abandonBlock(null, req); rpcProxy.abandonBlock(null, req);
} catch (ServiceException e) { } catch (ServiceException e) {
@ -365,8 +366,8 @@ public class ClientNamenodeProtocolTranslatorPB implements
} }
@Override @Override
public LocatedBlock getAdditionalDatanode(String src, ExtendedBlock blk, public LocatedBlock getAdditionalDatanode(String src, long fileId,
DatanodeInfo[] existings, String[] existingStorageIDs, ExtendedBlock blk, DatanodeInfo[] existings, String[] existingStorageIDs,
DatanodeInfo[] excludes, DatanodeInfo[] excludes,
int numAdditionalNodes, String clientName) throws AccessControlException, int numAdditionalNodes, String clientName) throws AccessControlException,
FileNotFoundException, SafeModeException, UnresolvedLinkException, FileNotFoundException, SafeModeException, UnresolvedLinkException,
@ -374,6 +375,7 @@ public class ClientNamenodeProtocolTranslatorPB implements
GetAdditionalDatanodeRequestProto req = GetAdditionalDatanodeRequestProto GetAdditionalDatanodeRequestProto req = GetAdditionalDatanodeRequestProto
.newBuilder() .newBuilder()
.setSrc(src) .setSrc(src)
.setFileId(fileId)
.setBlk(PBHelper.convert(blk)) .setBlk(PBHelper.convert(blk))
.addAllExistings(PBHelper.convert(existings)) .addAllExistings(PBHelper.convert(existings))
.addAllExistingStorageUuids(Arrays.asList(existingStorageIDs)) .addAllExistingStorageUuids(Arrays.asList(existingStorageIDs))
@ -750,11 +752,13 @@ public class ClientNamenodeProtocolTranslatorPB implements
} }
@Override @Override
public void fsync(String src, String client, long lastBlockLength) public void fsync(String src, long fileId, String client,
long lastBlockLength)
throws AccessControlException, FileNotFoundException, throws AccessControlException, FileNotFoundException,
UnresolvedLinkException, IOException { UnresolvedLinkException, IOException {
FsyncRequestProto req = FsyncRequestProto.newBuilder().setSrc(src) FsyncRequestProto req = FsyncRequestProto.newBuilder().setSrc(src)
.setClient(client).setLastBlockLength(lastBlockLength).build(); .setClient(client).setLastBlockLength(lastBlockLength)
.setFileId(fileId).build();
try { try {
rpcProxy.fsync(null, req); rpcProxy.fsync(null, req);
} catch (ServiceException e) { } catch (ServiceException e) {

View File

@ -1149,7 +1149,7 @@ public class FSEditLog implements LogsPurgeable {
* Finalize the current log segment. * Finalize the current log segment.
* Transitions from IN_SEGMENT state to BETWEEN_LOG_SEGMENTS state. * Transitions from IN_SEGMENT state to BETWEEN_LOG_SEGMENTS state.
*/ */
synchronized void endCurrentLogSegment(boolean writeEndTxn) { public synchronized void endCurrentLogSegment(boolean writeEndTxn) {
LOG.info("Ending log segment " + curSegmentTxId); LOG.info("Ending log segment " + curSegmentTxId);
Preconditions.checkState(isSegmentOpen(), Preconditions.checkState(isSegmentOpen(),
"Bad state: %s", state); "Bad state: %s", state);

View File

@ -542,7 +542,7 @@ public class FSImage implements Closeable {
} }
@VisibleForTesting @VisibleForTesting
void setEditLogForTesting(FSEditLog newLog) { public void setEditLogForTesting(FSEditLog newLog) {
editLog = newLog; editLog = newLog;
} }

View File

@ -2667,9 +2667,9 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
checkOperation(OperationCategory.READ); checkOperation(OperationCategory.READ);
src = FSDirectory.resolvePath(src, pathComponents, dir); src = FSDirectory.resolvePath(src, pathComponents, dir);
LocatedBlock[] onRetryBlock = new LocatedBlock[1]; LocatedBlock[] onRetryBlock = new LocatedBlock[1];
final INode[] inodes = analyzeFileState( final INodeFile pendingFile = analyzeFileState(
src, fileId, clientName, previous, onRetryBlock).getINodes(); src, fileId, clientName, previous, onRetryBlock);
final INodeFile pendingFile = inodes[inodes.length - 1].asFile(); src = pendingFile.getFullPathName();
if (onRetryBlock[0] != null && onRetryBlock[0].getLocations().length > 0) { if (onRetryBlock[0] != null && onRetryBlock[0].getLocations().length > 0) {
// This is a retry. Just return the last block if having locations. // This is a retry. Just return the last block if having locations.
@ -2703,10 +2703,8 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
// Run the full analysis again, since things could have changed // Run the full analysis again, since things could have changed
// while chooseTarget() was executing. // while chooseTarget() was executing.
LocatedBlock[] onRetryBlock = new LocatedBlock[1]; LocatedBlock[] onRetryBlock = new LocatedBlock[1];
INodesInPath inodesInPath = final INodeFile pendingFile =
analyzeFileState(src, fileId, clientName, previous, onRetryBlock); analyzeFileState(src, fileId, clientName, previous, onRetryBlock);
final INode[] inodes = inodesInPath.getINodes();
final INodeFile pendingFile = inodes[inodes.length - 1].asFile();
if (onRetryBlock[0] != null) { if (onRetryBlock[0] != null) {
if (onRetryBlock[0].getLocations().length > 0) { if (onRetryBlock[0].getLocations().length > 0) {
@ -2728,6 +2726,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
// allocate new block, record block locations in INode. // allocate new block, record block locations in INode.
newBlock = createNewBlock(); newBlock = createNewBlock();
INodesInPath inodesInPath = INodesInPath.fromINode(pendingFile);
saveAllocatedBlock(src, inodesInPath, newBlock, targets); saveAllocatedBlock(src, inodesInPath, newBlock, targets);
dir.persistNewBlock(src, pendingFile); dir.persistNewBlock(src, pendingFile);
@ -2741,7 +2740,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
return makeLocatedBlock(newBlock, targets, offset); return makeLocatedBlock(newBlock, targets, offset);
} }
INodesInPath analyzeFileState(String src, INodeFile analyzeFileState(String src,
long fileId, long fileId,
String clientName, String clientName,
ExtendedBlock previous, ExtendedBlock previous,
@ -2758,9 +2757,20 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
checkFsObjectLimit(); checkFsObjectLimit();
Block previousBlock = ExtendedBlock.getLocalBlock(previous); Block previousBlock = ExtendedBlock.getLocalBlock(previous);
final INodesInPath iip = dir.getINodesInPath4Write(src); INode inode;
final INodeFile pendingFile if (fileId == INodeId.GRANDFATHER_INODE_ID) {
= checkLease(src, fileId, clientName, iip.getLastINode()); // Older clients may not have given us an inode ID to work with.
// In this case, we have to try to resolve the path and hope it
// hasn't changed or been deleted since the file was opened for write.
final INodesInPath iip = dir.getINodesInPath4Write(src);
inode = iip.getLastINode();
} else {
// Newer clients pass the inode ID, so we can just get the inode
// directly.
inode = dir.getInode(fileId);
if (inode != null) src = inode.getFullPathName();
}
final INodeFile pendingFile = checkLease(src, clientName, inode, fileId);
BlockInfo lastBlockInFile = pendingFile.getLastBlock(); BlockInfo lastBlockInFile = pendingFile.getLastBlock();
if (!Block.matchingIdAndGenStamp(previousBlock, lastBlockInFile)) { if (!Block.matchingIdAndGenStamp(previousBlock, lastBlockInFile)) {
// The block that the client claims is the current last block // The block that the client claims is the current last block
@ -2818,7 +2828,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
onRetryBlock[0] = makeLocatedBlock(lastBlockInFile, onRetryBlock[0] = makeLocatedBlock(lastBlockInFile,
((BlockInfoUnderConstruction)lastBlockInFile).getExpectedStorageLocations(), ((BlockInfoUnderConstruction)lastBlockInFile).getExpectedStorageLocations(),
offset); offset);
return iip; return pendingFile;
} else { } else {
// Case 3 // Case 3
throw new IOException("Cannot allocate block in " + src + ": " + throw new IOException("Cannot allocate block in " + src + ": " +
@ -2831,7 +2841,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
if (!checkFileProgress(pendingFile, false)) { if (!checkFileProgress(pendingFile, false)) {
throw new NotReplicatedYetException("Not replicated yet: " + src); throw new NotReplicatedYetException("Not replicated yet: " + src);
} }
return iip; return pendingFile;
} }
LocatedBlock makeLocatedBlock(Block blk, DatanodeStorageInfo[] locs, LocatedBlock makeLocatedBlock(Block blk, DatanodeStorageInfo[] locs,
@ -2844,8 +2854,9 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
} }
/** @see ClientProtocol#getAdditionalDatanode */ /** @see ClientProtocol#getAdditionalDatanode */
LocatedBlock getAdditionalDatanode(String src, final ExtendedBlock blk, LocatedBlock getAdditionalDatanode(String src, long fileId,
final DatanodeInfo[] existings, final String[] storageIDs, final ExtendedBlock blk, final DatanodeInfo[] existings,
final String[] storageIDs,
final Set<Node> excludes, final Set<Node> excludes,
final int numAdditionalNodes, final String clientName final int numAdditionalNodes, final String clientName
) throws IOException { ) throws IOException {
@ -2865,7 +2876,17 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
src = FSDirectory.resolvePath(src, pathComponents, dir); src = FSDirectory.resolvePath(src, pathComponents, dir);
//check lease //check lease
final INodeFile file = checkLease(src, clientName); final INode inode;
if (fileId == INodeId.GRANDFATHER_INODE_ID) {
// Older clients may not have given us an inode ID to work with.
// In this case, we have to try to resolve the path and hope it
// hasn't changed or been deleted since the file was opened for write.
inode = dir.getINode(src);
} else {
inode = dir.getInode(fileId);
if (inode != null) src = inode.getFullPathName();
}
final INodeFile file = checkLease(src, clientName, inode, fileId);
clientnode = file.getFileUnderConstructionFeature().getClientNode(); clientnode = file.getFileUnderConstructionFeature().getClientNode();
preferredblocksize = file.getPreferredBlockSize(); preferredblocksize = file.getPreferredBlockSize();
@ -2889,7 +2910,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
/** /**
* The client would like to let go of the given block * The client would like to let go of the given block
*/ */
boolean abandonBlock(ExtendedBlock b, String src, String holder) boolean abandonBlock(ExtendedBlock b, long fileId, String src, String holder)
throws LeaseExpiredException, FileNotFoundException, throws LeaseExpiredException, FileNotFoundException,
UnresolvedLinkException, IOException { UnresolvedLinkException, IOException {
if(NameNode.stateChangeLog.isDebugEnabled()) { if(NameNode.stateChangeLog.isDebugEnabled()) {
@ -2901,13 +2922,24 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
writeLock(); writeLock();
try { try {
checkOperation(OperationCategory.WRITE); checkOperation(OperationCategory.WRITE);
checkNameNodeSafeMode("Cannot abandon block " + b + " for fle" + src); checkNameNodeSafeMode("Cannot abandon block " + b + " for file" + src);
src = FSDirectory.resolvePath(src, pathComponents, dir); src = FSDirectory.resolvePath(src, pathComponents, dir);
final INode inode;
if (fileId == INodeId.GRANDFATHER_INODE_ID) {
// Older clients may not have given us an inode ID to work with.
// In this case, we have to try to resolve the path and hope it
// hasn't changed or been deleted since the file was opened for write.
inode = dir.getINode(src);
} else {
inode = dir.getInode(fileId);
if (inode != null) src = inode.getFullPathName();
}
final INodeFile file = checkLease(src, holder, inode, fileId);
// //
// Remove the block from the pending creates list // Remove the block from the pending creates list
// //
INodeFile file = checkLease(src, holder);
boolean removed = dir.removeBlock(src, file, boolean removed = dir.removeBlock(src, file,
ExtendedBlock.getLocalBlock(b)); ExtendedBlock.getLocalBlock(b));
if (!removed) { if (!removed) {
@ -2925,39 +2957,39 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
return true; return true;
} }
/** make sure that we still have the lease on this file. */ private INodeFile checkLease(String src, String holder, INode inode,
private INodeFile checkLease(String src, String holder) long fileId)
throws LeaseExpiredException, UnresolvedLinkException, throws LeaseExpiredException, FileNotFoundException {
FileNotFoundException {
return checkLease(src, INodeId.GRANDFATHER_INODE_ID, holder,
dir.getINode(src));
}
private INodeFile checkLease(String src, long fileId, String holder,
INode inode) throws LeaseExpiredException, FileNotFoundException {
assert hasReadLock(); assert hasReadLock();
if (inode == null || !inode.isFile()) { final String ident = src + " (inode " + fileId + ")";
if (inode == null) {
Lease lease = leaseManager.getLease(holder); Lease lease = leaseManager.getLease(holder);
throw new LeaseExpiredException( throw new LeaseExpiredException(
"No lease on " + src + ": File does not exist. " "No lease on " + ident + ": File does not exist. "
+ (lease != null ? lease.toString() + (lease != null ? lease.toString()
: "Holder " + holder + " does not have any open files.")); : "Holder " + holder + " does not have any open files."));
} }
if (!inode.isFile()) {
Lease lease = leaseManager.getLease(holder);
throw new LeaseExpiredException(
"No lease on " + ident + ": INode is not a regular file. "
+ (lease != null ? lease.toString()
: "Holder " + holder + " does not have any open files."));
}
final INodeFile file = inode.asFile(); final INodeFile file = inode.asFile();
if (!file.isUnderConstruction()) { if (!file.isUnderConstruction()) {
Lease lease = leaseManager.getLease(holder); Lease lease = leaseManager.getLease(holder);
throw new LeaseExpiredException( throw new LeaseExpiredException(
"No lease on " + src + ": File is not open for writing. " "No lease on " + ident + ": File is not open for writing. "
+ (lease != null ? lease.toString() + (lease != null ? lease.toString()
: "Holder " + holder + " does not have any open files.")); : "Holder " + holder + " does not have any open files."));
} }
String clientName = file.getFileUnderConstructionFeature().getClientName(); String clientName = file.getFileUnderConstructionFeature().getClientName();
if (holder != null && !clientName.equals(holder)) { if (holder != null && !clientName.equals(holder)) {
throw new LeaseExpiredException("Lease mismatch on " + src + " owned by " throw new LeaseExpiredException("Lease mismatch on " + ident +
+ clientName + " but is accessed by " + holder); " owned by " + clientName + " but is accessed by " + holder);
} }
INodeId.checkId(fileId, file);
return file; return file;
} }
@ -3000,10 +3032,20 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
String holder, Block last, long fileId) throws SafeModeException, String holder, Block last, long fileId) throws SafeModeException,
UnresolvedLinkException, IOException { UnresolvedLinkException, IOException {
assert hasWriteLock(); assert hasWriteLock();
final INodesInPath iip = dir.getLastINodeInPath(src);
final INodeFile pendingFile; final INodeFile pendingFile;
try { try {
pendingFile = checkLease(src, fileId, holder, iip.getINode(0)); final INode inode;
if (fileId == INodeId.GRANDFATHER_INODE_ID) {
// Older clients may not have given us an inode ID to work with.
// In this case, we have to try to resolve the path and hope it
// hasn't changed or been deleted since the file was opened for write.
final INodesInPath iip = dir.getLastINodeInPath(src);
inode = iip.getINode(0);
} else {
inode = dir.getInode(fileId);
if (inode != null) src = inode.getFullPathName();
}
pendingFile = checkLease(src, holder, inode, fileId);
} catch (LeaseExpiredException lee) { } catch (LeaseExpiredException lee) {
final INode inode = dir.getINode(src); final INode inode = dir.getINode(src);
if (inode != null if (inode != null
@ -3018,9 +3060,9 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
final Block realLastBlock = inode.asFile().getLastBlock(); final Block realLastBlock = inode.asFile().getLastBlock();
if (Block.matchingIdAndGenStamp(last, realLastBlock)) { if (Block.matchingIdAndGenStamp(last, realLastBlock)) {
NameNode.stateChangeLog.info("DIR* completeFile: " + NameNode.stateChangeLog.info("DIR* completeFile: " +
"request from " + holder + " to complete " + src + "request from " + holder + " to complete inode " + fileId +
" which is already closed. But, it appears to be an RPC " + "(" + src + ") which is already closed. But, it appears to be " +
"retry. Returning success"); "an RPC retry. Returning success");
return true; return true;
} }
} }
@ -3040,7 +3082,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
} }
finalizeINodeFileUnderConstruction(src, pendingFile, finalizeINodeFileUnderConstruction(src, pendingFile,
iip.getLatestSnapshotId()); Snapshot.CURRENT_STATE_ID);
return true; return true;
} }
@ -3681,12 +3723,14 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
/** Persist all metadata about this file. /** Persist all metadata about this file.
* @param src The string representation of the path * @param src The string representation of the path
* @param fileId The inode ID that we're fsyncing. Older clients will pass
* INodeId.GRANDFATHER_INODE_ID here.
* @param clientName The string representation of the client * @param clientName The string representation of the client
* @param lastBlockLength The length of the last block * @param lastBlockLength The length of the last block
* under construction reported from client. * under construction reported from client.
* @throws IOException if path does not exist * @throws IOException if path does not exist
*/ */
void fsync(String src, String clientName, long lastBlockLength) void fsync(String src, long fileId, String clientName, long lastBlockLength)
throws IOException, UnresolvedLinkException { throws IOException, UnresolvedLinkException {
NameNode.stateChangeLog.info("BLOCK* fsync: " + src + " for " + clientName); NameNode.stateChangeLog.info("BLOCK* fsync: " + src + " for " + clientName);
checkOperation(OperationCategory.WRITE); checkOperation(OperationCategory.WRITE);
@ -3696,7 +3740,17 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
checkOperation(OperationCategory.WRITE); checkOperation(OperationCategory.WRITE);
checkNameNodeSafeMode("Cannot fsync file " + src); checkNameNodeSafeMode("Cannot fsync file " + src);
src = FSDirectory.resolvePath(src, pathComponents, dir); src = FSDirectory.resolvePath(src, pathComponents, dir);
INodeFile pendingFile = checkLease(src, clientName); final INode inode;
if (fileId == INodeId.GRANDFATHER_INODE_ID) {
// Older clients may not have given us an inode ID to work with.
// In this case, we have to try to resolve the path and hope it
// hasn't changed or been deleted since the file was opened for write.
inode = dir.getINode(src);
} else {
inode = dir.getInode(fileId);
if (inode != null) src = inode.getFullPathName();
}
final INodeFile pendingFile = checkLease(src, clientName, inode, fileId);
if (lastBlockLength > 0) { if (lastBlockLength > 0) {
pendingFile.getFileUnderConstructionFeature().updateLengthOfLastBlock( pendingFile.getFileUnderConstructionFeature().updateLengthOfLastBlock(
pendingFile, lastBlockLength); pendingFile, lastBlockLength);

View File

@ -46,6 +46,28 @@ public class INodesInPath {
: Arrays.equals(HdfsConstants.DOT_SNAPSHOT_DIR_BYTES, pathComponent); : Arrays.equals(HdfsConstants.DOT_SNAPSHOT_DIR_BYTES, pathComponent);
} }
static INodesInPath fromINode(INode inode) {
int depth = 0, index;
INode tmp = inode;
while (tmp != null) {
depth++;
tmp = tmp.getParent();
}
final byte[][] path = new byte[depth][];
final INode[] inodes = new INode[depth];
final INodesInPath iip = new INodesInPath(path, depth);
tmp = inode;
index = depth;
while (tmp != null) {
index--;
path[index] = tmp.getKey();
inodes[index] = tmp;
tmp = tmp.getParent();
}
iip.setINodes(inodes);
return iip;
}
/** /**
* Given some components, create a path name. * Given some components, create a path name.
* @param components The path components * @param components The path components
@ -341,6 +363,11 @@ public class INodesInPath {
private void addNode(INode node) { private void addNode(INode node) {
inodes[numNonNull++] = node; inodes[numNonNull++] = node;
} }
private void setINodes(INode inodes[]) {
this.inodes = inodes;
this.numNonNull = this.inodes.length;
}
void setINode(int i, INode inode) { void setINode(int i, INode inode) {
inodes[i >= 0? i: inodes.length + i] = inode; inodes[i >= 0? i: inodes.length + i] = inode;
@ -417,4 +444,4 @@ public class INodesInPath {
+ ", this=" + toString(false)); + ", this=" + toString(false));
} }
} }
} }

View File

@ -595,13 +595,15 @@ class NameNodeRpcServer implements NamenodeProtocols {
} }
@Override // ClientProtocol @Override // ClientProtocol
public LocatedBlock getAdditionalDatanode(final String src, final ExtendedBlock blk, public LocatedBlock getAdditionalDatanode(final String src,
final long fileId, final ExtendedBlock blk,
final DatanodeInfo[] existings, final String[] existingStorageIDs, final DatanodeInfo[] existings, final String[] existingStorageIDs,
final DatanodeInfo[] excludes, final DatanodeInfo[] excludes,
final int numAdditionalNodes, final String clientName final int numAdditionalNodes, final String clientName
) throws IOException { ) throws IOException {
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("getAdditionalDatanode: src=" + src LOG.debug("getAdditionalDatanode: src=" + src
+ ", fileId=" + fileId
+ ", blk=" + blk + ", blk=" + blk
+ ", existings=" + Arrays.asList(existings) + ", existings=" + Arrays.asList(existings)
+ ", excludes=" + Arrays.asList(excludes) + ", excludes=" + Arrays.asList(excludes)
@ -618,20 +620,20 @@ class NameNodeRpcServer implements NamenodeProtocols {
excludeSet.add(node); excludeSet.add(node);
} }
} }
return namesystem.getAdditionalDatanode(src, blk, existings, return namesystem.getAdditionalDatanode(src, fileId, blk, existings,
existingStorageIDs, excludeSet, numAdditionalNodes, clientName); existingStorageIDs, excludeSet, numAdditionalNodes, clientName);
} }
/** /**
* The client needs to give up on the block. * The client needs to give up on the block.
*/ */
@Override // ClientProtocol @Override // ClientProtocol
public void abandonBlock(ExtendedBlock b, String src, String holder) public void abandonBlock(ExtendedBlock b, long fileId, String src,
throws IOException { String holder) throws IOException {
if(stateChangeLog.isDebugEnabled()) { if(stateChangeLog.isDebugEnabled()) {
stateChangeLog.debug("*BLOCK* NameNode.abandonBlock: " stateChangeLog.debug("*BLOCK* NameNode.abandonBlock: "
+b+" of file "+src); +b+" of file "+src);
} }
if (!namesystem.abandonBlock(b, src, holder)) { if (!namesystem.abandonBlock(b, fileId, src, holder)) {
throw new IOException("Cannot abandon block during write to " + src); throw new IOException("Cannot abandon block during write to " + src);
} }
} }
@ -939,9 +941,10 @@ class NameNodeRpcServer implements NamenodeProtocols {
} }
@Override // ClientProtocol @Override // ClientProtocol
public void fsync(String src, String clientName, long lastBlockLength) public void fsync(String src, long fileId, String clientName,
long lastBlockLength)
throws IOException { throws IOException {
namesystem.fsync(src, clientName, lastBlockLength); namesystem.fsync(src, fileId, clientName, lastBlockLength);
} }
@Override // ClientProtocol @Override // ClientProtocol

View File

@ -117,6 +117,7 @@ message AbandonBlockRequestProto {
required ExtendedBlockProto b = 1; required ExtendedBlockProto b = 1;
required string src = 2; required string src = 2;
required string holder = 3; required string holder = 3;
optional uint64 fileId = 4 [default = 0]; // default to GRANDFATHER_INODE_ID
} }
message AbandonBlockResponseProto { // void response message AbandonBlockResponseProto { // void response
@ -143,6 +144,7 @@ message GetAdditionalDatanodeRequestProto {
required uint32 numAdditionalNodes = 5; required uint32 numAdditionalNodes = 5;
required string clientName = 6; required string clientName = 6;
repeated string existingStorageUuids = 7; repeated string existingStorageUuids = 7;
optional uint64 fileId = 8 [default = 0]; // default to GRANDFATHER_INODE_ID
} }
message GetAdditionalDatanodeResponseProto { message GetAdditionalDatanodeResponseProto {
@ -532,6 +534,7 @@ message FsyncRequestProto {
required string src = 1; required string src = 1;
required string client = 2; required string client = 2;
optional sint64 lastBlockLength = 3 [default = -1]; optional sint64 lastBlockLength = 3 [default = -1];
optional uint64 fileId = 4 [default = 0]; // default to GRANDFATHER_INODE_ID
} }
message FsyncResponseProto { // void response message FsyncResponseProto { // void response

View File

@ -54,8 +54,8 @@ public class DFSClientAdapter {
return dfs.dfs; return dfs.dfs;
} }
public static ExtendedBlock getPreviousBlock(DFSClient client, String file) { public static ExtendedBlock getPreviousBlock(DFSClient client, long fileId) {
return client.getPreviousBlock(file); return client.getPreviousBlock(fileId);
} }
public static long getFileId(DFSOutputStream out) { public static long getFileId(DFSOutputStream out) {

View File

@ -71,6 +71,7 @@ public class TestAbandonBlock {
fout.write(123); fout.write(123);
} }
fout.hflush(); fout.hflush();
long fileId = ((DFSOutputStream)fout.getWrappedStream()).getFileId();
// Now abandon the last block // Now abandon the last block
DFSClient dfsclient = DFSClientAdapter.getDFSClient(fs); DFSClient dfsclient = DFSClientAdapter.getDFSClient(fs);
@ -78,11 +79,11 @@ public class TestAbandonBlock {
dfsclient.getNamenode().getBlockLocations(src, 0, Integer.MAX_VALUE); dfsclient.getNamenode().getBlockLocations(src, 0, Integer.MAX_VALUE);
int orginalNumBlocks = blocks.locatedBlockCount(); int orginalNumBlocks = blocks.locatedBlockCount();
LocatedBlock b = blocks.getLastLocatedBlock(); LocatedBlock b = blocks.getLastLocatedBlock();
dfsclient.getNamenode().abandonBlock(b.getBlock(), src, dfsclient.getNamenode().abandonBlock(b.getBlock(), fileId, src,
dfsclient.clientName); dfsclient.clientName);
// call abandonBlock again to make sure the operation is idempotent // call abandonBlock again to make sure the operation is idempotent
dfsclient.getNamenode().abandonBlock(b.getBlock(), src, dfsclient.getNamenode().abandonBlock(b.getBlock(), fileId, src,
dfsclient.clientName); dfsclient.clientName);
// And close the file // And close the file

View File

@ -253,18 +253,9 @@ public class TestFileAppend3 {
assertTrue(fs.rename(p, pnew)); assertTrue(fs.rename(p, pnew));
//d. Close file handle that was opened in (b). //d. Close file handle that was opened in (b).
try { out.close();
out.close();
fail("close() should throw an exception");
} catch(Exception e) {
AppendTestUtil.LOG.info("GOOD!", e);
}
//wait for the lease recovery //check block sizes
cluster.setLeasePeriod(1000, 1000);
AppendTestUtil.sleep(5000);
//check block sizes
final long len = fs.getFileStatus(pnew).getLen(); final long len = fs.getFileStatus(pnew).getLen();
final LocatedBlocks locatedblocks = fs.dfs.getNamenode().getBlockLocations(pnew.toString(), 0L, len); final LocatedBlocks locatedblocks = fs.dfs.getNamenode().getBlockLocations(pnew.toString(), 0L, len);
final int numblock = locatedblocks.locatedBlockCount(); final int numblock = locatedblocks.locatedBlockCount();

View File

@ -74,6 +74,7 @@ import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.INodeId; import org.apache.hadoop.hdfs.server.namenode.INodeId;
import org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException;
import org.apache.hadoop.hdfs.server.namenode.LeaseManager; import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols; import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
import org.apache.hadoop.io.EnumSetWritable; import org.apache.hadoop.io.EnumSetWritable;
@ -384,7 +385,6 @@ public class TestFileCreation {
Path p = new Path("/testfile"); Path p = new Path("/testfile");
FSDataOutputStream stm1 = fs.create(p); FSDataOutputStream stm1 = fs.create(p);
stm1.write(1); stm1.write(1);
stm1.hflush();
// Create file again without overwrite // Create file again without overwrite
try { try {
@ -403,7 +403,8 @@ public class TestFileCreation {
stm1.close(); stm1.close();
fail("Should have exception closing stm1 since it was deleted"); fail("Should have exception closing stm1 since it was deleted");
} catch (IOException ioe) { } catch (IOException ioe) {
GenericTestUtils.assertExceptionContains("File is not open for writing", ioe); GenericTestUtils.assertExceptionContains("No lease on /testfile", ioe);
GenericTestUtils.assertExceptionContains("File does not exist.", ioe);
} }
} finally { } finally {
@ -1189,8 +1190,8 @@ public class TestFileCreation {
cluster.getNameNodeRpc() cluster.getNameNodeRpc()
.complete(f.toString(), client.clientName, null, someOtherFileId); .complete(f.toString(), client.clientName, null, someOtherFileId);
fail(); fail();
} catch(FileNotFoundException fnf) { } catch(LeaseExpiredException e) {
FileSystem.LOG.info("Caught Expected FileNotFoundException: ", fnf); FileSystem.LOG.info("Caught Expected LeaseExpiredException: ", e);
} }
} finally { } finally {
IOUtils.closeStream(dfs); IOUtils.closeStream(dfs);

View File

@ -242,11 +242,56 @@ public class TestLease {
Assert.assertTrue(pRenamedAgain+" not found", fs2.exists(pRenamedAgain)); Assert.assertTrue(pRenamedAgain+" not found", fs2.exists(pRenamedAgain));
Assert.assertTrue("no lease for "+pRenamedAgain, hasLease(cluster, pRenamedAgain)); Assert.assertTrue("no lease for "+pRenamedAgain, hasLease(cluster, pRenamedAgain));
Assert.assertEquals(1, leaseCount(cluster)); Assert.assertEquals(1, leaseCount(cluster));
out.close();
} finally { } finally {
cluster.shutdown(); cluster.shutdown();
} }
} }
/**
* Test that we can open up a file for write, move it to another location,
* and then create a new file in the previous location, without causing any
* lease conflicts. This is possible because we now use unique inode IDs
* to identify files to the NameNode.
*/
@Test
public void testLeaseAfterRenameAndRecreate() throws Exception {
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
try {
final Path path1 = new Path("/test-file");
final String contents1 = "contents1";
final Path path2 = new Path("/test-file-new-location");
final String contents2 = "contents2";
// open a file to get a lease
FileSystem fs = cluster.getFileSystem();
FSDataOutputStream out1 = fs.create(path1);
out1.writeBytes(contents1);
Assert.assertTrue(hasLease(cluster, path1));
Assert.assertEquals(1, leaseCount(cluster));
DistributedFileSystem fs2 = (DistributedFileSystem)
FileSystem.newInstance(fs.getUri(), fs.getConf());
fs2.rename(path1, path2);
FSDataOutputStream out2 = fs2.create(path1);
out2.writeBytes(contents2);
out2.close();
// The first file should still be open and valid
Assert.assertTrue(hasLease(cluster, path2));
out1.close();
// Contents should be as expected
DistributedFileSystem fs3 = (DistributedFileSystem)
FileSystem.newInstance(fs.getUri(), fs.getConf());
Assert.assertEquals(contents1, DFSTestUtil.readFile(fs3, path2));
Assert.assertEquals(contents2, DFSTestUtil.readFile(fs3, path1));
} finally {
cluster.shutdown();
}
}
@Test @Test
public void testLease() throws Exception { public void testLease() throws Exception {
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build(); MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();

View File

@ -107,8 +107,8 @@ public class TestLeaseRenewer {
// Set up a file so that we start renewing our lease. // Set up a file so that we start renewing our lease.
DFSOutputStream mockStream = Mockito.mock(DFSOutputStream.class); DFSOutputStream mockStream = Mockito.mock(DFSOutputStream.class);
String filePath = "/foo"; long fileId = 123L;
renewer.put(filePath, mockStream, MOCK_DFSCLIENT); renewer.put(fileId, mockStream, MOCK_DFSCLIENT);
// Wait for lease to get renewed // Wait for lease to get renewed
long failTime = Time.now() + 5000; long failTime = Time.now() + 5000;
@ -120,7 +120,7 @@ public class TestLeaseRenewer {
Assert.fail("Did not renew lease at all!"); Assert.fail("Did not renew lease at all!");
} }
renewer.closeFile(filePath, MOCK_DFSCLIENT); renewer.closeFile(fileId, MOCK_DFSCLIENT);
} }
/** /**
@ -138,8 +138,8 @@ public class TestLeaseRenewer {
// Set up a file so that we start renewing our lease. // Set up a file so that we start renewing our lease.
DFSOutputStream mockStream1 = Mockito.mock(DFSOutputStream.class); DFSOutputStream mockStream1 = Mockito.mock(DFSOutputStream.class);
String filePath = "/foo"; long fileId = 456L;
renewer.put(filePath, mockStream1, mockClient1); renewer.put(fileId, mockStream1, mockClient1);
// Second DFSClient does renew lease // Second DFSClient does renew lease
final DFSClient mockClient2 = createMockClient(); final DFSClient mockClient2 = createMockClient();
@ -149,7 +149,7 @@ public class TestLeaseRenewer {
// Set up a file so that we start renewing our lease. // Set up a file so that we start renewing our lease.
DFSOutputStream mockStream2 = Mockito.mock(DFSOutputStream.class); DFSOutputStream mockStream2 = Mockito.mock(DFSOutputStream.class);
renewer.put(filePath, mockStream2, mockClient2); renewer.put(fileId, mockStream2, mockClient2);
// Wait for lease to get renewed // Wait for lease to get renewed
@ -170,19 +170,19 @@ public class TestLeaseRenewer {
} }
}, 100, 10000); }, 100, 10000);
renewer.closeFile(filePath, mockClient1); renewer.closeFile(fileId, mockClient1);
renewer.closeFile(filePath, mockClient2); renewer.closeFile(fileId, mockClient2);
} }
@Test @Test
public void testThreadName() throws Exception { public void testThreadName() throws Exception {
DFSOutputStream mockStream = Mockito.mock(DFSOutputStream.class); DFSOutputStream mockStream = Mockito.mock(DFSOutputStream.class);
String filePath = "/foo"; long fileId = 789L;
Assert.assertFalse("Renewer not initially running", Assert.assertFalse("Renewer not initially running",
renewer.isRunning()); renewer.isRunning());
// Pretend to open a file // Pretend to open a file
renewer.put(filePath, mockStream, MOCK_DFSCLIENT); renewer.put(fileId, mockStream, MOCK_DFSCLIENT);
Assert.assertTrue("Renewer should have started running", Assert.assertTrue("Renewer should have started running",
renewer.isRunning()); renewer.isRunning());
@ -192,7 +192,7 @@ public class TestLeaseRenewer {
Assert.assertEquals("LeaseRenewer:myuser@hdfs://nn1/", threadName); Assert.assertEquals("LeaseRenewer:myuser@hdfs://nn1/", threadName);
// Pretend to close the file // Pretend to close the file
renewer.closeFile(filePath, MOCK_DFSCLIENT); renewer.closeFile(fileId, MOCK_DFSCLIENT);
renewer.setEmptyTime(Time.now()); renewer.setEmptyTime(Time.now());
// Should stop the renewer running within a few seconds // Should stop the renewer running within a few seconds

View File

@ -35,6 +35,7 @@ import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.LocatedBlock;
import org.apache.hadoop.hdfs.protocol.LocatedBlocks; import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
@ -159,12 +160,13 @@ public class TestPersistBlocks {
// Abandon the last block // Abandon the last block
DFSClient dfsclient = DFSClientAdapter.getDFSClient((DistributedFileSystem)fs); DFSClient dfsclient = DFSClientAdapter.getDFSClient((DistributedFileSystem)fs);
HdfsFileStatus fileStatus = dfsclient.getNamenode().getFileInfo(FILE_NAME);
LocatedBlocks blocks = dfsclient.getNamenode().getBlockLocations( LocatedBlocks blocks = dfsclient.getNamenode().getBlockLocations(
FILE_NAME, 0, BLOCK_SIZE * NUM_BLOCKS); FILE_NAME, 0, BLOCK_SIZE * NUM_BLOCKS);
assertEquals(NUM_BLOCKS, blocks.getLocatedBlocks().size()); assertEquals(NUM_BLOCKS, blocks.getLocatedBlocks().size());
LocatedBlock b = blocks.getLastLocatedBlock(); LocatedBlock b = blocks.getLastLocatedBlock();
dfsclient.getNamenode().abandonBlock(b.getBlock(), FILE_NAME, dfsclient.getNamenode().abandonBlock(b.getBlock(), fileStatus.getFileId(),
dfsclient.clientName); FILE_NAME, dfsclient.clientName);
// explicitly do NOT close the file. // explicitly do NOT close the file.
cluster.restartNameNode(); cluster.restartNameNode();

View File

@ -17,6 +17,8 @@
*/ */
package org.apache.hadoop.hdfs; package org.apache.hadoop.hdfs;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.mockito.Mockito.doNothing;
import static org.mockito.Mockito.spy;
import java.io.IOException; import java.io.IOException;
@ -26,11 +28,13 @@ import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.server.namenode.FSEditLog;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.LeaseManager; import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
import org.apache.hadoop.hdfs.server.namenode.NameNode; import org.apache.hadoop.hdfs.server.namenode.NameNode;
import org.apache.log4j.Level; import org.apache.log4j.Level;
import org.junit.Test; import org.junit.Test;
import org.mockito.Mockito;
public class TestRenameWhileOpen { public class TestRenameWhileOpen {
{ {
@ -57,6 +61,7 @@ public class TestRenameWhileOpen {
conf.setInt(DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY, 1000); conf.setInt(DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY, 1000);
conf.setInt(DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY, 1); conf.setInt(DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY, 1);
conf.setInt(DFSConfigKeys.DFS_NAMENODE_SAFEMODE_THRESHOLD_PCT_KEY, 1); conf.setInt(DFSConfigKeys.DFS_NAMENODE_SAFEMODE_THRESHOLD_PCT_KEY, 1);
conf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, TestFileCreation.blockSize);
// create cluster // create cluster
System.out.println("Test 1*****************************"); System.out.println("Test 1*****************************");
@ -65,6 +70,15 @@ public class TestRenameWhileOpen {
try { try {
cluster.waitActive(); cluster.waitActive();
fs = cluster.getFileSystem(); fs = cluster.getFileSystem();
// Normally, the in-progress edit log would be finalized by
// FSEditLog#endCurrentLogSegment. For testing purposes, we
// disable that here.
FSEditLog spyLog =
spy(cluster.getNameNode().getFSImage().getEditLog());
doNothing().when(spyLog).endCurrentLogSegment(Mockito.anyBoolean());
cluster.getNameNode().getFSImage().setEditLogForTesting(spyLog);
final int nnport = cluster.getNameNodePort(); final int nnport = cluster.getNameNodePort();
// create file1. // create file1.
@ -92,18 +106,21 @@ public class TestRenameWhileOpen {
// create file3 // create file3
Path file3 = new Path(dir3, "file3"); Path file3 = new Path(dir3, "file3");
FSDataOutputStream stm3 = TestFileCreation.createFile(fs, file3, 1); FSDataOutputStream stm3 = fs.create(file3);
TestFileCreation.writeFile(stm3); fs.rename(file3, new Path(dir3, "bozo"));
// rename file3 to some bad name // Get a new block for the file.
try { TestFileCreation.writeFile(stm3, TestFileCreation.blockSize + 1);
fs.rename(file3, new Path(dir3, "$ ")); stm3.hflush();
} catch(Exception e) {
e.printStackTrace(); // Stop the NameNode before closing the files.
} // This will ensure that the write leases are still active and present
// in the edit log. Simiarly, there should be a pending ADD_BLOCK_OP
// restart cluster with the same namenode port as before. // for file3, since we just added a block to that file.
// This ensures that leases are persisted in fsimage. cluster.getNameNode().stop();
// Restart cluster with the same namenode port as before.
cluster.shutdown(); cluster.shutdown();
try {Thread.sleep(2*MAX_IDLE_TIME);} catch (InterruptedException e) {} try {Thread.sleep(2*MAX_IDLE_TIME);} catch (InterruptedException e) {}
cluster = new MiniDFSCluster.Builder(conf).nameNodePort(nnport) cluster = new MiniDFSCluster.Builder(conf).nameNodePort(nnport)
.format(false) .format(false)
@ -111,7 +128,7 @@ public class TestRenameWhileOpen {
cluster.waitActive(); cluster.waitActive();
// restart cluster yet again. This triggers the code to read in // restart cluster yet again. This triggers the code to read in
// persistent leases from fsimage. // persistent leases from the edit log.
cluster.shutdown(); cluster.shutdown();
try {Thread.sleep(5000);} catch (InterruptedException e) {} try {Thread.sleep(5000);} catch (InterruptedException e) {}
cluster = new MiniDFSCluster.Builder(conf).nameNodePort(nnport) cluster = new MiniDFSCluster.Builder(conf).nameNodePort(nnport)

View File

@ -22,7 +22,7 @@ import java.io.IOException;
import org.junit.Test; import org.junit.Test;
public class TestSetrepDecreasing { public class TestSetrepDecreasing {
@Test @Test(timeout=120000)
public void testSetrepDecreasing() throws IOException { public void testSetrepDecreasing() throws IOException {
TestSetrepIncreasing.setrep(5, 3, false); TestSetrepIncreasing.setrep(5, 3, false);
} }

View File

@ -75,11 +75,11 @@ public class TestSetrepIncreasing {
} }
} }
@Test @Test(timeout=120000)
public void testSetrepIncreasing() throws IOException { public void testSetrepIncreasing() throws IOException {
setrep(3, 7, false); setrep(3, 7, false);
} }
@Test @Test(timeout=120000)
public void testSetrepIncreasingSimulatedStorage() throws IOException { public void testSetrepIncreasingSimulatedStorage() throws IOException {
setrep(3, 7, true); setrep(3, 7, true);
} }

View File

@ -468,8 +468,8 @@ public class TestINodeFile {
} }
} }
@Test @Test(timeout=120000)
public void testWriteToRenamedFile() throws IOException { public void testWriteToDeletedFile() throws IOException {
Configuration conf = new Configuration(); Configuration conf = new Configuration();
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1) MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1)
.build(); .build();
@ -486,18 +486,16 @@ public class TestINodeFile {
Path filePath = new Path("/test1/file"); Path filePath = new Path("/test1/file");
FSDataOutputStream fos = fs.create(filePath); FSDataOutputStream fos = fs.create(filePath);
// Rename /test1 to test2, and recreate /test1/file // Delete the file
Path renamedPath = new Path("/test2"); fs.delete(filePath, false);
fs.rename(path, renamedPath);
fs.create(filePath, (short) 1);
// Add new block should fail since /test1/file has a different fileId // Add new block should fail since /test1/file has been deleted.
try { try {
fos.write(data, 0, data.length); fos.write(data, 0, data.length);
// make sure addBlock() request gets to NN immediately // make sure addBlock() request gets to NN immediately
fos.hflush(); fos.hflush();
fail("Write should fail after rename"); fail("Write should fail after delete");
} catch (Exception e) { } catch (Exception e) {
/* Ignore */ /* Ignore */
} finally { } finally {

View File

@ -36,6 +36,7 @@ import org.apache.commons.logging.impl.Log4JLogger;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.FsPermission;
@ -819,10 +820,13 @@ public class TestHASafeMode {
null); null);
create.write(testData.getBytes()); create.write(testData.getBytes());
create.hflush(); create.hflush();
long fileId = ((DFSOutputStream)create.
getWrappedStream()).getFileId();
FileStatus fileStatus = dfs.getFileStatus(filePath);
DFSClient client = DFSClientAdapter.getClient(dfs); DFSClient client = DFSClientAdapter.getClient(dfs);
// add one dummy block at NN, but not write to DataNode // add one dummy block at NN, but not write to DataNode
ExtendedBlock previousBlock = DFSClientAdapter.getPreviousBlock(client, ExtendedBlock previousBlock =
pathString); DFSClientAdapter.getPreviousBlock(client, fileId);
DFSClientAdapter.getNamenode(client).addBlock( DFSClientAdapter.getNamenode(client).addBlock(
pathString, pathString,
client.getClientName(), client.getClientName(),