Merge HDFS-4525. Provide an API for knowing that whether file is closed or not. Contributed by SreeHari.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1465435 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
b66abebca2
commit
d208e6a15e
|
@ -38,6 +38,9 @@ Release 2.0.5-beta - UNRELEASED
|
||||||
|
|
||||||
HDFS-4618. Default transaction interval for checkpoints is too low. (todd)
|
HDFS-4618. Default transaction interval for checkpoints is too low. (todd)
|
||||||
|
|
||||||
|
HDFS-4525. Provide an API for knowing that whether file is closed or not.
|
||||||
|
(SreeHari via umamahesh)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
|
@ -1571,7 +1571,22 @@ public class DFSClient implements java.io.Closeable {
|
||||||
UnresolvedPathException.class);
|
UnresolvedPathException.class);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Close status of a file
|
||||||
|
* @return true if file is already closed
|
||||||
|
*/
|
||||||
|
public boolean isFileClosed(String src) throws IOException{
|
||||||
|
checkOpen();
|
||||||
|
try {
|
||||||
|
return namenode.isFileClosed(src);
|
||||||
|
} catch(RemoteException re) {
|
||||||
|
throw re.unwrapRemoteException(AccessControlException.class,
|
||||||
|
FileNotFoundException.class,
|
||||||
|
UnresolvedPathException.class);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the file info for a specific file or directory. If src
|
* Get the file info for a specific file or directory. If src
|
||||||
* refers to a symlink then the FileStatus of the link is returned.
|
* refers to a symlink then the FileStatus of the link is returned.
|
||||||
|
|
|
@ -961,4 +961,17 @@ public class DistributedFileSystem extends FileSystem {
|
||||||
public boolean isInSafeMode() throws IOException {
|
public boolean isInSafeMode() throws IOException {
|
||||||
return setSafeMode(SafeModeAction.SAFEMODE_GET, true);
|
return setSafeMode(SafeModeAction.SAFEMODE_GET, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the close status of a file
|
||||||
|
* @param src The path to the file
|
||||||
|
*
|
||||||
|
* @return return true if file is closed
|
||||||
|
* @throws FileNotFoundException if the file does not exist.
|
||||||
|
* @throws IOException If an I/O error occurred
|
||||||
|
*/
|
||||||
|
public boolean isFileClosed(Path src) throws IOException {
|
||||||
|
return dfs.isFileClosed(getPathName(src));
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -753,7 +753,21 @@ public interface ClientProtocol {
|
||||||
@Idempotent
|
@Idempotent
|
||||||
public HdfsFileStatus getFileInfo(String src) throws AccessControlException,
|
public HdfsFileStatus getFileInfo(String src) throws AccessControlException,
|
||||||
FileNotFoundException, UnresolvedLinkException, IOException;
|
FileNotFoundException, UnresolvedLinkException, IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the close status of a file
|
||||||
|
* @param src The string representation of the path to the file
|
||||||
|
*
|
||||||
|
* @return return true if file is closed
|
||||||
|
* @throws AccessControlException permission denied
|
||||||
|
* @throws FileNotFoundException file <code>src</code> is not found
|
||||||
|
* @throws UnresolvedLinkException if the path contains a symlink.
|
||||||
|
* @throws IOException If an I/O error occurred
|
||||||
|
*/
|
||||||
|
@Idempotent
|
||||||
|
public boolean isFileClosed(String src) throws AccessControlException,
|
||||||
|
FileNotFoundException, UnresolvedLinkException, IOException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the file info for a specific file or directory. If the path
|
* Get the file info for a specific file or directory. If the path
|
||||||
* refers to a symlink then the FileStatus of the symlink is returned.
|
* refers to a symlink then the FileStatus of the symlink is returned.
|
||||||
|
|
|
@ -76,6 +76,8 @@ import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetPre
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetPreferredBlockSizeResponseProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetPreferredBlockSizeResponseProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsResponseProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsResponseProto;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.IsFileClosedRequestProto;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.IsFileClosedResponseProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksResponseProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksResponseProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MetaSaveRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MetaSaveRequestProto;
|
||||||
|
@ -858,4 +860,17 @@ public class ClientNamenodeProtocolServerSideTranslatorPB implements
|
||||||
throw new ServiceException(e);
|
throw new ServiceException(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public IsFileClosedResponseProto isFileClosed(
|
||||||
|
RpcController controller, IsFileClosedRequestProto request)
|
||||||
|
throws ServiceException {
|
||||||
|
try {
|
||||||
|
boolean result = server.isFileClosed(request.getSrc());
|
||||||
|
return IsFileClosedResponseProto.newBuilder().setResult(result).build();
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new ServiceException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -76,6 +76,7 @@ import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetLis
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetListingResponseProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetListingResponseProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetPreferredBlockSizeRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetPreferredBlockSizeRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.GetServerDefaultsRequestProto;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.IsFileClosedRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.ListCorruptFileBlocksRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MetaSaveRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MetaSaveRequestProto;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MkdirsRequestProto;
|
import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.MkdirsRequestProto;
|
||||||
|
@ -846,6 +847,19 @@ public class ClientNamenodeProtocolTranslatorPB implements
|
||||||
throw ProtobufHelper.getRemoteException(e);
|
throw ProtobufHelper.getRemoteException(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean isFileClosed(String src) throws AccessControlException,
|
||||||
|
FileNotFoundException, UnresolvedLinkException, IOException {
|
||||||
|
IsFileClosedRequestProto req = IsFileClosedRequestProto.newBuilder()
|
||||||
|
.setSrc(src).build();
|
||||||
|
try {
|
||||||
|
return rpcProxy.isFileClosed(null, req).getResult();
|
||||||
|
} catch (ServiceException e) {
|
||||||
|
throw ProtobufHelper.getRemoteException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Object getUnderlyingProxyObject() {
|
public Object getUnderlyingProxyObject() {
|
||||||
|
|
|
@ -2904,6 +2904,33 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
|
||||||
logAuditEvent(true, "getfileinfo", src);
|
logAuditEvent(true, "getfileinfo", src);
|
||||||
return stat;
|
return stat;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns true if the file is closed
|
||||||
|
*/
|
||||||
|
boolean isFileClosed(String src)
|
||||||
|
throws AccessControlException, UnresolvedLinkException,
|
||||||
|
StandbyException, IOException {
|
||||||
|
FSPermissionChecker pc = getPermissionChecker();
|
||||||
|
checkOperation(OperationCategory.READ);
|
||||||
|
readLock();
|
||||||
|
try {
|
||||||
|
checkOperation(OperationCategory.READ);
|
||||||
|
if (isPermissionEnabled) {
|
||||||
|
checkTraverse(pc, src);
|
||||||
|
}
|
||||||
|
return !INodeFile.valueOf(dir.getINode(src), src).isUnderConstruction();
|
||||||
|
} catch (AccessControlException e) {
|
||||||
|
if (isAuditEnabled() && isExternalInvocation()) {
|
||||||
|
logAuditEvent(false, UserGroupInformation.getCurrentUser(),
|
||||||
|
getRemoteIp(),
|
||||||
|
"isFileClosed", src, null, null);
|
||||||
|
}
|
||||||
|
throw e;
|
||||||
|
} finally {
|
||||||
|
readUnlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create all the necessary directories
|
* Create all the necessary directories
|
||||||
|
|
|
@ -682,7 +682,12 @@ class NameNodeRpcServer implements NamenodeProtocols {
|
||||||
metrics.incrFileInfoOps();
|
metrics.incrFileInfoOps();
|
||||||
return namesystem.getFileInfo(src, true);
|
return namesystem.getFileInfo(src, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override // ClientProtocol
|
||||||
|
public boolean isFileClosed(String src) throws IOException{
|
||||||
|
return namesystem.isFileClosed(src);
|
||||||
|
}
|
||||||
|
|
||||||
@Override // ClientProtocol
|
@Override // ClientProtocol
|
||||||
public HdfsFileStatus getFileLinkInfo(String src) throws IOException {
|
public HdfsFileStatus getFileLinkInfo(String src) throws IOException {
|
||||||
metrics.incrFileInfoOps();
|
metrics.incrFileInfoOps();
|
||||||
|
|
|
@ -330,6 +330,14 @@ message GetFileInfoResponseProto {
|
||||||
optional HdfsFileStatusProto fs = 1;
|
optional HdfsFileStatusProto fs = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
message IsFileClosedRequestProto {
|
||||||
|
required string src = 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
message IsFileClosedResponseProto {
|
||||||
|
required bool result = 1;
|
||||||
|
}
|
||||||
|
|
||||||
message GetFileLinkInfoRequestProto {
|
message GetFileLinkInfoRequestProto {
|
||||||
required string src = 1;
|
required string src = 1;
|
||||||
}
|
}
|
||||||
|
@ -496,4 +504,6 @@ service ClientNamenodeProtocol {
|
||||||
returns(SetBalancerBandwidthResponseProto);
|
returns(SetBalancerBandwidthResponseProto);
|
||||||
rpc getDataEncryptionKey(GetDataEncryptionKeyRequestProto)
|
rpc getDataEncryptionKey(GetDataEncryptionKeyRequestProto)
|
||||||
returns(GetDataEncryptionKeyResponseProto);
|
returns(GetDataEncryptionKeyResponseProto);
|
||||||
|
rpc isFileClosed(IsFileClosedRequestProto)
|
||||||
|
returns(IsFileClosedResponseProto);
|
||||||
}
|
}
|
||||||
|
|
|
@ -735,4 +735,27 @@ public class TestDistributedFileSystem {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test(timeout=60000)
|
||||||
|
public void testFileCloseStatus() throws IOException {
|
||||||
|
Configuration conf = new HdfsConfiguration();
|
||||||
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
||||||
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
try {
|
||||||
|
// create a new file.
|
||||||
|
Path file = new Path("/simpleFlush.dat");
|
||||||
|
FSDataOutputStream output = fs.create(file);
|
||||||
|
// write to file
|
||||||
|
output.writeBytes("Some test data");
|
||||||
|
output.flush();
|
||||||
|
assertFalse("File status should be open", fs.isFileClosed(file));
|
||||||
|
output.close();
|
||||||
|
assertTrue("File status should be closed", fs.isFileClosed(file));
|
||||||
|
} finally {
|
||||||
|
if (cluster != null) {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue