HDFS-4507. Update quota verification for snapshots.

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-2802@1451081 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Tsz-wo Sze 2013-02-28 03:08:08 +00:00
parent e2a618e1cc
commit c7cf85ccb4
25 changed files with 515 additions and 273 deletions

View File

@ -179,3 +179,5 @@ Branch-2802 Snapshot (Unreleased)
HDFS-4523. Fix INodeFile replacement, TestQuota and javac errors from trunk
merge. (szetszwo)
HDFS-4507. Update quota verification for snapshots. (szetszwo)

View File

@ -20,7 +20,6 @@ package org.apache.hadoop.hdfs.protocol;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.fs.Path;
@InterfaceAudience.Private
@InterfaceStability.Evolving
@ -48,21 +47,29 @@ public abstract class FSLimitException extends QuotaExceededException {
class PathComponentTooLongException extends FSLimitException {
protected static final long serialVersionUID = 1L;
private String childName;
protected PathComponentTooLongException() {}
protected PathComponentTooLongException(String msg) {
super(msg);
}
public PathComponentTooLongException(long quota, long count) {
public PathComponentTooLongException(long quota, long count,
String parentPath, String childName) {
super(quota, count);
setPathName(parentPath);
this.childName = childName;
}
String getParentPath() {
return pathName;
}
@Override
public String getMessage() {
Path violator = new Path(pathName);
return "The maximum path component name limit of " + violator.getName() +
" in directory " + violator.getParent() +
return "The maximum path component name limit of " + childName +
" in directory " + getParentPath() +
" is exceeded: limit=" + quota + " length=" + count;
}
}

View File

@ -26,6 +26,8 @@ import org.apache.hadoop.classification.InterfaceStability;
public final class NSQuotaExceededException extends QuotaExceededException {
protected static final long serialVersionUID = 1L;
private String prefix;
public NSQuotaExceededException() {}
public NSQuotaExceededException(String msg) {
@ -40,11 +42,19 @@ public final class NSQuotaExceededException extends QuotaExceededException {
public String getMessage() {
String msg = super.getMessage();
if (msg == null) {
return "The NameSpace quota (directories and files)" +
msg = "The NameSpace quota (directories and files)" +
(pathName==null?"":(" of directory " + pathName)) +
" is exceeded: quota=" + quota + " file count=" + count;
} else {
return msg;
if (prefix != null) {
msg = prefix + ": " + msg;
}
}
return msg;
}
/** Set a prefix for the error message. */
public void setMessagePrefix(final String prefix) {
this.prefix = prefix;
}
}

View File

@ -52,6 +52,7 @@ import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
import org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus;
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.QuotaExceededException;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction;
@ -838,7 +839,8 @@ public class FSDirectory implements Closeable {
long dsDelta = (replication - oldRepl) * (fileNode.diskspaceConsumed()/oldRepl);
updateCount(inodesInPath, 0, dsDelta, true);
fileNode.setFileReplication(replication, inodesInPath.getLatestSnapshot());
fileNode = fileNode.setFileReplication(
replication, inodesInPath.getLatestSnapshot());
if (oldReplication != null) {
oldReplication[0] = oldRepl;
@ -877,7 +879,7 @@ public class FSDirectory implements Closeable {
void setPermission(String src, FsPermission permission)
throws FileNotFoundException, UnresolvedLinkException,
SnapshotAccessControlException {
NSQuotaExceededException, SnapshotAccessControlException {
writeLock();
try {
unprotectedSetPermission(src, permission);
@ -889,7 +891,7 @@ public class FSDirectory implements Closeable {
void unprotectedSetPermission(String src, FsPermission permissions)
throws FileNotFoundException, UnresolvedLinkException,
SnapshotAccessControlException {
NSQuotaExceededException, SnapshotAccessControlException {
assert hasWriteLock();
final INodesInPath inodesInPath = rootDir.getINodesInPath4Write(src, true);
final INode inode = inodesInPath.getLastINode();
@ -901,7 +903,7 @@ public class FSDirectory implements Closeable {
void setOwner(String src, String username, String groupname)
throws FileNotFoundException, UnresolvedLinkException,
SnapshotAccessControlException {
NSQuotaExceededException, SnapshotAccessControlException {
writeLock();
try {
unprotectedSetOwner(src, username, groupname);
@ -913,7 +915,7 @@ public class FSDirectory implements Closeable {
void unprotectedSetOwner(String src, String username, String groupname)
throws FileNotFoundException, UnresolvedLinkException,
SnapshotAccessControlException {
NSQuotaExceededException, SnapshotAccessControlException {
assert hasWriteLock();
final INodesInPath inodesInPath = rootDir.getINodesInPath4Write(src, true);
INode inode = inodesInPath.getLastINode();
@ -932,7 +934,8 @@ public class FSDirectory implements Closeable {
* Concat all the blocks from srcs to trg and delete the srcs files
*/
public void concat(String target, String [] srcs)
throws UnresolvedLinkException, SnapshotAccessControlException {
throws UnresolvedLinkException, NSQuotaExceededException,
SnapshotAccessControlException {
writeLock();
try {
// actual move
@ -956,7 +959,8 @@ public class FSDirectory implements Closeable {
* NOTE: - it does not update quota (not needed for concat)
*/
public void unprotectedConcat(String target, String [] srcs, long timestamp)
throws UnresolvedLinkException, SnapshotAccessControlException {
throws UnresolvedLinkException, NSQuotaExceededException,
SnapshotAccessControlException {
assert hasWriteLock();
if (NameNode.stateChangeLog.isDebugEnabled()) {
NameNode.stateChangeLog.debug("DIR* FSNamesystem.concat to "+target);
@ -1093,8 +1097,8 @@ public class FSDirectory implements Closeable {
* @param mtime the time the inode is removed
* @throws SnapshotAccessControlException if path is in RO snapshot
*/
void unprotectedDelete(String src, long mtime)
throws UnresolvedLinkException, SnapshotAccessControlException {
void unprotectedDelete(String src, long mtime) throws UnresolvedLinkException,
NSQuotaExceededException, SnapshotAccessControlException {
assert hasWriteLock();
BlocksMapUpdateInfo collectedBlocks = new BlocksMapUpdateInfo();
@ -1116,7 +1120,8 @@ public class FSDirectory implements Closeable {
* @return the number of inodes deleted; 0 if no inodes are deleted.
*/
int unprotectedDelete(INodesInPath inodesInPath,
BlocksMapUpdateInfo collectedBlocks, long mtime) {
BlocksMapUpdateInfo collectedBlocks, long mtime)
throws NSQuotaExceededException {
assert hasWriteLock();
// check if target node exists
@ -1812,37 +1817,62 @@ public class FSDirectory implements Closeable {
/**
* Verify that filesystem limit constraints are not violated
* @throws PathComponentTooLongException child's name is too long
* @throws MaxDirectoryItemsExceededException items per directory is exceeded
*/
protected <T extends INode> void verifyFsLimits(INode[] pathComponents,
int pos, T child) throws FSLimitException {
boolean includeChildName = false;
try {
if (maxComponentLength != 0) {
int length = child.getLocalName().length();
if (length > maxComponentLength) {
includeChildName = true;
throw new PathComponentTooLongException(maxComponentLength, length);
}
void verifyFsLimits(INode[] pathComponents, int pos, INode child)
throws FSLimitException {
verifyMaxComponentLength(child.getLocalName(), pathComponents, pos);
verifyMaxDirItems(pathComponents, pos);
}
/**
* Verify child's name for fs limit.
* @throws PathComponentTooLongException child's name is too long.
*/
public void verifyMaxComponentLength(String childName,
Object parentPath, int pos) throws PathComponentTooLongException {
if (maxComponentLength == 0) {
return;
}
final int length = childName.length();
if (length > maxComponentLength) {
final String p = parentPath instanceof INode[]?
getFullPathName((INode[])parentPath, pos - 1): (String)parentPath;
final PathComponentTooLongException e = new PathComponentTooLongException(
maxComponentLength, length, p, childName);
if (ready) {
throw e;
} else {
// Do not throw if edits log is still being processed
NameNode.LOG.error("FSDirectory.verifyMaxComponentLength: "
+ e.getLocalizedMessage());
}
if (maxDirItems != 0) {
INodeDirectory parent = (INodeDirectory)pathComponents[pos-1];
int count = parent.getChildrenList(null).size();
if (count >= maxDirItems) {
throw new MaxDirectoryItemsExceededException(maxDirItems, count);
}
}
}
/**
* Verify children size for fs limit.
* @throws MaxDirectoryItemsExceededException too many children.
*/
private void verifyMaxDirItems(INode[] pathComponents, int pos)
throws MaxDirectoryItemsExceededException {
if (maxDirItems == 0) {
return;
}
final INodeDirectory parent = (INodeDirectory)pathComponents[pos-1];
final int count = parent.getChildrenList(null).size();
if (count >= maxDirItems) {
final MaxDirectoryItemsExceededException e
= new MaxDirectoryItemsExceededException(maxDirItems, count);
if (ready) {
e.setPathName(getFullPathName(pathComponents, pos - 1));
throw e;
} else {
// Do not throw if edits log is still being processed
NameNode.LOG.error("FSDirectory.verifyMaxDirItems: "
+ e.getLocalizedMessage());
}
} catch (FSLimitException e) {
String badPath = getFullPathName(pathComponents, pos-1);
if (includeChildName) {
badPath += Path.SEPARATOR + child.getLocalName();
}
e.setPathName(badPath);
// Do not throw if edits log is still being processed
if (ready) throw(e);
// log pre-existing paths that exceed limits
NameNode.LOG.error("FSDirectory.verifyFsLimits - " + e.getLocalizedMessage());
}
}
@ -1900,20 +1930,22 @@ public class FSDirectory implements Closeable {
* Remove the last inode in the path from the namespace.
* Count of each ancestor with quota is also updated.
* @return the removed node; null if the removal fails.
* @throws NSQuotaExceededException
*/
private INode removeLastINode(final INodesInPath inodesInPath) {
private INode removeLastINode(final INodesInPath inodesInPath)
throws NSQuotaExceededException {
final Snapshot latestSnapshot = inodesInPath.getLatestSnapshot();
final INode[] inodes = inodesInPath.getINodes();
final int pos = inodes.length - 1;
final boolean removed = ((INodeDirectory)inodes[pos-1]).removeChild(
inodes[pos], inodesInPath.getLatestSnapshot());
if (removed) {
inodes[pos], latestSnapshot);
if (removed && latestSnapshot == null) {
inodesInPath.setINode(pos - 1, inodes[pos].getParent());
final Quota.Counts counts = inodes[pos].computeQuotaUsage();
updateCountNoQuotaCheck(inodesInPath, pos,
-counts.get(Quota.NAMESPACE), -counts.get(Quota.DISKSPACE));
return inodes[pos];
}
return null;
return removed? inodes[pos]: null;
}
/**
@ -1986,7 +2018,7 @@ public class FSDirectory implements Closeable {
final Snapshot latest = iip.getLatestSnapshot();
if (dirNode instanceof INodeDirectoryWithQuota) {
// a directory with quota; so set the quota to the new value
((INodeDirectoryWithQuota)dirNode).setQuota(nsQuota, dsQuota, latest);
((INodeDirectoryWithQuota)dirNode).setQuota(nsQuota, dsQuota);
if (!dirNode.isQuotaSet() && latest == null) {
// will not come here for root because root's nsQuota is always set
return dirNode.replaceSelf4INodeDirectory();
@ -2033,7 +2065,7 @@ public class FSDirectory implements Closeable {
* Sets the access time on the file/directory. Logs it in the transaction log.
*/
void setTimes(String src, INode inode, long mtime, long atime, boolean force,
Snapshot latest) {
Snapshot latest) throws NSQuotaExceededException {
boolean status = false;
writeLock();
try {
@ -2047,7 +2079,7 @@ public class FSDirectory implements Closeable {
}
boolean unprotectedSetTimes(String src, long mtime, long atime, boolean force)
throws UnresolvedLinkException {
throws UnresolvedLinkException, NSQuotaExceededException {
assert hasWriteLock();
final INodesInPath i = getLastINodeInPath(src);
return unprotectedSetTimes(src, i.getLastINode(), mtime, atime, force,
@ -2055,7 +2087,8 @@ public class FSDirectory implements Closeable {
}
private boolean unprotectedSetTimes(String src, INode inode, long mtime,
long atime, boolean force, Snapshot latest) {
long atime, boolean force, Snapshot latest)
throws NSQuotaExceededException {
assert hasWriteLock();
boolean status = false;
if (mtime != -1) {

View File

@ -17,6 +17,8 @@
*/
package org.apache.hadoop.hdfs.server.namenode;
import static org.apache.hadoop.util.Time.now;
import java.io.Closeable;
import java.io.File;
import java.io.IOException;
@ -28,28 +30,26 @@ import java.util.HashMap;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Stack;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.HAUtil;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.LayoutVersion;
import org.apache.hadoop.hdfs.protocol.LayoutVersion.Feature;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException;
import org.apache.hadoop.hdfs.server.common.Storage;
import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable;
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
import org.apache.hadoop.hdfs.server.common.Storage.StorageState;
import org.apache.hadoop.hdfs.server.common.Util;
import static org.apache.hadoop.util.Time.now;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeDirType;
import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeFile;
import org.apache.hadoop.hdfs.server.protocol.CheckpointCommand;
@ -62,9 +62,6 @@ import org.apache.hadoop.hdfs.util.MD5FileUtils;
import org.apache.hadoop.io.MD5Hash;
import org.apache.hadoop.util.IdGenerator;
import org.apache.hadoop.util.Time;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.HAUtil;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Joiner;
@ -726,7 +723,8 @@ public class FSImage implements Closeable {
return lastAppliedTxId - prevLastAppliedTxId;
}
/** Update the count of each directory with quota in the namespace
/**
* Update the count of each directory with quota in the namespace.
* A directory's count is defined as the total number inodes in the tree
* rooted at the directory.
*
@ -734,39 +732,22 @@ public class FSImage implements Closeable {
* throw QuotaExceededException.
*/
static void updateCountForQuota(INodeDirectoryWithQuota root) {
updateCountForINodeWithQuota(root, new Quota.Counts(), new Stack<INode>());
updateCountForQuotaRecursively(root, new Quota.Counts());
}
/**
* Update the count of the directory if it has a quota and return the count
*
* This does not throw a QuotaExceededException. This is just an update
* of of existing state and throwing QuotaExceededException does not help
* with fixing the state, if there is a problem.
*
* @param dir the root of the tree that represents the directory
* @param counters counters for name space and disk space
* @param stack INodes for the each of components in the path.
*/
private static void updateCountForINodeWithQuota(INodeDirectory dir,
Quota.Counts counts, Stack<INode> stack) {
// The stack is not needed since we could use the 'parent' field in INode.
// However, using 'parent' is not recommended.
stack.push(dir);
private static void updateCountForQuotaRecursively(INodeDirectory dir,
Quota.Counts counts) {
final long parentNamespace = counts.get(Quota.NAMESPACE);
final long parentDiskspace = counts.get(Quota.DISKSPACE);
counts.add(Quota.NAMESPACE, 1);
dir.computeQuotaUsage4CurrentDirectory(counts);
for (INode child : dir.getChildrenList(null)) {
if (child.isDirectory()) {
updateCountForINodeWithQuota((INodeDirectory)child, counts, stack);
updateCountForQuotaRecursively((INodeDirectory)child, counts);
} else {
// file or symlink: count here to reduce recursive calls.
counts.add(Quota.NAMESPACE, 1);
if (child.isFile()) {
counts.add(Quota.DISKSPACE, ((INodeFile)child).diskspaceConsumed());
}
child.computeQuotaUsage(counts, false);
}
}
@ -774,24 +755,20 @@ public class FSImage implements Closeable {
// check if quota is violated. It indicates a software bug.
final long namespace = counts.get(Quota.NAMESPACE) - parentNamespace;
if (Quota.isViolated(dir.getNsQuota(), namespace)) {
final INode[] inodes = stack.toArray(new INode[stack.size()]);
LOG.error("BUG: Namespace quota violation in image for "
+ FSDirectory.getFullPathName(inodes, inodes.length)
+ dir.getFullPathName()
+ " quota = " + dir.getNsQuota() + " < consumed = " + namespace);
}
final long diskspace = counts.get(Quota.DISKSPACE) - parentDiskspace;
if (Quota.isViolated(dir.getDsQuota(), diskspace)) {
final INode[] inodes = stack.toArray(new INode[stack.size()]);
LOG.error("BUG: Diskspace quota violation in image for "
+ FSDirectory.getFullPathName(inodes, inodes.length)
+ dir.getFullPathName()
+ " quota = " + dir.getDsQuota() + " < consumed = " + diskspace);
}
((INodeDirectoryWithQuota)dir).setSpaceConsumed(namespace, diskspace);
}
stack.pop();
}
/**

View File

@ -319,7 +319,7 @@ public class FSImageFormat {
long dsQuota = root.getDsQuota();
FSDirectory fsDir = namesystem.dir;
if (nsQuota != -1 || dsQuota != -1) {
fsDir.rootDir.setQuota(nsQuota, dsQuota, null);
fsDir.rootDir.setQuota(nsQuota, dsQuota);
}
fsDir.rootDir.cloneModificationTime(root);
fsDir.rootDir.clonePermissionStatus(root);
@ -503,7 +503,7 @@ public class FSImageFormat {
*/
private void addToParent(INodeDirectory parent, INode child) {
// NOTE: This does not update space counts for parents
if (!parent.addChild(child, false, null)) {
if (!parent.addChild(child)) {
return;
}
namesystem.dir.cacheName(child);

View File

@ -32,12 +32,14 @@ import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.fs.permission.PermissionStatus;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.namenode.INode.Content.CountsMap.Key;
import org.apache.hadoop.hdfs.server.namenode.snapshot.Snapshot;
import org.apache.hadoop.hdfs.server.namenode.snapshot.diff.Diff;
import org.apache.hadoop.util.StringUtils;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Preconditions;
import com.google.common.primitives.SignedBytes;
/**
@ -158,11 +160,8 @@ public abstract class INode implements Diff.Element<byte[]> {
public final PermissionStatus getPermissionStatus() {
return getPermissionStatus(null);
}
private INode updatePermissionStatus(PermissionStatusFormat f, long n,
Snapshot latest) {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.permission = f.combine(n, permission);
return nodeToUpdate;
private void updatePermissionStatus(PermissionStatusFormat f, long n) {
this.permission = f.combine(n, permission);
}
/**
* @param snapshot
@ -183,9 +182,16 @@ public abstract class INode implements Diff.Element<byte[]> {
return getUserName(null);
}
/** Set user */
protected INode setUser(String user, Snapshot latest) {
final void setUser(String user) {
int n = SerialNumberManager.INSTANCE.getUserSerialNumber(user);
return updatePermissionStatus(PermissionStatusFormat.USER, n, latest);
updatePermissionStatus(PermissionStatusFormat.USER, n);
}
/** Set user */
final INode setUser(String user, Snapshot latest)
throws NSQuotaExceededException {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.setUser(user);
return nodeToUpdate;
}
/**
* @param snapshot
@ -206,9 +212,16 @@ public abstract class INode implements Diff.Element<byte[]> {
return getGroupName(null);
}
/** Set group */
protected INode setGroup(String group, Snapshot latest) {
final void setGroup(String group) {
int n = SerialNumberManager.INSTANCE.getGroupSerialNumber(group);
return updatePermissionStatus(PermissionStatusFormat.GROUP, n, latest);
updatePermissionStatus(PermissionStatusFormat.GROUP, n);
}
/** Set group */
final INode setGroup(String group, Snapshot latest)
throws NSQuotaExceededException {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.setGroup(group);
return nodeToUpdate;
}
/**
* @param snapshot
@ -232,9 +245,16 @@ public abstract class INode implements Diff.Element<byte[]> {
return (short)PermissionStatusFormat.MODE.retrieve(permission);
}
/** Set the {@link FsPermission} of this {@link INode} */
INode setPermission(FsPermission permission, Snapshot latest) {
void setPermission(FsPermission permission) {
final short mode = permission.toShort();
return updatePermissionStatus(PermissionStatusFormat.MODE, mode, latest);
updatePermissionStatus(PermissionStatusFormat.MODE, mode);
}
/** Set the {@link FsPermission} of this {@link INode} */
INode setPermission(FsPermission permission, Snapshot latest)
throws NSQuotaExceededException {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.setPermission(permission);
return nodeToUpdate;
}
/**
@ -263,7 +283,8 @@ public abstract class INode implements Diff.Element<byte[]> {
* However, in some cases, this inode may be replaced with a new inode
* for maintaining snapshots. The current inode is then the new inode.
*/
abstract INode recordModification(final Snapshot latest);
abstract INode recordModification(final Snapshot latest)
throws NSQuotaExceededException;
/**
* Check whether it's a file.
@ -335,7 +356,7 @@ public abstract class INode implements Diff.Element<byte[]> {
* @return the number of deleted inodes in the subtree.
*/
public abstract int cleanSubtree(final Snapshot snapshot, Snapshot prior,
BlocksMapUpdateInfo collectedBlocks);
BlocksMapUpdateInfo collectedBlocks) throws NSQuotaExceededException;
/**
* Destroy self and clear everything! If the INode is a file, this method
@ -427,6 +448,16 @@ public abstract class INode implements Diff.Element<byte[]> {
*/
public abstract Content.Counts computeContentSummary(Content.Counts counts);
/**
* Check and add namespace consumed to itself and the ancestors.
* @throws NSQuotaExceededException if quote is violated.
*/
public void addNamespaceConsumed(int delta) throws NSQuotaExceededException {
if (parent != null) {
parent.addNamespaceConsumed(delta);
}
}
/**
* Get the quota set for this inode
* @return the quota if it is set; -1 otherwise
@ -447,7 +478,7 @@ public abstract class INode implements Diff.Element<byte[]> {
* Count subtree {@link Quota#NAMESPACE} and {@link Quota#DISKSPACE} usages.
*/
final Quota.Counts computeQuotaUsage() {
return computeQuotaUsage(new Quota.Counts());
return computeQuotaUsage(new Quota.Counts(), true);
}
/**
@ -456,7 +487,8 @@ public abstract class INode implements Diff.Element<byte[]> {
* @param counts The subtree counts for returning.
* @return The same objects as the counts parameter.
*/
abstract Quota.Counts computeQuotaUsage(Quota.Counts counts);
public abstract Quota.Counts computeQuotaUsage(Quota.Counts counts,
boolean useCache);
/**
* @return null if the local name is null; otherwise, return the local name.
@ -574,8 +606,9 @@ public abstract class INode implements Diff.Element<byte[]> {
}
/** Update modification time if it is larger than the current value. */
public final INode updateModificationTime(long mtime, Snapshot latest) {
assert isDirectory();
public final INode updateModificationTime(long mtime, Snapshot latest)
throws NSQuotaExceededException {
Preconditions.checkState(isDirectory());
if (mtime <= modificationTime) {
return this;
}
@ -586,12 +619,15 @@ public abstract class INode implements Diff.Element<byte[]> {
this.modificationTime = that.modificationTime;
}
/**
* Always set the last modification time of inode.
*/
public final INode setModificationTime(long modtime, Snapshot latest) {
/** Set the last modification time of inode. */
public final void setModificationTime(long modificationTime) {
this.modificationTime = modificationTime;
}
/** Set the last modification time of inode. */
public final INode setModificationTime(long modificationTime, Snapshot latest)
throws NSQuotaExceededException {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.modificationTime = modtime;
nodeToUpdate.setModificationTime(modificationTime);
return nodeToUpdate;
}
@ -617,9 +653,16 @@ public abstract class INode implements Diff.Element<byte[]> {
/**
* Set last access time of inode.
*/
public INode setAccessTime(long atime, Snapshot latest) {
public void setAccessTime(long accessTime) {
this.accessTime = accessTime;
}
/**
* Set last access time of inode.
*/
public INode setAccessTime(long accessTime, Snapshot latest)
throws NSQuotaExceededException {
final INode nodeToUpdate = recordModification(latest);
nodeToUpdate.accessTime = atime;
nodeToUpdate.setAccessTime(accessTime);
return nodeToUpdate;
}

View File

@ -29,6 +29,7 @@ import org.apache.hadoop.fs.UnresolvedLinkException;
import org.apache.hadoop.fs.permission.PermissionStatus;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.UnresolvedPathException;
import org.apache.hadoop.hdfs.server.namenode.INode.Content.CountsMap.Key;
import org.apache.hadoop.hdfs.server.namenode.snapshot.INodeDirectorySnapshottable;
@ -97,18 +98,12 @@ public class INodeDirectory extends INode {
return false;
}
private void assertChildrenNonNull() {
if (children == null) {
throw new AssertionError("children is null: " + this);
}
}
private int searchChildren(byte[] name) {
return Collections.binarySearch(children, name);
return children == null? -1: Collections.binarySearch(children, name);
}
protected int searchChildrenForExistingINode(final INode inode) {
assertChildrenNonNull();
private int searchChildrenForExistingINode(final INode inode) {
Preconditions.checkNotNull(children);
final byte[] name = inode.getLocalNameBytes();
final int i = searchChildren(name);
if (i < 0) {
@ -124,7 +119,8 @@ public class INodeDirectory extends INode {
* @param child the child inode to be removed
* @param latest See {@link INode#recordModification(Snapshot)}.
*/
public boolean removeChild(INode child, Snapshot latest) {
public boolean removeChild(INode child, Snapshot latest)
throws NSQuotaExceededException {
if (isInLatestSnapshot(latest)) {
return replaceSelf4INodeDirectoryWithSnapshot()
.removeChild(child, latest);
@ -142,7 +138,7 @@ public class INodeDirectory extends INode {
* @return true if the child is removed; false if the child is not found.
*/
protected final boolean removeChild(final INode child) {
assertChildrenNonNull();
Preconditions.checkNotNull(children);
final int i = searchChildren(child.getLocalNameBytes());
if (i < 0) {
return false;
@ -165,7 +161,7 @@ public class INodeDirectory extends INode {
* {@link INodeDirectoryWithSnapshot} depending on the latest snapshot.
*/
INodeDirectoryWithQuota replaceSelf4Quota(final Snapshot latest,
final long nsQuota, final long dsQuota) {
final long nsQuota, final long dsQuota) throws NSQuotaExceededException {
Preconditions.checkState(!(this instanceof INodeDirectoryWithQuota),
"this is already an INodeDirectoryWithQuota, this=%s", this);
@ -176,13 +172,13 @@ public class INodeDirectory extends INode {
return q;
} else {
final INodeDirectoryWithSnapshot s = new INodeDirectoryWithSnapshot(this);
s.setQuota(nsQuota, dsQuota, null);
s.setQuota(nsQuota, dsQuota);
return replaceSelf(s).saveSelf2Snapshot(latest, this);
}
}
/** Replace itself with an {@link INodeDirectorySnapshottable}. */
public INodeDirectorySnapshottable replaceSelf4INodeDirectorySnapshottable(
Snapshot latest) {
Snapshot latest) throws NSQuotaExceededException {
Preconditions.checkState(!(this instanceof INodeDirectorySnapshottable),
"this is already an INodeDirectorySnapshottable, this=%s", this);
final INodeDirectorySnapshottable s = new INodeDirectorySnapshottable(this);
@ -213,7 +209,7 @@ public class INodeDirectory extends INode {
}
public void replaceChild(final INode oldChild, final INode newChild) {
assertChildrenNonNull();
Preconditions.checkNotNull(children);
final int i = searchChildrenForExistingINode(newChild);
final INode removed = children.set(i, newChild);
Preconditions.checkState(removed == oldChild);
@ -247,7 +243,8 @@ public class INodeDirectory extends INode {
}
@Override
public INodeDirectory recordModification(Snapshot latest) {
public INodeDirectory recordModification(Snapshot latest)
throws NSQuotaExceededException {
return isInLatestSnapshot(latest)?
replaceSelf4INodeDirectoryWithSnapshot().recordModification(latest)
: this;
@ -259,7 +256,7 @@ public class INodeDirectory extends INode {
* @return the child inode, which may be replaced.
*/
public INode saveChild2Snapshot(final INode child, final Snapshot latest,
final INode snapshotCopy) {
final INode snapshotCopy) throws NSQuotaExceededException {
if (latest == null) {
return child;
}
@ -387,11 +384,11 @@ public class INodeDirectory extends INode {
if (index >= 0) {
existing.addNode(curNode);
}
if (curNode instanceof INodeDirectorySnapshottable) {
if (curNode instanceof INodeDirectoryWithSnapshot) {
//if the path is a non-snapshot path, update the latest snapshot.
if (!existing.isSnapshot()) {
existing.updateLatestSnapshot(
((INodeDirectorySnapshottable)curNode).getLastSnapshot());
((INodeDirectoryWithSnapshot)curNode).getLastSnapshot());
}
}
if (curNode.isSymlink() && (!lastComp || (lastComp && resolveLink))) {
@ -488,64 +485,64 @@ public class INodeDirectory extends INode {
* otherwise, return true;
*/
public boolean addChild(INode node, final boolean setModTime,
final Snapshot latest) {
if (isInLatestSnapshot(latest)) {
return replaceSelf4INodeDirectoryWithSnapshot()
.addChild(node, setModTime, latest);
}
if (children == null) {
children = new ArrayList<INode>(DEFAULT_FILES_PER_DIRECTORY);
}
final Snapshot latest) throws NSQuotaExceededException {
final int low = searchChildren(node.getLocalNameBytes());
if (low >= 0) {
return false;
}
node.parent = this;
children.add(-low - 1, node);
// update modification time of the parent directory
if (setModTime) {
updateModificationTime(node.getModificationTime(), latest);
if (isInLatestSnapshot(latest)) {
return replaceSelf4INodeDirectoryWithSnapshot()
.addChild(node, setModTime, latest);
}
if (node.getGroupName() == null) {
node.setGroup(getGroupName(), null);
addChild(node, low);
if (setModTime) {
// update modification time of the parent directory
updateModificationTime(node.getModificationTime(), latest);
}
return true;
}
/**
* Add new INode to the file tree.
* Find the parent and insert
*
* @param path file path
* @param newNode INode to be added
* @return false if the node already exists; otherwise, return true;
* @throws FileNotFoundException if parent does not exist or
* @throws UnresolvedLinkException if any path component is a symbolic link
* is not a directory.
*/
boolean addINode(String path, INode newNode
) throws FileNotFoundException, PathIsNotDirectoryException,
UnresolvedLinkException {
byte[][] pathComponents = getPathComponents(path);
if (pathComponents.length < 2) { // add root
/** The same as addChild(node, false, null, false) */
public boolean addChild(INode node) {
final int low = searchChildren(node.getLocalNameBytes());
if (low >= 0) {
return false;
}
newNode.setLocalName(pathComponents[pathComponents.length - 1]);
// insert into the parent children list
final INodesInPath iip = getExistingPathINodes(pathComponents, 2, false);
final INodeDirectory parent = INodeDirectory.valueOf(iip.getINode(0),
pathComponents);
return parent.addChild(newNode, true, iip.getLatestSnapshot());
addChild(node, low);
return true;
}
/**
* Add the node to the children list at the given insertion point.
* The basic add method which actually calls children.add(..).
*/
private void addChild(final INode node, final int insertionPoint) {
if (children == null) {
children = new ArrayList<INode>(DEFAULT_FILES_PER_DIRECTORY);
}
node.parent = this;
children.add(-insertionPoint - 1, node);
if (node.getGroupName() == null) {
node.setGroup(getGroupName());
}
}
@Override
Quota.Counts computeQuotaUsage(Quota.Counts counts) {
public Quota.Counts computeQuotaUsage(Quota.Counts counts, boolean useCache) {
if (children != null) {
for (INode child : children) {
child.computeQuotaUsage(counts);
child.computeQuotaUsage(counts, useCache);
}
}
return computeQuotaUsage4CurrentDirectory(counts);
}
/** Add quota usage for this inode excluding children. */
public Quota.Counts computeQuotaUsage4CurrentDirectory(Quota.Counts counts) {
counts.add(Quota.NAMESPACE, 1);
return counts;
}
@ -582,15 +579,15 @@ public class INodeDirectory extends INode {
: ReadOnlyList.Util.asReadOnlyList(children);
}
/** Set the children list. */
public void setChildren(List<INode> children) {
this.children = children;
/** Set the children list to null. */
public void clearChildren() {
this.children = null;
}
@Override
public void clearReferences() {
super.clearReferences();
setChildren(null);
clearChildren();
}
/**
@ -598,7 +595,8 @@ public class INodeDirectory extends INode {
* recursively down the subtree.
*/
public int cleanSubtreeRecursively(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
int total = 0;
// in case of deletion snapshot, since this call happens after we modify
// the diff list, the snapshot to be deleted has been combined or renamed
@ -624,7 +622,8 @@ public class INodeDirectory extends INode {
@Override
public int cleanSubtree(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
int total = 0;
if (prior == null && snapshot == null) {
// destroy the whole subtree and collect blocks that should be deleted
@ -870,7 +869,7 @@ public class INodeDirectory extends INode {
super.dumpTreeRecursively(out, prefix, snapshot);
out.print(", childrenSize=" + getChildrenList(snapshot).size());
if (this instanceof INodeDirectoryWithQuota) {
// out.print(((INodeDirectoryWithQuota)this).quotaString());
out.print(((INodeDirectoryWithQuota)this).quotaString());
}
out.println();

View File

@ -22,7 +22,6 @@ import org.apache.hadoop.hdfs.protocol.DSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.QuotaExceededException;
import org.apache.hadoop.hdfs.server.namenode.snapshot.Snapshot;
/**
* Directory INode class that has a quota restriction
@ -87,18 +86,21 @@ public class INodeDirectoryWithQuota extends INodeDirectory {
* @param nsQuota Namespace quota to be set
* @param dsQuota diskspace quota to be set
*/
public void setQuota(long nsQuota, long dsQuota, Snapshot latest) {
final INodeDirectoryWithQuota nodeToUpdate
= (INodeDirectoryWithQuota)recordModification(latest);
nodeToUpdate.nsQuota = nsQuota;
nodeToUpdate.dsQuota = dsQuota;
public void setQuota(long nsQuota, long dsQuota) {
this.nsQuota = nsQuota;
this.dsQuota = dsQuota;
}
@Override
public final Quota.Counts computeQuotaUsage(Quota.Counts counts) {
// use cache value
counts.add(Quota.NAMESPACE, namespace);
counts.add(Quota.DISKSPACE, diskspace);
public final Quota.Counts computeQuotaUsage(Quota.Counts counts,
boolean useCache) {
if (useCache) {
// use cache value
counts.add(Quota.NAMESPACE, namespace);
counts.add(Quota.DISKSPACE, diskspace);
} else {
super.computeQuotaUsage(counts, false);
}
return counts;
}
@ -121,9 +123,9 @@ public class INodeDirectoryWithQuota extends INodeDirectory {
}
private void checkDiskspace(final long computed) {
if (-1 != getDsQuota() && diskspaceConsumed() != computed) {
if (-1 != getDsQuota() && diskspace != computed) {
NameNode.LOG.error("BUG: Inconsistent diskspace for directory "
+ getFullPathName() + ". Cached = " + diskspaceConsumed()
+ getFullPathName() + ". Cached = " + diskspace
+ " != Computed = " + computed);
}
}
@ -135,10 +137,25 @@ public class INodeDirectoryWithQuota extends INodeDirectory {
return namespace;
}
long diskspaceConsumed() {
return diskspace;
@Override
public final void addNamespaceConsumed(final int delta)
throws NSQuotaExceededException {
if (isQuotaSet()) {
// The following steps are important:
// check quotas in this inode and all ancestors before changing counts
// so that no change is made if there is any quota violation.
// (1) verify quota in this inode
verifyNamespaceQuota(delta);
// (2) verify quota and then add count in ancestors
super.addNamespaceConsumed(delta);
// (3) add count in this inode
namespace += delta;
} else {
super.addNamespaceConsumed(delta);
}
}
/** Update the size of the tree
*
* @param nsDelta the change of the tree size
@ -162,13 +179,19 @@ public class INodeDirectoryWithQuota extends INodeDirectory {
this.diskspace = diskspace;
}
/** Verify if the namespace quota is violated after applying delta. */
void verifyNamespaceQuota(long delta) throws NSQuotaExceededException {
if (Quota.isViolated(nsQuota, namespace, delta)) {
throw new NSQuotaExceededException(nsQuota, namespace + delta);
}
}
/** Verify if the namespace count disk space satisfies the quota restriction
* @throws QuotaExceededException if the given quota is less than the count
*/
void verifyQuota(long nsDelta, long dsDelta) throws QuotaExceededException {
if (Quota.isViolated(nsQuota, namespace, nsDelta)) {
throw new NSQuotaExceededException(nsQuota, namespace + nsDelta);
}
verifyNamespaceQuota(nsDelta);
if (Quota.isViolated(dsQuota, diskspace, dsDelta)) {
throw new DSQuotaExceededException(dsQuota, diskspace + dsDelta);
}

View File

@ -25,6 +25,7 @@ import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.fs.permission.FsAction;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.fs.permission.PermissionStatus;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockCollection;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction;
@ -132,7 +133,8 @@ public class INodeFile extends INode implements BlockCollection {
}
@Override
public INodeFile recordModification(final Snapshot latest) {
public INodeFile recordModification(final Snapshot latest)
throws NSQuotaExceededException {
return isInLatestSnapshot(latest)?
parent.replaceChild4INodeFileWithSnapshot(this)
.recordModification(latest)
@ -145,7 +147,18 @@ public class INodeFile extends INode implements BlockCollection {
* the {@link FsAction#EXECUTE} action, if any, is ignored.
*/
@Override
final INode setPermission(FsPermission permission, Snapshot latest) {
final void setPermission(FsPermission permission) {
super.setPermission(permission.applyUMask(UMASK));
}
/**
* Set the {@link FsPermission} of this {@link INodeFile}.
* Since this is a file,
* the {@link FsAction#EXECUTE} action, if any, is ignored.
*/
@Override
final INode setPermission(FsPermission permission, Snapshot latest)
throws NSQuotaExceededException {
return super.setPermission(permission.applyUMask(UMASK), latest);
}
@ -170,15 +183,19 @@ public class INodeFile extends INode implements BlockCollection {
: getFileReplication(null);
}
public void setFileReplication(short replication, Snapshot latest) {
final INodeFile nodeToUpdate = recordModification(latest);
if (nodeToUpdate != this) {
nodeToUpdate.setFileReplication(replication, null);
return;
}
/** Set the replication factor of this file. */
public final void setFileReplication(short replication) {
header = HeaderFormat.combineReplication(header, replication);
}
/** Set the replication factor of this file. */
public final INodeFile setFileReplication(short replication, Snapshot latest)
throws NSQuotaExceededException {
final INodeFile nodeToUpdate = recordModification(latest);
nodeToUpdate.setFileReplication(replication);
return nodeToUpdate;
}
/** @return preferred block size (in bytes) of the file. */
@Override
public long getPreferredBlockSize() {
@ -253,7 +270,8 @@ public class INodeFile extends INode implements BlockCollection {
@Override
public int cleanSubtree(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
if (snapshot == null && prior == null) {
// this only happens when deleting the current file
return destroyAndCollectBlocks(collectedBlocks);
@ -282,7 +300,8 @@ public class INodeFile extends INode implements BlockCollection {
}
@Override
Quota.Counts computeQuotaUsage(Quota.Counts counts) {
public final Quota.Counts computeQuotaUsage(Quota.Counts counts,
boolean useCache) {
counts.add(Quota.NAMESPACE, this instanceof FileWithSnapshot?
((FileWithSnapshot)this).getDiffs().asList().size() + 1: 1);
counts.add(Quota.DISKSPACE, diskspaceConsumed());
@ -397,7 +416,7 @@ public class INodeFile extends INode implements BlockCollection {
return size;
}
long diskspaceConsumed() {
final long diskspaceConsumed() {
// use preferred block size for the last block if it is under construction
return computeFileSize(true, true) * getBlockReplication();
}

View File

@ -24,6 +24,7 @@ import java.util.Arrays;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.fs.permission.PermissionStatus;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction;
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor;
@ -131,7 +132,8 @@ public class INodeFileUnderConstruction extends INodeFile implements MutableBloc
}
@Override
public INodeFileUnderConstruction recordModification(final Snapshot latest) {
public INodeFileUnderConstruction recordModification(final Snapshot latest)
throws NSQuotaExceededException {
return isInLatestSnapshot(latest)?
parent.replaceChild4INodeFileUcWithSnapshot(this)
.recordModification(latest)

View File

@ -22,6 +22,7 @@ import java.io.PrintWriter;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.fs.permission.PermissionStatus;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.namenode.INode.Content.CountsMap.Key;
import org.apache.hadoop.hdfs.server.namenode.snapshot.Snapshot;
@ -44,7 +45,7 @@ public class INodeSymlink extends INode {
}
@Override
INode recordModification(Snapshot latest) {
INode recordModification(Snapshot latest) throws NSQuotaExceededException {
return isInLatestSnapshot(latest)?
parent.saveChild2Snapshot(this, latest, new INodeSymlink(this))
: this;
@ -77,7 +78,8 @@ public class INodeSymlink extends INode {
}
@Override
Quota.Counts computeQuotaUsage(final Quota.Counts counts) {
public Quota.Counts computeQuotaUsage(Quota.Counts counts,
boolean updateCache) {
counts.add(Quota.NAMESPACE, 1);
return counts;
}

View File

@ -22,6 +22,7 @@ import java.util.Collections;
import java.util.Iterator;
import java.util.List;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.namenode.INode;
import org.apache.hadoop.hdfs.server.namenode.INode.BlocksMapUpdateInfo;
@ -48,9 +49,11 @@ abstract class AbstractINodeDiffList<N extends INode,
return Collections.unmodifiableList(diffs);
}
/** clear the diff list */
void clear() {
/** clear the diff list, */
int clear() {
final int n = diffs.size();
diffs.clear();
return n;
}
/**
@ -102,7 +105,9 @@ abstract class AbstractINodeDiffList<N extends INode,
}
/** Add an {@link AbstractINodeDiff} for the given snapshot. */
final D addDiff(Snapshot latest, N currentINode) {
final D addDiff(Snapshot latest, N currentINode)
throws NSQuotaExceededException {
currentINode.addNamespaceConsumed(1);
return addLast(factory.createDiff(latest, currentINode));
}
@ -220,14 +225,25 @@ abstract class AbstractINodeDiffList<N extends INode,
* Check if the latest snapshot diff exists. If not, add it.
* @return the latest snapshot diff, which is never null.
*/
final D checkAndAddLatestSnapshotDiff(Snapshot latest, N currentINode) {
final D checkAndAddLatestSnapshotDiff(Snapshot latest, N currentINode)
throws NSQuotaExceededException {
final D last = getLast();
return last != null && last.snapshot.equals(latest)? last
: addDiff(latest, currentINode);
if (last != null
&& Snapshot.ID_COMPARATOR.compare(last.getSnapshot(), latest) >= 0) {
return last;
} else {
try {
return addDiff(latest, currentINode);
} catch(NSQuotaExceededException e) {
e.setMessagePrefix("Failed to record modification for snapshot");
throw e;
}
}
}
/** Save the snapshot copy to the latest snapshot. */
public void saveSelf2Snapshot(Snapshot latest, N currentINode, N snapshotCopy) {
public void saveSelf2Snapshot(Snapshot latest, N currentINode, N snapshotCopy)
throws NSQuotaExceededException {
if (latest != null) {
checkAndAddLatestSnapshotDiff(latest, currentINode).saveSnapshotCopy(
snapshotCopy, factory, currentINode);

View File

@ -30,6 +30,7 @@ import java.util.Map;
import org.apache.hadoop.HadoopIllegalArgumentException;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.SnapshotDiffReport;
import org.apache.hadoop.hdfs.protocol.SnapshotDiffReport.DiffReportEntry;
import org.apache.hadoop.hdfs.protocol.SnapshotDiffReport.DiffType;
@ -270,7 +271,8 @@ public class INodeDirectorySnapshottable extends INodeDirectoryWithSnapshot {
}
/** Add a snapshot. */
Snapshot addSnapshot(int id, String name) throws SnapshotException {
Snapshot addSnapshot(int id, String name)
throws SnapshotException, NSQuotaExceededException {
//check snapshot quota
final int n = getNumSnapshots();
if (n + 1 > snapshotQuota) {
@ -315,7 +317,11 @@ public class INodeDirectorySnapshottable extends INodeDirectoryWithSnapshot {
} else {
final Snapshot snapshot = snapshotsByNames.remove(i);
Snapshot prior = Snapshot.findLatestSnapshot(this, snapshot);
cleanSubtree(snapshot, prior, collectedBlocks);
try {
cleanSubtree(snapshot, prior, collectedBlocks);
} catch(NSQuotaExceededException e) {
LOG.error("BUG: removeSnapshot increases namespace usage.", e);
}
return snapshot;
}
}
@ -427,7 +433,7 @@ public class INodeDirectorySnapshottable extends INodeDirectoryWithSnapshot {
* Replace itself with {@link INodeDirectoryWithSnapshot} or
* {@link INodeDirectory} depending on the latest snapshot.
*/
void replaceSelf(final Snapshot latest) {
void replaceSelf(final Snapshot latest) throws NSQuotaExceededException {
if (latest == null) {
Preconditions.checkState(getLastSnapshot() == null,
"latest == null but getLastSnapshot() != null, this=%s", this);

View File

@ -25,10 +25,12 @@ import java.util.Iterator;
import java.util.List;
import java.util.Map;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.protocol.SnapshotDiffReport.DiffReportEntry;
import org.apache.hadoop.hdfs.protocol.SnapshotDiffReport.DiffType;
import org.apache.hadoop.hdfs.server.namenode.FSImageSerialization;
import org.apache.hadoop.hdfs.server.namenode.INode;
import org.apache.hadoop.hdfs.server.namenode.Quota;
import org.apache.hadoop.hdfs.server.namenode.INode.Content.CountsMap.Key;
import org.apache.hadoop.hdfs.server.namenode.INodeDirectory;
import org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota;
@ -79,7 +81,7 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
for (INode c : createdList) {
removedNum += c.destroyAndCollectBlocks(collectedBlocks);
// if c is also contained in the children list, remove it
currentINode.removeChild(c, null);
currentINode.removeChild(c);
}
createdList.clear();
return removedNum;
@ -353,7 +355,7 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
new INodeDirectoryWithQuota(currentDir, false,
currentDir.getNsQuota(), currentDir.getDsQuota())
: new INodeDirectory(currentDir, false);
copy.setChildren(null);
copy.clearChildren();
return copy;
}
}
@ -462,21 +464,23 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
}
@Override
public INodeDirectoryWithSnapshot recordModification(final Snapshot latest) {
public INodeDirectoryWithSnapshot recordModification(final Snapshot latest)
throws NSQuotaExceededException {
return isInLatestSnapshot(latest)?
saveSelf2Snapshot(latest, null): this;
}
/** Save the snapshot copy to the latest snapshot. */
public INodeDirectoryWithSnapshot saveSelf2Snapshot(
final Snapshot latest, final INodeDirectory snapshotCopy) {
final Snapshot latest, final INodeDirectory snapshotCopy)
throws NSQuotaExceededException {
diffs.saveSelf2Snapshot(latest, this, snapshotCopy);
return this;
}
@Override
public INode saveChild2Snapshot(final INode child, final Snapshot latest,
final INode snapshotCopy) {
final INode snapshotCopy) throws NSQuotaExceededException {
Preconditions.checkArgument(!child.isDirectory(),
"child is a directory, child=%s", child);
if (latest == null) {
@ -494,7 +498,8 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
}
@Override
public boolean addChild(INode inode, boolean setModTime, Snapshot latest) {
public boolean addChild(INode inode, boolean setModTime, Snapshot latest)
throws NSQuotaExceededException {
ChildrenDiff diff = null;
Integer undoInfo = null;
if (latest != null) {
@ -509,7 +514,8 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
}
@Override
public boolean removeChild(INode child, Snapshot latest) {
public boolean removeChild(INode child, Snapshot latest)
throws NSQuotaExceededException {
ChildrenDiff diff = null;
UndoInfo<INode> undoInfo = null;
if (latest != null) {
@ -605,7 +611,8 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
@Override
public int cleanSubtree(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
int n = 0;
if (snapshot == null) { // delete the current directory
recordModification(prior);
@ -637,11 +644,24 @@ public class INodeDirectoryWithSnapshot extends INodeDirectoryWithQuota {
for (DirectoryDiff diff : diffs) {
total += diff.destroyAndCollectBlocks(this, collectedBlocks);
}
diffs.clear();
total += diffs.clear();
total += super.destroyAndCollectBlocks(collectedBlocks);
return total;
}
@Override
public Quota.Counts computeQuotaUsage4CurrentDirectory(Quota.Counts counts) {
super.computeQuotaUsage4CurrentDirectory(counts);
for(DirectoryDiff d : diffs) {
for(INode deleted : d.getChildrenDiff().getDeletedList()) {
deleted.computeQuotaUsage(counts, false);
}
}
counts.add(Quota.NAMESPACE, diffs.asList().size());
return counts;
}
@Override
public Content.CountsMap computeContentSummary(
final Content.CountsMap countsMap) {

View File

@ -18,6 +18,7 @@
package org.apache.hadoop.hdfs.server.namenode.snapshot;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor;
import org.apache.hadoop.hdfs.server.namenode.INodeFile;
import org.apache.hadoop.hdfs.server.namenode.INodeFileUnderConstruction;
@ -75,8 +76,8 @@ public class INodeFileUnderConstructionWithSnapshot
assertAllBlocksComplete();
final long atime = getModificationTime();
final INodeFileWithSnapshot f = new INodeFileWithSnapshot(this, getDiffs());
f.setModificationTime(mtime, null);
f.setAccessTime(atime, null);
f.setModificationTime(mtime);
f.setAccessTime(atime);
return f;
}
@ -92,7 +93,7 @@ public class INodeFileUnderConstructionWithSnapshot
@Override
public INodeFileUnderConstructionWithSnapshot recordModification(
final Snapshot latest) {
final Snapshot latest) throws NSQuotaExceededException {
if (isInLatestSnapshot(latest)) {
diffs.saveSelf2Snapshot(latest, this, null);
}
@ -111,7 +112,8 @@ public class INodeFileUnderConstructionWithSnapshot
@Override
public int cleanSubtree(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
if (snapshot == null) { // delete the current file
recordModification(prior);
isCurrentFileDeleted = true;

View File

@ -18,6 +18,7 @@
package org.apache.hadoop.hdfs.server.namenode.snapshot;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor;
import org.apache.hadoop.hdfs.server.namenode.INodeFile;
@ -64,7 +65,8 @@ public class INodeFileWithSnapshot extends INodeFile
}
@Override
public INodeFileWithSnapshot recordModification(final Snapshot latest) {
public INodeFileWithSnapshot recordModification(final Snapshot latest)
throws NSQuotaExceededException {
if (isInLatestSnapshot(latest)) {
diffs.saveSelf2Snapshot(latest, this, null);
}
@ -83,7 +85,8 @@ public class INodeFileWithSnapshot extends INodeFile
@Override
public int cleanSubtree(final Snapshot snapshot, Snapshot prior,
final BlocksMapUpdateInfo collectedBlocks) {
final BlocksMapUpdateInfo collectedBlocks)
throws NSQuotaExceededException {
if (snapshot == null) { // delete the current file
recordModification(prior);
isCurrentFileDeleted = true;
@ -98,8 +101,8 @@ public class INodeFileWithSnapshot extends INodeFile
public int destroyAndCollectBlocks(
final BlocksMapUpdateInfo collectedBlocks) {
Preconditions.checkState(this.isCurrentFileDeleted);
diffs.clear();
return super.destroyAndCollectBlocks(collectedBlocks);
final int n = diffs.clear();
return n + super.destroyAndCollectBlocks(collectedBlocks);
}
@Override

View File

@ -129,6 +129,8 @@ public class SnapshotManager implements SnapshotStats {
final INodesInPath i = fsdir.getINodesInPath4Write(path);
final INodeDirectorySnapshottable srcRoot
= INodeDirectorySnapshottable.valueOf(i.getLastINode(), path);
fsdir.verifyMaxComponentLength(snapshotName, path, 0);
srcRoot.addSnapshot(snapshotCounter, snapshotName);
//create success, update id

View File

@ -18,9 +18,7 @@
package org.apache.hadoop.hdfs.server.namenode;
import java.io.File;
import java.io.FileWriter;
import java.io.IOException;
import java.io.PrintWriter;
import java.util.ArrayList;
import java.util.EnumSet;
import java.util.List;
@ -59,7 +57,6 @@ public class TestFSImageWithSnapshot {
static final int BLOCKSIZE = 1024;
static final long txid = 1;
private final Path rootDir = new Path("/");
private final Path dir = new Path("/TestSnapshot");
private static String testDir =
System.getProperty("test.build.data", "build/test/data");
@ -114,10 +111,7 @@ public class TestFSImageWithSnapshot {
*/
private File dumpTree2File(String fileSuffix) throws IOException {
File file = getDumpTreeFile(testDir, fileSuffix);
PrintWriter out = new PrintWriter(new FileWriter(file, false), true);
fsn.getFSDirectory().getINode(rootDir.toString())
.dumpTreeRecursively(out, new StringBuilder(), null);
out.close();
SnapshotTestHelper.dumpTree2File(fsn.getFSDirectory(), file);
return file;
}
@ -155,6 +149,8 @@ public class TestFSImageWithSnapshot {
fsn.getFSDirectory().writeLock();
try {
loader.load(imageFile);
FSImage.updateCountForQuota(
(INodeDirectoryWithQuota)fsn.getFSDirectory().getINode("/"));
} finally {
fsn.getFSDirectory().writeUnlock();
fsn.writeUnlock();

View File

@ -160,7 +160,7 @@ public class TestFsLimits {
Class<?> generated = null;
try {
fs.verifyFsLimits(inodes, 1, child);
rootInode.addChild(child, false, null);
rootInode.addChild(child);
} catch (QuotaExceededException e) {
generated = e.getClass();
}

View File

@ -149,11 +149,11 @@ public class TestINodeFile {
assertEquals("f", inf.getFullPathName());
assertEquals("", inf.getLocalParentDir());
dir.addChild(inf, false, null);
dir.addChild(inf);
assertEquals("d"+Path.SEPARATOR+"f", inf.getFullPathName());
assertEquals("d", inf.getLocalParentDir());
root.addChild(dir, false, null);
root.addChild(dir);
assertEquals(Path.SEPARATOR+"d"+Path.SEPARATOR+"f", inf.getFullPathName());
assertEquals(Path.SEPARATOR+"d", dir.getFullPathName());

View File

@ -23,6 +23,7 @@ import static org.junit.Assert.assertTrue;
import java.io.BufferedReader;
import java.io.File;
import java.io.FileReader;
import java.io.FileWriter;
import java.io.IOException;
import java.io.PrintWriter;
import java.util.ArrayList;
@ -48,6 +49,7 @@ import org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceStorage;
import org.apache.hadoop.hdfs.server.datanode.DataBlockScanner;
import org.apache.hadoop.hdfs.server.datanode.DataNode;
import org.apache.hadoop.hdfs.server.datanode.DirectoryScanner;
import org.apache.hadoop.hdfs.server.namenode.FSDirectory;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.INode;
import org.apache.hadoop.hdfs.server.namenode.INodeDirectory;
@ -128,6 +130,8 @@ public class SnapshotTestHelper {
assertTrue(hdfs.exists(snapshotRoot));
hdfs.allowSnapshot(snapshotRoot.toString());
hdfs.createSnapshot(snapshotRoot, snapshotName);
// set quota to a large value for testing counts
hdfs.setQuota(snapshotRoot, Long.MAX_VALUE-1, Long.MAX_VALUE-1);
return SnapshotTestHelper.getSnapshotRoot(snapshotRoot, snapshotName);
}
@ -192,6 +196,11 @@ public class SnapshotTestHelper {
}
private static void compareDumpedTreeInFile(File file1, File file2,
boolean print) throws IOException {
if (print) {
printFile(file1);
printFile(file2);
}
BufferedReader reader1 = new BufferedReader(new FileReader(file1));
BufferedReader reader2 = new BufferedReader(new FileReader(file2));
try {
@ -238,6 +247,25 @@ public class SnapshotTestHelper {
}
}
static void printFile(File f) throws IOException {
System.out.println();
System.out.println("File: " + f);
BufferedReader in = new BufferedReader(new FileReader(f));
try {
for(String line; (line = in.readLine()) != null; ) {
System.out.println(line);
}
} finally {
in.close();
}
}
public static void dumpTree2File(FSDirectory fsdir, File f) throws IOException{
final PrintWriter out = new PrintWriter(new FileWriter(f, false), true);
fsdir.getINode("/").dumpTreeRecursively(out, new StringBuilder(), null);
out.close();
}
/**
* Generate the path for a snapshot file.
*

View File

@ -31,8 +31,11 @@ import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.DFSUtil;
import org.apache.hadoop.hdfs.DistributedFileSystem;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.INodeDirectory;
import org.apache.hadoop.ipc.RemoteException;
import org.junit.AfterClass;
import org.junit.Assert;
import org.junit.BeforeClass;
@ -165,6 +168,65 @@ public class TestNestedSnapshots {
}
}
@Test
public void testSnapshotWithQuota() throws Exception {
final String dirStr = "/testSnapshotWithQuota/dir";
final Path dir = new Path(dirStr);
hdfs.mkdirs(dir, new FsPermission((short)0777));
hdfs.allowSnapshot(dirStr);
// set namespace quota
final int NS_QUOTA = 6;
hdfs.setQuota(dir, NS_QUOTA, HdfsConstants.QUOTA_DONT_SET);
// create object to use up the quota.
final Path foo = new Path(dir, "foo");
final Path f1 = new Path(foo, "f1");
DFSTestUtil.createFile(hdfs, f1, BLOCKSIZE, REPLICATION, SEED);
hdfs.createSnapshot(dir, "s0");
final Path f2 = new Path(foo, "f2");
DFSTestUtil.createFile(hdfs, f2, BLOCKSIZE, REPLICATION, SEED);
try {
// normal create file should fail with quota
final Path f3 = new Path(foo, "f3");
DFSTestUtil.createFile(hdfs, f3, BLOCKSIZE, REPLICATION, SEED);
Assert.fail();
} catch(NSQuotaExceededException e) {
SnapshotTestHelper.LOG.info("The exception is expected.", e);
}
try {
// createSnapshot should fail with quota
hdfs.createSnapshot(dir, "s1");
Assert.fail();
} catch(RemoteException re) {
final IOException ioe = re.unwrapRemoteException();
if (ioe instanceof NSQuotaExceededException) {
SnapshotTestHelper.LOG.info("The exception is expected.", ioe);
}
}
try {
// setPermission f1 should fail with quote since it cannot add diff.
hdfs.setPermission(f1, new FsPermission((short)0));
Assert.fail();
} catch(RemoteException re) {
final IOException ioe = re.unwrapRemoteException();
if (ioe instanceof NSQuotaExceededException) {
SnapshotTestHelper.LOG.info("The exception is expected.", ioe);
}
}
// setPermission f2 since it was created after the snapshot
hdfs.setPermission(f2, new FsPermission((short)0));
// increase quota and retry the commands.
hdfs.setQuota(dir, NS_QUOTA + 2, HdfsConstants.QUOTA_DONT_SET);
hdfs.createSnapshot(dir, "s1");
hdfs.setPermission(foo, new FsPermission((short)0444));
}
/**
* Test {@link Snapshot#ID_COMPARATOR}.
*/

View File

@ -21,8 +21,6 @@ import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
import java.io.File;
import java.io.FileWriter;
import java.io.PrintWriter;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.EnumSet;
@ -171,7 +169,7 @@ public class TestSnapshot {
private File getDumpTreeFile(String dir, String suffix) {
return new File(dir, String.format("dumptree_%s", suffix));
}
/**
* Restart the cluster to check edit log applying and fsimage saving/loading
*/
@ -180,10 +178,7 @@ public class TestSnapshot {
File fsnMiddle = getDumpTreeFile(testDir, "middle");
File fsnAfter = getDumpTreeFile(testDir, "after");
String rootDir = "/";
PrintWriter out = new PrintWriter(new FileWriter(fsnBefore, false), true);
fsdir.getINode(rootDir).dumpTreeRecursively(out, new StringBuilder(), null);
out.close();
SnapshotTestHelper.dumpTree2File(fsdir, fsnBefore);
cluster.shutdown();
cluster = new MiniDFSCluster.Builder(conf).format(false)
@ -191,11 +186,8 @@ public class TestSnapshot {
cluster.waitActive();
fsn = cluster.getNamesystem();
hdfs = cluster.getFileSystem();
// later check fsnMiddle to see if the edit log is recorded and applied
// correctly
out = new PrintWriter(new FileWriter(fsnMiddle, false), true);
fsdir.getINode(rootDir).dumpTreeRecursively(out, new StringBuilder(), null);
out.close();
// later check fsnMiddle to see if the edit log is applied correctly
SnapshotTestHelper.dumpTree2File(fsdir, fsnMiddle);
// save namespace and restart cluster
hdfs.setSafeMode(SafeModeAction.SAFEMODE_ENTER);
@ -208,9 +200,7 @@ public class TestSnapshot {
fsn = cluster.getNamesystem();
hdfs = cluster.getFileSystem();
// dump the namespace loaded from fsimage
out = new PrintWriter(new FileWriter(fsnAfter, false), true);
fsdir.getINode(rootDir).dumpTreeRecursively(out, new StringBuilder(), null);
out.close();
SnapshotTestHelper.dumpTree2File(fsdir, fsnAfter);
SnapshotTestHelper.compareDumpedTreeInFile(fsnBefore, fsnMiddle);
SnapshotTestHelper.compareDumpedTreeInFile(fsnBefore, fsnAfter);

View File

@ -301,7 +301,7 @@ public class TestDiff {
Assert.assertTrue(i >= 0);
final INodeDirectory oldinode = (INodeDirectory)current.get(i);
final INodeDirectory newinode = new INodeDirectory(oldinode, false);
newinode.updateModificationTime(oldinode.getModificationTime() + 1, null);
newinode.setModificationTime(oldinode.getModificationTime() + 1);
current.set(i, newinode);
if (diff != null) {