HDFS-15313. Ensure inodes in active filesytem are not deleted during snapshot delete. Contributed by Shashikant Banerjee.
This commit is contained in:
parent
b7b9bd32db
commit
1a11c4bc71
|
@ -17,12 +17,8 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.namenode;
|
package org.apache.hadoop.hdfs.server.namenode;
|
||||||
|
|
||||||
import java.io.PrintStream;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
import java.io.PrintWriter;
|
import com.google.common.base.Preconditions;
|
||||||
import java.io.StringWriter;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Map;
|
|
||||||
|
|
||||||
import com.google.common.collect.ImmutableMap;
|
import com.google.common.collect.ImmutableMap;
|
||||||
import com.google.common.collect.Maps;
|
import com.google.common.collect.Maps;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
|
@ -32,12 +28,12 @@ import org.apache.hadoop.fs.ContentSummary;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.fs.permission.PermissionStatus;
|
import org.apache.hadoop.fs.permission.PermissionStatus;
|
||||||
|
import org.apache.hadoop.hdfs.DFSUtil;
|
||||||
import org.apache.hadoop.hdfs.DFSUtilClient;
|
import org.apache.hadoop.hdfs.DFSUtilClient;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
||||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockUnderConstructionFeature;
|
import org.apache.hadoop.hdfs.server.blockmanagement.BlockUnderConstructionFeature;
|
||||||
import org.apache.hadoop.hdfs.DFSUtil;
|
|
||||||
import org.apache.hadoop.hdfs.server.namenode.INodeReference.DstReference;
|
import org.apache.hadoop.hdfs.server.namenode.INodeReference.DstReference;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.INodeReference.WithCount;
|
import org.apache.hadoop.hdfs.server.namenode.INodeReference.WithCount;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.INodeReference.WithName;
|
import org.apache.hadoop.hdfs.server.namenode.INodeReference.WithName;
|
||||||
|
@ -47,8 +43,11 @@ import org.apache.hadoop.security.AccessControlException;
|
||||||
import org.apache.hadoop.util.ChunkedArrayList;
|
import org.apache.hadoop.util.ChunkedArrayList;
|
||||||
import org.apache.hadoop.util.StringUtils;
|
import org.apache.hadoop.util.StringUtils;
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import java.io.PrintStream;
|
||||||
import com.google.common.base.Preconditions;
|
import java.io.PrintWriter;
|
||||||
|
import java.io.StringWriter;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* We keep an in-memory representation of the file/block hierarchy.
|
* We keep an in-memory representation of the file/block hierarchy.
|
||||||
|
@ -225,6 +224,27 @@ public abstract class INode implements INodeAttributes, Diff.Element<byte[]> {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/** Is this inode in the current state? */
|
||||||
|
public boolean isInCurrentState() {
|
||||||
|
if (isRoot()) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
final INodeDirectory parentDir = getParent();
|
||||||
|
if (parentDir == null) {
|
||||||
|
return false; // this inode is only referenced in snapshots
|
||||||
|
}
|
||||||
|
if (!parentDir.isInCurrentState()) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
final INode child = parentDir.getChild(getLocalNameBytes(),
|
||||||
|
Snapshot.CURRENT_STATE_ID);
|
||||||
|
if (this == child) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
return child != null && child.isReference() &&
|
||||||
|
this.equals(child.asReference().getReferredINode());
|
||||||
|
}
|
||||||
|
|
||||||
/** Is this inode in the latest snapshot? */
|
/** Is this inode in the latest snapshot? */
|
||||||
public final boolean isInLatestSnapshot(final int latestSnapshotId) {
|
public final boolean isInLatestSnapshot(final int latestSnapshotId) {
|
||||||
if (latestSnapshotId == Snapshot.CURRENT_STATE_ID ||
|
if (latestSnapshotId == Snapshot.CURRENT_STATE_ID ||
|
||||||
|
@ -234,6 +254,8 @@ public abstract class INode implements INodeAttributes, Diff.Element<byte[]> {
|
||||||
// if parent is a reference node, parent must be a renamed node. We can
|
// if parent is a reference node, parent must be a renamed node. We can
|
||||||
// stop the check at the reference node.
|
// stop the check at the reference node.
|
||||||
if (parent != null && parent.isReference()) {
|
if (parent != null && parent.isReference()) {
|
||||||
|
// TODO: Is it a bug to return true?
|
||||||
|
// Some ancestor nodes may not be in the latest snapshot.
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
final INodeDirectory parentDir = getParent();
|
final INodeDirectory parentDir = getParent();
|
||||||
|
|
|
@ -739,19 +739,22 @@ public class DirectoryWithSnapshotFeature implements INode.Feature {
|
||||||
// were created before "prior" will be covered by the later
|
// were created before "prior" will be covered by the later
|
||||||
// cleanSubtreeRecursively call.
|
// cleanSubtreeRecursively call.
|
||||||
if (priorCreated != null) {
|
if (priorCreated != null) {
|
||||||
if (currentINode.isLastReference() &&
|
// The nodes in priorCreated must be destroyed if
|
||||||
currentINode.getDiffs().getLastSnapshotId() == prior) {
|
// (1) this is the last reference, and
|
||||||
// If this is the last reference of the directory inode and it
|
// (2) prior is the last snapshot, and
|
||||||
// can not be accessed in any of the subsequent snapshots i.e,
|
// (3) currentINode is not in the current state.
|
||||||
// this is the latest snapshot diff and if this is the last
|
final boolean destroy = currentINode.isLastReference()
|
||||||
// reference, the created list can be
|
&& currentINode.getDiffs().getLastSnapshotId() == prior
|
||||||
// destroyed.
|
&& !currentINode.isInCurrentState();
|
||||||
priorDiff.getChildrenDiff().destroyCreatedList(
|
// we only check the node originally in prior's created list
|
||||||
reclaimContext, currentINode);
|
for (INode cNode : new ArrayList<>(priorDiff.
|
||||||
} else {
|
diff.getCreatedUnmodifiable())) {
|
||||||
// we only check the node originally in prior's created list
|
if (priorCreated.containsKey(cNode)) {
|
||||||
for (INode cNode : priorDiff.diff.getCreatedUnmodifiable()) {
|
if (destroy) {
|
||||||
if (priorCreated.containsKey(cNode)) {
|
cNode.destroyAndCollectBlocks(reclaimContext);
|
||||||
|
currentINode.removeChild(cNode);
|
||||||
|
priorDiff.diff.removeCreated(cNode);
|
||||||
|
} else {
|
||||||
cNode.cleanSubtree(reclaimContext, snapshot, NO_SNAPSHOT_ID);
|
cNode.cleanSubtree(reclaimContext, snapshot, NO_SNAPSHOT_ID);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,13 +17,13 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.util;
|
package org.apache.hadoop.hdfs.util;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
import com.google.common.base.Preconditions;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The difference between the current state and a previous state of a list.
|
* The difference between the current state and a previous state of a list.
|
||||||
*
|
*
|
||||||
|
@ -166,6 +166,17 @@ public class Diff<K, E extends Diff.Element<K>> {
|
||||||
return old;
|
return old;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public boolean removeCreated(final E element) {
|
||||||
|
if (created != null) {
|
||||||
|
final int i = search(created, element.getKey());
|
||||||
|
if (i >= 0 && created.get(i) == element) {
|
||||||
|
created.remove(i);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
public void clearCreated() {
|
public void clearCreated() {
|
||||||
if (created != null) {
|
if (created != null) {
|
||||||
created.clear();
|
created.clear();
|
||||||
|
|
|
@ -609,4 +609,44 @@ public class TestFSImageWithSnapshot {
|
||||||
output.println(b);
|
output.println(b);
|
||||||
return b;
|
return b;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test (timeout=60000)
|
||||||
|
public void testFSImageWithDoubleRename() throws Exception {
|
||||||
|
final Path dir1 = new Path("/dir1");
|
||||||
|
final Path dir2 = new Path("/dir2");
|
||||||
|
hdfs.mkdirs(dir1);
|
||||||
|
hdfs.mkdirs(dir2);
|
||||||
|
Path dira = new Path(dir1, "dira");
|
||||||
|
Path dirx = new Path(dir1, "dirx");
|
||||||
|
Path dirb = new Path(dira, "dirb");
|
||||||
|
hdfs.mkdirs(dira);
|
||||||
|
hdfs.mkdirs(dirb);
|
||||||
|
hdfs.mkdirs(dirx);
|
||||||
|
hdfs.allowSnapshot(dir1);
|
||||||
|
hdfs.createSnapshot(dir1, "s0");
|
||||||
|
Path file1 = new Path(dirb, "file1");
|
||||||
|
DFSTestUtil.createFile(hdfs, file1, BLOCKSIZE, (short) 1, seed);
|
||||||
|
Path rennamePath = new Path(dirx, "dirb");
|
||||||
|
// mv /dir1/dira/dirb to /dir1/dirx/dirb
|
||||||
|
hdfs.rename(dirb, rennamePath);
|
||||||
|
hdfs.createSnapshot(dir1, "s1");
|
||||||
|
DFSTestUtil.appendFile(hdfs, new Path("/dir1/dirx/dirb/file1"),
|
||||||
|
"more data");
|
||||||
|
Path renamePath1 = new Path(dir2, "dira");
|
||||||
|
hdfs.mkdirs(renamePath1);
|
||||||
|
//mv dirx/dirb to /dir2/dira/dirb
|
||||||
|
hdfs.rename(rennamePath, renamePath1);
|
||||||
|
hdfs.delete(renamePath1, true);
|
||||||
|
hdfs.deleteSnapshot(dir1, "s1");
|
||||||
|
// save namespace and restart cluster
|
||||||
|
hdfs.setSafeMode(SafeModeAction.SAFEMODE_ENTER);
|
||||||
|
hdfs.saveNamespace();
|
||||||
|
hdfs.setSafeMode(SafeModeAction.SAFEMODE_LEAVE);
|
||||||
|
cluster.shutdown();
|
||||||
|
cluster = new MiniDFSCluster.Builder(conf).format(false)
|
||||||
|
.numDataNodes(NUM_DATANODES).build();
|
||||||
|
cluster.waitActive();
|
||||||
|
fsn = cluster.getNamesystem();
|
||||||
|
hdfs = cluster.getFileSystem();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -304,7 +304,46 @@ public class TestRenameWithSnapshots {
|
||||||
assertTrue(existsInDiffReport(entries, DiffType.RENAME, sub2.getName(),
|
assertTrue(existsInDiffReport(entries, DiffType.RENAME, sub2.getName(),
|
||||||
sub3.getName()));
|
sub3.getName()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test (timeout=60000)
|
||||||
|
public void testRenameDirectoryAndFileInSnapshot() throws Exception {
|
||||||
|
final Path sub2 = new Path(sub1, "sub2");
|
||||||
|
final Path sub3 = new Path(sub1, "sub3");
|
||||||
|
final Path sub2file1 = new Path(sub2, "file1");
|
||||||
|
final Path sub2file2 = new Path(sub2, "file2");
|
||||||
|
final Path sub3file2 = new Path(sub3, "file2");
|
||||||
|
final Path sub3file3 = new Path(sub3, "file3");
|
||||||
|
final String sub1snap1 = "sub1snap1";
|
||||||
|
final String sub1snap2 = "sub1snap2";
|
||||||
|
final String sub1snap3 = "sub1snap3";
|
||||||
|
final String sub1snap4 = "sub1snap4";
|
||||||
|
hdfs.mkdirs(sub1);
|
||||||
|
hdfs.mkdirs(sub2);
|
||||||
|
DFSTestUtil.createFile(hdfs, sub2file1, BLOCKSIZE, REPL, SEED);
|
||||||
|
SnapshotTestHelper.createSnapshot(hdfs, sub1, sub1snap1);
|
||||||
|
hdfs.rename(sub2file1, sub2file2);
|
||||||
|
SnapshotTestHelper.createSnapshot(hdfs, sub1, sub1snap2);
|
||||||
|
|
||||||
|
// First rename the sub-directory.
|
||||||
|
hdfs.rename(sub2, sub3);
|
||||||
|
SnapshotTestHelper.createSnapshot(hdfs, sub1, sub1snap3);
|
||||||
|
hdfs.rename(sub3file2, sub3file3);
|
||||||
|
SnapshotTestHelper.createSnapshot(hdfs, sub1, sub1snap4);
|
||||||
|
hdfs.deleteSnapshot(sub1, sub1snap1);
|
||||||
|
hdfs.deleteSnapshot(sub1, sub1snap2);
|
||||||
|
hdfs.deleteSnapshot(sub1, sub1snap3);
|
||||||
|
// check the internal details
|
||||||
|
INode sub3file3Inode = fsdir.getINode4Write(sub3file3.toString());
|
||||||
|
INodeReference ref = sub3file3Inode
|
||||||
|
.asReference();
|
||||||
|
INodeReference.WithCount withCount = (WithCount) ref
|
||||||
|
.getReferredINode();
|
||||||
|
Assert.assertEquals(withCount.getReferenceCount(), 1);
|
||||||
|
// Ensure name list is empty for the reference sub3file3Inode
|
||||||
|
Assert.assertNull(withCount.getLastWithName());
|
||||||
|
Assert.assertTrue(sub3file3Inode.isInCurrentState());
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* After the following steps:
|
* After the following steps:
|
||||||
* <pre>
|
* <pre>
|
||||||
|
|
Loading…
Reference in New Issue