HDFS-4119. Complete the allowSnapshot code and add a test for it.

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-2802@1403087 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Tsz-wo Sze 2012-10-28 20:02:54 +00:00
parent 9e26fdcda7
commit f60a844e7a
6 changed files with 123 additions and 10 deletions

View File

@ -35,3 +35,5 @@ Branch-2802 Snapshot (Unreleased)
block list. (szetszwo) block list. (szetszwo)
HDFS-4111. Support snapshot of subtrees. (szetszwo via suresh) HDFS-4111. Support snapshot of subtrees. (szetszwo via suresh)
HDFS-4119. Complete the allowSnapshot code and add a test for it. (szetszwo)

View File

@ -879,15 +879,15 @@ public class FSEditLog implements LogsPurgeable {
logEdit(op); logEdit(op);
} }
void logAllowSnapshot(String snapName, String snapRoot) { void logAllowSnapshot(String path) {
AllowSnapshotOp op = AllowSnapshotOp.getInstance(cache.get()) AllowSnapshotOp op = AllowSnapshotOp.getInstance(cache.get())
.setSnapshotRoot(snapRoot); .setSnapshotRoot(path);
logEdit(op); logEdit(op);
} }
void logDisallowSnapshot(String snapName, String snapRoot) { void logDisallowSnapshot(String path) {
DisallowSnapshotOp op = DisallowSnapshotOp.getInstance(cache.get()) DisallowSnapshotOp op = DisallowSnapshotOp.getInstance(cache.get())
.setSnapshotRoot(snapRoot); .setSnapshotRoot(path);
logEdit(op); logEdit(op);
} }

View File

@ -114,6 +114,11 @@ public abstract class FSEditLogOp {
inst.put(OP_END_LOG_SEGMENT, inst.put(OP_END_LOG_SEGMENT,
new LogSegmentOp(OP_END_LOG_SEGMENT)); new LogSegmentOp(OP_END_LOG_SEGMENT));
inst.put(OP_UPDATE_BLOCKS, new UpdateBlocksOp()); inst.put(OP_UPDATE_BLOCKS, new UpdateBlocksOp());
inst.put(OP_ALLOW_SNAPSHOT, new AllowSnapshotOp());
inst.put(OP_DISALLOW_SNAPSHOT, new DisallowSnapshotOp());
inst.put(OP_CREATE_SNAPSHOT, new CreateSnapshotOp());
inst.put(OP_DELETE_SNAPSHOT, new DeleteSnapshotOp());
} }
public FSEditLogOp get(FSEditLogOpCodes opcode) { public FSEditLogOp get(FSEditLogOpCodes opcode) {

View File

@ -5463,6 +5463,10 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
public BlockManager getBlockManager() { public BlockManager getBlockManager() {
return blockManager; return blockManager;
} }
/** @return the FSDirectory. */
public FSDirectory getFSDirectory() {
return dir;
}
/** /**
* Verifies that the given identifier and password are valid and match. * Verifies that the given identifier and password are valid and match.
@ -5510,11 +5514,25 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
.isAvoidingStaleDataNodesForWrite(); .isAvoidingStaleDataNodesForWrite();
} }
// Allow snapshot on a directroy. /** Allow snapshot on a directroy. */
@VisibleForTesting public void allowSnapshot(String path) throws SafeModeException, IOException {
public void allowSnapshot(String snapshotRoot) writeLock();
throws SafeModeException, IOException { try {
// TODO: implement checkOperation(OperationCategory.WRITE);
if (isInSafeMode()) {
throw new SafeModeException("Cannot allow snapshot for " + path, safeMode);
}
checkOwner(path);
//TODO: do not hardcode snapshot quota value
snapshotManager.setSnapshottable(path, 256);
getEditLog().logAllowSnapshot(path);
} finally {
writeUnlock();
}
getEditLog().logSync();
//TODO: audit log
} }
// Disallow snapshot on a directory. // Disallow snapshot on a directory.
@ -5542,6 +5560,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
dir.writeLock(); dir.writeLock();
try { try {
snapshotManager.createSnapshot(snapshotName, path); snapshotManager.createSnapshot(snapshotName, path);
getEditLog().logCreateSnapshot(snapshotName, path);
} finally { } finally {
dir.writeUnlock(); dir.writeUnlock();
} }

View File

@ -81,7 +81,7 @@ public class SnapshotManager {
} }
/** /**
* Create a snapshot of subtrees for recursively coping the directory * Create a snapshot of subtrees by recursively coping the directory
* structure from the source directory to the snapshot destination directory. * structure from the source directory to the snapshot destination directory.
* This creation algorithm requires O(N) running time and O(N) memory, * This creation algorithm requires O(N) running time and O(N) memory,
* where N = # files + # directories + # symlinks. * where N = # files + # directories + # symlinks.

View File

@ -0,0 +1,87 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hdfs.server.namenode.snapshot;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.DistributedFileSystem;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.server.namenode.FSDirectory;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.namenode.INode;
import org.apache.hadoop.hdfs.server.namenode.INodeDirectory;
import org.junit.After;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Test;
/** Test snapshot related operations. */
public class TestSnapshot {
private static final long seed = 0;
private static final short REPLICATION = 3;
private final Path dir = new Path("/TestSnapshot");
private final Path sub1 = new Path(dir, "sub1");
private final Path file1 = new Path(sub1, "file1");
private final Path file2 = new Path(sub1, "file2");
private Configuration conf;
private MiniDFSCluster cluster;
private FSNamesystem fsn;
private FSDirectory fsdir;
private DistributedFileSystem hdfs;
@Before
public void setUp() throws Exception {
conf = new Configuration();
cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(REPLICATION)
.build();
cluster.waitActive();
fsn = cluster.getNamesystem();
fsdir = fsn.getFSDirectory();
hdfs = cluster.getFileSystem();
DFSTestUtil.createFile(hdfs, file1, 1024, REPLICATION, seed);
DFSTestUtil.createFile(hdfs, file2, 1024, REPLICATION, seed);
}
@After
public void tearDown() throws Exception {
if (cluster != null) {
cluster.shutdown();
}
}
/** Test allow-snapshot operation. */
@Test
public void testAllowSnapshot() throws Exception {
final String path = sub1.toString();
final INode before = fsdir.getINode(path);
Assert.assertTrue(before instanceof INodeDirectory);
Assert.assertFalse(before instanceof INodeDirectorySnapshottable);
hdfs.allowSnapshot(path);
final INode after = fsdir.getINode(path);
Assert.assertTrue(after instanceof INodeDirectorySnapshottable);
}
}