Revert "HDFS-10431 Refactor and speedup TestAsyncDFSRename. Contributed by Xiaobing Zhou"

This reverts commit f4b9bcd87c.
This commit is contained in:
Andrew Wang 2016-06-03 18:09:13 -07:00
parent 8cf47d8589
commit 5ee5912ebd
2 changed files with 484 additions and 314 deletions

View File

@ -29,16 +29,13 @@
import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry; import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry;
import static org.junit.Assert.assertArrayEquals; import static org.junit.Assert.assertArrayEquals;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.fail; import static org.junit.Assert.fail;
import java.io.IOException; import java.io.IOException;
import java.security.PrivilegedExceptionAction; import java.security.PrivilegedExceptionAction;
import java.util.Arrays;
import java.util.HashMap; import java.util.HashMap;
import java.util.List; import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.Random;
import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutionException;
import java.util.concurrent.Future; import java.util.concurrent.Future;
@ -46,21 +43,15 @@
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.CommonConfigurationKeys; import org.apache.hadoop.fs.CommonConfigurationKeys;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.Options.Rename;
import org.apache.hadoop.fs.permission.AclEntry; import org.apache.hadoop.fs.permission.AclEntry;
import org.apache.hadoop.fs.permission.AclStatus; import org.apache.hadoop.fs.permission.AclStatus;
import org.apache.hadoop.fs.permission.FsAction;
import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.TestDFSPermission.PermissionGenerator;
import org.apache.hadoop.hdfs.server.namenode.AclTestHelpers; import org.apache.hadoop.hdfs.server.namenode.AclTestHelpers;
import org.apache.hadoop.hdfs.server.namenode.FSAclBaseTest; import org.apache.hadoop.hdfs.server.namenode.FSAclBaseTest;
import org.apache.hadoop.ipc.AsyncCallLimitExceededException; import org.apache.hadoop.ipc.AsyncCallLimitExceededException;
import org.apache.hadoop.security.AccessControlException;
import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.Time;
import org.junit.After; import org.junit.After;
import org.junit.Before; import org.junit.Before;
import org.junit.Test; import org.junit.Test;
@ -72,28 +63,21 @@
* */ * */
public class TestAsyncDFS { public class TestAsyncDFS {
public static final Log LOG = LogFactory.getLog(TestAsyncDFS.class); public static final Log LOG = LogFactory.getLog(TestAsyncDFS.class);
private final short replFactor = 1; private static final int NUM_TESTS = 1000;
private final long blockSize = 512;
private long fileLen = blockSize * 3;
private final long seed = Time.now();
private final Random r = new Random(seed);
private final PermissionGenerator permGenerator = new PermissionGenerator(r);
private static final int NUM_TESTS = 50;
private static final int NUM_NN_HANDLER = 10; private static final int NUM_NN_HANDLER = 10;
private static final int ASYNC_CALL_LIMIT = 1000; private static final int ASYNC_CALL_LIMIT = 100;
private Configuration conf; private Configuration conf;
private MiniDFSCluster cluster; private MiniDFSCluster cluster;
private FileSystem fs; private FileSystem fs;
private AsyncDistributedFileSystem adfs;
@Before @Before
public void setup() throws IOException { public void setup() throws IOException {
conf = new HdfsConfiguration(); conf = new HdfsConfiguration();
// explicitly turn on acl // explicitly turn on acl
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true); conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true);
// explicitly turn on permission checking // explicitly turn on ACL
conf.setBoolean(DFSConfigKeys.DFS_PERMISSIONS_ENABLED_KEY, true); conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true);
// set the limit of max async calls // set the limit of max async calls
conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY, conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY,
ASYNC_CALL_LIMIT); ASYNC_CALL_LIMIT);
@ -102,7 +86,6 @@ public void setup() throws IOException {
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build(); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
cluster.waitActive(); cluster.waitActive();
fs = FileSystem.get(conf); fs = FileSystem.get(conf);
adfs = cluster.getFileSystem().getAsyncDistributedFileSystem();
} }
@After @After
@ -147,9 +130,13 @@ public void testBatchAsyncAcl() throws Exception {
final String basePath = "testBatchAsyncAcl"; final String basePath = "testBatchAsyncAcl";
final Path parent = new Path(String.format("/test/%s/", basePath)); final Path parent = new Path(String.format("/test/%s/", basePath));
AsyncDistributedFileSystem adfs = cluster.getFileSystem()
.getAsyncDistributedFileSystem();
// prepare test // prepare test
final Path[] paths = new Path[NUM_TESTS]; int count = NUM_TESTS;
for (int i = 0; i < NUM_TESTS; i++) { final Path[] paths = new Path[count];
for (int i = 0; i < count; i++) {
paths[i] = new Path(parent, "acl" + i); paths[i] = new Path(parent, "acl" + i);
FileSystem.mkdirs(fs, paths[i], FileSystem.mkdirs(fs, paths[i],
FsPermission.createImmutable((short) 0750)); FsPermission.createImmutable((short) 0750));
@ -166,7 +153,7 @@ public void testBatchAsyncAcl() throws Exception {
int start = 0, end = 0; int start = 0, end = 0;
try { try {
// test setAcl // test setAcl
for (int i = 0; i < NUM_TESTS; i++) { for (int i = 0; i < count; i++) {
for (;;) { for (;;) {
try { try {
Future<Void> retFuture = adfs.setAcl(paths[i], aclSpec); Future<Void> retFuture = adfs.setAcl(paths[i], aclSpec);
@ -179,12 +166,12 @@ public void testBatchAsyncAcl() throws Exception {
} }
} }
} }
waitForAclReturnValues(setAclRetFutures, end, NUM_TESTS); waitForAclReturnValues(setAclRetFutures, end, count);
// test getAclStatus // test getAclStatus
start = 0; start = 0;
end = 0; end = 0;
for (int i = 0; i < NUM_TESTS; i++) { for (int i = 0; i < count; i++) {
for (;;) { for (;;) {
try { try {
Future<AclStatus> retFuture = adfs.getAclStatus(paths[i]); Future<AclStatus> retFuture = adfs.getAclStatus(paths[i]);
@ -198,23 +185,13 @@ public void testBatchAsyncAcl() throws Exception {
} }
} }
} }
waitForAclReturnValues(getAclRetFutures, end, NUM_TESTS, paths, waitForAclReturnValues(getAclRetFutures, end, count, paths,
expectedAclSpec); expectedAclSpec);
} catch (Exception e) { } catch (Exception e) {
throw e; throw e;
} }
} }
static void waitForReturnValues(final Map<Integer, Future<Void>> retFutures,
final int start, final int end)
throws InterruptedException, ExecutionException {
LOG.info(String.format("calling waitForReturnValues [%d, %d)", start, end));
for (int i = start; i < end; i++) {
LOG.info("calling Future#get #" + i);
retFutures.get(i).get();
}
}
private void waitForAclReturnValues( private void waitForAclReturnValues(
final Map<Integer, Future<Void>> aclRetFutures, final int start, final Map<Integer, Future<Void>> aclRetFutures, final int start,
final int end) throws InterruptedException, ExecutionException { final int end) throws InterruptedException, ExecutionException {
@ -289,12 +266,9 @@ public void testAsyncAPIWithException() throws Exception {
final Path parent = new Path("/test/async_api_exception/"); final Path parent = new Path("/test/async_api_exception/");
final Path aclDir = new Path(parent, "aclDir"); final Path aclDir = new Path(parent, "aclDir");
final Path src = new Path(parent, "src"); fs.mkdirs(aclDir, FsPermission.createImmutable((short) 0770));
final Path dst = new Path(parent, "dst");
fs.mkdirs(aclDir, FsPermission.createImmutable((short) 0700));
fs.mkdirs(src);
AsyncDistributedFileSystem adfs1 = ugi1 AsyncDistributedFileSystem adfs = ugi1
.doAs(new PrivilegedExceptionAction<AsyncDistributedFileSystem>() { .doAs(new PrivilegedExceptionAction<AsyncDistributedFileSystem>() {
@Override @Override
public AsyncDistributedFileSystem run() throws Exception { public AsyncDistributedFileSystem run() throws Exception {
@ -303,36 +277,9 @@ public AsyncDistributedFileSystem run() throws Exception {
}); });
Future<Void> retFuture; Future<Void> retFuture;
// test rename
try {
retFuture = adfs1.rename(src, dst, Rename.OVERWRITE);
retFuture.get();
} catch (ExecutionException e) {
checkPermissionDenied(e, src, user1);
assertTrue("Permission denied messages must carry the path parent", e
.getMessage().contains(src.getParent().toUri().getPath()));
}
// test setPermission
FsPermission fsPerm = new FsPermission(permGenerator.next());
try {
retFuture = adfs1.setPermission(src, fsPerm);
retFuture.get();
} catch (ExecutionException e) {
checkPermissionDenied(e, src, user1);
}
// test setOwner
try {
retFuture = adfs1.setOwner(src, "user1", "group2");
retFuture.get();
} catch (ExecutionException e) {
checkPermissionDenied(e, src, user1);
}
// test setAcl // test setAcl
try { try {
retFuture = adfs1.setAcl(aclDir, retFuture = adfs.setAcl(aclDir,
Lists.newArrayList(aclEntry(ACCESS, USER, ALL))); Lists.newArrayList(aclEntry(ACCESS, USER, ALL)));
retFuture.get(); retFuture.get();
fail("setAcl should fail with permission denied"); fail("setAcl should fail with permission denied");
@ -342,7 +289,7 @@ public AsyncDistributedFileSystem run() throws Exception {
// test getAclStatus // test getAclStatus
try { try {
Future<AclStatus> aclRetFuture = adfs1.getAclStatus(aclDir); Future<AclStatus> aclRetFuture = adfs.getAclStatus(aclDir);
aclRetFuture.get(); aclRetFuture.get();
fail("getAclStatus should fail with permission denied"); fail("getAclStatus should fail with permission denied");
} catch (ExecutionException e) { } catch (ExecutionException e) {
@ -360,148 +307,4 @@ public static void checkPermissionDenied(final Exception e, final Path dir,
assertTrue("Permission denied messages must carry the name of the path", assertTrue("Permission denied messages must carry the name of the path",
e.getMessage().contains(dir.getName())); e.getMessage().contains(dir.getName()));
} }
@Test(timeout = 120000)
public void testConcurrentAsyncAPI() throws Exception {
String group1 = "group1";
String group2 = "group2";
String user1 = "user1";
// create fake mapping for the groups
Map<String, String[]> u2gMap = new HashMap<String, String[]>(1);
u2gMap.put(user1, new String[] {group1, group2});
DFSTestUtil.updateConfWithFakeGroupMapping(conf, u2gMap);
// prepare for test
final Path parent = new Path(
String.format("/test/%s/", "testConcurrentAsyncAPI"));
final Path[] srcs = new Path[NUM_TESTS];
final Path[] dsts = new Path[NUM_TESTS];
short[] permissions = new short[NUM_TESTS];
for (int i = 0; i < NUM_TESTS; i++) {
srcs[i] = new Path(parent, "src" + i);
dsts[i] = new Path(parent, "dst" + i);
DFSTestUtil.createFile(fs, srcs[i], fileLen, replFactor, 1);
DFSTestUtil.createFile(fs, dsts[i], fileLen, replFactor, 1);
assertTrue(fs.exists(srcs[i]));
assertTrue(fs.getFileStatus(srcs[i]).isFile());
assertTrue(fs.exists(dsts[i]));
assertTrue(fs.getFileStatus(dsts[i]).isFile());
permissions[i] = permGenerator.next();
}
Map<Integer, Future<Void>> renameRetFutures =
new HashMap<Integer, Future<Void>>();
Map<Integer, Future<Void>> permRetFutures =
new HashMap<Integer, Future<Void>>();
Map<Integer, Future<Void>> ownerRetFutures =
new HashMap<Integer, Future<Void>>();
int start = 0, end = 0;
// test rename
for (int i = 0; i < NUM_TESTS; i++) {
for (;;) {
try {
Future<Void> returnFuture = adfs.rename(srcs[i], dsts[i],
Rename.OVERWRITE);
renameRetFutures.put(i, returnFuture);
break;
} catch (AsyncCallLimitExceededException e) {
start = end;
end = i;
waitForReturnValues(renameRetFutures, start, end);
}
}
}
// wait for completing the calls
waitForAclReturnValues(renameRetFutures, end, NUM_TESTS);
// verify the src should not exist, dst should
for (int i = 0; i < NUM_TESTS; i++) {
assertFalse(fs.exists(srcs[i]));
assertTrue(fs.exists(dsts[i]));
}
// test permissions
for (int i = 0; i < NUM_TESTS; i++) {
for (;;) {
try {
Future<Void> retFuture = adfs.setPermission(dsts[i],
new FsPermission(permissions[i]));
permRetFutures.put(i, retFuture);
break;
} catch (AsyncCallLimitExceededException e) {
start = end;
end = i;
waitForReturnValues(permRetFutures, start, end);
}
}
}
// wait for completing the calls
waitForAclReturnValues(permRetFutures, end, NUM_TESTS);
// verify the permission
for (int i = 0; i < NUM_TESTS; i++) {
assertTrue(fs.exists(dsts[i]));
FsPermission fsPerm = new FsPermission(permissions[i]);
checkAccessPermissions(fs.getFileStatus(dsts[i]), fsPerm.getUserAction());
}
// test setOwner
start = 0;
end = 0;
for (int i = 0; i < NUM_TESTS; i++) {
for (;;) {
try {
Future<Void> retFuture = adfs.setOwner(dsts[i], "user1", "group2");
ownerRetFutures.put(i, retFuture);
break;
} catch (AsyncCallLimitExceededException e) {
start = end;
end = i;
waitForReturnValues(ownerRetFutures, start, end);
}
}
}
// wait for completing the calls
waitForAclReturnValues(ownerRetFutures, end, NUM_TESTS);
// verify the owner
for (int i = 0; i < NUM_TESTS; i++) {
assertTrue(fs.exists(dsts[i]));
assertTrue("user1".equals(fs.getFileStatus(dsts[i]).getOwner()));
assertTrue("group2".equals(fs.getFileStatus(dsts[i]).getGroup()));
}
}
static void checkAccessPermissions(FileStatus stat, FsAction mode)
throws IOException {
checkAccessPermissions(UserGroupInformation.getCurrentUser(), stat, mode);
}
static void checkAccessPermissions(final UserGroupInformation ugi,
FileStatus stat, FsAction mode) throws IOException {
FsPermission perm = stat.getPermission();
String user = ugi.getShortUserName();
List<String> groups = Arrays.asList(ugi.getGroupNames());
if (user.equals(stat.getOwner())) {
if (perm.getUserAction().implies(mode)) {
return;
}
} else if (groups.contains(stat.getGroup())) {
if (perm.getGroupAction().implies(mode)) {
return;
}
} else {
if (perm.getOtherAction().implies(mode)) {
return;
}
}
throw new AccessControlException(String.format(
"Permission denied: user=%s, path=\"%s\":%s:%s:%s%s", user, stat
.getPath(), stat.getOwner(), stat.getGroup(),
stat.isDirectory() ? "d" : "-", perm));
}
} }

View File

@ -19,11 +19,14 @@
import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.junit.Assert.assertEquals;
import java.io.IOException; import java.io.IOException;
import java.security.PrivilegedExceptionAction;
import java.util.Arrays;
import java.util.HashMap; import java.util.HashMap;
import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.Random;
import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutionException;
import java.util.concurrent.Future; import java.util.concurrent.Future;
@ -31,157 +34,521 @@
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.CommonConfigurationKeys; import org.apache.hadoop.fs.CommonConfigurationKeys;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.Options.Rename; import org.apache.hadoop.fs.Options.Rename;
import org.apache.hadoop.fs.permission.FsAction;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.TestDFSPermission.PermissionGenerator;
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockManager;
import org.apache.hadoop.hdfs.server.namenode.NameNodeAdapter;
import org.apache.hadoop.ipc.AsyncCallLimitExceededException; import org.apache.hadoop.ipc.AsyncCallLimitExceededException;
import org.junit.After; import org.apache.hadoop.security.AccessControlException;
import org.junit.Before; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.Time;
import org.junit.Test; import org.junit.Test;
public class TestAsyncDFSRename { public class TestAsyncDFSRename {
public static final Log LOG = LogFactory.getLog(TestAsyncDFSRename.class); public static final Log LOG = LogFactory.getLog(TestAsyncDFSRename.class);
private final short replFactor = 1; private final long seed = Time.now();
private final Random r = new Random(seed);
private final PermissionGenerator permGenerator = new PermissionGenerator(r);
private final short replFactor = 2;
private final long blockSize = 512; private final long blockSize = 512;
private long fileLen = blockSize * 3; private long fileLen = blockSize * 3;
private static final int NUM_TESTS = 50;
private static final int NUM_NN_HANDLER = 10;
private static final int ASYNC_CALL_LIMIT = 1000;
private Configuration conf; /**
private MiniDFSCluster cluster; * Check the blocks of dst file are cleaned after rename with overwrite
private FileSystem fs; * Restart NN to check the rename successfully
private AsyncDistributedFileSystem adfs; */
@Test(timeout = 60000)
@Before public void testAsyncRenameWithOverwrite() throws Exception {
public void setup() throws IOException { Configuration conf = new Configuration();
conf = new HdfsConfiguration(); MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(
// set the limit of max async calls replFactor).build();
conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY,
ASYNC_CALL_LIMIT);
// set server handlers
conf.setInt(DFSConfigKeys.DFS_NAMENODE_HANDLER_COUNT_KEY, NUM_NN_HANDLER);
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build();
cluster.waitActive(); cluster.waitActive();
fs = FileSystem.get(conf); DistributedFileSystem dfs = cluster.getFileSystem();
adfs = cluster.getFileSystem().getAsyncDistributedFileSystem(); AsyncDistributedFileSystem adfs = dfs.getAsyncDistributedFileSystem();
}
@After try {
public void tearDown() throws IOException { String src = "/foo/src";
if (fs != null) { String dst = "/foo/dst";
fs.close(); String src2 = "/foo/src2";
fs = null; String dst2 = "/foo/dst2";
} Path srcPath = new Path(src);
if (cluster != null) { Path dstPath = new Path(dst);
cluster.shutdown(); Path srcPath2 = new Path(src2);
cluster = null; Path dstPath2 = new Path(dst2);
DFSTestUtil.createFile(dfs, srcPath, fileLen, replFactor, 1);
DFSTestUtil.createFile(dfs, dstPath, fileLen, replFactor, 1);
DFSTestUtil.createFile(dfs, srcPath2, fileLen, replFactor, 1);
DFSTestUtil.createFile(dfs, dstPath2, fileLen, replFactor, 1);
LocatedBlocks lbs = NameNodeAdapter.getBlockLocations(
cluster.getNameNode(), dst, 0, fileLen);
LocatedBlocks lbs2 = NameNodeAdapter.getBlockLocations(
cluster.getNameNode(), dst2, 0, fileLen);
BlockManager bm = NameNodeAdapter.getNamesystem(cluster.getNameNode())
.getBlockManager();
assertTrue(bm.getStoredBlock(lbs.getLocatedBlocks().get(0).getBlock()
.getLocalBlock()) != null);
assertTrue(bm.getStoredBlock(lbs2.getLocatedBlocks().get(0).getBlock()
.getLocalBlock()) != null);
Future<Void> retVal1 = adfs.rename(srcPath, dstPath, Rename.OVERWRITE);
Future<Void> retVal2 = adfs.rename(srcPath2, dstPath2, Rename.OVERWRITE);
retVal1.get();
retVal2.get();
assertTrue(bm.getStoredBlock(lbs.getLocatedBlocks().get(0).getBlock()
.getLocalBlock()) == null);
assertTrue(bm.getStoredBlock(lbs2.getLocatedBlocks().get(0).getBlock()
.getLocalBlock()) == null);
// Restart NN and check the rename successfully
cluster.restartNameNodes();
assertFalse(dfs.exists(srcPath));
assertTrue(dfs.exists(dstPath));
assertFalse(dfs.exists(srcPath2));
assertTrue(dfs.exists(dstPath2));
} finally {
if (dfs != null) {
dfs.close();
}
if (cluster != null) {
cluster.shutdown();
}
} }
} }
@Test(timeout = 60000) @Test(timeout = 60000)
public void testCallGetReturnValueMultipleTimes() throws Exception { public void testCallGetReturnValueMultipleTimes() throws Exception {
final Path parent = new Path("/test/testCallGetReturnValueMultipleTimes/"); final Path renameDir = new Path(
assertTrue(fs.mkdirs(parent)); "/test/testCallGetReturnValueMultipleTimes/");
final Configuration conf = new HdfsConfiguration();
conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY, 200);
final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(2).build();
cluster.waitActive();
final DistributedFileSystem dfs = cluster.getFileSystem();
final AsyncDistributedFileSystem adfs = dfs.getAsyncDistributedFileSystem();
final int count = 100;
final Map<Integer, Future<Void>> returnFutures = new HashMap<Integer, Future<Void>>();
// prepare test assertTrue(dfs.mkdirs(renameDir));
final Path[] srcs = new Path[NUM_TESTS];
final Path[] dsts = new Path[NUM_TESTS];
for (int i = 0; i < NUM_TESTS; i++) {
srcs[i] = new Path(parent, "src" + i);
dsts[i] = new Path(parent, "dst" + i);
DFSTestUtil.createFile(fs, srcs[i], fileLen, replFactor, 1);
DFSTestUtil.createFile(fs, dsts[i], fileLen, replFactor, 1);
}
// concurrently invoking many rename try {
final Map<Integer, Future<Void>> reFutures = // concurrently invoking many rename
new HashMap<Integer, Future<Void>>(); for (int i = 0; i < count; i++) {
for (int i = 0; i < NUM_TESTS; i++) { Path src = new Path(renameDir, "src" + i);
Future<Void> retFuture = adfs.rename(srcs[i], dsts[i], Path dst = new Path(renameDir, "dst" + i);
Rename.OVERWRITE); DFSTestUtil.createFile(dfs, src, fileLen, replFactor, 1);
reFutures.put(i, retFuture); DFSTestUtil.createFile(dfs, dst, fileLen, replFactor, 1);
} Future<Void> returnFuture = adfs.rename(src, dst, Rename.OVERWRITE);
returnFutures.put(i, returnFuture);
}
assertEquals(NUM_TESTS, reFutures.size()); for (int i = 0; i < 5; i++) {
verifyCallGetReturnValueMultipleTimes(returnFutures, count, cluster,
for (int i = 0; i < 5; i++) { renameDir, dfs);
verifyCallGetReturnValueMultipleTimes(reFutures, srcs, dsts); }
} finally {
if (dfs != null) {
dfs.close();
}
if (cluster != null) {
cluster.shutdown();
}
} }
} }
private void verifyCallGetReturnValueMultipleTimes( private void verifyCallGetReturnValueMultipleTimes(
final Map<Integer, Future<Void>> reFutures, final Path[] srcs, Map<Integer, Future<Void>> returnFutures, int count,
final Path[] dsts) MiniDFSCluster cluster, Path renameDir, DistributedFileSystem dfs)
throws InterruptedException, ExecutionException, IOException { throws InterruptedException, ExecutionException, IOException {
// wait for completing the calls // wait for completing the calls
waitForReturnValues(reFutures, 0, NUM_TESTS); for (int i = 0; i < count; i++) {
returnFutures.get(i).get();
}
// verify the src dir should not exist, dst should // Restart NN and check the rename successfully
verifyRenames(srcs, dsts); cluster.restartNameNodes();
// very the src dir should not exist, dst should
for (int i = 0; i < count; i++) {
Path src = new Path(renameDir, "src" + i);
Path dst = new Path(renameDir, "dst" + i);
assertFalse(dfs.exists(src));
assertTrue(dfs.exists(dst));
}
}
@Test
public void testConservativeConcurrentAsyncRenameWithOverwrite()
throws Exception {
internalTestConcurrentAsyncRenameWithOverwrite(100,
"testAggressiveConcurrentAsyncRenameWithOverwrite");
} }
@Test(timeout = 60000) @Test(timeout = 60000)
public void testConcurrentAsyncRename() throws Exception { public void testAggressiveConcurrentAsyncRenameWithOverwrite()
final Path parent = new Path( throws Exception {
String.format("/test/%s/", "testConcurrentAsyncRename")); internalTestConcurrentAsyncRenameWithOverwrite(10000,
assertTrue(fs.mkdirs(parent)); "testConservativeConcurrentAsyncRenameWithOverwrite");
}
// prepare test private void internalTestConcurrentAsyncRenameWithOverwrite(
final Path[] srcs = new Path[NUM_TESTS]; final int asyncCallLimit, final String basePath) throws Exception {
final Path[] dsts = new Path[NUM_TESTS]; final Path renameDir = new Path(String.format("/test/%s/", basePath));
for (int i = 0; i < NUM_TESTS; i++) { Configuration conf = new HdfsConfiguration();
conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY,
asyncCallLimit);
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2)
.build();
cluster.waitActive();
DistributedFileSystem dfs = cluster.getFileSystem();
AsyncDistributedFileSystem adfs = dfs.getAsyncDistributedFileSystem();
int count = 1000;
int start = 0, end = 0;
Map<Integer, Future<Void>> returnFutures = new HashMap<Integer, Future<Void>>();
assertTrue(dfs.mkdirs(renameDir));
try {
// concurrently invoking many rename
for (int i = 0; i < count; i++) {
Path src = new Path(renameDir, "src" + i);
Path dst = new Path(renameDir, "dst" + i);
DFSTestUtil.createFile(dfs, src, fileLen, replFactor, 1);
DFSTestUtil.createFile(dfs, dst, fileLen, replFactor, 1);
for (;;) {
try {
LOG.info("rename #" + i);
Future<Void> returnFuture = adfs.rename(src, dst, Rename.OVERWRITE);
returnFutures.put(i, returnFuture);
break;
} catch (AsyncCallLimitExceededException e) {
/**
* reached limit of async calls, fetch results of finished async
* calls to let follow-on calls go
*/
LOG.error(e);
start = end;
end = i;
LOG.info(String.format("start=%d, end=%d, i=%d", start, end, i));
waitForReturnValues(returnFutures, start, end);
}
}
}
// wait for completing the calls
for (int i = start; i < count; i++) {
returnFutures.get(i).get();
}
// Restart NN and check the rename successfully
cluster.restartNameNodes();
// very the src dir should not exist, dst should
for (int i = 0; i < count; i++) {
Path src = new Path(renameDir, "src" + i);
Path dst = new Path(renameDir, "dst" + i);
assertFalse(dfs.exists(src));
assertTrue(dfs.exists(dst));
}
} finally {
if (dfs != null) {
dfs.close();
}
if (cluster != null) {
cluster.shutdown();
}
}
}
private void waitForReturnValues(
final Map<Integer, Future<Void>> returnFutures, final int start,
final int end) throws InterruptedException, ExecutionException {
LOG.info(String.format("calling waitForReturnValues [%d, %d)", start, end));
for (int i = start; i < end; i++) {
LOG.info("calling Future#get #" + i);
returnFutures.get(i).get();
}
}
@Test
public void testConservativeConcurrentAsyncAPI() throws Exception {
internalTestConcurrentAsyncAPI(100, "testConservativeConcurrentAsyncAPI");
}
@Test(timeout = 60000)
public void testAggressiveConcurrentAsyncAPI() throws Exception {
internalTestConcurrentAsyncAPI(10000, "testAggressiveConcurrentAsyncAPI");
}
private void internalTestConcurrentAsyncAPI(final int asyncCallLimit,
final String basePath) throws Exception {
Configuration conf = new HdfsConfiguration();
String group1 = "group1";
String group2 = "group2";
String user1 = "user1";
int count = 500;
// explicitly turn on permission checking
conf.setBoolean(DFSConfigKeys.DFS_PERMISSIONS_ENABLED_KEY, true);
// set the limit of max async calls
conf.setInt(CommonConfigurationKeys.IPC_CLIENT_ASYNC_CALLS_MAX_KEY,
asyncCallLimit);
// create fake mapping for the groups
Map<String, String[]> u2gMap = new HashMap<String, String[]>(1);
u2gMap.put(user1, new String[] {group1, group2});
DFSTestUtil.updateConfWithFakeGroupMapping(conf, u2gMap);
// start mini cluster
final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(3).build();
cluster.waitActive();
AsyncDistributedFileSystem adfs = cluster.getFileSystem()
.getAsyncDistributedFileSystem();
// prepare for test
FileSystem rootFs = FileSystem.get(conf);
final Path parent = new Path(String.format("/test/%s/", basePath));
final Path[] srcs = new Path[count];
final Path[] dsts = new Path[count];
short[] permissions = new short[count];
for (int i = 0; i < count; i++) {
srcs[i] = new Path(parent, "src" + i); srcs[i] = new Path(parent, "src" + i);
dsts[i] = new Path(parent, "dst" + i); dsts[i] = new Path(parent, "dst" + i);
DFSTestUtil.createFile(fs, srcs[i], fileLen, replFactor, 1); DFSTestUtil.createFile(rootFs, srcs[i], fileLen, replFactor, 1);
DFSTestUtil.createFile(fs, dsts[i], fileLen, replFactor, 1); DFSTestUtil.createFile(rootFs, dsts[i], fileLen, replFactor, 1);
assertTrue(rootFs.exists(srcs[i]));
assertTrue(rootFs.getFileStatus(srcs[i]).isFile());
assertTrue(rootFs.exists(dsts[i]));
assertTrue(rootFs.getFileStatus(dsts[i]).isFile());
permissions[i] = permGenerator.next();
} }
// concurrently invoking many rename Map<Integer, Future<Void>> renameRetFutures =
int start = 0, end = 0;
Map<Integer, Future<Void>> retFutures =
new HashMap<Integer, Future<Void>>(); new HashMap<Integer, Future<Void>>();
for (int i = 0; i < NUM_TESTS; i++) { Map<Integer, Future<Void>> permRetFutures =
new HashMap<Integer, Future<Void>>();
Map<Integer, Future<Void>> ownerRetFutures =
new HashMap<Integer, Future<Void>>();
int start = 0, end = 0;
// test rename
for (int i = 0; i < count; i++) {
for (;;) { for (;;) {
try { try {
LOG.info("rename #" + i); Future<Void> returnFuture = adfs.rename(srcs[i], dsts[i],
Future<Void> retFuture = adfs.rename(srcs[i], dsts[i],
Rename.OVERWRITE); Rename.OVERWRITE);
retFutures.put(i, retFuture); renameRetFutures.put(i, returnFuture);
break; break;
} catch (AsyncCallLimitExceededException e) { } catch (AsyncCallLimitExceededException e) {
/**
* reached limit of async calls, fetch results of finished async calls
* to let follow-on calls go
*/
LOG.error(e);
start = end; start = end;
end = i; end = i;
LOG.info(String.format("start=%d, end=%d, i=%d", start, end, i)); waitForReturnValues(renameRetFutures, start, end);
waitForReturnValues(retFutures, start, end);
} }
} }
} }
// wait for completing the calls // wait for completing the calls
waitForReturnValues(retFutures, end, NUM_TESTS); for (int i = start; i < count; i++) {
renameRetFutures.get(i).get();
}
// verify the src dir should not exist, dst should // Restart NN and check the rename successfully
verifyRenames(srcs, dsts); cluster.restartNameNodes();
}
private void verifyRenames(final Path[] srcs, final Path[] dsts) // very the src should not exist, dst should
throws IOException { for (int i = 0; i < count; i++) {
for (int i = 0; i < NUM_TESTS; i++) { assertFalse(rootFs.exists(srcs[i]));
assertFalse(fs.exists(srcs[i])); assertTrue(rootFs.exists(dsts[i]));
assertTrue(fs.exists(dsts[i])); }
// test permissions
try {
for (int i = 0; i < count; i++) {
for (;;) {
try {
Future<Void> retFuture = adfs.setPermission(dsts[i],
new FsPermission(permissions[i]));
permRetFutures.put(i, retFuture);
break;
} catch (AsyncCallLimitExceededException e) {
start = end;
end = i;
waitForReturnValues(permRetFutures, start, end);
}
}
}
// wait for completing the calls
for (int i = start; i < count; i++) {
permRetFutures.get(i).get();
}
// Restart NN and check permission then
cluster.restartNameNodes();
// verify the permission
for (int i = 0; i < count; i++) {
assertTrue(rootFs.exists(dsts[i]));
FsPermission fsPerm = new FsPermission(permissions[i]);
checkAccessPermissions(rootFs.getFileStatus(dsts[i]),
fsPerm.getUserAction());
}
// test setOwner
start = 0;
end = 0;
for (int i = 0; i < count; i++) {
for (;;) {
try {
Future<Void> retFuture = adfs.setOwner(dsts[i], "user1",
"group2");
ownerRetFutures.put(i, retFuture);
break;
} catch (AsyncCallLimitExceededException e) {
start = end;
end = i;
waitForReturnValues(ownerRetFutures, start, end);
}
}
}
// wait for completing the calls
for (int i = start; i < count; i++) {
ownerRetFutures.get(i).get();
}
// Restart NN and check owner then
cluster.restartNameNodes();
// verify the owner
for (int i = 0; i < count; i++) {
assertTrue(rootFs.exists(dsts[i]));
assertTrue(
"user1".equals(rootFs.getFileStatus(dsts[i]).getOwner()));
assertTrue(
"group2".equals(rootFs.getFileStatus(dsts[i]).getGroup()));
}
} catch (AccessControlException ace) {
throw ace;
} finally {
if (rootFs != null) {
rootFs.close();
}
if (cluster != null) {
cluster.shutdown();
}
} }
} }
void waitForReturnValues(final Map<Integer, Future<Void>> retFutures, static void checkAccessPermissions(FileStatus stat, FsAction mode)
final int start, final int end) throws IOException {
throws InterruptedException, ExecutionException { checkAccessPermissions(UserGroupInformation.getCurrentUser(), stat, mode);
TestAsyncDFS.waitForReturnValues(retFutures, start, end); }
static void checkAccessPermissions(final UserGroupInformation ugi,
FileStatus stat, FsAction mode) throws IOException {
FsPermission perm = stat.getPermission();
String user = ugi.getShortUserName();
List<String> groups = Arrays.asList(ugi.getGroupNames());
if (user.equals(stat.getOwner())) {
if (perm.getUserAction().implies(mode)) {
return;
}
} else if (groups.contains(stat.getGroup())) {
if (perm.getGroupAction().implies(mode)) {
return;
}
} else {
if (perm.getOtherAction().implies(mode)) {
return;
}
}
throw new AccessControlException(String.format(
"Permission denied: user=%s, path=\"%s\":%s:%s:%s%s", user, stat
.getPath(), stat.getOwner(), stat.getGroup(),
stat.isDirectory() ? "d" : "-", perm));
}
@Test(timeout = 60000)
public void testAsyncAPIWithException() throws Exception {
Configuration conf = new HdfsConfiguration();
String group1 = "group1";
String group2 = "group2";
String user1 = "user1";
UserGroupInformation ugi1;
// explicitly turn on permission checking
conf.setBoolean(DFSConfigKeys.DFS_PERMISSIONS_ENABLED_KEY, true);
// create fake mapping for the groups
Map<String, String[]> u2gMap = new HashMap<String, String[]>(1);
u2gMap.put(user1, new String[] {group1, group2});
DFSTestUtil.updateConfWithFakeGroupMapping(conf, u2gMap);
// Initiate all four users
ugi1 = UserGroupInformation.createUserForTesting(user1, new String[] {
group1, group2 });
final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(3).build();
cluster.waitActive();
FileSystem rootFs = FileSystem.get(conf);
final Path renameDir = new Path("/test/async_api_exception/");
final Path src = new Path(renameDir, "src");
final Path dst = new Path(renameDir, "dst");
rootFs.mkdirs(src);
AsyncDistributedFileSystem adfs = ugi1
.doAs(new PrivilegedExceptionAction<AsyncDistributedFileSystem>() {
@Override
public AsyncDistributedFileSystem run() throws Exception {
return cluster.getFileSystem().getAsyncDistributedFileSystem();
}
});
Future<Void> retFuture;
try {
retFuture = adfs.rename(src, dst, Rename.OVERWRITE);
retFuture.get();
} catch (ExecutionException e) {
TestAsyncDFS.checkPermissionDenied(e, src, user1);
assertTrue("Permission denied messages must carry the path parent", e
.getMessage().contains(src.getParent().toUri().getPath()));
}
FsPermission fsPerm = new FsPermission(permGenerator.next());
try {
retFuture = adfs.setPermission(src, fsPerm);
retFuture.get();
} catch (ExecutionException e) {
TestAsyncDFS.checkPermissionDenied(e, src, user1);
assertTrue("Permission denied messages must carry the name of the path",
e.getMessage().contains(src.getName()));
}
try {
retFuture = adfs.setOwner(src, "user1", "group2");
retFuture.get();
} catch (ExecutionException e) {
TestAsyncDFS.checkPermissionDenied(e, src, user1);
assertTrue("Permission denied messages must carry the name of the path",
e.getMessage().contains(src.getName()));
} finally {
if (rootFs != null) {
rootFs.close();
}
if (cluster != null) {
cluster.shutdown();
}
}
} }
} }