HDFS-13106. Need to exercise all HDFS APIs for EC. Contributed by Haibo Yan.

This commit is contained in:
Sean Mackrory 2018-02-06 16:33:11 -07:00
parent 4d1bf33d0f
commit b14db95c35
1 changed files with 546 additions and 0 deletions

View File

@ -0,0 +1,546 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hdfs;
import com.google.common.collect.Lists;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.crypto.key.JavaKeyStoreProvider;
import org.apache.hadoop.fs.*;
import org.apache.hadoop.fs.permission.AclEntry;
import org.apache.hadoop.fs.permission.AclStatus;
import org.apache.hadoop.fs.permission.FsAction;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.client.CreateEncryptionZoneFlag;
import org.apache.hadoop.hdfs.client.HdfsAdmin;
import org.apache.hadoop.hdfs.protocol.*;
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
import org.apache.hadoop.io.erasurecode.ECSchema;
import org.apache.hadoop.security.Credentials;
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.security.token.Token;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.DataOutputStream;
import java.io.IOException;
import java.nio.file.Paths;
import java.security.NoSuchAlgorithmException;
import java.security.PrivilegedExceptionAction;
import java.util.EnumSet;
import java.util.List;
import java.util.Map;
import static org.apache.hadoop.fs.permission.AclEntryScope.ACCESS;
import static org.apache.hadoop.fs.permission.AclEntryScope.DEFAULT;
import static org.apache.hadoop.fs.permission.AclEntryType.GROUP;
import static org.apache.hadoop.fs.permission.AclEntryType.OTHER;
import static org.apache.hadoop.fs.permission.AclEntryType.USER;
import static org.apache.hadoop.fs.permission.FsAction.ALL;
import static org.apache.hadoop.fs.permission.FsAction.NONE;
import static org.apache.hadoop.fs.permission.FsAction.READ_EXECUTE;
import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry;
import static org.junit.Assert.*;
/**
* Test after enable Erasure Coding on cluster, exercise Java API make sure they
* are working as expected.
*
*/
public class TestErasureCodingExerciseAPIs {
private Configuration conf;
private MiniDFSCluster cluster;
private DistributedFileSystem fs;
private HdfsAdmin dfsAdmin;
private FileSystemTestWrapper fsWrapper;
private static final int BLOCK_SIZE = 1 << 14; // 16k
private ErasureCodingPolicy ecPolicy;
private static ErasureCodingPolicy getEcPolicy() {
return StripedFileTestUtil.getDefaultECPolicy();
}
private static final Logger LOG =
LoggerFactory.getLogger(TestErasureCodingExerciseAPIs.class);
@Before
public void setupCluster() throws IOException {
ecPolicy = getEcPolicy();
conf = new HdfsConfiguration();
conf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, BLOCK_SIZE);
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true);
// Set up java key store
String testRootDir = Paths.get(new FileSystemTestHelper().getTestRootDir())
.toString();
String keyProviderURI = JavaKeyStoreProvider.SCHEME_NAME + "://file"
+ new Path(testRootDir, "test.jks").toUri();
conf.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_KEY_PROVIDER_PATH,
keyProviderURI);
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_DELEGATION_TOKEN_ALWAYS_USE_KEY,
true);
cluster = new MiniDFSCluster.Builder(conf).
numDataNodes(ecPolicy.getNumDataUnits() + ecPolicy.getNumParityUnits()).
build();
cluster.waitActive();
fs = cluster.getFileSystem();
fsWrapper = new FileSystemTestWrapper(fs);
dfsAdmin = new HdfsAdmin(cluster.getURI(), conf);
DFSTestUtil.enableAllECPolicies(fs);
fs.setErasureCodingPolicy(new Path("/"), ecPolicy.getName());
}
/**
* FileSystem.[access, setOwner, setTime] API call should succeed without
* failure.
* @throws IOException if any IO operation failed.
* @throws InterruptedException
*/
@Test
public void testAccess() throws IOException, InterruptedException {
final Path p1 = new Path("/p1");
final String userName = "user1";
final String groupName = "group1";
fs.mkdir(p1, new FsPermission((short) 0444));
fs.setOwner(p1, userName, groupName);
UserGroupInformation userGroupInfo = UserGroupInformation
.createUserForTesting(userName, new String[]{groupName});
FileSystem userFs = userGroupInfo.doAs(
(PrivilegedExceptionAction<FileSystem>) () -> FileSystem.get(conf));
userFs.setOwner(p1, userName, groupName);
userFs.access(p1, FsAction.READ);
long mtime = System.currentTimeMillis() - 1000L;
long atime = System.currentTimeMillis() - 2000L;
fs.setTimes(p1, mtime, atime);
FileStatus fileStatus = fs.getFileStatus(p1);
assertEquals(userName, fileStatus.getOwner());
assertEquals(groupName, fileStatus.getGroup());
assertEquals(new FsPermission(FsAction.READ, FsAction.READ, FsAction.READ),
fileStatus.getPermission());
assertEquals(mtime, fileStatus.getModificationTime());
assertEquals(atime, fileStatus.getAccessTime());
}
/**
* FileSystem.[setQuota, getQuotaUsage, getContentSummary,
* setQuotaByStorageType] API call should succeed without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testQuota() throws IOException {
final Path qDir = new Path("/quotaDir");
fs.mkdirs(qDir);
fs.setQuota(qDir, 6, HdfsConstants.QUOTA_DONT_SET);
QuotaUsage usage = fs.getQuotaUsage(qDir);
assertEquals(fs.getContentSummary(qDir), usage);
fs.setQuotaByStorageType(qDir, StorageType.DEFAULT, 10);
}
/**
* FileSystem.[addCachePool, modifyCachePool,removeCachePool] API call
* should without failure. FileSystem.[addCacheDirective,
* modifyCacheDirective, removeCacheDirective] are noop.
* @throws IOException if any IO operation failed.
*/
@Test
public void testCache() throws IOException {
fs.addCachePool(new CachePoolInfo("pool1"));
fs.modifyCachePool(new CachePoolInfo("pool1"));
fs.removeCachePool("pool1");
fs.addCachePool(new CachePoolInfo("pool1"));
// Below calls should be noop.
long id = fs.addCacheDirective(new CacheDirectiveInfo.Builder()
.setPool("pool1").setPath(new Path("/pool2"))
.build());
RemoteIterator<CacheDirectiveEntry> iter = fs.listCacheDirectives(
new CacheDirectiveInfo.Builder().setPool("pool1").build());
assertTrue(iter.hasNext());
assertEquals("pool1", iter.next().getInfo().getPool());
fs.modifyCacheDirective(new CacheDirectiveInfo.Builder()
.setId(id).setReplication((short) 2).build());
fs.removeCacheDirective(id);
}
/**
* FileSystem.[addErasureCodingPolicies, disableErasureCodingPolicy,
* getErasureCodingPolicy, removeErasureCodingPolicy, setErasureCodingPolicy
* unsetErasureCodingPolicy] API call still should be succeed without
* failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testErasureCodingPolicy() throws IOException {
final Path tDir = new Path("/ecpDir");
fs.mkdirs(tDir);
ErasureCodingPolicy defaultPolicy
= SystemErasureCodingPolicies.getPolicies().get(0);
fs.setErasureCodingPolicy(tDir, defaultPolicy.getName());
ErasureCodingPolicy fPolicy = fs.getErasureCodingPolicy(tDir);
assertEquals(defaultPolicy, fPolicy);
final int cellSize = 1024 * 1024;
final ECSchema schema = new ECSchema("rs", 5, 3);
ErasureCodingPolicy newPolicy =
new ErasureCodingPolicy(schema, cellSize);
fs.addErasureCodingPolicies(new ErasureCodingPolicy[]{newPolicy});
assertEquals(SystemErasureCodingPolicies.getPolicies().size() + 1,
fs.getAllErasureCodingPolicies().size());
fs.disableErasureCodingPolicy(
ErasureCodingPolicy.composePolicyName(schema, cellSize));
assertEquals(SystemErasureCodingPolicies.getPolicies().size() + 1,
fs.getAllErasureCodingPolicies().size());
fs.unsetErasureCodingPolicy(tDir);
fPolicy = fs.getErasureCodingPolicy(tDir);
assertNotNull(fPolicy);
fs.removeErasureCodingPolicy(
ErasureCodingPolicy.composePolicyName(schema, cellSize));
assertEquals(SystemErasureCodingPolicies.getPolicies().size() + 1,
fs.getAllErasureCodingPolicies().size());
}
/**
* FileSystem.[getAclStatus, modifyAclEntries, removeAclEntries, removeAcl
* removeDefaultAcl] API call should succeed without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testACLAPI() throws IOException {
Path p = new Path("/aclTest");
fs.mkdirs(p, FsPermission.createImmutable((short) 0750));
List<AclEntry> aclSpec = Lists.newArrayList(
aclEntry(DEFAULT, USER, ALL),
aclEntry(DEFAULT, USER, "foo", ALL),
aclEntry(DEFAULT, GROUP, READ_EXECUTE),
aclEntry(DEFAULT, OTHER, NONE));
fs.setAcl(p, aclSpec);
AclStatus as = fs.getAclStatus(p);
for (AclEntry entry : aclSpec) {
assertTrue(String.format("as: %s, entry: %s", as, entry),
as.getEntries().contains(entry));
}
List<AclEntry> maclSpec = Lists.newArrayList(
aclEntry(ACCESS, USER, "bar", READ_EXECUTE),
aclEntry(DEFAULT, USER, "bar", READ_EXECUTE));
fs.modifyAclEntries(p, maclSpec);
as = fs.getAclStatus(p);
for (AclEntry entry : maclSpec) {
assertTrue(String.format("as: %s, entry: %s", as, entry),
as.getEntries().contains(entry));
}
fs.removeAclEntries(p, maclSpec);
fs.removeDefaultAcl(p);
fs.removeAcl(p);
assertEquals(0, fs.getAclStatus(p).getEntries().size());
}
/**
* FileSystem.[setXAttr, getXAttr, getXAttrs, removeXAttr, listXAttrs] API
* call should succeed without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testAttr() throws IOException {
final Path p = new Path("/attrTest");
fs.mkdirs(p);
final Path filePath = new Path(p, "file");
try (DataOutputStream dos = fs.create(filePath)) {
dos.writeBytes("write something");
}
final String name = "user.a1";
final byte[] value = {0x31, 0x32, 0x33};
fs.setXAttr(filePath, name, value, EnumSet.of(XAttrSetFlag.CREATE));
Map<String, byte[]> xattrs = fs.getXAttrs(filePath);
assertEquals(1, xattrs.size());
assertArrayEquals(value, xattrs.get(name));
assertArrayEquals(value, fs.getXAttr(filePath, name));
List<String> listXAttrs = fs.listXAttrs(filePath);
assertEquals(1, listXAttrs.size());
fs.removeXAttr(filePath, name);
xattrs = fs.getXAttrs(filePath);
assertEquals(0, xattrs.size());
listXAttrs = fs.listXAttrs(filePath);
assertEquals(0, listXAttrs.size());
}
/**
* FileSystem.[allowSnapshot, createSnapshot, deleteSnapshot,
* renameSnapshot, getSnapshotDiffReport, disallowSnapshot] API call should
* succeed without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testSnapshotAPI() throws IOException {
Path p = new Path("/snapshotTest");
fs.mkdirs(p);
fs.allowSnapshot(p);
fs.createSnapshot(p, "s1");
Path f = new Path("/snapshotTest/f1");
try (DataOutputStream dos = fs.create(f)) {
dos.writeBytes("write something");
}
fs.createSnapshot(p, "s2");
fs.renameSnapshot(p, "s2", "s3");
SnapshotDiffReport report = fs.getSnapshotDiffReport(p, "s1",
"s3");
assertEquals("s1", report.getFromSnapshot());
assertEquals("s3", report.getLaterSnapshotName());
fs.deleteSnapshot(p, "s1");
fs.deleteSnapshot(p, "s3");
fs.disallowSnapshot(p);
}
/**
* FileSystem.[createSymlink, getFileLinkStatus] API call should succeed
* without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testSymbolicLink() throws IOException {
Path p = new Path("/slTest");
fs.mkdirs(p);
Path f = new Path("/slTest/file");
try (DataOutputStream dos = fs.create(f)) {
dos.writeBytes("write something");
}
Path sl = new Path("/slTest1/sl");
fs.createSymlink(f, sl, true);
assertEquals(fs.getLinkTarget(sl), f);
FileStatus linkStatus = fs.getFileLinkStatus(sl);
assertTrue(linkStatus.isSymlink());
}
/**
* FileSystem.[create, open, append, concat, getFileChecksum, rename,
* delete] API call should succeed without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testFileOpsAPI() throws IOException {
Path p = new Path("/fileTest");
fs.mkdirs(p);
Path f1 = new Path(p, "file1");
Path fa = new Path(p, "filea");
try (DataOutputStream dos = fs.create(f1)) {
dos.writeBytes("create with some content");
}
try (DataOutputStream dos = fs.create(fa)) {
dos.writeBytes("create with some content");
}
// setReplication is a noop
short replication = fs.getDefaultReplication();
fs.setReplication(f1, (short) 5);
assertEquals(replication, fs.getDefaultReplication(f1));
BlockLocation[] locations = fs.getFileBlockLocations(f1, 0, 1);
assertEquals(1, locations.length);
FileStatus status1 = fs.getFileStatus(f1);
assertFalse(status1.isDirectory());
assertTrue(status1.getPath().toString().contains(p.toString()));
FileStatus statusa = fs.getFileStatus(fa);
assertFalse(statusa.isDirectory());
assertTrue(statusa.getPath().toString().contains(fa.toString()));
FileStatus[] statuses = fs.listStatus(p);
assertEquals(2, statuses.length);
assertEquals(status1, statuses[0]);
assertEquals(statusa, statuses[1]);
RemoteIterator<FileStatus> iter = fs.listStatusIterator(p);
assertEquals(status1, iter.next());
assertEquals(statusa, iter.next());
assertFalse(iter.hasNext());
Path[] concatPs = new Path[]{
new Path(p, "c1"),
new Path(p, "c2"),
new Path(p, "c3"),
};
for (Path cp : concatPs) {
try (DataOutputStream dos = fs.create(cp)) {
dos.writeBytes("concat some content");
}
}
fs.concat(f1, concatPs);
FileChecksum checksum1 = fs.getFileChecksum(f1);
Path f2 = new Path("/fileTest/file2");
fs.rename(f1, f2);
FileStatus fileStatus = fs.getFileStatus(f2);
assertTrue(fileStatus.getPath().toString().contains("/fileTest/file2"));
FileChecksum checksum2 = fs.getFileChecksum(f2);
assertEquals(checksum1, checksum2);
fs.delete(f2, true);
RemoteIterator<Path> corruptFileBlocks = fs.listCorruptFileBlocks(f2);
assertFalse(corruptFileBlocks.hasNext());
}
/**
* FileSystem.[createEncryptionZone, getLocatedBlocks, getEZForPath,
* reencryptEncryptionZone, addDelegationTokens] API call should succeed
* without failure.
* @throws IOException if any IO operation failed.
* @throws NoSuchAlgorithmException
*/
@Test
public void testEncryptionZone() throws IOException,
NoSuchAlgorithmException {
final Path zoneRoot = new Path("ecRoot");
final Path zonePath = new Path(zoneRoot, "/ec");
fsWrapper.mkdir(zonePath, FsPermission.getDirDefault(), true);
final String testKey = "test_key";
DFSTestUtil.createKey(testKey, cluster, conf);
final EnumSet<CreateEncryptionZoneFlag> noTrash =
EnumSet.of(CreateEncryptionZoneFlag.NO_TRASH);
dfsAdmin.createEncryptionZone(zonePath, testKey, noTrash);
final Path fp = new Path(zonePath, "encFile");
DFSTestUtil.createFile(fs, fp, 1 << 13, (short) 1, 0xFEEE);
LocatedBlocks blocks = fs.getClient().getLocatedBlocks(fp.toString(), 0);
FileEncryptionInfo fei = blocks.getFileEncryptionInfo();
assertEquals(testKey, fei.getKeyName());
EncryptionZone ez = fs.getEZForPath(fp);
assertEquals(zonePath.toString(), ez.getPath());
dfsAdmin.reencryptEncryptionZone(zonePath,
HdfsConstants.ReencryptAction.START);
Credentials creds = new Credentials();
final Token<?>[] tokens = fs.addDelegationTokens("JobTracker", creds);
assertEquals(1, tokens.length);
}
/**
* FileSystem.[setStoragePolicy, unsetStoragePolicy] API call should succeed
* without failure.
* @throws IOException if any IO operation failed.
*/
@Test
public void testStoragePolicy() throws IOException {
Path p = new Path("/storagePolicyTest");
fs.mkdirs(p);
final Path sp = new Path(p, "/sp");
try (DataOutputStream dos = fs.create(sp)) {
dos.writeBytes("create with some content");
}
final BlockStoragePolicySuite suite = BlockStoragePolicySuite
.createDefaultSuite();
final BlockStoragePolicy hot = suite.getPolicy("HOT");
fs.setStoragePolicy(sp, hot.getName());
assertEquals(fs.getStoragePolicy(sp), hot);
fs.unsetStoragePolicy(sp);
assertEquals(fs.getStoragePolicy(sp), hot);
}
/**
* append is not supported in EC.
* @throws IOException if any IO operation failed.
*/
@Test
public void testAppend() throws IOException {
Path p = new Path("/fileTest");
fs.mkdirs(p);
Path f = new Path("/fileTest/appendFile");
try (DataOutputStream dos = fs.create(f)) {
dos.writeBytes("create with some content");
}
try {
fs.append(f);
fail("append is not supported on erasure coded file");
} catch (IOException ioe) {
//Work as expected.
}
}
/**
* truncate is not supported in EC.
* @throws IOException if any IO operation failed.
*/
@Test
public void testTruncate() throws IOException {
Path p = new Path("/truncateTest");
fs.mkdirs(p);
Path f = new Path("/truncateTest/truncatefile");
try (DataOutputStream dos = fs.create(f)) {
dos.writeBytes("create with some content");
}
try {
fs.truncate(f, 0);
fail("truncate is not supported on erasure coded file.");
} catch (IOException ex) {
//Work as expected.
}
}
@After
public void shutdownCluster() {
if (cluster != null) {
cluster.shutdown();
cluster = null;
}
}
}