HDFS-15533: Provide DFS API compatible class, but use ViewFileSystemOverloadScheme inside. (#2229). Contributed by Uma Maheswara Rao G.
This commit is contained in:
parent
82ec28f442
commit
dd013f2fdf
|
@ -45,5 +45,4 @@ public interface FsConstants {
|
|||
String FS_VIEWFS_OVERLOAD_SCHEME_TARGET_FS_IMPL_PATTERN =
|
||||
"fs.viewfs.overload.scheme.target.%s.impl";
|
||||
String VIEWFS_TYPE = "viewfs";
|
||||
String VIEWFSOS_TYPE = "viewfsOverloadScheme";
|
||||
}
|
||||
|
|
|
@ -599,9 +599,10 @@ abstract class InodeTree<T> {
|
|||
|
||||
if (!gotMountTableEntry) {
|
||||
if (!initingUriAsFallbackOnNoMounts) {
|
||||
throw new IOException(
|
||||
"ViewFs: Cannot initialize: Empty Mount table in config for "
|
||||
+ "viewfs://" + mountTableName + "/");
|
||||
throw new IOException(new StringBuilder(
|
||||
"ViewFs: Cannot initialize: Empty Mount table in config for ")
|
||||
.append(theUri.getScheme()).append("://").append(mountTableName)
|
||||
.append("/").toString());
|
||||
}
|
||||
StringBuilder msg =
|
||||
new StringBuilder("Empty mount table detected for ").append(theUri)
|
||||
|
|
|
@ -259,13 +259,14 @@ public class ViewFileSystem extends FileSystem {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns the ViewFileSystem type.
|
||||
* @return <code>viewfs</code>
|
||||
* Returns false as it does not support to add fallback link automatically on
|
||||
* no mounts.
|
||||
*/
|
||||
String getType() {
|
||||
return FsConstants.VIEWFS_TYPE;
|
||||
boolean supportAutoAddingFallbackOnNoMounts() {
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Called after a new FileSystem instance is constructed.
|
||||
* @param theUri a uri whose authority section names the host, port, etc. for
|
||||
|
@ -293,7 +294,7 @@ public class ViewFileSystem extends FileSystem {
|
|||
try {
|
||||
myUri = new URI(getScheme(), authority, "/", null, null);
|
||||
boolean initingUriAsFallbackOnNoMounts =
|
||||
!FsConstants.VIEWFS_TYPE.equals(getType());
|
||||
supportAutoAddingFallbackOnNoMounts();
|
||||
fsState = new InodeTree<FileSystem>(conf, tableName, myUri,
|
||||
initingUriAsFallbackOnNoMounts) {
|
||||
@Override
|
||||
|
|
|
@ -104,6 +104,7 @@ import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN
|
|||
@InterfaceStability.Evolving
|
||||
public class ViewFileSystemOverloadScheme extends ViewFileSystem {
|
||||
private URI myUri;
|
||||
private boolean supportAutoAddingFallbackOnNoMounts = true;
|
||||
public ViewFileSystemOverloadScheme() throws IOException {
|
||||
super();
|
||||
}
|
||||
|
@ -114,11 +115,19 @@ public class ViewFileSystemOverloadScheme extends ViewFileSystem {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns the ViewFileSystem type.
|
||||
* @return <code>viewfs</code>
|
||||
* By default returns false as ViewFileSystemOverloadScheme supports auto
|
||||
* adding fallback on no mounts.
|
||||
*/
|
||||
String getType() {
|
||||
return FsConstants.VIEWFSOS_TYPE;
|
||||
public boolean supportAutoAddingFallbackOnNoMounts() {
|
||||
return this.supportAutoAddingFallbackOnNoMounts;
|
||||
}
|
||||
|
||||
/**
|
||||
* Sets whether to add fallback automatically when no mount points found.
|
||||
*/
|
||||
public void setSupportAutoAddingFallbackOnNoMounts(
|
||||
boolean addAutoFallbackOnNoMounts) {
|
||||
this.supportAutoAddingFallbackOnNoMounts = addAutoFallbackOnNoMounts;
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -287,4 +296,62 @@ public class ViewFileSystemOverloadScheme extends ViewFileSystem {
|
|||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Gets the mount path info, which contains the target file system and
|
||||
* remaining path to pass to the target file system.
|
||||
*/
|
||||
public MountPathInfo<FileSystem> getMountPathInfo(Path path,
|
||||
Configuration conf) throws IOException {
|
||||
InodeTree.ResolveResult<FileSystem> res;
|
||||
try {
|
||||
res = fsState.resolve(getUriPath(path), true);
|
||||
FileSystem fs = res.isInternalDir() ?
|
||||
(fsState.getRootFallbackLink() != null ?
|
||||
((ChRootedFileSystem) fsState
|
||||
.getRootFallbackLink().targetFileSystem).getMyFs() :
|
||||
fsGetter().get(path.toUri(), conf)) :
|
||||
((ChRootedFileSystem) res.targetFileSystem).getMyFs();
|
||||
return new MountPathInfo<FileSystem>(res.remainingPath, res.resolvedPath,
|
||||
fs);
|
||||
} catch (FileNotFoundException e) {
|
||||
// No link configured with passed path.
|
||||
throw new NotInMountpointException(path,
|
||||
"No link found for the given path.");
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* A class to maintain the target file system and a path to pass to the target
|
||||
* file system.
|
||||
*/
|
||||
public static class MountPathInfo<T> {
|
||||
private Path pathOnTarget;
|
||||
private T targetFs;
|
||||
|
||||
public MountPathInfo(Path pathOnTarget, String resolvedPath, T targetFs) {
|
||||
this.pathOnTarget = pathOnTarget;
|
||||
this.targetFs = targetFs;
|
||||
}
|
||||
|
||||
public Path getPathOnTarget() {
|
||||
return this.pathOnTarget;
|
||||
}
|
||||
|
||||
public T getTargetFs() {
|
||||
return this.targetFs;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* @return Gets the fallback file system configured. Usually, this will be the
|
||||
* default cluster.
|
||||
*/
|
||||
public FileSystem getFallbackFileSystem() {
|
||||
if (fsState.getRootFallbackLink() == null) {
|
||||
return null;
|
||||
}
|
||||
return ((ChRootedFileSystem) fsState.getRootFallbackLink().targetFileSystem)
|
||||
.getMyFs();
|
||||
}
|
||||
|
||||
}
|
|
@ -186,7 +186,7 @@ public class DistributedFileSystem extends FileSystem
|
|||
throw new IOException("Incomplete HDFS URI, no host: "+ uri);
|
||||
}
|
||||
|
||||
this.dfs = new DFSClient(uri, conf, statistics);
|
||||
initDFSClient(uri, conf);
|
||||
this.uri = URI.create(uri.getScheme()+"://"+uri.getAuthority());
|
||||
this.workingDir = getHomeDirectory();
|
||||
|
||||
|
@ -200,6 +200,10 @@ public class DistributedFileSystem extends FileSystem
|
|||
});
|
||||
}
|
||||
|
||||
void initDFSClient(URI theUri, Configuration conf) throws IOException {
|
||||
this.dfs = new DFSClient(theUri, conf, statistics);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Path getWorkingDirectory() {
|
||||
return workingDir;
|
||||
|
@ -1510,10 +1514,14 @@ public class DistributedFileSystem extends FileSystem
|
|||
@Override
|
||||
public void close() throws IOException {
|
||||
try {
|
||||
dfs.closeOutputStreams(false);
|
||||
if (dfs != null) {
|
||||
dfs.closeOutputStreams(false);
|
||||
}
|
||||
super.close();
|
||||
} finally {
|
||||
dfs.close();
|
||||
if (dfs != null) {
|
||||
dfs.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -41,7 +41,6 @@ import org.apache.hadoop.ha.ServiceFailedException;
|
|||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.DFSUtil;
|
||||
import org.apache.hadoop.hdfs.DFSUtilClient;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.HAUtil;
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
|
@ -727,11 +726,6 @@ public class NameNode extends ReconfigurableBase implements
|
|||
intervals);
|
||||
}
|
||||
}
|
||||
// Currently NN uses FileSystem.get to initialize DFS in startTrashEmptier.
|
||||
// If fs.hdfs.impl was overridden by core-site.xml, we may get other
|
||||
// filesystem. To make sure we get DFS, we are setting fs.hdfs.impl to DFS.
|
||||
// HDFS-15450
|
||||
conf.set(FS_HDFS_IMPL_KEY, DistributedFileSystem.class.getName());
|
||||
|
||||
UserGroupInformation.setConfiguration(conf);
|
||||
loginAsNameNodeUser(conf);
|
||||
|
|
|
@ -37,8 +37,8 @@ public class TestViewFSOverloadSchemeWithMountTableConfigInHDFS
|
|||
|
||||
@Before
|
||||
@Override
|
||||
public void startCluster() throws IOException {
|
||||
super.startCluster();
|
||||
public void setUp() throws IOException {
|
||||
super.setUp();
|
||||
String mountTableDir =
|
||||
URI.create(getConf().get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY))
|
||||
.toString() + "/MountTable/";
|
||||
|
|
|
@ -36,13 +36,15 @@ import org.apache.hadoop.fs.FsConstants;
|
|||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.RawLocalFileSystem;
|
||||
import org.apache.hadoop.fs.UnsupportedFileSystemException;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.test.LambdaTestUtils;
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME;
|
||||
|
@ -58,7 +60,7 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
private static final String FS_IMPL_PATTERN_KEY = "fs.%s.impl";
|
||||
private static final String HDFS_SCHEME = "hdfs";
|
||||
private Configuration conf = null;
|
||||
private MiniDFSCluster cluster = null;
|
||||
private static MiniDFSCluster cluster = null;
|
||||
private URI defaultFSURI;
|
||||
private File localTargetDir;
|
||||
private static final String TEST_ROOT_DIR = PathUtils
|
||||
|
@ -66,33 +68,52 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
private static final String HDFS_USER_FOLDER = "/HDFSUser";
|
||||
private static final String LOCAL_FOLDER = "/local";
|
||||
|
||||
@BeforeClass
|
||||
public static void init() throws IOException {
|
||||
cluster =
|
||||
new MiniDFSCluster.Builder(new Configuration()).numDataNodes(2).build();
|
||||
cluster.waitClusterUp();
|
||||
}
|
||||
|
||||
/**
|
||||
* Sets up the configurations and starts the MiniDFSCluster.
|
||||
*/
|
||||
@Before
|
||||
public void startCluster() throws IOException {
|
||||
conf = new Configuration();
|
||||
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_DELEGATION_TOKEN_ALWAYS_USE_KEY,
|
||||
true);
|
||||
conf.setInt(
|
||||
public void setUp() throws IOException {
|
||||
Configuration config = getNewConf();
|
||||
config.setInt(
|
||||
CommonConfigurationKeysPublic.IPC_CLIENT_CONNECT_MAX_RETRIES_KEY, 1);
|
||||
conf.set(String.format(FS_IMPL_PATTERN_KEY, HDFS_SCHEME),
|
||||
config.set(String.format(FS_IMPL_PATTERN_KEY, HDFS_SCHEME),
|
||||
ViewFileSystemOverloadScheme.class.getName());
|
||||
conf.set(String.format(
|
||||
FsConstants.FS_VIEWFS_OVERLOAD_SCHEME_TARGET_FS_IMPL_PATTERN,
|
||||
HDFS_SCHEME), DistributedFileSystem.class.getName());
|
||||
conf.setBoolean(CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME,
|
||||
config.setBoolean(CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME,
|
||||
CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME_DEFAULT);
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||
cluster.waitClusterUp();
|
||||
setConf(config);
|
||||
defaultFSURI =
|
||||
URI.create(conf.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
URI.create(config.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
localTargetDir = new File(TEST_ROOT_DIR, "/root/");
|
||||
localTargetDir.mkdirs();
|
||||
Assert.assertEquals(HDFS_SCHEME, defaultFSURI.getScheme()); // hdfs scheme.
|
||||
}
|
||||
|
||||
@After
|
||||
public void tearDown() throws IOException {
|
||||
public void cleanUp() throws IOException {
|
||||
if (cluster != null) {
|
||||
FileSystem fs = new DistributedFileSystem();
|
||||
fs.initialize(defaultFSURI, conf);
|
||||
try {
|
||||
FileStatus[] statuses = fs.listStatus(new Path("/"));
|
||||
for (FileStatus st : statuses) {
|
||||
Assert.assertTrue(fs.delete(st.getPath(), true));
|
||||
}
|
||||
} finally {
|
||||
fs.close();
|
||||
}
|
||||
FileSystem.closeAll();
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDown() throws IOException {
|
||||
if (cluster != null) {
|
||||
FileSystem.closeAll();
|
||||
cluster.shutdown();
|
||||
|
@ -132,9 +153,9 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
// /local/test
|
||||
Path localDir = new Path(LOCAL_FOLDER + "/test");
|
||||
|
||||
try (ViewFileSystemOverloadScheme fs
|
||||
= (ViewFileSystemOverloadScheme) FileSystem.get(conf)) {
|
||||
Assert.assertEquals(2, fs.getMountPoints().length);
|
||||
try (FileSystem fs
|
||||
= FileSystem.get(conf)) {
|
||||
Assert.assertEquals(2, fs.getChildFileSystems().length);
|
||||
fs.createNewFile(hdfsFile); // /HDFSUser/testfile
|
||||
fs.mkdirs(localDir); // /local/test
|
||||
}
|
||||
|
@ -166,8 +187,13 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
* hdfs://localhost:xxx/HDFSUser --> nonexistent://NonExistent/User/
|
||||
* It should fail to add non existent fs link.
|
||||
*/
|
||||
@Test(expected = IOException.class, timeout = 30000)
|
||||
@Test(timeout = 30000)
|
||||
public void testMountLinkWithNonExistentLink() throws Exception {
|
||||
testMountLinkWithNonExistentLink(true);
|
||||
}
|
||||
|
||||
public void testMountLinkWithNonExistentLink(boolean expectFsInitFailure)
|
||||
throws Exception {
|
||||
final String userFolder = "/User";
|
||||
final Path nonExistTargetPath =
|
||||
new Path("nonexistent://NonExistent" + userFolder);
|
||||
|
@ -176,10 +202,17 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
* Below addLink will create following mount points
|
||||
* hdfs://localhost:xxx/User --> nonexistent://NonExistent/User/
|
||||
*/
|
||||
addMountLinks(defaultFSURI.getAuthority(), new String[] {userFolder },
|
||||
new String[] {nonExistTargetPath.toUri().toString() }, conf);
|
||||
FileSystem.get(conf);
|
||||
Assert.fail("Expected to fail with non existent link");
|
||||
addMountLinks(defaultFSURI.getAuthority(), new String[] {userFolder},
|
||||
new String[] {nonExistTargetPath.toUri().toString()}, conf);
|
||||
if (expectFsInitFailure) {
|
||||
LambdaTestUtils.intercept(IOException.class, () -> {
|
||||
FileSystem.get(conf);
|
||||
});
|
||||
} else {
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals("hdfs", fs.getScheme());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -271,14 +304,10 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
|
||||
// check for viewfs path without authority
|
||||
Path viewFsRootPath = new Path("viewfs:/");
|
||||
try {
|
||||
viewFsRootPath.getFileSystem(conf);
|
||||
Assert.fail(
|
||||
"Mount table with authority default should not be initialized");
|
||||
} catch (IOException e) {
|
||||
assertTrue(e.getMessage().contains(
|
||||
"Empty Mount table in config for viewfs://default/"));
|
||||
}
|
||||
LambdaTestUtils.intercept(IOException.class,
|
||||
"Empty Mount table in config for viewfs://default", () -> {
|
||||
viewFsRootPath.getFileSystem(conf);
|
||||
});
|
||||
|
||||
// set the name of the default mount table here and
|
||||
// subsequent calls should succeed.
|
||||
|
@ -334,18 +363,25 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
*
|
||||
* It cannot find any mount link. ViewFS expects a mount point from root.
|
||||
*/
|
||||
@Test(expected = NotInMountpointException.class, timeout = 30000)
|
||||
public void testCreateOnRootShouldFailWhenMountLinkConfigured()
|
||||
throws Exception {
|
||||
@Test(timeout = 30000)
|
||||
public void testCreateOnRoot() throws Exception {
|
||||
testCreateOnRoot(false);
|
||||
}
|
||||
|
||||
public void testCreateOnRoot(boolean fallbackExist) throws Exception {
|
||||
final Path hdfsTargetPath = new Path(defaultFSURI + HDFS_USER_FOLDER);
|
||||
addMountLinks(defaultFSURI.getAuthority(),
|
||||
new String[] {HDFS_USER_FOLDER, LOCAL_FOLDER },
|
||||
new String[] {HDFS_USER_FOLDER, LOCAL_FOLDER},
|
||||
new String[] {hdfsTargetPath.toUri().toString(),
|
||||
localTargetDir.toURI().toString() },
|
||||
conf);
|
||||
localTargetDir.toURI().toString()}, conf);
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
fs.createNewFile(new Path("/newFileOnRoot"));
|
||||
Assert.fail("It should fail as root is read only in viewFS.");
|
||||
if (fallbackExist) {
|
||||
Assert.assertTrue(fs.createNewFile(new Path("/newFileOnRoot")));
|
||||
} else {
|
||||
LambdaTestUtils.intercept(NotInMountpointException.class, () -> {
|
||||
fs.createNewFile(new Path("/newFileOnRoot"));
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -433,15 +469,13 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
conf);
|
||||
|
||||
// 1. Only 1 hdfs child file system should be there with cache.
|
||||
try (ViewFileSystemOverloadScheme vfs =
|
||||
(ViewFileSystemOverloadScheme) FileSystem.get(conf)) {
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(1, vfs.getChildFileSystems().length);
|
||||
}
|
||||
|
||||
// 2. Two hdfs file systems should be there if no cache.
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
try (ViewFileSystemOverloadScheme vfs =
|
||||
(ViewFileSystemOverloadScheme) FileSystem.get(conf)) {
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(2, vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
|
@ -466,8 +500,7 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
// Two hdfs file systems should be there if no cache.
|
||||
try (ViewFileSystemOverloadScheme vfs =
|
||||
(ViewFileSystemOverloadScheme) FileSystem.get(conf)) {
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(2, vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
|
@ -494,8 +527,7 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
// Only one local file system should be there if no InnerCache, but fs
|
||||
// cache should work.
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
try (ViewFileSystemOverloadScheme vfs =
|
||||
(ViewFileSystemOverloadScheme) FileSystem.get(conf)) {
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(1, vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
|
@ -656,4 +688,18 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
|||
public Configuration getConf() {
|
||||
return this.conf;
|
||||
}
|
||||
|
||||
/**
|
||||
* @return configuration.
|
||||
*/
|
||||
public Configuration getNewConf() {
|
||||
return new Configuration(cluster.getConfiguration(0));
|
||||
}
|
||||
|
||||
/**
|
||||
* sets configuration.
|
||||
*/
|
||||
public void setConf(Configuration config) {
|
||||
conf = config;
|
||||
}
|
||||
}
|
|
@ -142,7 +142,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
private boolean noXmlDefaults = false;
|
||||
|
||||
private HdfsConfiguration getTestConfiguration() {
|
||||
HdfsConfiguration getTestConfiguration() {
|
||||
HdfsConfiguration conf;
|
||||
if (noXmlDefaults) {
|
||||
conf = new HdfsConfiguration(false);
|
||||
|
@ -813,7 +813,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testStatistics2() throws IOException, NoSuchAlgorithmException {
|
||||
HdfsConfiguration conf = new HdfsConfiguration();
|
||||
HdfsConfiguration conf = getTestConfiguration();
|
||||
conf.set(DFSConfigKeys.DFS_STORAGE_POLICY_SATISFIER_MODE_KEY,
|
||||
StoragePolicySatisfierMode.EXTERNAL.toString());
|
||||
File tmpDir = GenericTestUtils.getTestDir(UUID.randomUUID().toString());
|
||||
|
@ -1475,7 +1475,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test(timeout=60000)
|
||||
public void testFileCloseStatus() throws IOException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
DistributedFileSystem fs = cluster.getFileSystem();
|
||||
try {
|
||||
|
@ -1495,7 +1495,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testCreateWithStoragePolicy() throws Throwable {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.storageTypes(
|
||||
new StorageType[] {StorageType.DISK, StorageType.ARCHIVE,
|
||||
|
@ -1534,7 +1534,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test(timeout=60000)
|
||||
public void testListFiles() throws IOException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
|
||||
try {
|
||||
|
@ -1557,7 +1557,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testListStatusOfSnapshotDirs() throws IOException {
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(new HdfsConfiguration())
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(getTestConfiguration())
|
||||
.build();
|
||||
try {
|
||||
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||
|
@ -1577,7 +1577,7 @@ public class TestDistributedFileSystem {
|
|||
@Test(timeout=10000)
|
||||
public void testDFSClientPeerReadTimeout() throws IOException {
|
||||
final int timeout = 1000;
|
||||
final Configuration conf = new HdfsConfiguration();
|
||||
final Configuration conf = getTestConfiguration();
|
||||
conf.setInt(HdfsClientConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY, timeout);
|
||||
|
||||
// only need cluster to create a dfs client to get a peer
|
||||
|
@ -1611,7 +1611,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test(timeout=60000)
|
||||
public void testGetServerDefaults() throws IOException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
try {
|
||||
cluster.waitActive();
|
||||
|
@ -1626,7 +1626,7 @@ public class TestDistributedFileSystem {
|
|||
@Test(timeout=10000)
|
||||
public void testDFSClientPeerWriteTimeout() throws IOException {
|
||||
final int timeout = 1000;
|
||||
final Configuration conf = new HdfsConfiguration();
|
||||
final Configuration conf = getTestConfiguration();
|
||||
conf.setInt(HdfsClientConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY, timeout);
|
||||
|
||||
// only need cluster to create a dfs client to get a peer
|
||||
|
@ -1663,7 +1663,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test(timeout = 30000)
|
||||
public void testTotalDfsUsed() throws Exception {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
MiniDFSCluster cluster = null;
|
||||
try {
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build();
|
||||
|
@ -1850,7 +1850,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testSuperUserPrivilege() throws Exception {
|
||||
HdfsConfiguration conf = new HdfsConfiguration();
|
||||
HdfsConfiguration conf = getTestConfiguration();
|
||||
File tmpDir = GenericTestUtils.getTestDir(UUID.randomUUID().toString());
|
||||
final Path jksPath = new Path(tmpDir.toString(), "test.jks");
|
||||
conf.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_KEY_PROVIDER_PATH,
|
||||
|
@ -1903,7 +1903,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testListingStoragePolicyNonSuperUser() throws Exception {
|
||||
HdfsConfiguration conf = new HdfsConfiguration();
|
||||
HdfsConfiguration conf = getTestConfiguration();
|
||||
try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build()) {
|
||||
cluster.waitActive();
|
||||
final DistributedFileSystem dfs = cluster.getFileSystem();
|
||||
|
@ -2055,7 +2055,7 @@ public class TestDistributedFileSystem {
|
|||
@Test
|
||||
public void testStorageFavouredNodes()
|
||||
throws IOException, InterruptedException, TimeoutException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.storageTypes(new StorageType[] {StorageType.SSD, StorageType.DISK})
|
||||
.numDataNodes(3).storagesPerDatanode(2).build()) {
|
||||
|
@ -2080,7 +2080,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testGetECTopologyResultForPolicies() throws Exception {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
Configuration conf = getTestConfiguration();
|
||||
try (MiniDFSCluster cluster = DFSTestUtil.setupCluster(conf, 9, 3, 0)) {
|
||||
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||
dfs.enableErasureCodingPolicy("RS-6-3-1024k");
|
||||
|
@ -2111,7 +2111,7 @@ public class TestDistributedFileSystem {
|
|||
|
||||
@Test
|
||||
public void testECCloseCommittedBlock() throws Exception {
|
||||
HdfsConfiguration conf = new HdfsConfiguration();
|
||||
HdfsConfiguration conf = getTestConfiguration();
|
||||
conf.setInt(DFS_NAMENODE_FILE_CLOSE_NUM_COMMITTED_ALLOWED_KEY, 1);
|
||||
try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
|
||||
.numDataNodes(3).build()) {
|
||||
|
|
|
@ -0,0 +1,47 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||
import org.apache.hadoop.test.Whitebox;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestViewDistributedFileSystem extends TestDistributedFileSystem{
|
||||
@Override
|
||||
HdfsConfiguration getTestConfiguration() {
|
||||
HdfsConfiguration conf = super.getTestConfiguration();
|
||||
conf.set("fs.hdfs.impl", ViewDistributedFileSystem.class.getName());
|
||||
return conf;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void testStatistics() throws IOException {
|
||||
FileSystem.getStatistics(HdfsConstants.HDFS_URI_SCHEME,
|
||||
ViewDistributedFileSystem.class).reset();
|
||||
@SuppressWarnings("unchecked")
|
||||
ThreadLocal<FileSystem.Statistics.StatisticsData> data =
|
||||
(ThreadLocal<FileSystem.Statistics.StatisticsData>) Whitebox
|
||||
.getInternalState(FileSystem
|
||||
.getStatistics(HdfsConstants.HDFS_URI_SCHEME,
|
||||
ViewDistributedFileSystem.class), "threadData");
|
||||
data.set(null);
|
||||
super.testStatistics();
|
||||
}
|
||||
}
|
|
@ -0,0 +1,94 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.FileSystemContractBaseTest;
|
||||
import org.apache.hadoop.fs.viewfs.ConfigUtil;
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.apache.hadoop.test.LambdaTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
|
||||
public class TestViewDistributedFileSystemContract
|
||||
extends TestHDFSFileSystemContract {
|
||||
private static MiniDFSCluster cluster;
|
||||
private static String defaultWorkingDirectory;
|
||||
private static Configuration conf = new HdfsConfiguration();
|
||||
|
||||
@BeforeClass
|
||||
public static void init() throws IOException {
|
||||
final File basedir = GenericTestUtils.getRandomizedTestDir();
|
||||
conf.set(CommonConfigurationKeys.FS_PERMISSIONS_UMASK_KEY,
|
||||
FileSystemContractBaseTest.TEST_UMASK);
|
||||
cluster = new MiniDFSCluster.Builder(conf, basedir)
|
||||
.numDataNodes(2)
|
||||
.build();
|
||||
defaultWorkingDirectory =
|
||||
"/user/" + UserGroupInformation.getCurrentUser().getShortUserName();
|
||||
}
|
||||
|
||||
@Before
|
||||
public void setUp() throws Exception {
|
||||
conf.set("fs.hdfs.impl", ViewDistributedFileSystem.class.getName());
|
||||
URI defaultFSURI =
|
||||
URI.create(conf.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
ConfigUtil.addLink(conf, defaultFSURI.getHost(), "/user",
|
||||
defaultFSURI);
|
||||
ConfigUtil.addLinkFallback(conf, defaultFSURI.getHost(),
|
||||
defaultFSURI);
|
||||
fs = FileSystem.get(conf);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDownAfter() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
cluster = null;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
protected String getDefaultWorkingDirectory() {
|
||||
return defaultWorkingDirectory;
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testRenameRootDirForbidden() throws Exception {
|
||||
LambdaTestUtils.intercept(AccessControlException.class,
|
||||
"InternalDir of ViewFileSystem is readonly", () -> {
|
||||
super.testRenameRootDirForbidden();
|
||||
});
|
||||
}
|
||||
|
||||
@Ignore("Ignore this test until HDFS-15532")
|
||||
@Override
|
||||
public void testLSRootDir() throws Throwable {
|
||||
}
|
||||
}
|
|
@ -0,0 +1,64 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.viewfs.ConfigUtil;
|
||||
import org.apache.hadoop.fs.viewfs.TestViewFileSystemOverloadSchemeWithHdfsScheme;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME;
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME_DEFAULT;
|
||||
|
||||
public class TestViewDistributedFileSystemWithMountLinks extends
|
||||
TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
||||
@Override
|
||||
public void setUp() throws IOException {
|
||||
super.setUp();
|
||||
Configuration conf = getConf();
|
||||
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_DELEGATION_TOKEN_ALWAYS_USE_KEY,
|
||||
true);
|
||||
conf.setInt(
|
||||
CommonConfigurationKeysPublic.IPC_CLIENT_CONNECT_MAX_RETRIES_KEY, 1);
|
||||
conf.set("fs.hdfs.impl",
|
||||
ViewDistributedFileSystem.class.getName());
|
||||
conf.setBoolean(CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME,
|
||||
CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME_DEFAULT);
|
||||
URI defaultFSURI =
|
||||
URI.create(conf.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
ConfigUtil.addLinkFallback(conf, defaultFSURI.getAuthority(),
|
||||
new Path(defaultFSURI.toString()).toUri());
|
||||
setConf(conf);
|
||||
}
|
||||
|
||||
@Test(timeout = 30000)
|
||||
public void testCreateOnRoot() throws Exception {
|
||||
testCreateOnRoot(true);
|
||||
}
|
||||
|
||||
@Test(timeout = 30000)
|
||||
public void testMountLinkWithNonExistentLink() throws Exception {
|
||||
testMountLinkWithNonExistentLink(false);
|
||||
}
|
||||
}
|
|
@ -132,17 +132,23 @@ public class TestCacheDirectives {
|
|||
conf.setInt(DFSConfigKeys.DFS_NAMENODE_LIST_CACHE_POOLS_NUM_RESPONSES, 2);
|
||||
conf.setInt(DFSConfigKeys.DFS_NAMENODE_LIST_CACHE_DIRECTIVES_NUM_RESPONSES,
|
||||
2);
|
||||
|
||||
return conf;
|
||||
}
|
||||
|
||||
/**
|
||||
* @return the configuration.
|
||||
*/
|
||||
Configuration getConf() {
|
||||
return this.conf;
|
||||
}
|
||||
|
||||
@Before
|
||||
public void setup() throws Exception {
|
||||
conf = createCachingConf();
|
||||
cluster =
|
||||
new MiniDFSCluster.Builder(conf).numDataNodes(NUM_DATANODES).build();
|
||||
cluster.waitActive();
|
||||
dfs = cluster.getFileSystem();
|
||||
dfs = getDFS();
|
||||
proto = cluster.getNameNodeRpc();
|
||||
namenode = cluster.getNameNode();
|
||||
prevCacheManipulator = NativeIO.POSIX.getCacheManipulator();
|
||||
|
@ -150,6 +156,13 @@ public class TestCacheDirectives {
|
|||
BlockReaderTestUtil.enableHdfsCachingTracing();
|
||||
}
|
||||
|
||||
/**
|
||||
* @return the dfs instance.
|
||||
*/
|
||||
DistributedFileSystem getDFS() throws IOException {
|
||||
return (DistributedFileSystem) FileSystem.get(conf);
|
||||
}
|
||||
|
||||
@After
|
||||
public void teardown() throws Exception {
|
||||
// Remove cache directives left behind by tests so that we release mmaps.
|
||||
|
@ -1613,6 +1626,14 @@ public class TestCacheDirectives {
|
|||
"testAddingCacheDirectiveInfosWhenCachingIsDisabled:2");
|
||||
}
|
||||
|
||||
/**
|
||||
* @return the dfs instance for nnIdx.
|
||||
*/
|
||||
DistributedFileSystem getDFS(MiniDFSCluster cluster, int nnIdx)
|
||||
throws IOException {
|
||||
return cluster.getFileSystem(0);
|
||||
}
|
||||
|
||||
@Test(timeout=120000)
|
||||
public void testExpiryTimeConsistency() throws Exception {
|
||||
conf.setInt(DFSConfigKeys.DFS_HA_LOGROLL_PERIOD_KEY, 1);
|
||||
|
@ -1623,7 +1644,7 @@ public class TestCacheDirectives {
|
|||
.build();
|
||||
dfsCluster.transitionToActive(0);
|
||||
|
||||
DistributedFileSystem fs = dfsCluster.getFileSystem(0);
|
||||
DistributedFileSystem fs = getDFS(dfsCluster, 0);
|
||||
final NameNode ann = dfsCluster.getNameNode(0);
|
||||
|
||||
final Path filename = new Path("/file");
|
||||
|
|
|
@ -0,0 +1,56 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
* <p>
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
* <p>
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hdfs.server.namenode;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.viewfs.ConfigUtil;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.ViewDistributedFileSystem;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
|
||||
public class TestCacheDirectivesWithViewDFS extends TestCacheDirectives {
|
||||
|
||||
@Override
|
||||
public DistributedFileSystem getDFS() throws IOException {
|
||||
Configuration conf = getConf();
|
||||
conf.set("fs.hdfs.impl", ViewDistributedFileSystem.class.getName());
|
||||
URI defaultFSURI =
|
||||
URI.create(conf.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
ConfigUtil.addLinkFallback(conf, defaultFSURI.getHost(),
|
||||
new Path(defaultFSURI.toString()).toUri());
|
||||
ConfigUtil.addLink(conf, defaultFSURI.getHost(), "/tmp",
|
||||
new Path(defaultFSURI.toString()).toUri());
|
||||
return super.getDFS();
|
||||
}
|
||||
|
||||
@Override
|
||||
public DistributedFileSystem getDFS(MiniDFSCluster cluster, int nnIdx)
|
||||
throws IOException {
|
||||
Configuration conf = cluster.getConfiguration(nnIdx);
|
||||
conf.set("fs.hdfs.impl", ViewDistributedFileSystem.class.getName());
|
||||
URI uri = cluster.getURI(0);
|
||||
ConfigUtil.addLinkFallback(conf, uri.getHost(), uri);
|
||||
ConfigUtil.addLink(conf, uri.getHost(), "/tmp", uri);
|
||||
return cluster.getFileSystem(0);
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue