diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java index 7af5fa74e76..cb3fb862325 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java @@ -23,6 +23,7 @@ import java.lang.reflect.Constructor; import java.net.URI; import java.net.URISyntaxException; import java.util.ArrayList; +import java.util.Collection; import java.util.EnumSet; import java.util.HashMap; import java.util.List; @@ -1221,6 +1222,32 @@ public abstract class AbstractFileSystem { + " doesn't support deleteSnapshot"); } + /** + * Set the storage policy for a given file or directory. + * + * @param path file or directory path. + * @param policyName the name of the target storage policy. The list + * of supported Storage policies can be retrieved + * via {@link #getAllStoragePolicies}. + */ + public void setStoragePolicy(final Path path, final String policyName) + throws IOException { + throw new UnsupportedOperationException(getClass().getSimpleName() + + " doesn't support setStoragePolicy"); + } + + /** + * Retrieve all the storage policies supported by this file system. + * + * @return all storage policies supported by this filesystem. + * @throws IOException + */ + public Collection getAllStoragePolicies() + throws IOException { + throw new UnsupportedOperationException(getClass().getSimpleName() + + " doesn't support getAllStoragePolicies"); + } + @Override //Object public int hashCode() { return myUri.hashCode(); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/BlockStoragePolicySpi.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/BlockStoragePolicySpi.java new file mode 100644 index 00000000000..1d6502e3aec --- /dev/null +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/BlockStoragePolicySpi.java @@ -0,0 +1,72 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + +/** + * A storage policy specifies the placement of block replicas on specific + * storage types. + */ +@InterfaceAudience.Public +@InterfaceStability.Stable +public interface BlockStoragePolicySpi { + + /** + * Return the name of the storage policy. Policies are uniquely + * identified by name. + * + * @return the name of the storage policy. + */ + String getName(); + + /** + * Return the preferred storage types associated with this policy. These + * storage types are used sequentially for successive block replicas. + * + * @return preferred storage types used for placing block replicas. + */ + StorageType[] getStorageTypes(); + + /** + * Get the fallback storage types for creating new block replicas. Fallback + * storage types are used if the preferred storage types are not available. + * + * @return fallback storage types for new block replicas.. + */ + StorageType[] getCreationFallbacks(); + + /** + * Get the fallback storage types for replicating existing block replicas. + * Fallback storage types are used if the preferred storage types are not + * available. + * + * @return fallback storage types for replicating existing block replicas. + */ + StorageType[] getReplicationFallbacks(); + + /** + * Returns true if the policy is inherit-only and cannot be changed for + * an existing file. + * + * @return true if the policy is inherit-only. + */ + boolean isCopyOnCreateFile(); +} diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java index 16cb59124ef..122ddf68c78 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java @@ -24,6 +24,7 @@ import java.io.OutputStream; import java.net.URI; import java.security.PrivilegedExceptionAction; import java.util.ArrayList; +import java.util.Collection; import java.util.EnumSet; import java.util.HashSet; import java.util.IdentityHashMap; @@ -2665,4 +2666,36 @@ public class FileContext { } }.resolve(this, absF); } + + /** + * Set the storage policy for a given file or directory. + * + * @param path file or directory path. + * @param policyName the name of the target storage policy. The list + * of supported Storage policies can be retrieved + * via {@link #getAllStoragePolicies}. + */ + public void setStoragePolicy(final Path path, final String policyName) + throws IOException { + final Path absF = fixRelativePart(path); + new FSLinkResolver() { + @Override + public Void next(final AbstractFileSystem fs, final Path p) + throws IOException { + fs.setStoragePolicy(path, policyName); + return null; + } + }.resolve(this, absF); + } + + /** + * Retrieve all the storage policies supported by this file system. + * + * @return all storage policies supported by this filesystem. + * @throws IOException + */ + public Collection getAllStoragePolicies() + throws IOException { + return defaultFS.getAllStoragePolicies(); + } } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java index b508727229d..33d7c881cac 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java @@ -26,6 +26,7 @@ import java.net.URISyntaxException; import java.security.PrivilegedExceptionAction; import java.util.ArrayList; import java.util.Arrays; +import java.util.Collection; import java.util.EnumSet; import java.util.HashMap; import java.util.HashSet; @@ -2609,6 +2610,33 @@ public abstract class FileSystem extends Configured implements Closeable { + " doesn't support removeXAttr"); } + /** + * Set the storage policy for a given file or directory. + * + * @param src file or directory path. + * @param policyName the name of the target storage policy. The list + * of supported Storage policies can be retrieved + * via {@link #getAllStoragePolicies}. + * @throws IOException + */ + public void setStoragePolicy(final Path src, final String policyName) + throws IOException { + throw new UnsupportedOperationException(getClass().getSimpleName() + + " doesn't support setStoragePolicy"); + } + + /** + * Retrieve all the storage policies supported by this file system. + * + * @return all storage policies supported by this filesystem. + * @throws IOException + */ + public Collection getAllStoragePolicies() + throws IOException { + throw new UnsupportedOperationException(getClass().getSimpleName() + + " doesn't support getAllStoragePolicies"); + } + // making it volatile to be able to do a double checked locking private volatile static boolean FILE_SYSTEMS_LOADED = false; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFileSystem.java index ec056a4c610..11f3b23c33c 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFileSystem.java @@ -21,6 +21,7 @@ package org.apache.hadoop.fs; import java.io.*; import java.net.URI; import java.net.URISyntaxException; +import java.util.Collection; import java.util.EnumSet; import java.util.List; import java.util.Map; @@ -619,4 +620,16 @@ public class FilterFileSystem extends FileSystem { public void removeXAttr(Path path, String name) throws IOException { fs.removeXAttr(path, name); } + + @Override + public void setStoragePolicy(Path src, String policyName) + throws IOException { + fs.setStoragePolicy(src, policyName); + } + + @Override + public Collection getAllStoragePolicies() + throws IOException { + return fs.getAllStoragePolicies(); + } } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFs.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFs.java index 2ba63182355..539b26e166a 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFs.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFs.java @@ -20,6 +20,7 @@ import java.io.FileNotFoundException; import java.io.IOException; import java.net.URI; import java.net.URISyntaxException; +import java.util.Collection; import java.util.EnumSet; import java.util.List; import java.util.Map; @@ -397,4 +398,16 @@ public abstract class FilterFs extends AbstractFileSystem { throws IOException { myFs.deleteSnapshot(path, snapshotName); } + + @Override + public void setStoragePolicy(Path path, String policyName) + throws IOException { + myFs.setStoragePolicy(path, policyName); + } + + @Override + public Collection getAllStoragePolicies() + throws IOException { + return myFs.getAllStoragePolicies(); + } } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java index a05a7000f5b..4e5a0d55e1f 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java @@ -20,6 +20,7 @@ import java.io.FileNotFoundException; import java.io.IOException; import java.net.URI; import java.net.URISyntaxException; +import java.util.Collection; import java.util.EnumSet; import java.util.List; import java.util.Map; @@ -28,6 +29,7 @@ import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.AbstractFileSystem; import org.apache.hadoop.fs.BlockLocation; +import org.apache.hadoop.fs.BlockStoragePolicySpi; import org.apache.hadoop.fs.CreateFlag; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; @@ -377,6 +379,18 @@ class ChRootedFs extends AbstractFileSystem { myFs.deleteSnapshot(fullPath(snapshotDir), snapshotName); } + @Override + public void setStoragePolicy(Path path, String policyName) + throws IOException { + myFs.setStoragePolicy(fullPath(path), policyName); + } + + @Override + public Collection getAllStoragePolicies() + throws IOException { + return myFs.getAllStoragePolicies(); + } + @Override public void setVerifyChecksum(final boolean verifyChecksum) throws IOException, UnresolvedLinkException { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java index a23aa869ee3..bec292cd5b9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java @@ -740,6 +740,14 @@ public class ViewFs extends AbstractFileSystem { res.targetFileSystem.deleteSnapshot(res.remainingPath, snapshotName); } + @Override + public void setStoragePolicy(final Path path, final String policyName) + throws IOException { + InodeTree.ResolveResult res = + fsState.resolve(getUriPath(path), true); + res.targetFileSystem.setStoragePolicy(res.remainingPath, policyName); + } + /* * An instance of this class represents an internal dir of the viewFs * ie internal dir of the mount table. @@ -1070,5 +1078,11 @@ public class ViewFs extends AbstractFileSystem { checkPathIsSlash(path); throw readOnlyMountTable("deleteSnapshot", path); } + + @Override + public void setStoragePolicy(Path path, String policyName) + throws IOException { + throw readOnlyMountTable("setStoragePolicy", path); + } } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHarFileSystem.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHarFileSystem.java index 374bb2e8eb3..46f24fc6804 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHarFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHarFileSystem.java @@ -34,6 +34,7 @@ import org.junit.Test; import java.io.IOException; import java.lang.reflect.Method; import java.lang.reflect.Modifier; +import java.util.Collection; import java.util.EnumSet; import java.util.Iterator; import java.util.List; @@ -205,6 +206,12 @@ public class TestHarFileSystem { public AclStatus getAclStatus(Path path) throws IOException; public void access(Path path, FsAction mode) throws IOException; + + public void setStoragePolicy(Path src, String policyName) + throws IOException; + + public Collection getAllStoragePolicies() + throws IOException; } @Test diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/BlockStoragePolicy.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/BlockStoragePolicy.java index 9ecf6e8684b..2624960d93a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/BlockStoragePolicy.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/BlockStoragePolicy.java @@ -25,6 +25,7 @@ import java.util.List; import com.google.common.annotations.VisibleForTesting; import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.fs.BlockStoragePolicySpi; import org.apache.hadoop.fs.StorageType; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -34,7 +35,7 @@ import org.slf4j.LoggerFactory; * for the replicas of a block. */ @InterfaceAudience.Private -public class BlockStoragePolicy { +public class BlockStoragePolicy implements BlockStoragePolicySpi { public static final Logger LOG = LoggerFactory.getLogger(BlockStoragePolicy .class); @@ -239,18 +240,22 @@ public class BlockStoragePolicy { return id; } + @Override public String getName() { return name; } + @Override public StorageType[] getStorageTypes() { return this.storageTypes; } + @Override public StorageType[] getCreationFallbacks() { return this.creationFallbacks; } + @Override public StorageType[] getReplicationFallbacks() { return this.replicationFallbacks; } @@ -265,6 +270,7 @@ public class BlockStoragePolicy { return null; } + @Override public boolean isCopyOnCreateFile() { return copyOnCreateFile; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 8d0c5b6f106..4270a9c3b35 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -567,6 +567,9 @@ Release 2.8.0 - UNRELEASED HDFS-8157. Writes to RAM DISK reserve locked memory for block files. (Arpit Agarwal) + HDFS-8345. Storage policy APIs must be exposed via the FileSystem + interface. (Arpit Agarwal) + OPTIMIZATIONS HDFS-8026. Trace FSOutputSummer#writeChecksumChunks rather than diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/fs/Hdfs.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/fs/Hdfs.java index b7768493cbd..3f78b3182a0 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/fs/Hdfs.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/fs/Hdfs.java @@ -22,6 +22,8 @@ import java.io.IOException; import java.net.URI; import java.net.URISyntaxException; import java.util.ArrayList; +import java.util.Arrays; +import java.util.Collection; import java.util.EnumSet; import java.util.List; import java.util.Map; @@ -465,6 +467,17 @@ public class Hdfs extends AbstractFileSystem { dfs.checkAccess(getUriPath(path), mode); } + @Override + public void setStoragePolicy(Path path, String policyName) throws IOException { + dfs.setStoragePolicy(getUriPath(path), policyName); + } + + @Override + public Collection getAllStoragePolicies() + throws IOException { + return Arrays.asList(dfs.getStoragePolicies()); + } + /** * Renew an existing delegation token. * diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java index 4ca6d5731ed..902636ce609 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java @@ -23,6 +23,8 @@ import java.io.IOException; import java.net.InetSocketAddress; import java.net.URI; import java.util.ArrayList; +import java.util.Arrays; +import java.util.Collection; import java.util.EnumSet; import java.util.List; import java.util.Map; @@ -32,6 +34,7 @@ import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.BlockLocation; import org.apache.hadoop.fs.BlockStorageLocation; +import org.apache.hadoop.fs.BlockStoragePolicySpi; import org.apache.hadoop.fs.CacheFlag; import org.apache.hadoop.fs.ContentSummary; import org.apache.hadoop.fs.CreateFlag; @@ -532,6 +535,7 @@ public class DistributedFileSystem extends FileSystem { * @param src The source path referring to either a directory or a file. * @param policyName The name of the storage policy. */ + @Override public void setStoragePolicy(final Path src, final String policyName) throws IOException { statistics.incrementWriteOps(1); @@ -546,19 +550,24 @@ public class DistributedFileSystem extends FileSystem { @Override public Void next(final FileSystem fs, final Path p) throws IOException { - if (fs instanceof DistributedFileSystem) { - ((DistributedFileSystem) fs).setStoragePolicy(p, policyName); - return null; - } else { - throw new UnsupportedOperationException( - "Cannot perform setStoragePolicy on a non-DistributedFileSystem: " - + src + " -> " + p); - } + fs.setStoragePolicy(p, policyName); + return null; } }.resolve(this, absF); } - /** Get all the existing storage policies */ + @Override + public Collection getAllStoragePolicies() + throws IOException { + return Arrays.asList(dfs.getStoragePolicies()); + } + + /** + * Deprecated. Prefer {@link FileSystem#getAllStoragePolicies()} + * @return + * @throws IOException + */ + @Deprecated public BlockStoragePolicy[] getStoragePolicies() throws IOException { statistics.incrementReadOps(1); return dfs.getStoragePolicies(); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java index 0710f3e0cc5..8715ce4cd94 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java @@ -27,6 +27,7 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; +import org.apache.hadoop.fs.BlockStoragePolicySpi; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.StorageType; import org.apache.hadoop.hdfs.HdfsConfiguration; @@ -152,8 +153,8 @@ public class Mover { } private void initStoragePolicies() throws IOException { - BlockStoragePolicy[] policies = dispatcher.getDistributedFileSystem() - .getStoragePolicies(); + Collection policies = + dispatcher.getDistributedFileSystem().getAllStoragePolicies(); for (BlockStoragePolicy policy : policies) { this.blockStoragePolicies[policy.getId()] = policy; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/StoragePolicyAdmin.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/StoragePolicyAdmin.java index e3bdffaa4ee..98c8a6b8d3f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/StoragePolicyAdmin.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/StoragePolicyAdmin.java @@ -30,6 +30,7 @@ import org.apache.hadoop.util.Tool; import java.io.IOException; import java.util.Arrays; +import java.util.Collection; import java.util.LinkedList; import java.util.List; @@ -97,7 +98,7 @@ public class StoragePolicyAdmin extends Configured implements Tool { public int run(Configuration conf, List args) throws IOException { final DistributedFileSystem dfs = AdminHelper.getDFS(conf); try { - BlockStoragePolicy[] policies = dfs.getStoragePolicies(); + Collection policies = dfs.getAllStoragePolicies(); System.out.println("Block Storage Policies:"); for (BlockStoragePolicy policy : policies) { if (policy != null) { @@ -155,7 +156,7 @@ public class StoragePolicyAdmin extends Configured implements Tool { System.out.println("The storage policy of " + path + " is unspecified"); return 0; } - BlockStoragePolicy[] policies = dfs.getStoragePolicies(); + Collection policies = dfs.getAllStoragePolicies(); for (BlockStoragePolicy p : policies) { if (p.getId() == storagePolicyId) { System.out.println("The storage policy of " + path + ":\n" + p); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestBlockStoragePolicy.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestBlockStoragePolicy.java index 5e3b55f5e26..ea69f976a6f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestBlockStoragePolicy.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestBlockStoragePolicy.java @@ -25,7 +25,9 @@ import java.io.IOException; import java.util.*; import com.google.common.collect.Lists; +import com.google.common.collect.Sets; import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.BlockStoragePolicySpi; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -1150,30 +1152,6 @@ public class TestBlockStoragePolicy { Assert.assertEquals(3, targets.length); } - /** - * Test getting all the storage policies from the namenode - */ - @Test - public void testGetAllStoragePolicies() throws Exception { - final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf) - .numDataNodes(0).build(); - cluster.waitActive(); - final DistributedFileSystem fs = cluster.getFileSystem(); - try { - BlockStoragePolicy[] policies = fs.getStoragePolicies(); - Assert.assertEquals(6, policies.length); - Assert.assertEquals(POLICY_SUITE.getPolicy(COLD).toString(), - policies[0].toString()); - Assert.assertEquals(POLICY_SUITE.getPolicy(WARM).toString(), - policies[1].toString()); - Assert.assertEquals(POLICY_SUITE.getPolicy(HOT).toString(), - policies[2].toString()); - } finally { - IOUtils.cleanup(null, fs); - cluster.shutdown(); - } - } - @Test public void testGetFileStoragePolicyAfterRestartNN() throws Exception { //HDFS8219 @@ -1217,4 +1195,42 @@ public class TestBlockStoragePolicy { cluster.shutdown(); } } + + /** + * Verify that {@link FileSystem#getAllStoragePolicies} returns all + * known storage policies for DFS. + * + * @throws IOException + */ + @Test + public void testGetAllStoragePoliciesFromFs() throws IOException { + final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf) + .numDataNodes(REPLICATION) + .storageTypes( + new StorageType[] {StorageType.DISK, StorageType.ARCHIVE}) + .build(); + try { + cluster.waitActive(); + + // Get policies via {@link FileSystem#getAllStoragePolicies} + Set policyNamesSet1 = new HashSet<>(); + for (BlockStoragePolicySpi policy : + cluster.getFileSystem().getAllStoragePolicies()) { + policyNamesSet1.add(policy.getName()); + } + + // Get policies from the default BlockStoragePolicySuite. + BlockStoragePolicySuite suite = BlockStoragePolicySuite.createDefaultSuite(); + Set policyNamesSet2 = new HashSet<>(); + for (BlockStoragePolicy policy : suite.getAllPolicies()) { + policyNamesSet2.add(policy.getName()); + } + + // Ensure that we got the same set of policies in both cases. + Assert.assertTrue(Sets.difference(policyNamesSet1, policyNamesSet2).isEmpty()); + Assert.assertTrue(Sets.difference(policyNamesSet2, policyNamesSet1).isEmpty()); + } finally { + cluster.shutdown(); + } + } }