diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java index 2d5f540e2d7..aae5cf749b4 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java @@ -19,6 +19,7 @@ import java.io.IOException; +import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -33,6 +34,9 @@ @InterfaceAudience.Public @InterfaceStability.Stable public class Trash extends Configured { + private static final org.apache.commons.logging.Log LOG = + LogFactory.getLog(Trash.class); + private TrashPolicy trashPolicy; // configured trash policy instance /** @@ -84,6 +88,7 @@ public static boolean moveToAppropriateTrash(FileSystem fs, Path p, } catch (Exception e) { // If we can not determine that trash is enabled server side then // bail rather than potentially deleting a file when trash is enabled. + LOG.warn("Failed to get server trash configuration", e); throw new IOException("Failed to get server trash configuration", e); } Trash trash = new Trash(fullyResolvedFs, conf); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java index cfb51e24ec7..d6a9b4b6222 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java @@ -134,11 +134,11 @@ public boolean moveToTrash(Path path) throws IOException { for (int i = 0; i < 2; i++) { try { if (!fs.mkdirs(baseTrashPath, PERMISSION)) { // create current - LOG.warn("Can't create(mkdir) trash directory: "+baseTrashPath); + LOG.warn("Can't create(mkdir) trash directory: " + baseTrashPath); return false; } } catch (IOException e) { - LOG.warn("Can't create trash directory: "+baseTrashPath); + LOG.warn("Can't create trash directory: " + baseTrashPath, e); cause = e; break; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 1acb9069962..0da2516f5bd 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -243,6 +243,9 @@ Release 2.7.1 - UNRELEASED HDFS-8153. Error Message points to wrong parent directory in case of path component name length error (Anu Engineer via jitendra) + HDFS-8179. DFSClient#getServerDefaults returns null within 1 + hour of system start. (Xiaoyu Yao via Arpit Agarwal) + Release 2.7.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java index 6e4819c666b..227372c6f60 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java @@ -653,10 +653,12 @@ public long getBlockSize(String f) throws IOException { */ public FsServerDefaults getServerDefaults() throws IOException { long now = Time.monotonicNow(); - if (now - serverDefaultsLastUpdate > SERVER_DEFAULTS_VALIDITY_PERIOD) { + if ((serverDefaults == null) || + (now - serverDefaultsLastUpdate > SERVER_DEFAULTS_VALIDITY_PERIOD)) { serverDefaults = namenode.getServerDefaults(); serverDefaultsLastUpdate = now; } + assert serverDefaults != null; return serverDefaults; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystem.java index 564c759b574..2c7f1a3ccac 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystem.java @@ -52,6 +52,7 @@ import org.apache.hadoop.fs.CreateFlag; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; +import org.apache.hadoop.fs.FsServerDefaults; import org.apache.hadoop.fs.FileChecksum; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -1028,4 +1029,18 @@ public void testDFSClientPeerTimeout() throws IOException { cluster.shutdown(); } } + + @Test(timeout=60000) + public void testGetServerDefaults() throws IOException { + Configuration conf = new HdfsConfiguration(); + MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build(); + try { + cluster.waitActive(); + DistributedFileSystem dfs = cluster.getFileSystem(); + FsServerDefaults fsServerDefaults = dfs.getServerDefaults(); + Assert.assertNotNull(fsServerDefaults); + } finally { + cluster.shutdown(); + } + } }