From 2c155afe2736a5571bbb3bdfb2fe6f9709227229 Mon Sep 17 00:00:00 2001 From: Kihwal Lee Date: Thu, 14 Apr 2016 14:25:11 -0500 Subject: [PATCH] HDFS-10292. Add block id when client got Unable to close file exception. Contributed by Brahma Reddy Battula. --- .../src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java index dc88e08991c..0f8279943d2 100755 --- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java +++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java @@ -811,7 +811,7 @@ public class DFSOutputStream extends FSOutputSummer try { if (retries == 0) { throw new IOException("Unable to close file because the last block" - + " does not have enough number of replicas."); + + last + " does not have enough number of replicas."); } retries--; Thread.sleep(sleeptime);