From a126a01fa197beebe955837c8f2efbd3257f7aa5 Mon Sep 17 00:00:00 2001 From: Tsz-wo Sze Date: Wed, 26 Mar 2014 19:03:16 +0000 Subject: [PATCH] HADOOP-10437. Fix the javac warnings in the conf and the util package. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1582015 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-common-project/hadoop-common/CHANGES.txt | 3 +++ .../java/org/apache/hadoop/conf/Configuration.java | 12 ++++++------ .../apache/hadoop/conf/ReconfigurationServlet.java | 5 ++--- .../main/java/org/apache/hadoop/io/WritableName.java | 6 +++--- .../org/apache/hadoop/util/AsyncDiskService.java | 4 ++-- .../main/java/org/apache/hadoop/util/ClassUtil.java | 6 +++--- .../java/org/apache/hadoop/util/DiskChecker.java | 4 ---- .../org/apache/hadoop/util/GenericOptionsParser.java | 7 ++++--- .../java/org/apache/hadoop/util/HostsFileReader.java | 5 ++--- .../main/java/org/apache/hadoop/util/LineReader.java | 4 ++-- .../java/org/apache/hadoop/util/ProgramDriver.java | 3 ++- .../main/java/org/apache/hadoop/util/Progress.java | 4 ++-- .../src/main/java/org/apache/hadoop/util/RunJar.java | 2 +- .../java/org/apache/hadoop/util/StringUtils.java | 2 +- .../java/org/apache/hadoop/util/VersionInfo.java | 9 ++------- 15 files changed, 35 insertions(+), 41 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 8a3901e6bce..01a762ef622 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -457,6 +457,9 @@ Release 2.4.0 - UNRELEASED HADOOP-10440. HarFsInputStream.read(byte[]) updates position incorrectly. (guodongdong via szetszwo) + HADOOP-10437. Fix the javac warnings in the conf and the util package. + (szetszwo) + BREAKDOWN OF HADOOP-10184 SUBTASKS AND RELATED JIRAS HADOOP-10185. FileSystem API for ACLs. (cnauroth) diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java index 480511efdf6..0312e545392 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java @@ -2259,13 +2259,13 @@ private Resource loadResource(Properties properties, Resource wrapper, boolean q root = (Element)resource; } - if (doc == null && root == null) { - if (quiet) - return null; - throw new RuntimeException(resource + " not found"); - } - if (root == null) { + if (doc == null) { + if (quiet) { + return null; + } + throw new RuntimeException(resource + " not found"); + } root = doc.getDocumentElement(); } Properties toAddTo = properties; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java index 1a24e655c5e..3fa162b87af 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java @@ -131,15 +131,14 @@ private void printConf(PrintWriter out, Reconfigurable reconf) { @SuppressWarnings("unchecked") private Enumeration getParams(HttpServletRequest req) { - return (Enumeration) req.getParameterNames(); + return req.getParameterNames(); } /** * Apply configuratio changes after admin has approved them. */ private void applyChanges(PrintWriter out, Reconfigurable reconf, - HttpServletRequest req) - throws IOException, ReconfigurationException { + HttpServletRequest req) throws ReconfigurationException { Configuration oldConf = reconf.getConf(); Configuration newConf = new Configuration(); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java index 05aabb974bb..43d396edad7 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java @@ -47,18 +47,18 @@ private WritableName() {} // no public ctor /** Set the name that a class should be known as to something other than the * class name. */ - public static synchronized void setName(Class writableClass, String name) { + public static synchronized void setName(Class writableClass, String name) { CLASS_TO_NAME.put(writableClass, name); NAME_TO_CLASS.put(name, writableClass); } /** Add an alternate name for a class. */ - public static synchronized void addName(Class writableClass, String name) { + public static synchronized void addName(Class writableClass, String name) { NAME_TO_CLASS.put(name, writableClass); } /** Return the name for a class. Default is {@link Class#getName()}. */ - public static synchronized String getName(Class writableClass) { + public static synchronized String getName(Class writableClass) { String name = CLASS_TO_NAME.get(writableClass); if (name != null) return name; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java index a1e20d242d6..df151666d47 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java @@ -17,7 +17,6 @@ */ package org.apache.hadoop.util; -import java.io.IOException; import java.util.ArrayList; import java.util.HashMap; import java.util.List; @@ -26,6 +25,7 @@ import java.util.concurrent.ThreadFactory; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; + import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -68,7 +68,7 @@ public class AsyncDiskService { * * @param volumes The roots of the file system volumes. */ - public AsyncDiskService(String[] volumes) throws IOException { + public AsyncDiskService(String[] volumes) { threadFactory = new ThreadFactory() { @Override diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java index 53a5de17325..6f949891508 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java @@ -36,13 +36,13 @@ public class ClassUtil { * @return a jar file that contains the class, or null. * @throws IOException */ - public static String findContainingJar(Class clazz) { + public static String findContainingJar(Class clazz) { ClassLoader loader = clazz.getClassLoader(); String classFile = clazz.getName().replaceAll("\\.", "/") + ".class"; try { - for (Enumeration itr = loader.getResources(classFile); + for(final Enumeration itr = loader.getResources(classFile); itr.hasMoreElements();) { - URL url = (URL) itr.nextElement(); + final URL url = itr.nextElement(); if ("jar".equals(url.getProtocol())) { String toReturn = url.getPath(); if (toReturn.startsWith("file:")) { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java index b02bd14f5ba..72a4d1b70e9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java @@ -27,7 +27,6 @@ import org.apache.hadoop.fs.LocalFileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.util.Shell; /** * Class that provides utility functions for checking disk problem @@ -35,9 +34,6 @@ @InterfaceAudience.Private @InterfaceStability.Unstable public class DiskChecker { - - private static final long SHELL_TIMEOUT = 10 * 1000; - public static class DiskErrorException extends IOException { public DiskErrorException(String msg) { super(msg); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java index 67818555393..cb6f91c7c67 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java @@ -25,7 +25,6 @@ import java.net.URL; import java.net.URLClassLoader; import java.util.ArrayList; -import java.util.Arrays; import java.util.List; import org.apache.commons.cli.CommandLine; @@ -397,7 +396,8 @@ private String validateFiles(String files, Configuration conf) if (!localFs.exists(path)) { throw new FileNotFoundException("File " + tmp + " does not exist."); } - finalPath = path.makeQualified(localFs).toString(); + finalPath = path.makeQualified(localFs.getUri(), + localFs.getWorkingDirectory()).toString(); } else { // check if the file exists in this file system @@ -408,7 +408,8 @@ private String validateFiles(String files, Configuration conf) if (!fs.exists(path)) { throw new FileNotFoundException("File " + tmp + " does not exist."); } - finalPath = path.makeQualified(fs).toString(); + finalPath = path.makeQualified(fs.getUri(), + fs.getWorkingDirectory()).toString(); } finalArr[i] = finalPath; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java index b4d370e9eba..b012add42c5 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java @@ -169,9 +169,8 @@ public synchronized void setExcludesFile(String excludesFile) { this.excludesFile = excludesFile; } - public synchronized void updateFileNames(String includesFile, - String excludesFile) - throws IOException { + public synchronized void updateFileNames(String includesFile, + String excludesFile) { setIncludesFile(includesFile); setExcludesFile(excludesFile); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java index 774f95b76ea..3188cb5fe95 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java @@ -245,7 +245,7 @@ private int readDefaultLine(Text str, int maxLineLength, int maxBytesToConsume) } } while (newlineLength == 0 && bytesConsumed < maxBytesToConsume); - if (bytesConsumed > (long)Integer.MAX_VALUE) { + if (bytesConsumed > Integer.MAX_VALUE) { throw new IOException("Too many bytes before newline: " + bytesConsumed); } return (int)bytesConsumed; @@ -343,7 +343,7 @@ private int readCustomLine(Text str, int maxLineLength, int maxBytesToConsume) } } while (delPosn < recordDelimiterBytes.length && bytesConsumed < maxBytesToConsume); - if (bytesConsumed > (long) Integer.MAX_VALUE) { + if (bytesConsumed > Integer.MAX_VALUE) { throw new IOException("Too many bytes before delimiter: " + bytesConsumed); } return (int) bytesConsumed; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java index e52dacccbb4..fb018cf8631 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java @@ -99,7 +99,8 @@ private static void printUsage(Map programs) { * @throws NoSuchMethodException * @throws SecurityException */ - public void addClass (String name, Class mainClass, String description) throws Throwable { + public void addClass(String name, Class mainClass, String description) + throws Throwable { programs.put(name , new ProgramDescription(mainClass, description)); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java index 0af0844134d..e3445b5da6b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java @@ -64,7 +64,7 @@ public Progress addPhase(String status) { public synchronized Progress addPhase() { Progress phase = addNewPhase(); // set equal weightage for all phases - progressPerPhase = 1.0f / (float)phases.size(); + progressPerPhase = 1.0f / phases.size(); fixedWeightageForAllPhases = true; return phase; } @@ -110,7 +110,7 @@ public synchronized void addPhases(int n) { addNewPhase(); } // set equal weightage for all phases - progressPerPhase = 1.0f / (float)phases.size(); + progressPerPhase = 1.0f / phases.size(); fixedWeightageForAllPhases = true; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java index ec1c6d84c0e..08b4fd15d84 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java @@ -78,7 +78,7 @@ public static void unJar(File jarFile, File toDir, Pattern unpackRegex) try { Enumeration entries = jar.entries(); while (entries.hasMoreElements()) { - JarEntry entry = (JarEntry)entries.nextElement(); + final JarEntry entry = entries.nextElement(); if (!entry.isDirectory() && unpackRegex.matcher(entry.getName()).matches()) { InputStream in = jar.getInputStream(entry); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java index f96353732d4..9ffbe63457b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java @@ -431,7 +431,7 @@ public static String[] split( ArrayList strList = new ArrayList(); int startIndex = 0; int nextIndex = 0; - while ((nextIndex = str.indexOf((int)separator, startIndex)) != -1) { + while ((nextIndex = str.indexOf(separator, startIndex)) != -1) { strList.add(str.substring(startIndex, nextIndex)); startIndex = nextIndex + 1; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java index 1768567bd66..9296d54912e 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java @@ -19,18 +19,13 @@ package org.apache.hadoop.util; import java.io.IOException; -import java.net.URL; -import java.net.URLDecoder; -import java.util.Enumeration; +import java.io.InputStream; +import java.util.Properties; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; - -import java.io.IOException; -import java.io.InputStream; -import java.util.Properties; import org.apache.hadoop.io.IOUtils; /**