diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 6705374cfe2..4a739a334af 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -156,6 +156,9 @@ Release 2.4.0 - UNRELEASED HADOOP-10440. HarFsInputStream.read(byte[]) updates position incorrectly. (guodongdong via szetszwo) + HADOOP-10437. Fix the javac warnings in the conf and the util package. + (szetszwo) + BREAKDOWN OF HADOOP-10184 SUBTASKS AND RELATED JIRAS HADOOP-10185. FileSystem API for ACLs. (cnauroth) diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java index 21d74f5b5ca..290c236aa25 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java @@ -2251,13 +2251,13 @@ public class Configuration implements Iterable>, root = (Element)resource; } - if (doc == null && root == null) { - if (quiet) - return null; - throw new RuntimeException(resource + " not found"); - } - if (root == null) { + if (doc == null) { + if (quiet) { + return null; + } + throw new RuntimeException(resource + " not found"); + } root = doc.getDocumentElement(); } Properties toAddTo = properties; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java index 1a24e655c5e..3fa162b87af 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java @@ -131,15 +131,14 @@ public class ReconfigurationServlet extends HttpServlet { @SuppressWarnings("unchecked") private Enumeration getParams(HttpServletRequest req) { - return (Enumeration) req.getParameterNames(); + return req.getParameterNames(); } /** * Apply configuratio changes after admin has approved them. */ private void applyChanges(PrintWriter out, Reconfigurable reconf, - HttpServletRequest req) - throws IOException, ReconfigurationException { + HttpServletRequest req) throws ReconfigurationException { Configuration oldConf = reconf.getConf(); Configuration newConf = new Configuration(); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java index 05aabb974bb..43d396edad7 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/WritableName.java @@ -47,18 +47,18 @@ public class WritableName { /** Set the name that a class should be known as to something other than the * class name. */ - public static synchronized void setName(Class writableClass, String name) { + public static synchronized void setName(Class writableClass, String name) { CLASS_TO_NAME.put(writableClass, name); NAME_TO_CLASS.put(name, writableClass); } /** Add an alternate name for a class. */ - public static synchronized void addName(Class writableClass, String name) { + public static synchronized void addName(Class writableClass, String name) { NAME_TO_CLASS.put(name, writableClass); } /** Return the name for a class. Default is {@link Class#getName()}. */ - public static synchronized String getName(Class writableClass) { + public static synchronized String getName(Class writableClass) { String name = CLASS_TO_NAME.get(writableClass); if (name != null) return name; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java index a1e20d242d6..df151666d47 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/AsyncDiskService.java @@ -17,7 +17,6 @@ */ package org.apache.hadoop.util; -import java.io.IOException; import java.util.ArrayList; import java.util.HashMap; import java.util.List; @@ -26,6 +25,7 @@ import java.util.concurrent.LinkedBlockingQueue; import java.util.concurrent.ThreadFactory; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; + import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -68,7 +68,7 @@ public class AsyncDiskService { * * @param volumes The roots of the file system volumes. */ - public AsyncDiskService(String[] volumes) throws IOException { + public AsyncDiskService(String[] volumes) { threadFactory = new ThreadFactory() { @Override diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java index 53a5de17325..6f949891508 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ClassUtil.java @@ -36,13 +36,13 @@ public class ClassUtil { * @return a jar file that contains the class, or null. * @throws IOException */ - public static String findContainingJar(Class clazz) { + public static String findContainingJar(Class clazz) { ClassLoader loader = clazz.getClassLoader(); String classFile = clazz.getName().replaceAll("\\.", "/") + ".class"; try { - for (Enumeration itr = loader.getResources(classFile); + for(final Enumeration itr = loader.getResources(classFile); itr.hasMoreElements();) { - URL url = (URL) itr.nextElement(); + final URL url = itr.nextElement(); if ("jar".equals(url.getProtocol())) { String toReturn = url.getPath(); if (toReturn.startsWith("file:")) { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java index b02bd14f5ba..72a4d1b70e9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java @@ -27,7 +27,6 @@ import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.LocalFileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.util.Shell; /** * Class that provides utility functions for checking disk problem @@ -35,9 +34,6 @@ import org.apache.hadoop.util.Shell; @InterfaceAudience.Private @InterfaceStability.Unstable public class DiskChecker { - - private static final long SHELL_TIMEOUT = 10 * 1000; - public static class DiskErrorException extends IOException { public DiskErrorException(String msg) { super(msg); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java index 67818555393..cb6f91c7c67 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java @@ -25,7 +25,6 @@ import java.net.URISyntaxException; import java.net.URL; import java.net.URLClassLoader; import java.util.ArrayList; -import java.util.Arrays; import java.util.List; import org.apache.commons.cli.CommandLine; @@ -397,7 +396,8 @@ public class GenericOptionsParser { if (!localFs.exists(path)) { throw new FileNotFoundException("File " + tmp + " does not exist."); } - finalPath = path.makeQualified(localFs).toString(); + finalPath = path.makeQualified(localFs.getUri(), + localFs.getWorkingDirectory()).toString(); } else { // check if the file exists in this file system @@ -408,7 +408,8 @@ public class GenericOptionsParser { if (!fs.exists(path)) { throw new FileNotFoundException("File " + tmp + " does not exist."); } - finalPath = path.makeQualified(fs).toString(); + finalPath = path.makeQualified(fs.getUri(), + fs.getWorkingDirectory()).toString(); } finalArr[i] = finalPath; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java index b4d370e9eba..b012add42c5 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java @@ -169,9 +169,8 @@ public class HostsFileReader { this.excludesFile = excludesFile; } - public synchronized void updateFileNames(String includesFile, - String excludesFile) - throws IOException { + public synchronized void updateFileNames(String includesFile, + String excludesFile) { setIncludesFile(includesFile); setExcludesFile(excludesFile); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java index 774f95b76ea..3188cb5fe95 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/LineReader.java @@ -245,7 +245,7 @@ public class LineReader implements Closeable { } } while (newlineLength == 0 && bytesConsumed < maxBytesToConsume); - if (bytesConsumed > (long)Integer.MAX_VALUE) { + if (bytesConsumed > Integer.MAX_VALUE) { throw new IOException("Too many bytes before newline: " + bytesConsumed); } return (int)bytesConsumed; @@ -343,7 +343,7 @@ public class LineReader implements Closeable { } } while (delPosn < recordDelimiterBytes.length && bytesConsumed < maxBytesToConsume); - if (bytesConsumed > (long) Integer.MAX_VALUE) { + if (bytesConsumed > Integer.MAX_VALUE) { throw new IOException("Too many bytes before delimiter: " + bytesConsumed); } return (int) bytesConsumed; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java index e52dacccbb4..fb018cf8631 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ProgramDriver.java @@ -99,7 +99,8 @@ public class ProgramDriver { * @throws NoSuchMethodException * @throws SecurityException */ - public void addClass (String name, Class mainClass, String description) throws Throwable { + public void addClass(String name, Class mainClass, String description) + throws Throwable { programs.put(name , new ProgramDescription(mainClass, description)); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java index 0af0844134d..e3445b5da6b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Progress.java @@ -64,7 +64,7 @@ public class Progress { public synchronized Progress addPhase() { Progress phase = addNewPhase(); // set equal weightage for all phases - progressPerPhase = 1.0f / (float)phases.size(); + progressPerPhase = 1.0f / phases.size(); fixedWeightageForAllPhases = true; return phase; } @@ -110,7 +110,7 @@ public class Progress { addNewPhase(); } // set equal weightage for all phases - progressPerPhase = 1.0f / (float)phases.size(); + progressPerPhase = 1.0f / phases.size(); fixedWeightageForAllPhases = true; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java index ec1c6d84c0e..08b4fd15d84 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java @@ -78,7 +78,7 @@ public class RunJar { try { Enumeration entries = jar.entries(); while (entries.hasMoreElements()) { - JarEntry entry = (JarEntry)entries.nextElement(); + final JarEntry entry = entries.nextElement(); if (!entry.isDirectory() && unpackRegex.matcher(entry.getName()).matches()) { InputStream in = jar.getInputStream(entry); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java index f96353732d4..9ffbe63457b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java @@ -431,7 +431,7 @@ public class StringUtils { ArrayList strList = new ArrayList(); int startIndex = 0; int nextIndex = 0; - while ((nextIndex = str.indexOf((int)separator, startIndex)) != -1) { + while ((nextIndex = str.indexOf(separator, startIndex)) != -1) { strList.add(str.substring(startIndex, nextIndex)); startIndex = nextIndex + 1; } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java index 1768567bd66..9296d54912e 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/VersionInfo.java @@ -19,18 +19,13 @@ package org.apache.hadoop.util; import java.io.IOException; -import java.net.URL; -import java.net.URLDecoder; -import java.util.Enumeration; +import java.io.InputStream; +import java.util.Properties; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; - -import java.io.IOException; -import java.io.InputStream; -import java.util.Properties; import org.apache.hadoop.io.IOUtils; /**