diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index eea5078f128..c299bb0017f 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -67,6 +67,9 @@ Release 2.1.1-beta - UNRELEASED HADOOP-9789. Support server advertised kerberos principals (daryn) + HADOOP-8814. Replace string equals "" by String#isEmpty(). + (Brandon Li via suresh) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java index d609f5c2bbe..9b32c42d966 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java @@ -1084,7 +1084,7 @@ public class Configuration implements Iterable>, */ public boolean getBoolean(String name, boolean defaultValue) { String valueString = getTrimmed(name); - if (null == valueString || "".equals(valueString)) { + if (null == valueString || valueString.isEmpty()) { return defaultValue; } @@ -1238,7 +1238,7 @@ public class Configuration implements Iterable>, */ public Pattern getPattern(String name, Pattern defaultValue) { String valString = get(name); - if (null == valString || "".equals(valString)) { + if (null == valString || valString.isEmpty()) { return defaultValue; } try { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java index 452d29f7b7a..1a24e655c5e 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java @@ -153,9 +153,9 @@ public class ReconfigurationServlet extends HttpServlet { StringEscapeUtils.unescapeHtml(req.getParameter(rawParam)); if (value != null) { if (value.equals(newConf.getRaw(param)) || value.equals("default") || - value.equals("null") || value.equals("")) { + value.equals("null") || value.isEmpty()) { if ((value.equals("default") || value.equals("null") || - value.equals("")) && + value.isEmpty()) && oldConf.getRaw(param) != null) { out.println("

Changed \"" + StringEscapeUtils.escapeHtml(param) + "\" from \"" + @@ -163,7 +163,7 @@ public class ReconfigurationServlet extends HttpServlet { "\" to default

"); reconf.reconfigureProperty(param, null); } else if (!value.equals("default") && !value.equals("null") && - !value.equals("") && + !value.isEmpty() && (oldConf.getRaw(param) == null || !oldConf.getRaw(param).equals(value))) { // change from default or value to different value diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Path.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Path.java index 8723376faf4..4b50882eae8 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Path.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Path.java @@ -104,7 +104,7 @@ public class Path implements Comparable { // Add a slash to parent's path so resolution is compatible with URI's URI parentUri = parent.uri; String parentPath = parentUri.getPath(); - if (!(parentPath.equals("/") || parentPath.equals(""))) { + if (!(parentPath.equals("/") || parentPath.isEmpty())) { try { parentUri = new URI(parentUri.getScheme(), parentUri.getAuthority(), parentUri.getPath()+"/", null, parentUri.getFragment()); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RawLocalFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RawLocalFileSystem.java index 26742d07e9d..3326df4ba1d 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RawLocalFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RawLocalFileSystem.java @@ -524,7 +524,7 @@ public class RawLocalFileSystem extends FileSystem { * onwer.equals(""). */ private boolean isPermissionLoaded() { - return !super.getOwner().equals(""); + return !super.getOwner().isEmpty(); } RawLocalFileStatus(File f, long defaultBlockSize, FileSystem fs) { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/local/RawLocalFs.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/local/RawLocalFs.java index da006595004..605bade09a8 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/local/RawLocalFs.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/local/RawLocalFs.java @@ -110,7 +110,7 @@ public class RawLocalFs extends DelegateToFileSystem { try { FileStatus fs = getFileStatus(f); // If f refers to a regular file or directory - if ("".equals(target)) { + if (target.isEmpty()) { return fs; } // Otherwise f refers to a symlink @@ -132,7 +132,7 @@ public class RawLocalFs extends DelegateToFileSystem { * the readBasicFileAttributes method in java.nio.file.attributes * when available. */ - if (!"".equals(target)) { + if (!target.isEmpty()) { return new FileStatus(0, false, 0, 0, 0, 0, FsPermission.getDefault(), "", "", new Path(target), f); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java index af1c7a9885e..efc415f89e3 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3native/NativeS3FileSystem.java @@ -300,7 +300,7 @@ public class NativeS3FileSystem extends FileSystem { } private static String pathToKey(Path path) { - if (path.toUri().getScheme() != null && "".equals(path.toUri().getPath())) { + if (path.toUri().getScheme() != null && path.toUri().getPath().isEmpty()) { // allow uris without trailing slash after bucket to refer to root, // like s3n://mybucket return ""; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java index 857335c0f03..68ddc64fa31 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java @@ -285,7 +285,7 @@ public class ActiveStandbyElector implements StatCallback, StringCallback { String pathParts[] = znodeWorkingDir.split("/"); Preconditions.checkArgument(pathParts.length >= 1 && - "".equals(pathParts[0]), + pathParts[0].isEmpty(), "Invalid path: %s", znodeWorkingDir); StringBuilder sb = new StringBuilder(); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java index 343693e95c1..0f5465194b9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java @@ -241,7 +241,7 @@ public class SshFenceByTcpPort extends Configured sshPort = DEFAULT_SSH_PORT; // Parse optional user and ssh port - if (arg != null && !"".equals(arg)) { + if (arg != null && !arg.isEmpty()) { Matcher m = USER_PORT_RE.matcher(arg); if (!m.matches()) { throw new BadFencingConfigurationException( diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/DefaultStringifier.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/DefaultStringifier.java index 2b8e259464b..d32d58b6008 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/DefaultStringifier.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/DefaultStringifier.java @@ -192,7 +192,7 @@ public class DefaultStringifier implements Stringifier { String[] parts = itemStr.split(SEPARATOR); for (String part : parts) { - if (!part.equals("")) + if (!part.isEmpty()) list.add(stringifier.fromString(part)); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/TFile.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/TFile.java index 9a57581c90c..c11678dd5d0 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/TFile.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/TFile.java @@ -2105,7 +2105,7 @@ public class TFile { } public boolean isSorted() { - return !strComparator.equals(""); + return !strComparator.isEmpty(); } public String getComparatorString() { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsDynamicMBeanBase.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsDynamicMBeanBase.java index 61522de7bc9..57014d5781b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsDynamicMBeanBase.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsDynamicMBeanBase.java @@ -129,7 +129,7 @@ public abstract class MetricsDynamicMBeanBase implements DynamicMBean { @Override public Object getAttribute(String attributeName) throws AttributeNotFoundException, MBeanException, ReflectionException { - if (attributeName == null || attributeName.equals("")) + if (attributeName == null || attributeName.isEmpty()) throw new IllegalArgumentException(); updateMbeanInfoIfMetricsListChanged(); @@ -197,7 +197,7 @@ public abstract class MetricsDynamicMBeanBase implements DynamicMBean { public Object invoke(String actionName, Object[] parms, String[] signature) throws MBeanException, ReflectionException { - if (actionName == null || actionName.equals("")) + if (actionName == null || actionName.isEmpty()) throw new IllegalArgumentException(); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordInput.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordInput.java index e9fa0c35b89..a3d12b692ea 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordInput.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordInput.java @@ -144,7 +144,7 @@ public class CsvRecordInput implements RecordInput { @Override public void startRecord(String tag) throws IOException { - if (tag != null && !"".equals(tag)) { + if (tag != null && !tag.isEmpty()) { char c1 = (char) stream.read(); char c2 = (char) stream.read(); if (c1 != 's' || c2 != '{') { @@ -156,7 +156,7 @@ public class CsvRecordInput implements RecordInput { @Override public void endRecord(String tag) throws IOException { char c = (char) stream.read(); - if (tag == null || "".equals(tag)) { + if (tag == null || tag.isEmpty()) { if (c != '\n' && c != '\r') { throw new IOException("Error deserializing record."); } else { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordOutput.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordOutput.java index d770f47cf5c..18cf23e78a4 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordOutput.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/CsvRecordOutput.java @@ -115,7 +115,7 @@ public class CsvRecordOutput implements RecordOutput { @Override public void startRecord(Record r, String tag) throws IOException { - if (tag != null && !"".equals(tag)) { + if (tag != null && ! tag.isEmpty()) { printCommaUnlessFirst(); stream.print("s{"); isFirst = true; @@ -124,7 +124,7 @@ public class CsvRecordOutput implements RecordOutput { @Override public void endRecord(Record r, String tag) throws IOException { - if (tag == null || "".equals(tag)) { + if (tag == null || tag.isEmpty()) { stream.print("\n"); isFirst = true; } else { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java index dabf0be79c0..416a442f127 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java @@ -217,7 +217,7 @@ public class SecurityUtil { private static String replacePattern(String[] components, String hostname) throws IOException { String fqdn = hostname; - if (fqdn == null || fqdn.equals("") || fqdn.equals("0.0.0.0")) { + if (fqdn == null || fqdn.isEmpty() || fqdn.equals("0.0.0.0")) { fqdn = getLocalHostName(); } return components[0] + "/" + fqdn.toLowerCase(Locale.US) + "@" + components[2]; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java index c1689bd94cd..df2e432eb2e 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java @@ -1046,7 +1046,7 @@ public class UserGroupInformation { @InterfaceAudience.Public @InterfaceStability.Evolving public static UserGroupInformation createRemoteUser(String user) { - if (user == null || "".equals(user)) { + if (user == null || user.isEmpty()) { throw new IllegalArgumentException("Null user"); } Subject subject = new Subject(); @@ -1118,7 +1118,7 @@ public class UserGroupInformation { @InterfaceStability.Evolving public static UserGroupInformation createProxyUser(String user, UserGroupInformation realUser) { - if (user == null || "".equals(user)) { + if (user == null || user.isEmpty()) { throw new IllegalArgumentException("Null user"); } if (realUser == null) { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/ServiceAuthorizationManager.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/ServiceAuthorizationManager.java index d17d065bf81..8523f38dec7 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/ServiceAuthorizationManager.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/ServiceAuthorizationManager.java @@ -88,7 +88,7 @@ public class ServiceAuthorizationManager { String clientPrincipal = null; if (krbInfo != null) { String clientKey = krbInfo.clientPrincipal(); - if (clientKey != null && !clientKey.equals("")) { + if (clientKey != null && !clientKey.isEmpty()) { try { clientPrincipal = SecurityUtil.getServerPrincipal( conf.get(clientKey), addr); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenIdentifier.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenIdentifier.java index 6ec3b7e606b..089cd99fd5c 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenIdentifier.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenIdentifier.java @@ -87,12 +87,12 @@ extends TokenIdentifier { */ @Override public UserGroupInformation getUser() { - if ( (owner == null) || ("".equals(owner.toString()))) { + if ( (owner == null) || (owner.toString().isEmpty())) { return null; } final UserGroupInformation realUgi; final UserGroupInformation ugi; - if ((realUser == null) || ("".equals(realUser.toString())) + if ((realUser == null) || (realUser.toString().isEmpty()) || realUser.equals(owner)) { ugi = realUgi = UserGroupInformation.createRemoteUser(owner.toString()); } else { diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenSecretManager.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenSecretManager.java index 1783ec9e2ad..42085acb178 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenSecretManager.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/AbstractDelegationTokenSecretManager.java @@ -335,7 +335,7 @@ extends AbstractDelegationTokenIdentifier> throw new InvalidToken("User " + renewer + " tried to renew an expired token"); } - if ((id.getRenewer() == null) || ("".equals(id.getRenewer().toString()))) { + if ((id.getRenewer() == null) || (id.getRenewer().toString().isEmpty())) { throw new AccessControlException("User " + renewer + " tried to renew a token without " + "a renewer"); @@ -392,7 +392,7 @@ extends AbstractDelegationTokenIdentifier> HadoopKerberosName cancelerKrbName = new HadoopKerberosName(canceller); String cancelerShortName = cancelerKrbName.getShortName(); if (!canceller.equals(owner) - && (renewer == null || "".equals(renewer.toString()) || !cancelerShortName + && (renewer == null || renewer.toString().isEmpty() || !cancelerShortName .equals(renewer.toString()))) { throw new AccessControlException(canceller + " is not authorized to cancel the token"); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java index 77fbe4f166b..d5573093d10 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java @@ -82,13 +82,13 @@ public class HostsFileReader { public synchronized void refresh() throws IOException { LOG.info("Refreshing hosts (include/exclude) list"); - if (!includesFile.equals("")) { + if (!includesFile.isEmpty()) { Set newIncludes = new HashSet(); readFileToSet("included", includesFile, newIncludes); // switch the new hosts that are to be included includes = newIncludes; } - if (!excludesFile.equals("")) { + if (!excludesFile.isEmpty()) { Set newExcludes = new HashSet(); readFileToSet("excluded", excludesFile, newExcludes); // switch the excluded hosts diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java index f51d02e300d..284a042e83b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java @@ -352,7 +352,7 @@ public class StringUtils { * @return an array of String values */ public static String[] getTrimmedStrings(String str){ - if (null == str || "".equals(str.trim())) { + if (null == str || str.trim().isEmpty()) { return emptyStringArray; } @@ -412,7 +412,7 @@ public class StringUtils { String str, char separator) { // String.split returns a single empty result for splitting the empty // string. - if ("".equals(str)) { + if (str.isEmpty()) { return new String[]{""}; } ArrayList strList = new ArrayList(); diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/nativeio/TestNativeIO.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/nativeio/TestNativeIO.java index 84e2094f6c9..4d71e15c4b3 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/nativeio/TestNativeIO.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/nativeio/TestNativeIO.java @@ -84,7 +84,7 @@ public class TestNativeIO { } assertEquals(expectedOwner, owner); assertNotNull(stat.getGroup()); - assertTrue(!"".equals(stat.getGroup())); + assertTrue(!stat.getGroup().isEmpty()); assertEquals("Stat mode field should indicate a regular file", NativeIO.POSIX.Stat.S_IFREG, stat.getMode() & NativeIO.POSIX.Stat.S_IFMT); @@ -118,7 +118,7 @@ public class TestNativeIO { NativeIO.POSIX.Stat stat = NativeIO.POSIX.getFstat(fos.getFD()); assertEquals(System.getProperty("user.name"), stat.getOwner()); assertNotNull(stat.getGroup()); - assertTrue(!"".equals(stat.getGroup())); + assertTrue(!stat.getGroup().isEmpty()); assertEquals("Stat mode field should indicate a regular file", NativeIO.POSIX.Stat.S_IFREG, stat.getMode() & NativeIO.POSIX.Stat.S_IFMT); diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java index 7ee893831a3..1e96ed1ddbc 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestSaslRPC.java @@ -175,7 +175,7 @@ public class TestSaslRPC { } @Override public UserGroupInformation getUser() { - if ("".equals(realUser.toString())) { + if (realUser.toString().isEmpty()) { return UserGroupInformation.createRemoteUser(tokenid.toString()); } else { UserGroupInformation realUgi = UserGroupInformation diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestStringUtils.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestStringUtils.java index f58ec5f86fd..4f06a31649e 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestStringUtils.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestStringUtils.java @@ -294,7 +294,8 @@ public class TestStringUtils extends UnitTestcaseTimeLimit { assertArrayEquals(expectedArray, StringUtils.getTrimmedStrings(pathologicalDirList2)); assertArrayEquals(emptyArray, StringUtils.getTrimmedStrings(emptyList1)); - assertArrayEquals(emptyArray, StringUtils.getTrimmedStrings(emptyList2)); + String[] estring = StringUtils.getTrimmedStrings(emptyList2); + assertArrayEquals(emptyArray, estring); } @Test (timeout = 30000)