HDFS-4740. Merge r1476596 from trunk

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1486174 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Suresh Srinivas 2013-05-24 19:14:04 +00:00
parent bc0eb9d993
commit c0fbc823f0
4 changed files with 28 additions and 17 deletions

View File

@ -391,6 +391,9 @@ Release 2.0.5-beta - UNRELEASED
HDFS-4743. TestNNStorageRetentionManager fails on Windows. HDFS-4743. TestNNStorageRetentionManager fails on Windows.
(Chris Nauroth via suresh) (Chris Nauroth via suresh)
HDFS-4740. Fixes for a few test failures on Windows.
(Arpit Agarwal via suresh)
Release 2.0.4-alpha - 2013-04-25 Release 2.0.4-alpha - 2013-04-25
INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES

View File

@ -29,11 +29,11 @@ import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_SECONDARY_HTTP_A
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_SERVICE_RPC_ADDRESS_KEY; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_SERVICE_RPC_ADDRESS_KEY;
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMESERVICES; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMESERVICES;
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMESERVICE_ID; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMESERVICE_ID;
import static org.junit.Assert.assertEquals; import org.apache.hadoop.util.Shell;
import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertNull; import static org.junit.Assert.*;
import static org.junit.Assert.assertTrue; import org.junit.Assume;
import static org.junit.Assert.fail; import static org.hamcrest.CoreMatchers.*;
import java.io.IOException; import java.io.IOException;
import java.net.InetSocketAddress; import java.net.InetSocketAddress;
@ -619,19 +619,25 @@ public class TestDFSUtil {
assertEquals(1, uris.size()); assertEquals(1, uris.size());
assertTrue(uris.contains(new URI("hdfs://" + NN1_SRVC_ADDR))); assertTrue(uris.contains(new URI("hdfs://" + NN1_SRVC_ADDR)));
}
@Test (timeout=15000)
public void testLocalhostReverseLookup() {
// 127.0.0.1 -> localhost reverse resolution does not happen on Windows.
Assume.assumeTrue(!Shell.WINDOWS);
// Make sure when config FS_DEFAULT_NAME_KEY using IP address, // Make sure when config FS_DEFAULT_NAME_KEY using IP address,
// it will automatically convert it to hostname // it will automatically convert it to hostname
conf = new HdfsConfiguration(); HdfsConfiguration conf = new HdfsConfiguration();
conf.set(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY, "hdfs://127.0.0.1:8020"); conf.set(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY, "hdfs://127.0.0.1:8020");
uris = DFSUtil.getNameServiceUris(conf); Collection<URI> uris = DFSUtil.getNameServiceUris(conf);
assertEquals(1, uris.size()); assertEquals(1, uris.size());
for (URI uri : uris) { for (URI uri : uris) {
assertFalse(uri.getHost().equals("127.0.0.1")); assertThat(uri.getHost(), not("127.0.0.1"));
} }
} }
@Test @Test (timeout=15000)
public void testIsValidName() { public void testIsValidName() {
assertFalse(DFSUtil.isValidName("/foo/../bar")); assertFalse(DFSUtil.isValidName("/foo/../bar"));
assertFalse(DFSUtil.isValidName("/foo//bar")); assertFalse(DFSUtil.isValidName("/foo//bar"));

View File

@ -155,10 +155,12 @@ public class TestLargeBlock {
} }
/** /**
* Test for block size of 2GB + 512B * Test for block size of 2GB + 512B. This test can take a rather long time to
* complete on Windows (reading the file back can be slow) so we use a larger
* timeout here.
* @throws IOException in case of errors * @throws IOException in case of errors
*/ */
@Test(timeout = 120000) @Test (timeout = 900000)
public void testLargeBlockSize() throws IOException { public void testLargeBlockSize() throws IOException {
final long blockSize = 2L * 1024L * 1024L * 1024L + 512L; // 2GB + 512B final long blockSize = 2L * 1024L * 1024L * 1024L + 512L; // 2GB + 512B
runTest(blockSize); runTest(blockSize);

View File

@ -54,7 +54,7 @@ public class TestWebHdfsUrl {
UserGroupInformation.setConfiguration(new Configuration()); UserGroupInformation.setConfiguration(new Configuration());
} }
@Test(timeout=4000) @Test(timeout=60000)
public void testSimpleAuthParamsInUrl() throws IOException { public void testSimpleAuthParamsInUrl() throws IOException {
Configuration conf = new Configuration(); Configuration conf = new Configuration();
@ -75,7 +75,7 @@ public class TestWebHdfsUrl {
fileStatusUrl); fileStatusUrl);
} }
@Test(timeout=4000) @Test(timeout=60000)
public void testSimpleProxyAuthParamsInUrl() throws IOException { public void testSimpleProxyAuthParamsInUrl() throws IOException {
Configuration conf = new Configuration(); Configuration conf = new Configuration();
@ -98,7 +98,7 @@ public class TestWebHdfsUrl {
fileStatusUrl); fileStatusUrl);
} }
@Test(timeout=4000) @Test(timeout=60000)
public void testSecureAuthParamsInUrl() throws IOException { public void testSecureAuthParamsInUrl() throws IOException {
Configuration conf = new Configuration(); Configuration conf = new Configuration();
// fake turning on security so api thinks it should use tokens // fake turning on security so api thinks it should use tokens
@ -178,7 +178,7 @@ public class TestWebHdfsUrl {
fileStatusUrl); fileStatusUrl);
} }
@Test(timeout=4000) @Test(timeout=60000)
public void testSecureProxyAuthParamsInUrl() throws IOException { public void testSecureProxyAuthParamsInUrl() throws IOException {
Configuration conf = new Configuration(); Configuration conf = new Configuration();
// fake turning on security so api thinks it should use tokens // fake turning on security so api thinks it should use tokens
@ -290,7 +290,7 @@ public class TestWebHdfsUrl {
return (WebHdfsFileSystem) FileSystem.get(uri, conf); return (WebHdfsFileSystem) FileSystem.get(uri, conf);
} }
@Test(timeout=4000) @Test(timeout=60000)
public void testSelectHdfsDelegationToken() throws Exception { public void testSelectHdfsDelegationToken() throws Exception {
SecurityUtilTestHelper.setTokenServiceUseIp(true); SecurityUtilTestHelper.setTokenServiceUseIp(true);