Test PR: Ran rewrite plugin
This commit is contained in:
parent
b90389ae98
commit
a8b5d1d648
@ -61,11 +61,27 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<artifactId>zookeeper</artifactId>
|
||||
<type>test-jar</type>
|
||||
<scope>test</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>org.junit.vintage</groupId>
|
||||
<artifactId>junit-vintage-engine</artifactId>
|
||||
</exclusion>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>io.dropwizard.metrics</groupId>
|
||||
<artifactId>metrics-core</artifactId>
|
||||
<scope>provided</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.xerial.snappy</groupId>
|
||||
@ -154,12 +170,6 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
|
||||
<dependency>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
<scope>test</scope>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
<artifactId>hadoop-minikdc</artifactId>
|
||||
@ -169,11 +179,23 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<groupId>org.mockito</groupId>
|
||||
<artifactId>mockito-core</artifactId>
|
||||
<scope>test</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.slf4j</groupId>
|
||||
<artifactId>slf4j-log4j12</artifactId>
|
||||
<scope>provided</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>io.netty</groupId>
|
||||
@ -184,6 +206,16 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<groupId>io.netty</groupId>
|
||||
<artifactId>netty-all</artifactId>
|
||||
<scope>compile</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>org.junit.vintage</groupId>
|
||||
<artifactId>junit-vintage-engine</artifactId>
|
||||
</exclusion>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
@ -209,11 +241,27 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<dependency>
|
||||
<groupId>com.fasterxml.jackson.core</groupId>
|
||||
<artifactId>jackson-databind</artifactId>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.apache.curator</groupId>
|
||||
<artifactId>curator-test</artifactId>
|
||||
<scope>test</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>org.junit.vintage</groupId>
|
||||
<artifactId>junit-vintage-engine</artifactId>
|
||||
</exclusion>
|
||||
<exclusion>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.assertj</groupId>
|
||||
@ -447,6 +495,23 @@ https://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
</filesets>
|
||||
</configuration>
|
||||
</plugin>
|
||||
<plugin>
|
||||
<groupId>org.openrewrite.maven</groupId>
|
||||
<artifactId>rewrite-maven-plugin</artifactId>
|
||||
<version>4.9.0</version>
|
||||
<configuration>
|
||||
<activeRecipes>
|
||||
<recipe>org.openrewrite.java.testing.junit5.JUnit5BestPractices</recipe>
|
||||
</activeRecipes>
|
||||
</configuration>
|
||||
<dependencies>
|
||||
<dependency>
|
||||
<groupId>org.openrewrite.recipe</groupId>
|
||||
<artifactId>rewrite-testing-frameworks</artifactId>
|
||||
<version>1.7.1</version>
|
||||
</dependency>
|
||||
</dependencies>
|
||||
</plugin>
|
||||
</plugins>
|
||||
</build>
|
||||
|
||||
|
@ -18,10 +18,8 @@
|
||||
|
||||
package org.apache.hadoop;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
@ -31,11 +29,7 @@
|
||||
|
||||
import org.apache.hadoop.ipc.RefreshRegistry;
|
||||
import org.apache.hadoop.ipc.RefreshResponse;
|
||||
import org.junit.Test;
|
||||
import org.junit.Before;
|
||||
import org.junit.After;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.jupiter.api.*;
|
||||
import org.mockito.Mockito;
|
||||
|
||||
/**
|
||||
@ -51,7 +45,7 @@ public class TestGenericRefresh {
|
||||
private static RefreshHandler firstHandler;
|
||||
private static RefreshHandler secondHandler;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setUpBeforeClass() throws Exception {
|
||||
config = new Configuration();
|
||||
config.set("hadoop.security.authorization", "true");
|
||||
@ -61,14 +55,14 @@ public static void setUpBeforeClass() throws Exception {
|
||||
cluster.waitActive();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDownBeforeClass() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
// Register Handlers, first one just sends an ok response
|
||||
firstHandler = Mockito.mock(RefreshHandler.class);
|
||||
@ -85,7 +79,7 @@ public void setUp() throws Exception {
|
||||
RefreshRegistry.defaultRegistry().register("secondHandler", secondHandler);
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
RefreshRegistry.defaultRegistry().unregisterAll("firstHandler");
|
||||
RefreshRegistry.defaultRegistry().unregisterAll("secondHandler");
|
||||
@ -96,7 +90,7 @@ public void testInvalidCommand() throws Exception {
|
||||
DFSAdmin admin = new DFSAdmin(config);
|
||||
String [] args = new String[]{"-refresh", "nn"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should fail due to bad args", -1, exitCode);
|
||||
assertEquals(-1, exitCode, "DFSAdmin should fail due to bad args");
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -105,7 +99,7 @@ public void testInvalidIdentifier() throws Exception {
|
||||
String [] args = new String[]{"-refresh", "localhost:" +
|
||||
cluster.getNameNodePort(), "unregisteredIdentity"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should fail due to no handler registered", -1, exitCode);
|
||||
assertEquals(-1, exitCode, "DFSAdmin should fail due to no handler registered");
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -114,7 +108,7 @@ public void testValidIdentifier() throws Exception {
|
||||
String[] args = new String[]{"-refresh",
|
||||
"localhost:" + cluster.getNameNodePort(), "firstHandler"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should succeed", 0, exitCode);
|
||||
assertEquals(0, exitCode, "DFSAdmin should succeed");
|
||||
|
||||
Mockito.verify(firstHandler).handleRefresh("firstHandler", new String[]{});
|
||||
// Second handler was never called
|
||||
@ -128,11 +122,11 @@ public void testVariableArgs() throws Exception {
|
||||
String[] args = new String[]{"-refresh", "localhost:" +
|
||||
cluster.getNameNodePort(), "secondHandler", "one"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should return 2", 2, exitCode);
|
||||
assertEquals(2, exitCode, "DFSAdmin should return 2");
|
||||
|
||||
exitCode = admin.run(new String[]{"-refresh", "localhost:" +
|
||||
cluster.getNameNodePort(), "secondHandler", "one", "two"});
|
||||
assertEquals("DFSAdmin should now return 3", 3, exitCode);
|
||||
assertEquals(3, exitCode, "DFSAdmin should now return 3");
|
||||
|
||||
Mockito.verify(secondHandler).handleRefresh("secondHandler", new String[]{"one"});
|
||||
Mockito.verify(secondHandler).handleRefresh("secondHandler", new String[]{"one", "two"});
|
||||
@ -147,7 +141,7 @@ public void testUnregistration() throws Exception {
|
||||
String[] args = new String[]{"-refresh", "localhost:" +
|
||||
cluster.getNameNodePort(), "firstHandler"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should return -1", -1, exitCode);
|
||||
assertEquals(-1, exitCode, "DFSAdmin should return -1");
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -18,10 +18,7 @@
|
||||
|
||||
package org.apache.hadoop;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.BindException;
|
||||
@ -40,8 +37,8 @@
|
||||
import org.apache.hadoop.ipc.FairCallQueue;
|
||||
import org.apache.hadoop.metrics2.MetricsException;
|
||||
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
|
||||
import org.junit.After;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestRefreshCallQueue {
|
||||
private MiniDFSCluster cluster;
|
||||
@ -77,7 +74,7 @@ private void setUp(Class<?> queueClass) throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws IOException {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -115,9 +112,9 @@ public void testRefresh() throws Exception {
|
||||
mockQueuePuts = 0;
|
||||
setUp(MockCallQueue.class);
|
||||
|
||||
assertTrue("Mock queue should have been constructed",
|
||||
mockQueueConstructions > 0);
|
||||
assertTrue("Puts are routed through MockQueue", canPutInMockQueue());
|
||||
assertTrue(
|
||||
mockQueueConstructions > 0, "Mock queue should have been constructed");
|
||||
assertTrue(canPutInMockQueue(), "Puts are routed through MockQueue");
|
||||
int lastMockQueueConstructions = mockQueueConstructions;
|
||||
|
||||
// Replace queue with the queue specified in core-site.xml, which would be
|
||||
@ -125,13 +122,13 @@ public void testRefresh() throws Exception {
|
||||
DFSAdmin admin = new DFSAdmin(config);
|
||||
String [] args = new String[]{"-refreshCallQueue"};
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should return 0", 0, exitCode);
|
||||
assertEquals(0, exitCode, "DFSAdmin should return 0");
|
||||
|
||||
assertEquals("Mock queue should have no additional constructions",
|
||||
lastMockQueueConstructions, mockQueueConstructions);
|
||||
assertEquals(
|
||||
lastMockQueueConstructions, mockQueueConstructions, "Mock queue should have no additional constructions");
|
||||
try {
|
||||
assertFalse("Puts are routed through LBQ instead of MockQueue",
|
||||
canPutInMockQueue());
|
||||
assertFalse(
|
||||
canPutInMockQueue(), "Puts are routed through LBQ instead of MockQueue");
|
||||
} catch (IOException ioe) {
|
||||
fail("Could not put into queue at all");
|
||||
}
|
||||
|
@ -22,9 +22,9 @@
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestAclCLI extends CLITestHelperDFS {
|
||||
private MiniDFSCluster cluster = null;
|
||||
@ -38,7 +38,7 @@ protected void initConf() {
|
||||
DFSConfigKeys.DFS_NAMENODE_POSIX_ACL_INHERITANCE_ENABLED_KEY, false);
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -49,7 +49,7 @@ public void setUp() throws Exception {
|
||||
username = System.getProperty("user.name");
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
super.tearDown();
|
||||
|
@ -17,9 +17,9 @@
|
||||
*/
|
||||
package org.apache.hadoop.cli;
|
||||
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_POSIX_ACL_INHERITANCE_ENABLED_KEY;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import org.junit.Test;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_POSIX_ACL_INHERITANCE_ENABLED_KEY;
|
||||
|
||||
/**
|
||||
* Test ACL CLI with POSIX ACL inheritance enabled.
|
||||
|
@ -18,7 +18,7 @@
|
||||
|
||||
package org.apache.hadoop.cli;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
@ -37,9 +37,9 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.tools.CacheAdmin;
|
||||
import org.apache.hadoop.security.authorize.PolicyProvider;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.xml.sax.SAXException;
|
||||
|
||||
public class TestCacheAdminCLI extends CLITestHelper {
|
||||
@ -51,7 +51,7 @@ public class TestCacheAdminCLI extends CLITestHelper {
|
||||
protected FileSystem fs = null;
|
||||
protected String namenode = null;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -68,11 +68,11 @@ public void setUp() throws Exception {
|
||||
username = System.getProperty("user.name");
|
||||
|
||||
fs = dfsCluster.getFileSystem();
|
||||
assertTrue("Not a HDFS: "+fs.getUri(),
|
||||
fs instanceof DistributedFileSystem);
|
||||
assertTrue(
|
||||
fs instanceof DistributedFileSystem, "Not a HDFS: " + fs.getUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -23,7 +23,7 @@
|
||||
import java.security.NoSuchAlgorithmException;
|
||||
import java.util.UUID;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.apache.hadoop.cli.util.CLICommand;
|
||||
import org.apache.hadoop.cli.util.CLICommandCryptoAdmin;
|
||||
@ -45,9 +45,9 @@
|
||||
import org.apache.hadoop.hdfs.tools.CryptoAdmin;
|
||||
import org.apache.hadoop.security.authorize.PolicyProvider;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.xml.sax.SAXException;
|
||||
|
||||
public class TestCryptoAdminCLI extends CLITestHelperDFS {
|
||||
@ -56,7 +56,7 @@ public class TestCryptoAdminCLI extends CLITestHelperDFS {
|
||||
protected String namenode = null;
|
||||
private static File tmpDir;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -78,11 +78,11 @@ public void setUp() throws Exception {
|
||||
username = System.getProperty("user.name");
|
||||
|
||||
fs = dfsCluster.getFileSystem();
|
||||
assertTrue("Not an HDFS: " + fs.getUri(),
|
||||
fs instanceof DistributedFileSystem);
|
||||
assertTrue(
|
||||
fs instanceof DistributedFileSystem, "Not an HDFS: " + fs.getUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -18,7 +18,7 @@
|
||||
|
||||
package org.apache.hadoop.cli;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.apache.hadoop.cli.util.CLICommand;
|
||||
import org.apache.hadoop.cli.util.CommandExecutor.Result;
|
||||
@ -27,16 +27,16 @@
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestDeleteCLI extends CLITestHelperDFS {
|
||||
protected MiniDFSCluster dfsCluster = null;
|
||||
protected FileSystem fs = null;
|
||||
protected String namenode = null;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -49,11 +49,11 @@ public void setUp() throws Exception {
|
||||
namenode = conf.get(DFSConfigKeys.FS_DEFAULT_NAME_KEY, "file:///");
|
||||
|
||||
fs = dfsCluster.getFileSystem();
|
||||
assertTrue("Not an HDFS: " + fs.getUri(),
|
||||
fs instanceof DistributedFileSystem);
|
||||
assertTrue(
|
||||
fs instanceof DistributedFileSystem, "Not an HDFS: " + fs.getUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -24,10 +24,10 @@
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Rule;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.rules.Timeout;
|
||||
import org.xml.sax.SAXException;
|
||||
|
||||
@ -40,7 +40,7 @@ public class TestErasureCodingCLI extends CLITestHelper {
|
||||
@Rule
|
||||
public Timeout globalTimeout = new Timeout(300000);
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -62,7 +62,7 @@ protected String getTestFile() {
|
||||
return "testErasureCodingConf.xml";
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -18,7 +18,7 @@
|
||||
|
||||
package org.apache.hadoop.cli;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.apache.hadoop.cli.util.CLICommand;
|
||||
import org.apache.hadoop.cli.util.CommandExecutor.Result;
|
||||
@ -28,9 +28,9 @@
|
||||
import org.apache.hadoop.hdfs.HDFSPolicyProvider;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.authorize.PolicyProvider;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestHDFSCLI extends CLITestHelperDFS {
|
||||
|
||||
@ -38,7 +38,7 @@ public class TestHDFSCLI extends CLITestHelperDFS {
|
||||
protected FileSystem fs = null;
|
||||
protected String namenode = null;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -63,8 +63,8 @@ public void setUp() throws Exception {
|
||||
username = System.getProperty("user.name");
|
||||
|
||||
fs = dfsCluster.getFileSystem();
|
||||
assertTrue("Not a HDFS: "+fs.getUri(),
|
||||
fs instanceof DistributedFileSystem);
|
||||
assertTrue(
|
||||
fs instanceof DistributedFileSystem, "Not a HDFS: " + fs.getUri());
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -72,7 +72,7 @@ protected String getTestFile() {
|
||||
return "testHDFSConf.xml";
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -18,7 +18,7 @@
|
||||
|
||||
package org.apache.hadoop.cli;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import org.apache.hadoop.cli.util.CLICommand;
|
||||
import org.apache.hadoop.cli.util.CommandExecutor.Result;
|
||||
@ -28,16 +28,16 @@
|
||||
import org.apache.hadoop.hdfs.HDFSPolicyProvider;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.authorize.PolicyProvider;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestXAttrCLI extends CLITestHelperDFS {
|
||||
protected MiniDFSCluster dfsCluster = null;
|
||||
protected FileSystem fs = null;
|
||||
protected String namenode = null;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
@ -53,8 +53,8 @@ public void setUp() throws Exception {
|
||||
username = System.getProperty("user.name");
|
||||
|
||||
fs = dfsCluster.getFileSystem();
|
||||
assertTrue("Not a HDFS: "+fs.getUri(),
|
||||
fs instanceof DistributedFileSystem);
|
||||
assertTrue(
|
||||
fs instanceof DistributedFileSystem, "Not a HDFS: " + fs.getUri());
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -62,7 +62,7 @@ protected String getTestFile() {
|
||||
return "testXAttrConf.xml";
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
|
@ -65,12 +65,11 @@
|
||||
import org.apache.hadoop.net.unix.TemporarySocketDirectory;
|
||||
import org.apache.hadoop.security.token.SecretManager.InvalidToken;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Assume;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Assumptions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import java.util.function.Supplier;
|
||||
|
||||
@ -85,7 +84,7 @@ public class TestEnhancedByteBufferAccess {
|
||||
|
||||
static private CacheManipulator prevCacheManipulator;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void init() {
|
||||
sockDir = new TemporarySocketDirectory();
|
||||
DomainSocket.disableBindPathValidation();
|
||||
@ -99,7 +98,7 @@ public void mlock(String identifier,
|
||||
});
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardown() {
|
||||
// Restore the original CacheManipulator
|
||||
NativeIO.POSIX.setCacheManipulator(prevCacheManipulator);
|
||||
@ -116,8 +115,8 @@ private static byte[] byteBufferToArray(ByteBuffer buf) {
|
||||
(int) NativeIO.POSIX.getCacheManipulator().getOperatingSystemPageSize();
|
||||
|
||||
public static HdfsConfiguration initZeroCopyTest() {
|
||||
Assume.assumeTrue(NativeIO.isAvailable());
|
||||
Assume.assumeTrue(SystemUtils.IS_OS_UNIX);
|
||||
Assumptions.assumeTrue(NativeIO.isAvailable());
|
||||
Assumptions.assumeTrue(SystemUtils.IS_OS_UNIX);
|
||||
HdfsConfiguration conf = new HdfsConfiguration();
|
||||
conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true);
|
||||
conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, BLOCK_SIZE);
|
||||
@ -152,10 +151,10 @@ public void testZeroCopyReads() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -165,13 +164,13 @@ public void testZeroCopyReads() throws Exception {
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
ByteBuffer result = fsIn.read(null, BLOCK_SIZE,
|
||||
EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assertions.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
HdfsDataInputStream dfsIn = (HdfsDataInputStream)fsIn;
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalBytesRead());
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalZeroCopyBytesRead());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
byteBufferToArray(result));
|
||||
fsIn.releaseBuffer(result);
|
||||
} finally {
|
||||
@ -198,10 +197,10 @@ public void testShortZeroCopyReads() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -214,20 +213,20 @@ public void testShortZeroCopyReads() throws Exception {
|
||||
HdfsDataInputStream dfsIn = (HdfsDataInputStream)fsIn;
|
||||
ByteBuffer result =
|
||||
dfsIn.read(null, 2 * BLOCK_SIZE, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalBytesRead());
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalZeroCopyBytesRead());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
byteBufferToArray(result));
|
||||
dfsIn.releaseBuffer(result);
|
||||
|
||||
// Try to read (1 + ${BLOCK_SIZE}), but only get ${BLOCK_SIZE} because of the block size.
|
||||
result =
|
||||
dfsIn.read(null, 1 + BLOCK_SIZE, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, BLOCK_SIZE, 2 * BLOCK_SIZE),
|
||||
Assertions.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, BLOCK_SIZE, 2 * BLOCK_SIZE),
|
||||
byteBufferToArray(result));
|
||||
dfsIn.releaseBuffer(result);
|
||||
} finally {
|
||||
@ -255,10 +254,10 @@ public void testZeroCopyReadsNoFallback() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -270,17 +269,17 @@ public void testZeroCopyReadsNoFallback() throws Exception {
|
||||
ByteBuffer result;
|
||||
try {
|
||||
result = dfsIn.read(null, BLOCK_SIZE + 1, EnumSet.noneOf(ReadOption.class));
|
||||
Assert.fail("expected UnsupportedOperationException");
|
||||
Assertions.fail("expected UnsupportedOperationException");
|
||||
} catch (UnsupportedOperationException e) {
|
||||
// expected
|
||||
}
|
||||
result = dfsIn.read(null, BLOCK_SIZE, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE, result.remaining());
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalBytesRead());
|
||||
Assert.assertEquals(BLOCK_SIZE,
|
||||
Assertions.assertEquals(BLOCK_SIZE,
|
||||
dfsIn.getReadStatistics().getTotalZeroCopyBytesRead());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0, BLOCK_SIZE),
|
||||
byteBufferToArray(result));
|
||||
} finally {
|
||||
if (fsIn != null) fsIn.close();
|
||||
@ -311,16 +310,16 @@ public void visit(int numOutstandingMmaps,
|
||||
LinkedMap evictable,
|
||||
LinkedMap evictableMmapped) {
|
||||
if (expectedNumOutstandingMmaps >= 0) {
|
||||
Assert.assertEquals(expectedNumOutstandingMmaps, numOutstandingMmaps);
|
||||
Assertions.assertEquals(expectedNumOutstandingMmaps, numOutstandingMmaps);
|
||||
}
|
||||
if (expectedNumReplicas >= 0) {
|
||||
Assert.assertEquals(expectedNumReplicas, replicas.size());
|
||||
Assertions.assertEquals(expectedNumReplicas, replicas.size());
|
||||
}
|
||||
if (expectedNumEvictable >= 0) {
|
||||
Assert.assertEquals(expectedNumEvictable, evictable.size());
|
||||
Assertions.assertEquals(expectedNumEvictable, evictable.size());
|
||||
}
|
||||
if (expectedNumMmapedEvictable >= 0) {
|
||||
Assert.assertEquals(expectedNumMmapedEvictable, evictableMmapped.size());
|
||||
Assertions.assertEquals(expectedNumMmapedEvictable, evictableMmapped.size());
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -346,10 +345,10 @@ public void testZeroCopyMmapCache() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -378,10 +377,10 @@ public void visit(int numOutstandingMmaps,
|
||||
LinkedMap evictableMmapped) {
|
||||
ShortCircuitReplica replica = replicas.get(
|
||||
new ExtendedBlockId(firstBlock.getBlockId(), firstBlock.getBlockPoolId()));
|
||||
Assert.assertNotNull(replica);
|
||||
Assert.assertTrue(replica.hasMmap());
|
||||
Assertions.assertNotNull(replica);
|
||||
Assertions.assertTrue(replica.hasMmap());
|
||||
// The replica should not yet be evictable, since we have it open.
|
||||
Assert.assertNull(replica.getEvictableTimeNs());
|
||||
Assertions.assertNull(replica.getEvictableTimeNs());
|
||||
}
|
||||
});
|
||||
|
||||
@ -449,10 +448,10 @@ public void testHdfsFallbackReads() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -493,22 +492,22 @@ private static void testFallbackImpl(InputStream stream,
|
||||
stream instanceof ByteBufferReadable);
|
||||
|
||||
ByteBuffer result = ByteBufferUtil.fallbackRead(stream, bufferPool, 10);
|
||||
Assert.assertEquals(10, result.remaining());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0, 10),
|
||||
Assertions.assertEquals(10, result.remaining());
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0, 10),
|
||||
byteBufferToArray(result));
|
||||
|
||||
result = ByteBufferUtil.fallbackRead(stream, bufferPool, 5000);
|
||||
Assert.assertEquals(5000, result.remaining());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 10, 5010),
|
||||
Assertions.assertEquals(5000, result.remaining());
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 10, 5010),
|
||||
byteBufferToArray(result));
|
||||
|
||||
result = ByteBufferUtil.fallbackRead(stream, bufferPool, 9999999);
|
||||
Assert.assertEquals(11375, result.remaining());
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 5010, 16385),
|
||||
Assertions.assertEquals(11375, result.remaining());
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 5010, 16385),
|
||||
byteBufferToArray(result));
|
||||
|
||||
result = ByteBufferUtil.fallbackRead(stream, bufferPool, 10);
|
||||
Assert.assertNull(result);
|
||||
Assertions.assertNull(result);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -533,10 +532,10 @@ public void testFallbackRead() throws Exception {
|
||||
try {
|
||||
DFSTestUtil.waitReplication(fs, TEST_PATH, (short)1);
|
||||
} catch (InterruptedException e) {
|
||||
Assert.fail("unexpected InterruptedException during " +
|
||||
Assertions.fail("unexpected InterruptedException during " +
|
||||
"waitReplication: " + e);
|
||||
} catch (TimeoutException e) {
|
||||
Assert.fail("unexpected TimeoutException during " +
|
||||
Assertions.fail("unexpected TimeoutException during " +
|
||||
"waitReplication: " + e);
|
||||
}
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
@ -618,7 +617,7 @@ public void testZeroCopyReadOfCachedData() throws Exception {
|
||||
try {
|
||||
result = fsIn.read(null, TEST_FILE_LENGTH / 2,
|
||||
EnumSet.noneOf(ReadOption.class));
|
||||
Assert.fail("expected UnsupportedOperationException");
|
||||
Assertions.fail("expected UnsupportedOperationException");
|
||||
} catch (UnsupportedOperationException e) {
|
||||
// expected
|
||||
}
|
||||
@ -637,9 +636,9 @@ public void testZeroCopyReadOfCachedData() throws Exception {
|
||||
result = fsIn.read(null, TEST_FILE_LENGTH,
|
||||
EnumSet.noneOf(ReadOption.class));
|
||||
} catch (UnsupportedOperationException e) {
|
||||
Assert.fail("expected to be able to read cached file via zero-copy");
|
||||
Assertions.fail("expected to be able to read cached file via zero-copy");
|
||||
}
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0,
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0,
|
||||
BLOCK_SIZE), byteBufferToArray(result));
|
||||
// Test that files opened after the cache operation has finished
|
||||
// still get the benefits of zero-copy (regression test for HDFS-6086)
|
||||
@ -648,9 +647,9 @@ public void testZeroCopyReadOfCachedData() throws Exception {
|
||||
result2 = fsIn2.read(null, TEST_FILE_LENGTH,
|
||||
EnumSet.noneOf(ReadOption.class));
|
||||
} catch (UnsupportedOperationException e) {
|
||||
Assert.fail("expected to be able to read cached file via zero-copy");
|
||||
Assertions.fail("expected to be able to read cached file via zero-copy");
|
||||
}
|
||||
Assert.assertArrayEquals(Arrays.copyOfRange(original, 0,
|
||||
Assertions.assertArrayEquals(Arrays.copyOfRange(original, 0,
|
||||
BLOCK_SIZE), byteBufferToArray(result2));
|
||||
fsIn2.releaseBuffer(result2);
|
||||
fsIn2.close();
|
||||
@ -688,10 +687,10 @@ public void visit(int numOutstandingMmaps,
|
||||
Map<ExtendedBlockId, InvalidToken> failedLoads,
|
||||
LinkedMap evictable,
|
||||
LinkedMap evictableMmapped) {
|
||||
Assert.assertEquals(expectedOutstandingMmaps, numOutstandingMmaps);
|
||||
Assertions.assertEquals(expectedOutstandingMmaps, numOutstandingMmaps);
|
||||
ShortCircuitReplica replica =
|
||||
replicas.get(ExtendedBlockId.fromExtendedBlock(block));
|
||||
Assert.assertNotNull(replica);
|
||||
Assertions.assertNotNull(replica);
|
||||
Slot slot = replica.getSlot();
|
||||
if ((expectedIsAnchorable != slot.isAnchorable()) ||
|
||||
(expectedIsAnchored != slot.isAnchored())) {
|
||||
@ -734,7 +733,7 @@ public void testClientMmapDisable() throws Exception {
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
try {
|
||||
fsIn.read(null, 1, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.fail("expected zero-copy read to fail when client mmaps " +
|
||||
Assertions.fail("expected zero-copy read to fail when client mmaps " +
|
||||
"were disabled.");
|
||||
} catch (UnsupportedOperationException e) {
|
||||
}
|
||||
@ -764,7 +763,7 @@ public void testClientMmapDisable() throws Exception {
|
||||
// Test EOF behavior
|
||||
IOUtils.skipFully(fsIn, TEST_FILE_LENGTH - 1);
|
||||
buf = fsIn.read(null, 1, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(null, buf);
|
||||
Assertions.assertEquals(null, buf);
|
||||
} finally {
|
||||
if (fsIn != null) fsIn.close();
|
||||
if (fs != null) fs.close();
|
||||
@ -774,7 +773,7 @@ public void testClientMmapDisable() throws Exception {
|
||||
|
||||
@Test
|
||||
public void test2GBMmapLimit() throws Exception {
|
||||
Assume.assumeTrue(BlockReaderTestUtil.shouldTestLargeFiles());
|
||||
Assumptions.assumeTrue(BlockReaderTestUtil.shouldTestLargeFiles());
|
||||
HdfsConfiguration conf = initZeroCopyTest();
|
||||
final long TEST_FILE_LENGTH = 2469605888L;
|
||||
conf.set(DFSConfigKeys.DFS_CHECKSUM_TYPE_KEY, "NULL");
|
||||
@ -795,20 +794,20 @@ public void test2GBMmapLimit() throws Exception {
|
||||
|
||||
fsIn = fs.open(TEST_PATH);
|
||||
buf1 = fsIn.read(null, 1, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(1, buf1.remaining());
|
||||
Assertions.assertEquals(1, buf1.remaining());
|
||||
fsIn.releaseBuffer(buf1);
|
||||
buf1 = null;
|
||||
fsIn.seek(2147483640L);
|
||||
buf1 = fsIn.read(null, 1024, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(7, buf1.remaining());
|
||||
Assert.assertEquals(Integer.MAX_VALUE, buf1.limit());
|
||||
Assertions.assertEquals(7, buf1.remaining());
|
||||
Assertions.assertEquals(Integer.MAX_VALUE, buf1.limit());
|
||||
fsIn.releaseBuffer(buf1);
|
||||
buf1 = null;
|
||||
Assert.assertEquals(2147483647L, fsIn.getPos());
|
||||
Assertions.assertEquals(2147483647L, fsIn.getPos());
|
||||
try {
|
||||
buf1 = fsIn.read(null, 1024,
|
||||
EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.fail("expected UnsupportedOperationException");
|
||||
Assertions.fail("expected UnsupportedOperationException");
|
||||
} catch (UnsupportedOperationException e) {
|
||||
// expected; can't read past 2GB boundary.
|
||||
}
|
||||
@ -825,13 +824,13 @@ public void test2GBMmapLimit() throws Exception {
|
||||
fsIn2 = fs.open(TEST_PATH2);
|
||||
fsIn2.seek(2147483640L);
|
||||
buf2 = fsIn2.read(null, 1024, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(8, buf2.remaining());
|
||||
Assert.assertEquals(2147483648L, fsIn2.getPos());
|
||||
Assertions.assertEquals(8, buf2.remaining());
|
||||
Assertions.assertEquals(2147483648L, fsIn2.getPos());
|
||||
fsIn2.releaseBuffer(buf2);
|
||||
buf2 = null;
|
||||
buf2 = fsIn2.read(null, 1024, EnumSet.of(ReadOption.SKIP_CHECKSUMS));
|
||||
Assert.assertEquals(1024, buf2.remaining());
|
||||
Assert.assertEquals(2147484672L, fsIn2.getPos());
|
||||
Assertions.assertEquals(1024, buf2.remaining());
|
||||
Assertions.assertEquals(2147484672L, fsIn2.getPos());
|
||||
fsIn2.releaseBuffer(buf2);
|
||||
buf2 = null;
|
||||
} finally {
|
||||
|
@ -27,10 +27,10 @@
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
public class TestFcHdfsCreateMkdir extends
|
||||
FileContextCreateMkdirBaseTest {
|
||||
@ -44,7 +44,7 @@ protected FileContextTestHelper createFileContextHelper() {
|
||||
}
|
||||
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining()
|
||||
throws IOException, LoginException, URISyntaxException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
@ -56,7 +56,7 @@ public static void clusterSetupAtBegining()
|
||||
}
|
||||
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -64,13 +64,13 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
}
|
||||
|
||||
@Override
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
super.tearDown();
|
||||
}
|
||||
|
@ -27,10 +27,10 @@
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
public class TestFcHdfsPermission extends FileContextPermissionBase {
|
||||
|
||||
@ -51,7 +51,7 @@ protected FileContext getFileContext() {
|
||||
return fc;
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining()
|
||||
throws IOException, LoginException, URISyntaxException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
@ -63,7 +63,7 @@ public static void clusterSetupAtBegining()
|
||||
}
|
||||
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -71,13 +71,13 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
}
|
||||
|
||||
@Override
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
super.tearDown();
|
||||
}
|
||||
|
@ -30,12 +30,13 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import static org.apache.hadoop.fs.FileContextTestHelper.*;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestFcHdfsSetUMask {
|
||||
|
||||
@ -78,7 +79,7 @@ public class TestFcHdfsSetUMask {
|
||||
private static final FsPermission WIDE_OPEN_TEST_UMASK = FsPermission
|
||||
.createImmutable((short) (0777 ^ 0777));
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining()
|
||||
throws IOException, LoginException, URISyntaxException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
@ -91,20 +92,20 @@ public static void clusterSetupAtBegining()
|
||||
fc.mkdir(defaultWorkingDirectory, FileContext.DEFAULT_PERM, true);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fc.setUMask(WIDE_OPEN_TEST_UMASK);
|
||||
fc.mkdir(fileContextTestHelper.getTestRootPath(fc), FileContext.DEFAULT_PERM, true);
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fc.delete(fileContextTestHelper.getTestRootPath(fc), true);
|
||||
}
|
||||
@ -194,8 +195,8 @@ public void testMkdirWithExistingDir(FsPermission umask,
|
||||
fc.setUMask(umask);
|
||||
fc.mkdir(f, FileContext.DEFAULT_PERM, true);
|
||||
Assert.assertTrue(isDir(fc, f));
|
||||
Assert.assertEquals("permissions on directory are wrong",
|
||||
expectedPerms, fc.getFileStatus(f).getPermission());
|
||||
Assertions.assertEquals(
|
||||
expectedPerms, fc.getFileStatus(f).getPermission(), "permissions on directory are wrong");
|
||||
}
|
||||
|
||||
public void testMkdirRecursiveWithNonExistingDir(FsPermission umask,
|
||||
@ -205,11 +206,11 @@ public void testMkdirRecursiveWithNonExistingDir(FsPermission umask,
|
||||
fc.setUMask(umask);
|
||||
fc.mkdir(f, FileContext.DEFAULT_PERM, true);
|
||||
Assert.assertTrue(isDir(fc, f));
|
||||
Assert.assertEquals("permissions on directory are wrong",
|
||||
expectedPerms, fc.getFileStatus(f).getPermission());
|
||||
Assertions.assertEquals(
|
||||
expectedPerms, fc.getFileStatus(f).getPermission(), "permissions on directory are wrong");
|
||||
Path fParent = fileContextTestHelper.getTestRootPath(fc, "NonExistant2");
|
||||
Assert.assertEquals("permissions on parent directory are wrong",
|
||||
expectedParentPerms, fc.getFileStatus(fParent).getPermission());
|
||||
Assertions.assertEquals(
|
||||
expectedParentPerms, fc.getFileStatus(fParent).getPermission(), "permissions on parent directory are wrong");
|
||||
}
|
||||
|
||||
|
||||
@ -219,8 +220,8 @@ public void testCreateRecursiveWithExistingDir(FsPermission umask,
|
||||
fc.setUMask(umask);
|
||||
createFile(fc, f);
|
||||
Assert.assertTrue(isFile(fc, f));
|
||||
Assert.assertEquals("permissions on file are wrong",
|
||||
expectedPerms , fc.getFileStatus(f).getPermission());
|
||||
Assertions.assertEquals(
|
||||
expectedPerms, fc.getFileStatus(f).getPermission(), "permissions on file are wrong");
|
||||
}
|
||||
|
||||
|
||||
@ -233,10 +234,10 @@ public void testCreateRecursiveWithNonExistingDir(FsPermission umask,
|
||||
fc.setUMask(umask);
|
||||
createFile(fc, f);
|
||||
Assert.assertTrue(isFile(fc, f));
|
||||
Assert.assertEquals("permissions on file are wrong",
|
||||
expectedFilePerms, fc.getFileStatus(f).getPermission());
|
||||
Assert.assertEquals("permissions on parent directory are wrong",
|
||||
expectedDirPerms, fc.getFileStatus(fParent).getPermission());
|
||||
Assertions.assertEquals(
|
||||
expectedFilePerms, fc.getFileStatus(f).getPermission(), "permissions on file are wrong");
|
||||
Assertions.assertEquals(
|
||||
expectedDirPerms, fc.getFileStatus(fParent).getPermission(), "permissions on parent directory are wrong");
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -18,7 +18,7 @@
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.apache.hadoop.test.PlatformAssumptions.assumeNotWindows;
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
@ -26,6 +26,7 @@
|
||||
import java.util.regex.Pattern;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.collect.Ordering;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.apache.commons.lang3.StringUtils;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.permission.FsPermission;
|
||||
@ -34,7 +35,10 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.INodeId;
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.*;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestGlobPaths {
|
||||
|
||||
@ -65,7 +69,7 @@ public boolean accept(Path path) {
|
||||
static private String USER_DIR;
|
||||
private final Path[] path = new Path[NUM_OF_PATHS];
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setUp() throws Exception {
|
||||
final Configuration conf = new HdfsConfiguration();
|
||||
dfsCluster = new MiniDFSCluster.Builder(conf).build();
|
||||
@ -81,7 +85,7 @@ public static void setUp() throws Exception {
|
||||
USER_DIR = fs.getHomeDirectory().toUri().getPath().toString();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() throws Exception {
|
||||
if(dfsCluster!=null) {
|
||||
dfsCluster.shutdown();
|
||||
@ -102,8 +106,8 @@ public void testCRInPathGlob() throws IOException {
|
||||
fs.createNewFile(fNormal);
|
||||
fs.createNewFile(fWithCR);
|
||||
statuses = fs.globStatus(new Path(d1, "f1*"));
|
||||
assertEquals("Expected both normal and CR-carrying files in result: ",
|
||||
2, statuses.length);
|
||||
assertEquals(
|
||||
2, statuses.length, "Expected both normal and CR-carrying files in result: ");
|
||||
cleanupDFS();
|
||||
}
|
||||
|
||||
@ -892,14 +896,14 @@ void run() throws Exception {
|
||||
// Test simple glob
|
||||
FileStatus[] statuses = wrap.globStatus(new Path(USER_DIR + "/alpha/*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alpha/beta", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alpha/beta", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
// Test glob through symlink
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alphaLink/*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alphaLink/beta", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alphaLink/beta", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
// If the terminal path component in a globbed path is a symlink,
|
||||
// we don't dereference that link.
|
||||
@ -907,20 +911,20 @@ void run() throws Exception {
|
||||
+ "/alphaLink/betaLink"), false);
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alpha/betaLi*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alpha/betaLink", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alpha/betaLink", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
// todo: test symlink-to-symlink-to-dir, etc.
|
||||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobWithSymlinksOnFS() throws Exception {
|
||||
testOnFileSystem(new TestGlobWithSymlinks(false));
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobWithSymlinksOnFC() throws Exception {
|
||||
testOnFileContext(new TestGlobWithSymlinks(true));
|
||||
@ -951,20 +955,20 @@ void run() throws Exception {
|
||||
// Test glob through symlink to a symlink to a directory
|
||||
FileStatus statuses[] = wrap.globStatus(new Path(USER_DIR
|
||||
+ "/alphaLinkLink"), new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alphaLinkLink", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alphaLinkLink", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alphaLinkLink/*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alphaLinkLink/beta", statuses[0]
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alphaLinkLink/beta", statuses[0]
|
||||
.getPath().toUri().getPath());
|
||||
// Test glob of dangling symlink (theta does not actually exist)
|
||||
wrap.createSymlink(new Path(USER_DIR + "theta"), new Path(USER_DIR
|
||||
+ "/alpha/kappa"), false);
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alpha/kappa/kappa"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertNull(statuses);
|
||||
Assertions.assertNull(statuses);
|
||||
// Test glob of symlinks
|
||||
wrap.createFile(USER_DIR + "/alpha/beta/gamma");
|
||||
wrap.createSymlink(new Path(USER_DIR + "gamma"), new Path(USER_DIR
|
||||
@ -975,8 +979,8 @@ void run() throws Exception {
|
||||
USER_DIR + "/alpha/beta/gammaLinkLinkLink"), false);
|
||||
statuses = wrap.globStatus(new Path(USER_DIR
|
||||
+ "/alpha/*/gammaLinkLinkLink"), new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alpha/beta/gammaLinkLinkLink",
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alpha/beta/gammaLinkLinkLink",
|
||||
statuses[0].getPath().toUri().getPath());
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alpha/beta/*"),
|
||||
new AcceptAllPathFilter());
|
||||
@ -992,17 +996,17 @@ void run() throws Exception {
|
||||
statuses = wrap.globStatus(
|
||||
new Path(USER_DIR + "/tweedledee/unobtainium"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertNull(statuses);
|
||||
Assertions.assertNull(statuses);
|
||||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobWithSymlinksToSymlinksOnFS() throws Exception {
|
||||
testOnFileSystem(new TestGlobWithSymlinksToSymlinks(false));
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobWithSymlinksToSymlinksOnFC() throws Exception {
|
||||
testOnFileContext(new TestGlobWithSymlinksToSymlinks(true));
|
||||
@ -1032,11 +1036,11 @@ void run() throws Exception {
|
||||
// PathFilter
|
||||
FileStatus statuses[] = wrap.globStatus(
|
||||
new Path(USER_DIR + "/alpha/beta"), new AcceptPathsEndingInZ());
|
||||
Assert.assertNull(statuses);
|
||||
Assertions.assertNull(statuses);
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/alphaLinkz/betaz"),
|
||||
new AcceptPathsEndingInZ());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals(USER_DIR + "/alphaLinkz/betaz", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(USER_DIR + "/alphaLinkz/betaz", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
statuses = wrap.globStatus(new Path(USER_DIR + "/*/*"),
|
||||
new AcceptPathsEndingInZ());
|
||||
@ -1050,13 +1054,13 @@ void run() throws Exception {
|
||||
}
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobSymlinksWithCustomPathFilterOnFS() throws Exception {
|
||||
testOnFileSystem(new TestGlobSymlinksWithCustomPathFilter(false));
|
||||
}
|
||||
|
||||
@Ignore
|
||||
@Disabled
|
||||
@Test
|
||||
public void testGlobSymlinksWithCustomPathFilterOnFC() throws Exception {
|
||||
testOnFileContext(new TestGlobSymlinksWithCustomPathFilter(true));
|
||||
@ -1078,22 +1082,22 @@ void run() throws Exception {
|
||||
+ "/alphaLink"), false);
|
||||
FileStatus statuses[] = wrap.globStatus(
|
||||
new Path(USER_DIR + "/alphaLink"), new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Path path = statuses[0].getPath();
|
||||
Assert.assertEquals(USER_DIR + "/alpha", path.toUri().getPath());
|
||||
Assert.assertEquals("hdfs", path.toUri().getScheme());
|
||||
Assertions.assertEquals(USER_DIR + "/alpha", path.toUri().getPath());
|
||||
Assertions.assertEquals("hdfs", path.toUri().getScheme());
|
||||
|
||||
// FileContext can list a file:/// URI.
|
||||
// Since everyone should have the root directory, we list that.
|
||||
statuses = fc.util().globStatus(new Path("file:///"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Path filePath = statuses[0].getPath();
|
||||
Assert.assertEquals("file", filePath.toUri().getScheme());
|
||||
Assert.assertEquals("/", filePath.toUri().getPath());
|
||||
Assertions.assertEquals("file", filePath.toUri().getScheme());
|
||||
Assertions.assertEquals("/", filePath.toUri().getPath());
|
||||
|
||||
// The FileSystem should have scheme 'hdfs'
|
||||
Assert.assertEquals("hdfs", fs.getScheme());
|
||||
Assertions.assertEquals("hdfs", fs.getScheme());
|
||||
}
|
||||
}
|
||||
|
||||
@ -1176,7 +1180,7 @@ void run() throws Exception {
|
||||
try {
|
||||
wrap.globStatus(new Path("/no*/*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.fail("expected to get an AccessControlException when " +
|
||||
Assertions.fail("expected to get an AccessControlException when " +
|
||||
"globbing through a directory we don't have permissions " +
|
||||
"to list.");
|
||||
} catch (AccessControlException ioe) {
|
||||
@ -1240,8 +1244,8 @@ void run() throws Exception {
|
||||
privWrap.setOwner(new Path("/"), newOwner, null);
|
||||
FileStatus[] status =
|
||||
wrap.globStatus(rootPath, new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, status.length);
|
||||
Assert.assertEquals(newOwner, status[0].getOwner());
|
||||
Assertions.assertEquals(1, status.length);
|
||||
Assertions.assertEquals(newOwner, status[0].getOwner());
|
||||
privWrap.setOwner(new Path("/"), oldRootStatus.getOwner(), null);
|
||||
}
|
||||
}
|
||||
@ -1273,8 +1277,8 @@ void run() throws Exception {
|
||||
FileStatus[] statuses =
|
||||
wrap.globStatus(new Path("/filed*/alpha"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals("/filed_away/alpha", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals("/filed_away/alpha", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
privWrap.mkdir(new Path("/filed_away/alphabet"),
|
||||
new FsPermission((short)0777), true);
|
||||
@ -1282,8 +1286,8 @@ void run() throws Exception {
|
||||
new FsPermission((short)0777), true);
|
||||
statuses = wrap.globStatus(new Path("/filed*/alph*/*b*"),
|
||||
new AcceptAllPathFilter());
|
||||
Assert.assertEquals(1, statuses.length);
|
||||
Assert.assertEquals("/filed_away/alphabet/abc", statuses[0].getPath()
|
||||
Assertions.assertEquals(1, statuses.length);
|
||||
Assertions.assertEquals("/filed_away/alphabet/abc", statuses[0].getPath()
|
||||
.toUri().getPath());
|
||||
} finally {
|
||||
privWrap.delete(new Path("/filed"), true);
|
||||
@ -1308,12 +1312,12 @@ public void testLocalFilesystem() throws Exception {
|
||||
FileSystem fs = FileSystem.getLocal(conf);
|
||||
String localTmp = System.getProperty("java.io.tmpdir");
|
||||
Path base = new Path(new Path(localTmp), UUID.randomUUID().toString());
|
||||
Assert.assertTrue(fs.mkdirs(base));
|
||||
Assert.assertTrue(fs.mkdirs(new Path(base, "e")));
|
||||
Assert.assertTrue(fs.mkdirs(new Path(base, "c")));
|
||||
Assert.assertTrue(fs.mkdirs(new Path(base, "a")));
|
||||
Assert.assertTrue(fs.mkdirs(new Path(base, "d")));
|
||||
Assert.assertTrue(fs.mkdirs(new Path(base, "b")));
|
||||
Assertions.assertTrue(fs.mkdirs(base));
|
||||
Assertions.assertTrue(fs.mkdirs(new Path(base, "e")));
|
||||
Assertions.assertTrue(fs.mkdirs(new Path(base, "c")));
|
||||
Assertions.assertTrue(fs.mkdirs(new Path(base, "a")));
|
||||
Assertions.assertTrue(fs.mkdirs(new Path(base, "d")));
|
||||
Assertions.assertTrue(fs.mkdirs(new Path(base, "b")));
|
||||
fs.deleteOnExit(base);
|
||||
FileStatus[] status = fs.globStatus(new Path(base, "*"));
|
||||
ArrayList list = new ArrayList();
|
||||
@ -1321,7 +1325,7 @@ public void testLocalFilesystem() throws Exception {
|
||||
list.add(f.getPath().toString());
|
||||
}
|
||||
boolean sorted = Ordering.natural().isOrdered(list);
|
||||
Assert.assertTrue(sorted);
|
||||
Assertions.assertTrue(sorted);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -19,7 +19,7 @@
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.apache.hadoop.fs.FileContextTestHelper.exists;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
@ -36,12 +36,13 @@
|
||||
import org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
|
||||
import org.apache.hadoop.ipc.RemoteException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
public class TestHDFSFileContextMainOperations extends
|
||||
FileContextMainOperationsBaseTest {
|
||||
@ -54,7 +55,7 @@ protected FileContextTestHelper createFileContextHelper() {
|
||||
return new FileContextTestHelper("/tmp/TestHDFSFileContextMainOperations");
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
cluster = new MiniDFSCluster.Builder(CONF).numDataNodes(2).build();
|
||||
@ -80,7 +81,7 @@ private static void restartCluster() throws IOException, LoginException {
|
||||
fc.mkdir(defaultWorkingDirectory, FileContext.DEFAULT_PERM, true);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -89,13 +90,13 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
super.setUp();
|
||||
}
|
||||
|
||||
@Override
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
super.tearDown();
|
||||
}
|
||||
@ -134,16 +135,16 @@ public void testTruncate() throws Exception {
|
||||
|
||||
boolean isReady = fc.truncate(file, newLength);
|
||||
|
||||
Assert.assertTrue("Recovery is not expected.", isReady);
|
||||
Assertions.assertTrue(isReady, "Recovery is not expected.");
|
||||
|
||||
FileStatus fileStatus = fc.getFileStatus(file);
|
||||
Assert.assertEquals(fileStatus.getLen(), newLength);
|
||||
Assertions.assertEquals(fileStatus.getLen(), newLength);
|
||||
AppendTestUtil.checkFullFile(fs, file, newLength, data, file.toString());
|
||||
|
||||
ContentSummary cs = fs.getContentSummary(dir);
|
||||
Assert.assertEquals("Bad disk space usage", cs.getSpaceConsumed(),
|
||||
newLength * repl);
|
||||
Assert.assertTrue(fs.delete(dir, true));
|
||||
Assertions.assertEquals(cs.getSpaceConsumed(),
|
||||
newLength * repl, "Bad disk space usage");
|
||||
Assertions.assertTrue(fs.delete(dir, true));
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -280,8 +281,8 @@ public void testEditsLogOldRename() throws Exception {
|
||||
fs = cluster.getFileSystem();
|
||||
src1 = getTestRootPath(fc, "testEditsLogOldRename/srcdir/src1");
|
||||
dst1 = getTestRootPath(fc, "testEditsLogOldRename/dstdir/dst1");
|
||||
Assert.assertFalse(fs.exists(src1)); // ensure src1 is already renamed
|
||||
Assert.assertTrue(fs.exists(dst1)); // ensure rename dst exists
|
||||
Assertions.assertFalse(fs.exists(src1)); // ensure src1 is already renamed
|
||||
Assertions.assertTrue(fs.exists(dst1)); // ensure rename dst exists
|
||||
}
|
||||
|
||||
/**
|
||||
@ -309,8 +310,8 @@ public void testEditsLogRename() throws Exception {
|
||||
fs = cluster.getFileSystem();
|
||||
src1 = getTestRootPath(fc, "testEditsLogRename/srcdir/src1");
|
||||
dst1 = getTestRootPath(fc, "testEditsLogRename/dstdir/dst1");
|
||||
Assert.assertFalse(fs.exists(src1)); // ensure src1 is already renamed
|
||||
Assert.assertTrue(fs.exists(dst1)); // ensure rename dst exists
|
||||
Assertions.assertFalse(fs.exists(src1)); // ensure src1 is already renamed
|
||||
Assertions.assertTrue(fs.exists(dst1)); // ensure rename dst exists
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -323,8 +324,8 @@ public void testIsValidNameInvalidNames() {
|
||||
};
|
||||
|
||||
for (String invalidName: invalidNames) {
|
||||
Assert.assertFalse(invalidName + " is not valid",
|
||||
fc.getDefaultFileSystem().isValidName(invalidName));
|
||||
Assert.assertFalse(
|
||||
fc.getDefaultFileSystem().isValidName(invalidName), invalidName + " is not valid");
|
||||
}
|
||||
}
|
||||
|
||||
@ -332,9 +333,9 @@ private void oldRename(Path src, Path dst, boolean renameSucceeds,
|
||||
boolean exception) throws Exception {
|
||||
DistributedFileSystem fs = cluster.getFileSystem();
|
||||
try {
|
||||
Assert.assertEquals(renameSucceeds, fs.rename(src, dst));
|
||||
Assertions.assertEquals(renameSucceeds, fs.rename(src, dst));
|
||||
} catch (Exception ex) {
|
||||
Assert.assertTrue(exception);
|
||||
Assertions.assertTrue(exception);
|
||||
}
|
||||
Assert.assertEquals(renameSucceeds, !exists(fc, src));
|
||||
Assert.assertEquals(renameSucceeds, exists(fc, dst));
|
||||
|
@ -17,12 +17,12 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import org.junit.Test;
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.apache.hadoop.util.NativeCodeLoader;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestHdfsNativeCodeLoader {
|
||||
static final Logger LOG =
|
||||
|
@ -19,14 +19,14 @@
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import java.io.File;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
import java.io.IOException;
|
||||
import java.util.EnumSet;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hdfs.DFSClient;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
@ -39,10 +39,10 @@
|
||||
import org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier;
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* Tests whether FileContext can resolve an hdfs path that has a symlink to
|
||||
@ -53,7 +53,7 @@ public class TestResolveHdfsSymlink {
|
||||
private static final FileContextTestHelper helper = new FileContextTestHelper();
|
||||
private static MiniDFSCluster cluster = null;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setUp() throws IOException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
conf.setBoolean(
|
||||
@ -63,7 +63,7 @@ public static void setUp() throws IOException {
|
||||
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -100,11 +100,11 @@ public void testFcResolveAfs() throws IOException, InterruptedException {
|
||||
|
||||
Set<AbstractFileSystem> afsList = fcHdfs
|
||||
.resolveAbstractFileSystems(alphaHdfsPathViaLink);
|
||||
Assert.assertEquals(2, afsList.size());
|
||||
Assertions.assertEquals(2, afsList.size());
|
||||
for (AbstractFileSystem afs : afsList) {
|
||||
if ((!afs.equals(fcHdfs.getDefaultFileSystem()))
|
||||
&& (!afs.equals(fcLocal.getDefaultFileSystem()))) {
|
||||
Assert.fail("Failed to resolve AFS correctly");
|
||||
Assertions.fail("Failed to resolve AFS correctly");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -25,8 +25,8 @@
|
||||
import org.apache.hadoop.security.ssl.KeyStoreTestUtil;
|
||||
import org.apache.hadoop.security.ssl.SSLFactory;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import javax.security.auth.login.LoginException;
|
||||
import java.io.File;
|
||||
@ -57,7 +57,7 @@ public class TestSWebHdfsFileContextMainOperations
|
||||
protected static final byte[] data = getFileData(numBlocks,
|
||||
getDefaultBlockSize());
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning()
|
||||
throws IOException, LoginException, URISyntaxException {
|
||||
|
||||
@ -104,7 +104,7 @@ public URI getWebhdfsUrl() {
|
||||
return webhdfsUrl;
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -17,8 +17,8 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
@ -38,9 +38,9 @@
|
||||
import org.apache.hadoop.hdfs.web.WebHdfsTestUtil;
|
||||
import org.apache.hadoop.ipc.RemoteException;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.slf4j.event.Level;
|
||||
|
||||
/**
|
||||
@ -84,7 +84,7 @@ protected IOException unwrapException(IOException e) {
|
||||
return e;
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void beforeClassSetup() throws Exception {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
conf.set(FsPermission.UMASK_LABEL, "000");
|
||||
@ -94,7 +94,7 @@ public static void beforeClassSetup() throws Exception {
|
||||
dfs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void afterClassTeardown() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
|
@ -17,18 +17,18 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
public class TestSymlinkHdfsFileContext extends TestSymlinkHdfs {
|
||||
|
||||
private static FileContext fc;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void testSetup() throws Exception {
|
||||
fc = FileContext.getFileContext(cluster.getURI(0));
|
||||
wrapper = new FileContextTestWrapper(fc, "/tmp/TestSymlinkHdfsFileContext");
|
||||
|
@ -17,28 +17,27 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
|
||||
public class TestSymlinkHdfsFileSystem extends TestSymlinkHdfs {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void testSetup() throws Exception {
|
||||
wrapper = new FileSystemTestWrapper(dfs, "/tmp/TestSymlinkHdfsFileSystem");
|
||||
}
|
||||
|
||||
@Override
|
||||
@Ignore("FileSystem adds missing authority in absolute URIs")
|
||||
@Disabled("FileSystem adds missing authority in absolute URIs")
|
||||
@Test(timeout=10000)
|
||||
public void testCreateWithPartQualPathFails() throws IOException {}
|
||||
|
||||
@Ignore("FileSystem#create creates parent directories," +
|
||||
@Disabled("FileSystem#create creates parent directories," +
|
||||
" so dangling links to directories are created")
|
||||
@Override
|
||||
@Test(timeout=10000)
|
||||
@ -56,7 +55,7 @@ public void testRecoverLease() throws IOException {
|
||||
wrapper.createSymlink(file, link, false);
|
||||
// Attempt recoverLease through a symlink
|
||||
boolean closed = dfs.recoverLease(link);
|
||||
assertTrue("Expected recoverLease to return true", closed);
|
||||
assertTrue(closed, "Expected recoverLease to return true");
|
||||
}
|
||||
|
||||
@Test(timeout=10000)
|
||||
@ -69,7 +68,7 @@ public void testIsFileClosed() throws IOException {
|
||||
wrapper.createSymlink(file, link, false);
|
||||
// Attempt recoverLease through a symlink
|
||||
boolean closed = dfs.isFileClosed(link);
|
||||
assertTrue("Expected isFileClosed to return true", closed);
|
||||
assertTrue(closed, "Expected isFileClosed to return true");
|
||||
}
|
||||
|
||||
@Test(timeout=10000)
|
||||
|
@ -26,9 +26,9 @@
|
||||
import org.apache.hadoop.hdfs.PeerCache;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Rule;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.rules.ExpectedException;
|
||||
import org.mockito.Mockito;
|
||||
|
||||
@ -72,19 +72,19 @@ public void testUnbufferClosesSockets() throws Exception {
|
||||
// Read a byte. This will trigger the creation of a block reader.
|
||||
stream.seek(2);
|
||||
int b = stream.read();
|
||||
Assert.assertTrue(-1 != b);
|
||||
Assertions.assertTrue(-1 != b);
|
||||
|
||||
// The Peer cache should start off empty.
|
||||
PeerCache cache = dfs.getClient().getClientContext().getPeerCache();
|
||||
Assert.assertEquals(0, cache.size());
|
||||
Assertions.assertEquals(0, cache.size());
|
||||
|
||||
// Unbuffer should clear the block reader and return the socket to the
|
||||
// cache.
|
||||
stream.unbuffer();
|
||||
stream.seek(2);
|
||||
Assert.assertEquals(1, cache.size());
|
||||
Assertions.assertEquals(1, cache.size());
|
||||
int b2 = stream.read();
|
||||
Assert.assertEquals(b, b2);
|
||||
Assertions.assertEquals(b, b2);
|
||||
} finally {
|
||||
if (stream != null) {
|
||||
IOUtils.cleanupWithLogger(null, stream);
|
||||
@ -117,7 +117,7 @@ public void testOpenManyFilesViaTcp() throws Exception {
|
||||
for (int i = 0; i < NUM_OPENS; i++) {
|
||||
streams[i] = dfs.open(TEST_PATH);
|
||||
LOG.info("opening file " + i + "...");
|
||||
Assert.assertTrue(-1 != streams[i].read());
|
||||
Assertions.assertTrue(-1 != streams[i].read());
|
||||
streams[i].unbuffer();
|
||||
}
|
||||
} finally {
|
||||
|
@ -17,9 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertNull;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
@ -34,8 +32,8 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* Test of the URL stream handler.
|
||||
@ -48,7 +46,7 @@ public class TestUrlStreamHandler {
|
||||
private static final FsUrlStreamHandlerFactory HANDLER_FACTORY
|
||||
= new FsUrlStreamHandlerFactory();
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setupHandler() {
|
||||
|
||||
// Setup our own factory
|
||||
@ -166,20 +164,20 @@ public void testFileUrls() throws IOException, URISyntaxException {
|
||||
|
||||
@Test
|
||||
public void testHttpDefaultHandler() throws Throwable {
|
||||
assertNull("Handler for HTTP is the Hadoop one",
|
||||
HANDLER_FACTORY.createURLStreamHandler("http"));
|
||||
assertNull(
|
||||
HANDLER_FACTORY.createURLStreamHandler("http"), "Handler for HTTP is the Hadoop one");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testHttpsDefaultHandler() throws Throwable {
|
||||
assertNull("Handler for HTTPS is the Hadoop one",
|
||||
HANDLER_FACTORY.createURLStreamHandler("https"));
|
||||
assertNull(
|
||||
HANDLER_FACTORY.createURLStreamHandler("https"), "Handler for HTTPS is the Hadoop one");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testUnknownProtocol() throws Throwable {
|
||||
assertNull("Unknown protocols are not handled",
|
||||
HANDLER_FACTORY.createURLStreamHandler("gopher"));
|
||||
assertNull(
|
||||
HANDLER_FACTORY.createURLStreamHandler("gopher"), "Unknown protocols are not handled");
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -24,10 +24,10 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import javax.security.auth.login.LoginException;
|
||||
import java.io.IOException;
|
||||
@ -38,8 +38,8 @@
|
||||
import static org.apache.hadoop.fs.CreateFlag.CREATE;
|
||||
import static org.apache.hadoop.fs.FileContextTestHelper.getDefaultBlockSize;
|
||||
import static org.apache.hadoop.fs.FileContextTestHelper.getFileData;
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertArrayEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
/**
|
||||
* Test of FileContext apis on Webhdfs.
|
||||
@ -71,7 +71,7 @@ public URI getWebhdfsUrl() {
|
||||
return webhdfsUrl;
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning()
|
||||
throws IOException, LoginException, URISyntaxException {
|
||||
|
||||
@ -85,7 +85,7 @@ public static void clusterSetupAtBeginning()
|
||||
fc.mkdir(defaultWorkingDirectory, FileContext.DEFAULT_PERM, true);
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
URI webhdfsUrlReal = getWebhdfsUrl();
|
||||
Path testBuildData = new Path(
|
||||
@ -153,7 +153,7 @@ public void testSetVerifyChecksum() throws IOException {
|
||||
assertArrayEquals(data, bb);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -27,7 +27,7 @@
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.Assert;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -73,14 +73,14 @@ public static MiniDFSCluster getCluster() {
|
||||
@Override
|
||||
public void init() throws IOException {
|
||||
super.init();
|
||||
Assert.assertTrue("contract options not loaded",
|
||||
isSupported(ContractOptions.IS_CASE_SENSITIVE, false));
|
||||
Assertions.assertTrue(
|
||||
isSupported(ContractOptions.IS_CASE_SENSITIVE, false), "contract options not loaded");
|
||||
}
|
||||
|
||||
@Override
|
||||
public FileSystem getTestFileSystem() throws IOException {
|
||||
//assumes cluster is not null
|
||||
Assert.assertNotNull("cluster not created", cluster);
|
||||
//assumes cluster is not null
|
||||
Assertions.assertNotNull(cluster, "cluster not created");
|
||||
return cluster.getFileSystem();
|
||||
}
|
||||
|
||||
|
@ -17,19 +17,19 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractAppendTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractAppend extends AbstractContractAppendTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractConcatTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,14 +31,14 @@
|
||||
*/
|
||||
public class TestHDFSContractConcat extends AbstractContractConcatTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
// perform a simple operation on the cluster to verify it is up
|
||||
HDFSContract.getCluster().getFileSystem().getDefaultBlockSize();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,19 +21,19 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractCreateTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractCreate extends AbstractContractCreateTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractDeleteTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,12 +31,12 @@
|
||||
*/
|
||||
public class TestHDFSContractDelete extends AbstractContractDeleteTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,20 +21,20 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractGetFileStatusTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractGetFileStatus extends
|
||||
AbstractContractGetFileStatusTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractMkdirTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,12 +31,12 @@
|
||||
*/
|
||||
public class TestHDFSContractMkdir extends AbstractContractMkdirTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -19,14 +19,14 @@
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractMultipartUploaderTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
|
||||
/**
|
||||
* Test MultipartUploader tests on HDFS.
|
||||
@ -37,12 +37,12 @@ public class TestHDFSContractMultipartUploader extends
|
||||
protected static final Logger LOG =
|
||||
LoggerFactory.getLogger(TestHDFSContractMultipartUploader.class);
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractOpenTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,12 +31,12 @@
|
||||
*/
|
||||
public class TestHDFSContractOpen extends AbstractContractOpenTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,11 +21,13 @@
|
||||
import org.apache.hadoop.fs.Options;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractPathHandleTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import static org.junit.jupiter.api.Assertions.super;
|
||||
|
||||
/**
|
||||
* Verify HDFS compliance with {@link org.apache.hadoop.fs.PathHandle}
|
||||
* semantics.
|
||||
@ -35,15 +37,15 @@ public class TestHDFSContractPathHandle
|
||||
|
||||
public TestHDFSContractPathHandle(String testname, Options.HandleOpt[] opts,
|
||||
boolean serialized) {
|
||||
super(testname, opts, serialized);
|
||||
super(opts, serialized, testname);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,19 +21,19 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractRenameTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractRename extends AbstractContractRenameTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractRootDirectoryTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -32,12 +32,12 @@
|
||||
public class TestHDFSContractRootDirectory extends
|
||||
AbstractContractRootDirectoryTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,8 +21,8 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractSeekTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,12 +31,12 @@
|
||||
*/
|
||||
public class TestHDFSContractSeek extends AbstractContractSeekTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,19 +21,19 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractSetTimesTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractSetTimes extends AbstractContractSetTimesTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -21,20 +21,19 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.contract.AbstractContractUnbufferTest;
|
||||
import org.apache.hadoop.fs.contract.AbstractFSContract;
|
||||
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class TestHDFSContractUnbuffer extends AbstractContractUnbufferTest {
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void createCluster() throws IOException {
|
||||
HDFSContract.createCluster();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardownCluster() throws IOException {
|
||||
HDFSContract.destroyCluster();
|
||||
}
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.fs.loadGenerator;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.File;
|
||||
@ -33,7 +33,8 @@
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.apache.hadoop.util.Tool;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This class tests if a balancer schedules tasks correctly.
|
||||
*/
|
||||
|
@ -22,10 +22,7 @@
|
||||
import static org.apache.hadoop.fs.permission.AclEntryType.USER;
|
||||
import static org.apache.hadoop.fs.permission.FsAction.ALL;
|
||||
import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Arrays;
|
||||
@ -45,10 +42,10 @@
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@ -66,7 +63,7 @@ public class TestStickyBit {
|
||||
private static FileSystem hdfsAsUser1;
|
||||
private static FileSystem hdfsAsUser2;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void init() throws Exception {
|
||||
conf = new HdfsConfiguration();
|
||||
conf.setBoolean(DFSConfigKeys.DFS_PERMISSIONS_ENABLED_KEY, true);
|
||||
@ -85,7 +82,7 @@ private static void initCluster(boolean format) throws Exception {
|
||||
assertTrue(hdfsAsUser2 instanceof DistributedFileSystem);
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setup() throws Exception {
|
||||
if (hdfs != null) {
|
||||
for (FileStatus stat: hdfs.listStatus(new Path("/"))) {
|
||||
@ -94,7 +91,7 @@ public void setup() throws Exception {
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void shutdown() throws Exception {
|
||||
IOUtils.cleanupWithLogger(null, hdfs, hdfsAsUser1, hdfsAsUser2);
|
||||
if (cluster != null) {
|
||||
|
@ -18,8 +18,8 @@
|
||||
|
||||
package org.apache.hadoop.fs.shell;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
@ -33,9 +33,9 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
|
||||
/**
|
||||
@ -48,7 +48,7 @@ public class TestHdfsTextCommand {
|
||||
private static MiniDFSCluster cluster;
|
||||
private static FileSystem fs;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws IOException{
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||
@ -56,7 +56,7 @@ public void setUp() throws IOException{
|
||||
fs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws IOException{
|
||||
if(fs != null){
|
||||
fs.close();
|
||||
|
@ -24,9 +24,9 @@
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.junit.After;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
/**
|
||||
* Tests that the NN startup is successful with ViewFSOverloadScheme.
|
||||
@ -37,7 +37,7 @@ public class TestNNStartupWhenViewFSOverloadSchemeEnabled {
|
||||
private static final String HDFS_SCHEME = "hdfs";
|
||||
private static final Configuration CONF = new Configuration();
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setUp() {
|
||||
CONF.setInt(DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY, 1);
|
||||
CONF.setInt(DFSConfigKeys.DFS_HA_TAILEDITS_PERIOD_KEY, 1);
|
||||
@ -78,7 +78,7 @@ public void testNameNodeAndDataNodeStartup() throws Exception {
|
||||
cluster.waitActive();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void shutdownCluster() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -25,7 +25,7 @@
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeys;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.junit.Before;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
/**
|
||||
* Tests ViewFileSystemOverloadScheme with configured mount links.
|
||||
@ -35,7 +35,7 @@ public class TestViewFSOverloadSchemeWithMountTableConfigInHDFS
|
||||
private Path oldVersionMountTablePath;
|
||||
private Path newVersionMountTablePath;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
@Override
|
||||
public void setUp() throws IOException {
|
||||
super.setUp();
|
||||
|
@ -29,9 +29,9 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
/**
|
||||
* Make sure that ViewFileSystem works when the root of an FS is mounted to a
|
||||
@ -48,7 +48,7 @@ protected FileSystemTestHelper createFileSystemHelper() {
|
||||
return new FileSystemTestHelper("/tmp/TestViewFileSystemAtHdfsRoot");
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
SupportsBlocks = true;
|
||||
@ -63,7 +63,7 @@ public static void clusterSetupAtBegining() throws IOException,
|
||||
fHdfs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -71,7 +71,7 @@ public static void clusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fHdfs;
|
||||
super.setUp();
|
||||
|
@ -56,12 +56,9 @@
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.FS_TRASH_INTERVAL_KEY;
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IPC_CLIENT_CONNECT_MAX_RETRIES_KEY;
|
||||
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import static org.junit.Assert.*;
|
||||
import org.junit.jupiter.api.*;
|
||||
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
@ -85,7 +82,7 @@ protected FileSystemTestHelper createFileSystemHelper() {
|
||||
return new FileSystemTestHelper("/tmp/TestViewFileSystemHdfs");
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
|
||||
@ -129,7 +126,7 @@ public static void clusterSetupAtBegining() throws IOException,
|
||||
}
|
||||
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -137,7 +134,7 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
// create the test root on local_fs
|
||||
fsTarget = fHdfs;
|
||||
@ -147,7 +144,7 @@ public void setUp() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
super.tearDown();
|
||||
}
|
||||
@ -253,15 +250,15 @@ public void testFileChecksum() throws IOException {
|
||||
viewFs.getFileChecksum(mountDataFilePath);
|
||||
FileChecksum fileChecksumViaTargetFs =
|
||||
fsTarget.getFileChecksum(fsTargetFilePath);
|
||||
assertTrue("File checksum not matching!",
|
||||
fileChecksumViaViewFs.equals(fileChecksumViaTargetFs));
|
||||
assertTrue(
|
||||
fileChecksumViaViewFs.equals(fileChecksumViaTargetFs), "File checksum not matching!");
|
||||
|
||||
fileChecksumViaViewFs =
|
||||
viewFs.getFileChecksum(mountDataFilePath, fileLength / 2);
|
||||
fileChecksumViaTargetFs =
|
||||
fsTarget.getFileChecksum(fsTargetFilePath, fileLength / 2);
|
||||
assertTrue("File checksum not matching!",
|
||||
fileChecksumViaViewFs.equals(fileChecksumViaTargetFs));
|
||||
assertTrue(
|
||||
fileChecksumViaViewFs.equals(fileChecksumViaTargetFs), "File checksum not matching!");
|
||||
}
|
||||
|
||||
//Rename should fail on across different fileSystems
|
||||
@ -276,7 +273,7 @@ public void testRenameAccorssFilesystem() throws IOException {
|
||||
fsView.create(filePath);
|
||||
try {
|
||||
fsView.rename(filePath, hdfFilepath);
|
||||
ContractTestUtils.fail("Should thrown IOE on Renames across filesytems");
|
||||
Assertions.fail("Should thrown IOE on Renames across filesytems");
|
||||
} catch (IOException e) {
|
||||
GenericTestUtils
|
||||
.assertExceptionContains("Renames across Mount points not supported",
|
||||
@ -325,7 +322,7 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
// 1. test mkdirs
|
||||
final Path testDir = new Path("testdir1/sub1/sub3");
|
||||
final Path testDir_tmp = new Path("testdir1/sub1/sub3_temp");
|
||||
assertTrue(testDir + ": Failed to create!", nfly.mkdirs(testDir));
|
||||
assertTrue(nfly.mkdirs(testDir), testDir + ": Failed to create!");
|
||||
|
||||
// Test renames
|
||||
assertTrue(nfly.rename(testDir, testDir_tmp));
|
||||
@ -333,7 +330,7 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
|
||||
for (final URI testUri : testUris) {
|
||||
final FileSystem fs = FileSystem.get(testUri, testConf);
|
||||
assertTrue(testDir + " should exist!", fs.exists(testDir));
|
||||
assertTrue(fs.exists(testDir), testDir + " should exist!");
|
||||
}
|
||||
|
||||
// 2. test write
|
||||
@ -349,7 +346,7 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
final FileSystem fs = FileSystem.get(testUri, testConf);
|
||||
final FSDataInputStream fsdis = fs.open(testFile);
|
||||
try {
|
||||
assertEquals("Wrong file content", testString, fsdis.readUTF());
|
||||
assertEquals(testString, fsdis.readUTF(), "Wrong file content");
|
||||
} finally {
|
||||
fsdis.close();
|
||||
}
|
||||
@ -364,7 +361,7 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
FSDataInputStream fsDis = null;
|
||||
try {
|
||||
fsDis = nfly.open(testFile);
|
||||
assertEquals("Wrong file content", testString, fsDis.readUTF());
|
||||
assertEquals(testString, fsDis.readUTF(), "Wrong file content");
|
||||
} finally {
|
||||
IOUtils.cleanupWithLogger(LOG, fsDis);
|
||||
cluster.restartNameNode(i);
|
||||
@ -378,7 +375,7 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
FSDataInputStream fsDis = null;
|
||||
try {
|
||||
fsDis = nfly.open(testFile);
|
||||
assertEquals("Wrong file content", testString, fsDis.readUTF());
|
||||
assertEquals(testString, fsDis.readUTF(), "Wrong file content");
|
||||
assertTrue(fs1.exists(testFile));
|
||||
} finally {
|
||||
IOUtils.cleanupWithLogger(LOG, fsDis);
|
||||
@ -393,18 +390,18 @@ private void testNflyRepair(NflyFSystem.NflyKey repairKey)
|
||||
for (final URI testUri : testUris) {
|
||||
final FileSystem fs = FileSystem.get(testUri, conf);
|
||||
fs.setTimes(testFile, 1L, 1L);
|
||||
assertEquals(testUri + "Set mtime failed!", 1L,
|
||||
fs.getFileStatus(testFile).getModificationTime());
|
||||
assertEquals("nfly file status wrong", expectedMtime,
|
||||
nfly.getFileStatus(testFile).getModificationTime());
|
||||
assertEquals(1L,
|
||||
fs.getFileStatus(testFile).getModificationTime(), testUri + "Set mtime failed!");
|
||||
assertEquals(expectedMtime,
|
||||
nfly.getFileStatus(testFile).getModificationTime(), "nfly file status wrong");
|
||||
FSDataInputStream fsDis2 = null;
|
||||
try {
|
||||
fsDis2 = nfly.open(testFile);
|
||||
assertEquals("Wrong file content", testString, fsDis2.readUTF());
|
||||
// repair is done, now trying via normal fs
|
||||
//
|
||||
assertEquals("Repair most recent failed!", expectedMtime,
|
||||
fs.getFileStatus(testFile).getModificationTime());
|
||||
assertEquals(testString, fsDis2.readUTF(), "Wrong file content");
|
||||
// repair is done, now trying via normal fs
|
||||
//
|
||||
assertEquals(expectedMtime,
|
||||
fs.getFileStatus(testFile).getModificationTime(), "Repair most recent failed!");
|
||||
} finally {
|
||||
IOUtils.cleanupWithLogger(LOG, fsDis2);
|
||||
}
|
||||
@ -476,7 +473,7 @@ public Object run() throws IOException {
|
||||
FileSystem otherfs = map.get("user1");
|
||||
otherfs.mkdirs(user1Path);
|
||||
String owner = otherfs.getFileStatus(user1Path).getOwner();
|
||||
assertEquals("The owner did not match ", owner, userUgi.getShortUserName());
|
||||
assertEquals(owner, userUgi.getShortUserName(), "The owner did not match ");
|
||||
otherfs.delete(user1Path, false);
|
||||
}
|
||||
}
|
||||
|
@ -18,11 +18,7 @@
|
||||
package org.apache.hadoop.fs.viewfs;
|
||||
|
||||
import static org.apache.hadoop.test.LambdaTestUtils.intercept;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
@ -45,11 +41,11 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@ -79,7 +75,7 @@ protected FileSystemTestHelper createFileSystemHelper() {
|
||||
return new FileSystemTestHelper(TEST_BASE_PATH);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
SupportsBlocks = true;
|
||||
@ -100,7 +96,7 @@ public static void clusterSetupAtBeginning() throws IOException,
|
||||
Constants.CONFIG_VIEWFS_DEFAULT_MOUNT_TABLE, "/", null, null);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -108,7 +104,7 @@ public static void clusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fsDefault;
|
||||
super.setUp();
|
||||
@ -179,10 +175,10 @@ public void testConfLinkFallback() throws Exception {
|
||||
FileStatus baseFileRelStat = vfs.getFileStatus(new Path(viewFsUri.toString()
|
||||
+ testBaseFileRelative.toUri().toString()));
|
||||
LOG.info("BaseFileRelStat: " + baseFileRelStat);
|
||||
Assert.assertEquals("Unexpected file length for " + testBaseFile,
|
||||
1, baseFileStat.getLen());
|
||||
Assert.assertEquals("Unexpected file length for " + testBaseFileRelative,
|
||||
baseFileStat.getLen(), baseFileRelStat.getLen());
|
||||
Assertions.assertEquals(
|
||||
1, baseFileStat.getLen(), "Unexpected file length for " + testBaseFile);
|
||||
Assertions.assertEquals(
|
||||
baseFileStat.getLen(), baseFileRelStat.getLen(), "Unexpected file length for " + testBaseFileRelative);
|
||||
FileStatus level2FileStat = vfs.getFileStatus(new Path(viewFsUri.toString()
|
||||
+ testLevel2File.toUri().toString()));
|
||||
LOG.info("Level2FileStat: " + level2FileStat);
|
||||
@ -228,8 +224,8 @@ public void testConfLinkFallbackWithRegularLinks() throws Exception {
|
||||
FileStatus baseFileStat = vfs.getFileStatus(
|
||||
new Path(viewFsUri.toString() + testBaseFile.toUri().toString()));
|
||||
LOG.info("BaseFileStat: " + baseFileStat);
|
||||
Assert.assertEquals("Unexpected file length for " + testBaseFile,
|
||||
0, baseFileStat.getLen());
|
||||
Assertions.assertEquals(
|
||||
0, baseFileStat.getLen(), "Unexpected file length for " + testBaseFile);
|
||||
FileStatus level2FileStat = vfs.getFileStatus(new Path(viewFsUri.toString()
|
||||
+ testLevel2File.toUri().toString()));
|
||||
LOG.info("Level2FileStat: " + level2FileStat);
|
||||
@ -240,8 +236,8 @@ public void testConfLinkFallbackWithRegularLinks() throws Exception {
|
||||
|
||||
FileStatus level2FileStatAfterWrite = vfs.getFileStatus(
|
||||
new Path(viewFsUri.toString() + testLevel2File.toUri().toString()));
|
||||
Assert.assertTrue("Unexpected file length for " + testLevel2File,
|
||||
level2FileStatAfterWrite.getLen() > level2FileStat.getLen());
|
||||
Assertions.assertTrue(
|
||||
level2FileStatAfterWrite.getLen() > level2FileStat.getLen(), "Unexpected file length for " + testLevel2File);
|
||||
|
||||
vfs.close();
|
||||
}
|
||||
@ -265,8 +261,8 @@ public void testConfLinkFallbackWithMountPoint() throws Exception {
|
||||
FileSystem.get(viewFsUri, conf);
|
||||
fail("Shouldn't allow linkMergeSlash to take extra mount points!");
|
||||
} catch (IOException e) {
|
||||
assertTrue("Unexpected error: " + e.getMessage(),
|
||||
e.getMessage().contains(expectedErrorMsg));
|
||||
assertTrue(
|
||||
e.getMessage().contains(expectedErrorMsg), "Unexpected error: " + e.getMessage());
|
||||
}
|
||||
}
|
||||
|
||||
@ -299,13 +295,13 @@ public void testListingWithFallbackLink() throws Exception {
|
||||
afterFallback.add(stat.getPath());
|
||||
}
|
||||
afterFallback.removeAll(beforeFallback);
|
||||
assertTrue("Listing didn't include fallback link",
|
||||
afterFallback.size() == 1);
|
||||
assertTrue(
|
||||
afterFallback.size() == 1, "Listing didn't include fallback link");
|
||||
Path[] fallbackArray = new Path[afterFallback.size()];
|
||||
afterFallback.toArray(fallbackArray);
|
||||
Path expected = new Path(viewFsUri.toString(), "dir1");
|
||||
assertEquals("Path did not match",
|
||||
expected, fallbackArray[0]);
|
||||
assertEquals(
|
||||
expected, fallbackArray[0], "Path did not match");
|
||||
|
||||
// Create a directory using the returned fallback path and verify
|
||||
Path childDir = new Path(fallbackArray[0], "child");
|
||||
@ -349,13 +345,13 @@ public void testListingWithFallbackLinkWithSameMountDirectories()
|
||||
afterFallback.add(stat.getPath());
|
||||
}
|
||||
afterFallback.removeAll(beforeFallback);
|
||||
assertEquals("The same directory name in fallback link should be shaded",
|
||||
1, afterFallback.size());
|
||||
assertEquals(
|
||||
1, afterFallback.size(), "The same directory name in fallback link should be shaded");
|
||||
Path[] fallbackArray = new Path[afterFallback.size()];
|
||||
// Only user1 should be listed as fallback link
|
||||
Path expected = new Path(viewFsDefaultClusterUri.toString(), "user1");
|
||||
assertEquals("Path did not match",
|
||||
expected, afterFallback.toArray(fallbackArray)[0]);
|
||||
assertEquals(
|
||||
expected, afterFallback.toArray(fallbackArray)[0], "Path did not match");
|
||||
|
||||
// Create a directory using the returned fallback path and verify
|
||||
Path childDir = new Path(fallbackArray[0], "child");
|
||||
@ -430,8 +426,8 @@ public void testListingWithFallbackLinkWithSameMountDirectoryTree()
|
||||
}
|
||||
//viewfs://default/user1/hive/warehouse
|
||||
afterFallback.removeAll(beforeFallback);
|
||||
assertEquals("The same directory name in fallback link should be shaded",
|
||||
1, afterFallback.size());
|
||||
assertEquals(
|
||||
1, afterFallback.size(), "The same directory name in fallback link should be shaded");
|
||||
}
|
||||
}
|
||||
|
||||
@ -502,8 +498,8 @@ public void testLSOnLinkParentWithFallbackLinkWithSameMountDirectoryTree()
|
||||
}
|
||||
}
|
||||
afterFallback.removeAll(beforeFallback);
|
||||
assertEquals("Just to make sure paths are same.", 0,
|
||||
afterFallback.size());
|
||||
assertEquals(0,
|
||||
afterFallback.size(), "Just to make sure paths are same.");
|
||||
}
|
||||
}
|
||||
|
||||
@ -559,14 +555,14 @@ public void testLSOnRootWithFallbackLinkWithSameMountDirectories()
|
||||
assertEquals(FsPermission.valueOf("-rwxr--rw-"),
|
||||
stat.getPermission());
|
||||
} else {
|
||||
assertEquals("Path is: " + stat.getPath(),
|
||||
FsPermission.valueOf("-rwxr--r--"), stat.getPermission());
|
||||
assertEquals(
|
||||
FsPermission.valueOf("-rwxr--r--"), stat.getPermission(), "Path is: " + stat.getPath());
|
||||
}
|
||||
}
|
||||
afterFallback.removeAll(beforeFallback);
|
||||
assertEquals(1, afterFallback.size());
|
||||
assertEquals("/user2 dir from fallback should be listed.", "user2",
|
||||
afterFallback.iterator().next().getName());
|
||||
assertEquals("user2",
|
||||
afterFallback.iterator().next().getName(), "/user2 dir from fallback should be listed.");
|
||||
}
|
||||
}
|
||||
|
||||
@ -908,7 +904,7 @@ public void testCreateFileSameAsInternalDirPath() throws Exception {
|
||||
assertFalse(fsTarget.exists(Path.mergePaths(fallbackTarget, vfsTestDir)));
|
||||
try {
|
||||
vfs.create(vfsTestDir);
|
||||
Assert.fail("Should fail to create file as this is an internal dir.");
|
||||
Assertions.fail("Should fail to create file as this is an internal dir.");
|
||||
} catch (NotInMountpointException e){
|
||||
// This tree is part of internal tree. The above exception will be
|
||||
// thrown from getDefaultReplication, getDefaultBlockSize APIs which was
|
||||
|
@ -34,17 +34,11 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import javax.security.auth.login.LoginException;
|
||||
|
||||
@ -74,7 +68,7 @@ protected FileSystemTestHelper createFileSystemHelper() {
|
||||
return new FileSystemTestHelper(TEST_TEMP_PATH);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
SupportsBlocks = true;
|
||||
@ -93,7 +87,7 @@ public static void clusterSetupAtBeginning() throws IOException,
|
||||
fsDefault = FS_HDFS[FS_INDEX_DEFAULT];
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -101,7 +95,7 @@ public static void clusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fsDefault;
|
||||
super.setUp();
|
||||
@ -191,9 +185,9 @@ public void testConfLinkMergeSlashWithRegularLinks() throws Exception {
|
||||
fail("Shouldn't allow both merge slash link and regular link on same "
|
||||
+ "mount table.");
|
||||
} catch (IOException e) {
|
||||
assertTrue("Unexpected error message: " + e.getMessage(),
|
||||
e.getMessage().contains(expectedErrorMsg1) || e.getMessage()
|
||||
.contains(expectedErrorMsg2));
|
||||
assertTrue(
|
||||
e.getMessage().contains(expectedErrorMsg1) || e.getMessage()
|
||||
.contains(expectedErrorMsg2), "Unexpected error message: " + e.getMessage());
|
||||
}
|
||||
}
|
||||
|
||||
@ -226,9 +220,9 @@ public void testChildFileSystems() throws Exception {
|
||||
LINK_MERGE_SLASH_CLUSTER_1_NAME, "/", null, null);
|
||||
FileSystem fs = FileSystem.get(viewFsUri, conf);
|
||||
FileSystem[] childFs = fs.getChildFileSystems();
|
||||
Assert.assertEquals("Unexpected number of child filesystems!",
|
||||
1, childFs.length);
|
||||
Assert.assertEquals("Unexpected child filesystem!",
|
||||
DistributedFileSystem.class, childFs[0].getClass());
|
||||
Assertions.assertEquals(
|
||||
1, childFs.length, "Unexpected number of child filesystems!");
|
||||
Assertions.assertEquals(
|
||||
DistributedFileSystem.class, childFs[0].getClass(), "Unexpected child filesystem!");
|
||||
}
|
||||
}
|
||||
|
@ -36,16 +36,12 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import static org.apache.hadoop.fs.viewfs.RegexMountPoint.INTERCEPTOR_INTERNAL_SEP;
|
||||
import static org.junit.Assert.assertSame;
|
||||
import static org.junit.jupiter.api.Assertions.assertSame;
|
||||
|
||||
/**
|
||||
* Test linkRegex node type for view file system.
|
||||
@ -73,7 +69,7 @@ protected FileSystemTestHelper createFileSystemHelper() {
|
||||
return new FileSystemTestHelper(TEST_BASE_PATH);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
SupportsBlocks = true;
|
||||
clusterConfig = ViewFileSystemTestSetup.createConfig();
|
||||
@ -91,7 +87,7 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fsDefault = FS_HDFS[FS_INDEX_DEFAULT];
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -99,7 +95,7 @@ public static void clusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fsDefault;
|
||||
super.setUp();
|
||||
@ -157,7 +153,7 @@ public String linkInterceptorSettings(
|
||||
private void createDirWithChildren(
|
||||
FileSystem fileSystem, Path dir, List<Path> childrenFiles)
|
||||
throws IOException {
|
||||
Assert.assertTrue(fileSystem.mkdirs(dir));
|
||||
Assertions.assertTrue(fileSystem.mkdirs(dir));
|
||||
int index = 0;
|
||||
for (Path childFile : childrenFiles) {
|
||||
createFile(fileSystem, childFile, index, true);
|
||||
@ -224,11 +220,11 @@ private void testRegexMountpoint(
|
||||
URI viewFsUri = new URI(
|
||||
FsConstants.VIEWFS_SCHEME, CLUSTER_NAME, "/", null, null);
|
||||
try (FileSystem vfs = FileSystem.get(viewFsUri, config)) {
|
||||
Assert.assertEquals(expectedResolveResult.toString(),
|
||||
Assertions.assertEquals(expectedResolveResult.toString(),
|
||||
vfs.resolvePath(dirPathBeforeMountPoint).toString());
|
||||
Assert.assertTrue(
|
||||
Assertions.assertTrue(
|
||||
vfs.getFileStatus(dirPathBeforeMountPoint).isDirectory());
|
||||
Assert.assertEquals(
|
||||
Assertions.assertEquals(
|
||||
childrenFilesCnt, vfs.listStatus(dirPathBeforeMountPoint).length);
|
||||
|
||||
// Test Inner cache, the resolved result's filesystem should be the same.
|
||||
|
@ -19,8 +19,6 @@
|
||||
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME;
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME_DEFAULT;
|
||||
import static org.junit.Assume.assumeTrue;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
@ -39,11 +37,11 @@
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
|
||||
/**
|
||||
* Tests ViewFileSystemOverloadScheme with file system contract tests.
|
||||
@ -55,7 +53,7 @@ public class TestViewFileSystemOverloadSchemeHdfsFileSystemContract
|
||||
private static String defaultWorkingDirectory;
|
||||
private static Configuration conf = new HdfsConfiguration();
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void init() throws IOException {
|
||||
final File basedir = GenericTestUtils.getRandomizedTestDir();
|
||||
conf.set(CommonConfigurationKeys.FS_PERMISSIONS_UMASK_KEY,
|
||||
@ -67,7 +65,7 @@ public static void init() throws IOException {
|
||||
"/user/" + UserGroupInformation.getCurrentUser().getShortUserName();
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
conf.set(String.format("fs.%s.impl", "hdfs"),
|
||||
ViewFileSystemOverloadScheme.class.getName());
|
||||
@ -89,7 +87,7 @@ public void setUp() throws Exception {
|
||||
fs = FileSystem.get(conf);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDownAfter() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -129,7 +127,7 @@ public void testListStatusRootDir() throws Throwable {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Ignore // This test same as above in this case.
|
||||
@Disabled // This test same as above in this case.
|
||||
public void testLSRootDir() throws Throwable {
|
||||
}
|
||||
}
|
||||
|
@ -39,16 +39,16 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.test.LambdaTestUtils;
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME;
|
||||
import static org.apache.hadoop.fs.viewfs.Constants.CONFIG_VIEWFS_IGNORE_PORT_IN_MOUNT_TABLE_NAME_DEFAULT;
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
|
||||
/**
|
||||
@ -67,7 +67,7 @@ public class TestViewFileSystemOverloadSchemeWithHdfsScheme {
|
||||
private static final String HDFS_USER_FOLDER = "/HDFSUser";
|
||||
private static final String LOCAL_FOLDER = "/local";
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void init() throws IOException {
|
||||
cluster =
|
||||
new MiniDFSCluster.Builder(new Configuration()).numDataNodes(2).build();
|
||||
@ -77,7 +77,7 @@ public static void init() throws IOException {
|
||||
/**
|
||||
* Sets up the configurations and starts the MiniDFSCluster.
|
||||
*/
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws IOException {
|
||||
Configuration config = getNewConf();
|
||||
config.setInt(
|
||||
@ -91,10 +91,10 @@ public void setUp() throws IOException {
|
||||
URI.create(config.get(CommonConfigurationKeys.FS_DEFAULT_NAME_KEY));
|
||||
localTargetDir = new File(TEST_ROOT_DIR, "/root/");
|
||||
localTargetDir.mkdirs();
|
||||
Assert.assertEquals(HDFS_SCHEME, defaultFSURI.getScheme()); // hdfs scheme.
|
||||
Assertions.assertEquals(HDFS_SCHEME, defaultFSURI.getScheme()); // hdfs scheme.
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void cleanUp() throws IOException {
|
||||
if (cluster != null) {
|
||||
FileSystem fs = new DistributedFileSystem();
|
||||
@ -102,7 +102,7 @@ public void cleanUp() throws IOException {
|
||||
try {
|
||||
FileStatus[] statuses = fs.listStatus(new Path("/"));
|
||||
for (FileStatus st : statuses) {
|
||||
Assert.assertTrue(fs.delete(st.getPath(), true));
|
||||
Assertions.assertTrue(fs.delete(st.getPath(), true));
|
||||
}
|
||||
} finally {
|
||||
fs.close();
|
||||
@ -111,7 +111,7 @@ public void cleanUp() throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() throws IOException {
|
||||
if (cluster != null) {
|
||||
FileSystem.closeAll();
|
||||
@ -154,7 +154,7 @@ public void testMountLinkWithLocalAndHDFS() throws Exception {
|
||||
|
||||
try (FileSystem fs
|
||||
= FileSystem.get(conf)) {
|
||||
Assert.assertEquals(2, fs.getChildFileSystems().length);
|
||||
Assertions.assertEquals(2, fs.getChildFileSystems().length);
|
||||
fs.createNewFile(hdfsFile); // /HDFSUser/testfile
|
||||
fs.mkdirs(localDir); // /local/test
|
||||
}
|
||||
@ -162,20 +162,20 @@ public void testMountLinkWithLocalAndHDFS() throws Exception {
|
||||
// Initialize HDFS and test files exist in ls or not
|
||||
try (DistributedFileSystem dfs = new DistributedFileSystem()) {
|
||||
dfs.initialize(defaultFSURI, conf);
|
||||
Assert.assertTrue(dfs.exists(
|
||||
Assertions.assertTrue(dfs.exists(
|
||||
new Path(Path.getPathWithoutSchemeAndAuthority(hdfsTargetPath),
|
||||
hdfsFile.getName()))); // should be in hdfs.
|
||||
Assert.assertFalse(dfs.exists(
|
||||
Assertions.assertFalse(dfs.exists(
|
||||
new Path(Path.getPathWithoutSchemeAndAuthority(localTragetPath),
|
||||
localDir.getName()))); // should not be in local fs.
|
||||
}
|
||||
|
||||
try (RawLocalFileSystem lfs = new RawLocalFileSystem()) {
|
||||
lfs.initialize(localTragetPath.toUri(), conf);
|
||||
Assert.assertFalse(lfs.exists(
|
||||
Assertions.assertFalse(lfs.exists(
|
||||
new Path(Path.getPathWithoutSchemeAndAuthority(hdfsTargetPath),
|
||||
hdfsFile.getName()))); // should not be in hdfs.
|
||||
Assert.assertTrue(lfs.exists(
|
||||
Assertions.assertTrue(lfs.exists(
|
||||
new Path(Path.getPathWithoutSchemeAndAuthority(localTragetPath),
|
||||
localDir.getName()))); // should be in local fs.
|
||||
}
|
||||
@ -210,7 +210,7 @@ public void testMountLinkWithNonExistentLink(boolean expectFsInitFailure)
|
||||
});
|
||||
} else {
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals("hdfs", fs.getScheme());
|
||||
Assertions.assertEquals("hdfs", fs.getScheme());
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -241,14 +241,14 @@ public void testListStatusOnRootShouldListAllMountLinks() throws Exception {
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
fs.mkdirs(hdfsTargetPath);
|
||||
FileStatus[] ls = fs.listStatus(new Path("/"));
|
||||
Assert.assertEquals(2, ls.length);
|
||||
Assertions.assertEquals(2, ls.length);
|
||||
String lsPath1 =
|
||||
Path.getPathWithoutSchemeAndAuthority(ls[0].getPath()).toString();
|
||||
String lsPath2 =
|
||||
Path.getPathWithoutSchemeAndAuthority(ls[1].getPath()).toString();
|
||||
Assert.assertTrue(
|
||||
Assertions.assertTrue(
|
||||
HDFS_USER_FOLDER.equals(lsPath1) || LOCAL_FOLDER.equals(lsPath1));
|
||||
Assert.assertTrue(
|
||||
Assertions.assertTrue(
|
||||
HDFS_USER_FOLDER.equals(lsPath2) || LOCAL_FOLDER.equals(lsPath2));
|
||||
}
|
||||
}
|
||||
@ -270,7 +270,7 @@ public void testListStatusOnNonMountedPath() throws Exception {
|
||||
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
fs.listStatus(new Path("/nonMount"));
|
||||
Assert.fail("It should fail as no mount link with /nonMount");
|
||||
Assertions.fail("It should fail as no mount link with /nonMount");
|
||||
}
|
||||
}
|
||||
|
||||
@ -349,8 +349,8 @@ public void testWithLinkFallBack() throws Exception {
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
fs.createNewFile(new Path("/nonMount/myfile"));
|
||||
FileStatus[] ls = fs.listStatus(new Path("/nonMount"));
|
||||
Assert.assertEquals(1, ls.length);
|
||||
Assert.assertEquals(
|
||||
Assertions.assertEquals(1, ls.length);
|
||||
Assertions.assertEquals(
|
||||
Path.getPathWithoutSchemeAndAuthority(ls[0].getPath()).getName(),
|
||||
"myfile");
|
||||
}
|
||||
@ -376,7 +376,7 @@ public void testCreateOnRoot(boolean fallbackExist) throws Exception {
|
||||
localTargetDir.toURI().toString()}, conf);
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
if (fallbackExist) {
|
||||
Assert.assertTrue(fs.createNewFile(new Path("/newFileOnRoot")));
|
||||
Assertions.assertTrue(fs.createNewFile(new Path("/newFileOnRoot")));
|
||||
} else {
|
||||
LambdaTestUtils.intercept(NotInMountpointException.class, () -> {
|
||||
fs.createNewFile(new Path("/newFileOnRoot"));
|
||||
@ -422,7 +422,7 @@ public void testInvalidOverloadSchemeTargetFS() throws Exception {
|
||||
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
fs.createNewFile(new Path("/onRootWhenFallBack"));
|
||||
Assert.fail("OverloadScheme target fs should be valid.");
|
||||
Assertions.fail("OverloadScheme target fs should be valid.");
|
||||
}
|
||||
}
|
||||
|
||||
@ -446,7 +446,7 @@ public void testViewFsOverloadSchemeWhenInnerCacheDisabled()
|
||||
try (FileSystem fs = FileSystem.get(conf)) {
|
||||
Path testFile = new Path(HDFS_USER_FOLDER + "/testFile");
|
||||
fs.createNewFile(testFile);
|
||||
Assert.assertTrue(fs.exists(testFile));
|
||||
Assertions.assertTrue(fs.exists(testFile));
|
||||
}
|
||||
}
|
||||
|
||||
@ -470,13 +470,13 @@ public void testViewFsOverloadSchemeWithInnerCache()
|
||||
|
||||
// 1. Only 1 hdfs child file system should be there with cache.
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(1, vfs.getChildFileSystems().length);
|
||||
Assertions.assertEquals(1, vfs.getChildFileSystems().length);
|
||||
}
|
||||
|
||||
// 2. Two hdfs file systems should be there if no cache.
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(isFallBackExist(conf) ? 3 : 2,
|
||||
Assertions.assertEquals(isFallBackExist(conf) ? 3 : 2,
|
||||
vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
@ -509,7 +509,7 @@ public void testViewFsOverloadSchemeWithNoInnerCacheAndHdfsTargets()
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
// Two hdfs file systems should be there if no cache.
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(isFallBackExist(conf) ? 3 : 2,
|
||||
Assertions.assertEquals(isFallBackExist(conf) ? 3 : 2,
|
||||
vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
@ -537,7 +537,7 @@ public void testViewFsOverloadSchemeWithNoInnerCacheAndLocalSchemeTargets()
|
||||
// cache should work.
|
||||
conf.setBoolean(Constants.CONFIG_VIEWFS_ENABLE_INNER_CACHE, false);
|
||||
try (FileSystem vfs = FileSystem.get(conf)) {
|
||||
Assert.assertEquals(isFallBackExist(conf) ? 2 : 1,
|
||||
Assertions.assertEquals(isFallBackExist(conf) ? 2 : 1,
|
||||
vfs.getChildFileSystems().length);
|
||||
}
|
||||
}
|
||||
@ -561,7 +561,7 @@ public void testNflyRename() throws Exception {
|
||||
|
||||
final Path testDir = new Path("/nflyroot/testdir1/sub1/sub3");
|
||||
final Path testDirTmp = new Path("/nflyroot/testdir1/sub1/sub3_temp");
|
||||
assertTrue(testDir + ": Failed to create!", nfly.mkdirs(testDir));
|
||||
assertTrue(nfly.mkdirs(testDir), testDir + ": Failed to create!");
|
||||
|
||||
// Test renames
|
||||
assertTrue(nfly.rename(testDir, testDirTmp));
|
||||
@ -570,7 +570,7 @@ public void testNflyRename() throws Exception {
|
||||
final URI[] testUris = new URI[] {uri1, uri2 };
|
||||
for (final URI testUri : testUris) {
|
||||
final FileSystem fs = FileSystem.get(testUri, conf);
|
||||
assertTrue(testDir + " should exist!", fs.exists(testDir));
|
||||
assertTrue(fs.exists(testDir), testDir + " should exist!");
|
||||
}
|
||||
}
|
||||
|
||||
@ -688,7 +688,7 @@ private void writeString(final FileSystem nfly, final String testString,
|
||||
private void readString(final FileSystem nfly, final Path testFile,
|
||||
final String testString, final URI testUri) throws IOException {
|
||||
try (FSDataInputStream fsDis = nfly.open(testFile)) {
|
||||
assertEquals("Wrong file content", testString, fsDis.readUTF());
|
||||
assertEquals(testString, fsDis.readUTF(), "Wrong file content");
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -28,11 +28,7 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.util.Lists;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
@ -42,8 +38,8 @@
|
||||
import static org.apache.hadoop.fs.permission.AclEntryType.*;
|
||||
import static org.apache.hadoop.fs.permission.FsAction.*;
|
||||
import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry;
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertArrayEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
/**
|
||||
* Verify ACL through ViewFileSystem functionality.
|
||||
@ -61,7 +57,7 @@ public class TestViewFileSystemWithAcls {
|
||||
private FileSystemTestHelper fileSystemTestHelper =
|
||||
new FileSystemTestHelper("/tmp/TestViewFileSystemWithAcls");
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
clusterConf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true);
|
||||
cluster = new MiniDFSCluster.Builder(clusterConf)
|
||||
@ -74,14 +70,14 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fHdfs2 = cluster.getFileSystem(1);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fHdfs;
|
||||
fsTarget2 = fHdfs2;
|
||||
@ -105,7 +101,7 @@ private void setupMountPoints() {
|
||||
ConfigUtil.addLink(fsViewConf, mountOnNn2.toString(), targetTestRoot2.toUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fsTarget.delete(fileSystemTestHelper.getTestRootPath(fsTarget), true);
|
||||
fsTarget2.delete(fileSystemTestHelper.getTestRootPath(fsTarget2), true);
|
||||
|
@ -30,13 +30,13 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
/**
|
||||
* Verify truncate through ViewFileSystem functionality.
|
||||
@ -53,7 +53,7 @@ public class TestViewFileSystemWithTruncate {
|
||||
private FileSystemTestHelper fileSystemTestHelper =
|
||||
new FileSystemTestHelper("/tmp/TestViewFileSystemWithXAttrs");
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
cluster = new MiniDFSCluster.Builder(clusterConf)
|
||||
.nnTopology(MiniDFSNNTopology.simpleFederatedTopology(2))
|
||||
@ -63,14 +63,14 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fHdfs = cluster.getFileSystem(0);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fHdfs;
|
||||
targetTestRoot = fileSystemTestHelper.getAbsoluteTestRootPath(fsTarget);
|
||||
@ -89,7 +89,7 @@ private void setupMountPoints() {
|
||||
.addLink(fsViewConf, mountOnNn1.toString(), targetTestRoot.toUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fsTarget.delete(fileSystemTestHelper.getTestRootPath(fsTarget), true);
|
||||
}
|
||||
|
@ -24,16 +24,12 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertArrayEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
/**
|
||||
* Verify XAttrs through ViewFileSystem functionality.
|
||||
@ -57,7 +53,7 @@ public class TestViewFileSystemWithXAttrs {
|
||||
protected static final String name2 = "user.a2";
|
||||
protected static final byte[] value2 = {0x37, 0x38, 0x39};
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
cluster = new MiniDFSCluster.Builder(clusterConf)
|
||||
.nnTopology(MiniDFSNNTopology.simpleFederatedTopology(2))
|
||||
@ -69,14 +65,14 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fHdfs2 = cluster.getFileSystem(1);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fHdfs;
|
||||
fsTarget2 = fHdfs2;
|
||||
@ -102,7 +98,7 @@ private void setupMountPoints() {
|
||||
targetTestRoot2.toUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fsTarget.delete(fileSystemTestHelper.getTestRootPath(fsTarget), true);
|
||||
fsTarget2.delete(fileSystemTestHelper.getTestRootPath(fsTarget2), true);
|
||||
|
@ -30,9 +30,9 @@
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
/**
|
||||
* Make sure that ViewFs works when the root of an FS is mounted to a ViewFs
|
||||
@ -49,7 +49,7 @@ protected FileContextTestHelper createFileContextHelper() {
|
||||
return new FileContextTestHelper("/tmp/TestViewFsAtHdfsRoot");
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
SupportsBlocks = true;
|
||||
@ -62,7 +62,7 @@ public static void clusterSetupAtBegining() throws IOException,
|
||||
}
|
||||
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -70,7 +70,7 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
// create the test root on local_fs
|
||||
fcTarget = fc;
|
||||
|
@ -28,9 +28,7 @@
|
||||
import static org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_DEFAULT;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
@ -49,9 +47,9 @@
|
||||
import org.apache.hadoop.fs.StorageType;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* Tests for viewfs implementation of default fs level values.
|
||||
@ -74,7 +72,7 @@ public class TestViewFsDefaultValue {
|
||||
// Use NotInMountpoint path to trigger the exception
|
||||
private static Path notInMountpointPath;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
|
||||
@ -218,7 +216,7 @@ public void testGetQuotaUsageWithQuotaDefined() throws IOException {
|
||||
assertTrue(qu.getSpaceConsumed() > 0);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void cleanup() throws IOException {
|
||||
fHdfs.delete(new Path(testFileName), true);
|
||||
fHdfs.delete(notInMountpointPath, true);
|
||||
|
@ -23,8 +23,8 @@
|
||||
* Since viewfs has overlayed ViewFsFileStatus, we ran into
|
||||
* serialization problems. THis test is test the fix.
|
||||
*/
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertFalse;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
@ -43,9 +43,9 @@
|
||||
import org.apache.hadoop.io.DataInputBuffer;
|
||||
import org.apache.hadoop.io.DataOutputBuffer;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestViewFsFileStatusHdfs {
|
||||
|
||||
@ -59,7 +59,7 @@ public class TestViewFsFileStatusHdfs {
|
||||
private static FileSystem fHdfs;
|
||||
private static FileSystem vfs;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
cluster = new MiniDFSCluster.Builder(CONF).numDataNodes(2).build();
|
||||
@ -108,15 +108,15 @@ public void testGetFileChecksum() throws IOException, URISyntaxException {
|
||||
// Get checksum of different file in HDFS
|
||||
FileChecksum otherHdfsFileCheckSum = fHdfs.getFileChecksum(
|
||||
new Path(someFile+"other"));
|
||||
// Checksums of the same file (got through HDFS and ViewFS should be same)
|
||||
assertEquals("HDFS and ViewFS checksums were not the same", viewFSCheckSum,
|
||||
hdfsCheckSum);
|
||||
// Checksum of different files should be different.
|
||||
assertFalse("Some other HDFS file which should not have had the same " +
|
||||
"checksum as viewFS did!", viewFSCheckSum.equals(otherHdfsFileCheckSum));
|
||||
// Checksums of the same file (got through HDFS and ViewFS should be same)
|
||||
assertEquals(viewFSCheckSum,
|
||||
hdfsCheckSum, "HDFS and ViewFS checksums were not the same");
|
||||
// Checksum of different files should be different.
|
||||
assertFalse(viewFSCheckSum.equals(otherHdfsFileCheckSum), "Some other HDFS file which should not have had the same " +
|
||||
"checksum as viewFS did!");
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void cleanup() throws IOException {
|
||||
fHdfs.delete(new Path(testfilename), true);
|
||||
fHdfs.delete(new Path(someFile), true);
|
||||
|
@ -36,13 +36,13 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
public class TestViewFsHdfs extends ViewFsBaseTest {
|
||||
|
||||
@ -56,7 +56,7 @@ protected FileContextTestHelper createFileContextHelper() {
|
||||
}
|
||||
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
SupportsBlocks = true;
|
||||
@ -72,7 +72,7 @@ public static void clusterSetupAtBegining() throws IOException,
|
||||
}
|
||||
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -80,7 +80,7 @@ public static void ClusterShutdownAtEnd() throws Exception {
|
||||
}
|
||||
|
||||
@Override
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
// create the test root on local_fs
|
||||
fcTarget = fc;
|
||||
@ -160,7 +160,7 @@ public Object run() throws IOException {
|
||||
FileContext otherfs = map.get("user1");
|
||||
otherfs.mkdir(user1Path, FileContext.DEFAULT_PERM, false);
|
||||
String owner = otherfs.getFileStatus(user1Path).getOwner();
|
||||
assertEquals("The owner did not match ", owner, userUgi.getShortUserName());
|
||||
assertEquals(owner, userUgi.getShortUserName(), "The owner did not match ");
|
||||
otherfs.delete(user1Path, false);
|
||||
}
|
||||
|
||||
|
@ -18,10 +18,7 @@
|
||||
package org.apache.hadoop.fs.viewfs;
|
||||
|
||||
import static org.apache.hadoop.fs.CreateFlag.CREATE;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
import java.io.IOException;
|
||||
@ -46,11 +43,11 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.test.LambdaTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
/**
|
||||
* Test for viewfs with LinkFallback mount table entries.
|
||||
@ -62,7 +59,7 @@ public class TestViewFsLinkFallback {
|
||||
private static URI viewFsDefaultClusterUri;
|
||||
private Path targetTestRoot;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning()
|
||||
throws IOException, URISyntaxException {
|
||||
int nameSpacesCount = 3;
|
||||
@ -88,14 +85,14 @@ public static void clusterSetupAtBeginning()
|
||||
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void clusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fsTarget = fsDefault;
|
||||
initializeTargetTestRoot();
|
||||
@ -295,7 +292,7 @@ public void testMkdirShouldFailWhenFallbackFSNotAvailable()
|
||||
// attempt to create in fallback.
|
||||
vfs.mkdir(nextLevelToInternalDir, FsPermission.getDirDefault(),
|
||||
false);
|
||||
Assert.fail("It should throw IOE when fallback fs not available.");
|
||||
Assertions.fail("It should throw IOE when fallback fs not available.");
|
||||
} catch (IOException e) {
|
||||
cluster.restartNameNodes();
|
||||
// should succeed when fallback fs is back to normal.
|
||||
@ -570,7 +567,7 @@ private void verifyRename(AbstractFileSystem fs, Path src, Path dst)
|
||||
fs.rename(src, dst, Options.Rename.OVERWRITE);
|
||||
LambdaTestUtils
|
||||
.intercept(FileNotFoundException.class, () -> fs.getFileStatus(src));
|
||||
Assert.assertNotNull(fs.getFileStatus(dst));
|
||||
Assertions.assertNotNull(fs.getFileStatus(dst));
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -29,11 +29,8 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.apache.hadoop.util.Lists;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
import java.io.IOException;
|
||||
@ -44,8 +41,8 @@
|
||||
import static org.apache.hadoop.fs.permission.FsAction.*;
|
||||
import static org.apache.hadoop.fs.permission.FsAction.NONE;
|
||||
import static org.apache.hadoop.hdfs.server.namenode.AclTestHelpers.aclEntry;
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertArrayEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
/**
|
||||
* Verify ACL through ViewFs functionality.
|
||||
@ -61,7 +58,7 @@ public class TestViewFsWithAcls {
|
||||
private FileContextTestHelper fileContextTestHelper =
|
||||
new FileContextTestHelper("/tmp/TestViewFsWithAcls");
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
clusterConf.setBoolean(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, true);
|
||||
cluster = new MiniDFSCluster.Builder(clusterConf)
|
||||
@ -74,14 +71,14 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fc2 = FileContext.getFileContext(cluster.getURI(1), clusterConf);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fcTarget = fc;
|
||||
fcTarget2 = fc2;
|
||||
@ -105,7 +102,7 @@ private void setupMountPoints() {
|
||||
ConfigUtil.addLink(fsViewConf, mountOnNn2.toString(), targetTestRoot2.toUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fcTarget.delete(fileContextTestHelper.getTestRootPath(fcTarget), true);
|
||||
fcTarget2.delete(fileContextTestHelper.getTestRootPath(fcTarget2), true);
|
||||
|
@ -25,16 +25,12 @@
|
||||
import org.apache.hadoop.fs.permission.FsPermission;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||
import org.junit.After;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Before;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.*;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertArrayEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
/**
|
||||
* Verify XAttrs through ViewFs functionality.
|
||||
@ -56,7 +52,7 @@ public class TestViewFsWithXAttrs {
|
||||
protected static final String name2 = "user.a2";
|
||||
protected static final byte[] value2 = {0x37, 0x38, 0x39};
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void clusterSetupAtBeginning() throws IOException {
|
||||
cluster = new MiniDFSCluster.Builder(clusterConf)
|
||||
.nnTopology(MiniDFSNNTopology.simpleFederatedTopology(2))
|
||||
@ -68,14 +64,14 @@ public static void clusterSetupAtBeginning() throws IOException {
|
||||
fc2 = FileContext.getFileContext(cluster.getURI(1), clusterConf);
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void ClusterShutdownAtEnd() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
fcTarget = fc;
|
||||
fcTarget2 = fc2;
|
||||
@ -99,7 +95,7 @@ private void setupMountPoints() {
|
||||
ConfigUtil.addLink(fsViewConf, mountOnNn2.toString(), targetTestRoot2.toUri());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
fcTarget.delete(fileContextTestHelper.getTestRootPath(fcTarget), true);
|
||||
fcTarget2.delete(fileContextTestHelper.getTestRootPath(fcTarget2), true);
|
||||
|
@ -17,8 +17,8 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
@ -29,6 +29,7 @@
|
||||
import java.util.Random;
|
||||
|
||||
import org.apache.hadoop.util.Lists;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
@ -46,8 +47,7 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
||||
import org.apache.hadoop.hdfs.server.namenode.NameNodeAdapter;
|
||||
import org.apache.hadoop.hdfs.util.HostsFileWriter;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
/**
|
||||
* This class provide utilities for testing of the admin operations of nodes.
|
||||
@ -81,7 +81,7 @@ protected MiniDFSCluster getCluster() {
|
||||
return cluster;
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setup() throws IOException {
|
||||
// Set up the hosts/exclude files.
|
||||
hostsFileWriter = new HostsFileWriter();
|
||||
@ -108,7 +108,7 @@ public void setup() throws IOException {
|
||||
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void teardown() throws IOException {
|
||||
hostsFileWriter.cleanup();
|
||||
shutdownCluster();
|
||||
@ -381,7 +381,7 @@ protected DFSClient getDfsClient(final int nnIndex) throws IOException {
|
||||
protected static void validateCluster(DFSClient client, int numDNs)
|
||||
throws IOException {
|
||||
DatanodeInfo[] info = client.datanodeReport(DatanodeReportType.LIVE);
|
||||
assertEquals("Number of Datanodes ", numDNs, info.length);
|
||||
assertEquals(numDNs, info.length, "Number of Datanodes ");
|
||||
}
|
||||
|
||||
/** Start a MiniDFSCluster.
|
||||
|
@ -17,9 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
@ -97,9 +95,9 @@ public static int[] randomFilePartition(int n, int parts) {
|
||||
}
|
||||
|
||||
LOG.info("partition=" + Arrays.toString(p));
|
||||
assertTrue("i=0", p[0] > 0 && p[0] < n);
|
||||
assertTrue(p[0] > 0 && p[0] < n, "i=0");
|
||||
for(int i = 1; i < p.length; i++) {
|
||||
assertTrue("i=" + i, p[i] > p[i - 1] && p[i] < n);
|
||||
assertTrue(p[i] > p[i - 1] && p[i] < n, "i=" + i);
|
||||
}
|
||||
return p;
|
||||
}
|
||||
@ -217,8 +215,8 @@ public static void checkFullFile(FileSystem fs, Path name, int len,
|
||||
boolean checkFileStatus) throws IOException {
|
||||
if (checkFileStatus) {
|
||||
final FileStatus status = fs.getFileStatus(name);
|
||||
assertEquals("len=" + len + " but status.getLen()=" + status.getLen(),
|
||||
len, status.getLen());
|
||||
assertEquals(
|
||||
len, status.getLen(), "len=" + len + " but status.getLen()=" + status.getLen());
|
||||
}
|
||||
|
||||
FSDataInputStream stm = fs.open(name);
|
||||
@ -231,9 +229,9 @@ public static void checkFullFile(FileSystem fs, Path name, int len,
|
||||
private static void checkData(final byte[] actual, int from,
|
||||
final byte[] expected, String message) {
|
||||
for (int idx = 0; idx < actual.length; idx++) {
|
||||
assertEquals(message+" byte "+(from+idx)+" differs. expected "+
|
||||
expected[from+idx]+" actual "+actual[idx],
|
||||
expected[from+idx], actual[idx]);
|
||||
assertEquals(
|
||||
expected[from + idx], actual[idx], message + " byte " + (from + idx) + " differs. expected " +
|
||||
expected[from + idx] + " actual " + actual[idx]);
|
||||
actual[idx] = 0;
|
||||
}
|
||||
}
|
||||
|
@ -25,10 +25,7 @@
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HA_NAMENODES_KEY_PREFIX;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_RPC_ADDRESS_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_SERVICE_RPC_ADDRESS_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.BufferedOutputStream;
|
||||
import java.io.BufferedReader;
|
||||
@ -190,8 +187,8 @@
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.apache.hadoop.util.Tool;
|
||||
import org.apache.hadoop.util.VersionInfo;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Assume;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Assumptions;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.annotations.VisibleForTesting;
|
||||
@ -1672,15 +1669,15 @@ public static long roundUpToMultiple(long val, int factor) {
|
||||
}
|
||||
|
||||
public static void checkComponentsEquals(byte[][] expected, byte[][] actual) {
|
||||
assertEquals("expected: " + DFSUtil.byteArray2PathString(expected)
|
||||
+ ", actual: " + DFSUtil.byteArray2PathString(actual), expected.length,
|
||||
actual.length);
|
||||
assertEquals(expected.length,
|
||||
actual.length, "expected: " + DFSUtil.byteArray2PathString(expected)
|
||||
+ ", actual: " + DFSUtil.byteArray2PathString(actual));
|
||||
int i = 0;
|
||||
for (byte[] e : expected) {
|
||||
byte[] actualComponent = actual[i++];
|
||||
assertTrue("expected: " + DFSUtil.bytes2String(e) + ", actual: "
|
||||
+ DFSUtil.bytes2String(actualComponent),
|
||||
Arrays.equals(e, actualComponent));
|
||||
assertTrue(
|
||||
Arrays.equals(e, actualComponent), "expected: " + DFSUtil.bytes2String(e) + ", actual: "
|
||||
+ DFSUtil.bytes2String(actualComponent));
|
||||
}
|
||||
}
|
||||
|
||||
@ -1699,7 +1696,7 @@ public ShortCircuitTestContext(String testName) {
|
||||
this.sockDir = new TemporarySocketDirectory();
|
||||
DomainSocket.disableBindPathValidation();
|
||||
formerTcpReadsDisabled = DFSInputStream.tcpReadsDisabledForTesting;
|
||||
Assume.assumeTrue(DomainSocket.getLoadingFailureReason() == null);
|
||||
Assumptions.assumeTrue(DomainSocket.getLoadingFailureReason() == null);
|
||||
}
|
||||
|
||||
public Configuration newConfiguration() {
|
||||
@ -1737,7 +1734,7 @@ public static void verifyFilesEqual(FileSystem fs, Path p1, Path p2, int len)
|
||||
try (FSDataInputStream in1 = fs.open(p1);
|
||||
FSDataInputStream in2 = fs.open(p2)) {
|
||||
for (int i = 0; i < len; i++) {
|
||||
assertEquals("Mismatch at byte " + i, in1.read(), in2.read());
|
||||
assertEquals(in1.read(), in2.read(), "Mismatch at byte " + i);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1813,32 +1810,32 @@ public static void verifyClientStats(Configuration conf,
|
||||
client.getReplicatedBlockStats();
|
||||
ECBlockGroupStats ecBlockGroupStats = client.getECBlockGroupStats();
|
||||
|
||||
assertEquals("Under replicated stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_LOW_REDUNDANCY_IDX],
|
||||
aggregatedStats[ClientProtocol.GET_STATS_UNDER_REPLICATED_IDX]);
|
||||
assertEquals("Low redundancy stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_LOW_REDUNDANCY_IDX],
|
||||
replicatedBlockStats.getLowRedundancyBlocks() +
|
||||
ecBlockGroupStats.getLowRedundancyBlockGroups());
|
||||
assertEquals("Corrupt blocks stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_CORRUPT_BLOCKS_IDX],
|
||||
replicatedBlockStats.getCorruptBlocks() +
|
||||
ecBlockGroupStats.getCorruptBlockGroups());
|
||||
assertEquals("Missing blocks stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_MISSING_BLOCKS_IDX],
|
||||
replicatedBlockStats.getMissingReplicaBlocks() +
|
||||
ecBlockGroupStats.getMissingBlockGroups());
|
||||
assertEquals("Missing blocks with replication factor one not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_MISSING_REPL_ONE_BLOCKS_IDX],
|
||||
replicatedBlockStats.getMissingReplicationOneBlocks());
|
||||
assertEquals("Bytes in future blocks stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_BYTES_IN_FUTURE_BLOCKS_IDX],
|
||||
replicatedBlockStats.getBytesInFutureBlocks() +
|
||||
ecBlockGroupStats.getBytesInFutureBlockGroups());
|
||||
assertEquals("Pending deletion blocks stats not matching!",
|
||||
aggregatedStats[ClientProtocol.GET_STATS_PENDING_DELETION_BLOCKS_IDX],
|
||||
replicatedBlockStats.getPendingDeletionBlocks() +
|
||||
ecBlockGroupStats.getPendingDeletionBlocks());
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_LOW_REDUNDANCY_IDX],
|
||||
aggregatedStats[ClientProtocol.GET_STATS_UNDER_REPLICATED_IDX], "Under replicated stats not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_LOW_REDUNDANCY_IDX],
|
||||
replicatedBlockStats.getLowRedundancyBlocks() +
|
||||
ecBlockGroupStats.getLowRedundancyBlockGroups(), "Low redundancy stats not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_CORRUPT_BLOCKS_IDX],
|
||||
replicatedBlockStats.getCorruptBlocks() +
|
||||
ecBlockGroupStats.getCorruptBlockGroups(), "Corrupt blocks stats not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_MISSING_BLOCKS_IDX],
|
||||
replicatedBlockStats.getMissingReplicaBlocks() +
|
||||
ecBlockGroupStats.getMissingBlockGroups(), "Missing blocks stats not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_MISSING_REPL_ONE_BLOCKS_IDX],
|
||||
replicatedBlockStats.getMissingReplicationOneBlocks(), "Missing blocks with replication factor one not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_BYTES_IN_FUTURE_BLOCKS_IDX],
|
||||
replicatedBlockStats.getBytesInFutureBlocks() +
|
||||
ecBlockGroupStats.getBytesInFutureBlockGroups(), "Bytes in future blocks stats not matching!");
|
||||
assertEquals(
|
||||
aggregatedStats[ClientProtocol.GET_STATS_PENDING_DELETION_BLOCKS_IDX],
|
||||
replicatedBlockStats.getPendingDeletionBlocks() +
|
||||
ecBlockGroupStats.getPendingDeletionBlocks(), "Pending deletion blocks stats not matching!");
|
||||
}
|
||||
|
||||
/**
|
||||
@ -1884,8 +1881,8 @@ public static DatanodeDescriptor getExpectedPrimaryNode(NameNode nn,
|
||||
ExtendedBlock blk) {
|
||||
BlockManager bm0 = nn.getNamesystem().getBlockManager();
|
||||
BlockInfo storedBlock = bm0.getStoredBlock(blk.getLocalBlock());
|
||||
assertTrue("Block " + blk + " should be under construction, " +
|
||||
"got: " + storedBlock, !storedBlock.isComplete());
|
||||
assertTrue(!storedBlock.isComplete(), "Block " + blk + " should be under construction, " +
|
||||
"got: " + storedBlock);
|
||||
// We expect that the replica with the most recent heart beat will be
|
||||
// the one to be in charge of the synchronization / recovery protocol.
|
||||
final DatanodeStorageInfo[] storages = storedBlock
|
||||
@ -1933,8 +1930,8 @@ public static void toolRun(Tool tool, String cmd, int retcode, String contain)
|
||||
}
|
||||
assertEquals(retcode, ret);
|
||||
if (contain != null) {
|
||||
assertTrue("The real output is: " + output + ".\n It should contain: "
|
||||
+ contain, output.contains(contain));
|
||||
assertTrue(output.contains(contain), "The real output is: " + output + ".\n It should contain: "
|
||||
+ contain);
|
||||
}
|
||||
}
|
||||
|
||||
@ -2338,23 +2335,23 @@ public static void verifyDelete(FsShell shell, FileSystem fs, Path path,
|
||||
|
||||
public static void verifyDelete(FsShell shell, FileSystem fs, Path path,
|
||||
Path trashPath, boolean shouldExistInTrash) throws Exception {
|
||||
assertTrue(path + " file does not exist", fs.exists(path));
|
||||
assertTrue(fs.exists(path), path + " file does not exist");
|
||||
|
||||
// Verify that trashPath has a path component named ".Trash"
|
||||
Path checkTrash = trashPath;
|
||||
while (!checkTrash.isRoot() && !checkTrash.getName().equals(".Trash")) {
|
||||
checkTrash = checkTrash.getParent();
|
||||
}
|
||||
assertEquals("No .Trash component found in trash path " + trashPath,
|
||||
".Trash", checkTrash.getName());
|
||||
assertEquals(
|
||||
".Trash", checkTrash.getName(), "No .Trash component found in trash path " + trashPath);
|
||||
|
||||
String[] argv = new String[]{"-rm", "-r", path.toString()};
|
||||
int res = ToolRunner.run(shell, argv);
|
||||
assertEquals("rm failed", 0, res);
|
||||
assertEquals(0, res, "rm failed");
|
||||
if (shouldExistInTrash) {
|
||||
assertTrue("File not in trash : " + trashPath, fs.exists(trashPath));
|
||||
assertTrue(fs.exists(trashPath), "File not in trash : " + trashPath);
|
||||
} else {
|
||||
assertFalse("File in trash : " + trashPath, fs.exists(trashPath));
|
||||
assertFalse(fs.exists(trashPath), "File in trash : " + trashPath);
|
||||
}
|
||||
}
|
||||
|
||||
@ -2563,7 +2560,7 @@ public static NameNodeConnector getNameNodeConnector(Configuration conf,
|
||||
Path filePath, int namenodeCount, boolean createMoverPath)
|
||||
throws IOException {
|
||||
final Collection<URI> namenodes = DFSUtil.getInternalNsRpcUris(conf);
|
||||
Assert.assertEquals(namenodeCount, namenodes.size());
|
||||
Assertions.assertEquals(namenodeCount, namenodes.size());
|
||||
NameNodeConnector.checkOtherInstanceRunning(createMoverPath);
|
||||
while (true) {
|
||||
try {
|
||||
|
@ -26,9 +26,9 @@
|
||||
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/** This is a comprehensive append test that tries
|
||||
* all combinations of file length and number of appended bytes
|
||||
@ -59,7 +59,7 @@ private static void init(Configuration conf) {
|
||||
conf.setInt(HdfsClientConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_KEY, PACKET_SIZE);
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void startUp () throws IOException {
|
||||
conf = new HdfsConfiguration();
|
||||
init(conf);
|
||||
@ -67,7 +67,7 @@ public static void startUp () throws IOException {
|
||||
fs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -27,7 +27,7 @@
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
|
||||
import static org.junit.Assume.assumeTrue;
|
||||
import static org.junit.jupiter.api.Assumptions.assumeTrue;
|
||||
|
||||
/**
|
||||
* Test striped file write operation with data node failures with parameterized
|
||||
@ -62,9 +62,9 @@ public void runTestWithSingleFailure() {
|
||||
}
|
||||
final int i = base;
|
||||
final Integer length = getLength(i);
|
||||
assumeTrue("Skip test " + i + " since length=null.", length != null);
|
||||
assumeTrue("Test " + i + ", length=" + length
|
||||
+ ", is not chosen to run.", RANDOM.nextInt(16) != 0);
|
||||
assumeTrue(length != null, "Skip test " + i + " since length=null.");
|
||||
assumeTrue(RANDOM.nextInt(16) != 0, "Test " + i + ", length=" + length
|
||||
+ ", is not chosen to run.");
|
||||
System.out.println("Run test " + i + ", length=" + length);
|
||||
runTest(length);
|
||||
}
|
||||
|
@ -30,7 +30,7 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.NameNode;
|
||||
import org.apache.hadoop.hdfs.util.StripedBlockUtil;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.Assert;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.slf4j.event.Level;
|
||||
@ -200,11 +200,11 @@ public static void testReadWithBlockCorrupted(MiniDFSCluster cluster,
|
||||
+ ", parityBlkDelNum = " + parityBlkDelNum
|
||||
+ ", deleteBlockFile? " + deleteBlockFile);
|
||||
int recoverBlkNum = dataBlkDelNum + parityBlkDelNum;
|
||||
Assert.assertTrue("dataBlkDelNum and parityBlkDelNum should be positive",
|
||||
dataBlkDelNum >= 0 && parityBlkDelNum >= 0);
|
||||
Assert.assertTrue("The sum of dataBlkDelNum and parityBlkDelNum " +
|
||||
"should be between 1 ~ " + NUM_PARITY_UNITS, recoverBlkNum <=
|
||||
NUM_PARITY_UNITS);
|
||||
Assertions.assertTrue(
|
||||
dataBlkDelNum >= 0 && parityBlkDelNum >= 0, "dataBlkDelNum and parityBlkDelNum should be positive");
|
||||
Assertions.assertTrue(recoverBlkNum <=
|
||||
NUM_PARITY_UNITS, "The sum of dataBlkDelNum and parityBlkDelNum " +
|
||||
"should be between 1 ~ " + NUM_PARITY_UNITS);
|
||||
|
||||
// write a file with the length of writeLen
|
||||
Path srcPath = new Path(src);
|
||||
@ -232,10 +232,10 @@ public static void corruptBlocks(MiniDFSCluster cluster,
|
||||
|
||||
int[] delDataBlkIndices = StripedFileTestUtil.randomArray(0, NUM_DATA_UNITS,
|
||||
dataBlkDelNum);
|
||||
Assert.assertNotNull(delDataBlkIndices);
|
||||
Assertions.assertNotNull(delDataBlkIndices);
|
||||
int[] delParityBlkIndices = StripedFileTestUtil.randomArray(NUM_DATA_UNITS,
|
||||
NUM_DATA_UNITS + NUM_PARITY_UNITS, parityBlkDelNum);
|
||||
Assert.assertNotNull(delParityBlkIndices);
|
||||
Assertions.assertNotNull(delParityBlkIndices);
|
||||
|
||||
int[] delBlkIndices = new int[recoverBlkNum];
|
||||
System.arraycopy(delDataBlkIndices, 0,
|
||||
|
@ -37,7 +37,7 @@
|
||||
import org.apache.hadoop.io.erasurecode.CodecUtil;
|
||||
import org.apache.hadoop.io.erasurecode.ErasureCoderOptions;
|
||||
import org.apache.hadoop.io.erasurecode.rawcoder.RawErasureEncoder;
|
||||
import org.junit.Assert;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@ -55,7 +55,7 @@
|
||||
import java.util.concurrent.TimeoutException;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
public class StripedFileTestUtil {
|
||||
public static final Logger LOG =
|
||||
@ -77,7 +77,7 @@ static byte getByte(long pos) {
|
||||
static void verifyLength(FileSystem fs, Path srcPath, int fileLength)
|
||||
throws IOException {
|
||||
FileStatus status = fs.getFileStatus(srcPath);
|
||||
assertEquals("File length should be the same", fileLength, status.getLen());
|
||||
assertEquals(fileLength, status.getLen(), "File length should be the same");
|
||||
}
|
||||
|
||||
static void verifyPread(DistributedFileSystem fs, Path srcPath,
|
||||
@ -109,9 +109,9 @@ static void verifyPread(FileSystem fs, Path srcPath, int fileLength,
|
||||
offset += target;
|
||||
}
|
||||
for (int i = 0; i < fileLength - startOffset; i++) {
|
||||
assertEquals("Byte at " + (startOffset + i) + " is different, "
|
||||
+ "the startOffset is " + startOffset, expected[startOffset + i],
|
||||
result[i]);
|
||||
assertEquals(expected[startOffset + i],
|
||||
result[i], "Byte at " + (startOffset + i) + " is different, "
|
||||
+ "the startOffset is " + startOffset);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -127,8 +127,8 @@ static void verifyStatefulRead(FileSystem fs, Path srcPath, int fileLength,
|
||||
System.arraycopy(buf, 0, result, readLen, ret);
|
||||
readLen += ret;
|
||||
}
|
||||
assertEquals("The length of file should be the same to write size", fileLength, readLen);
|
||||
Assert.assertArrayEquals(expected, result);
|
||||
assertEquals(fileLength, readLen, "The length of file should be the same to write size");
|
||||
Assertions.assertArrayEquals(expected, result);
|
||||
}
|
||||
}
|
||||
|
||||
@ -144,8 +144,8 @@ static void verifyStatefulRead(FileSystem fs, Path srcPath, int fileLength,
|
||||
result.put(buf);
|
||||
buf.clear();
|
||||
}
|
||||
assertEquals("The length of file should be the same to write size", fileLength, readLen);
|
||||
Assert.assertArrayEquals(expected, result.array());
|
||||
assertEquals(fileLength, readLen, "The length of file should be the same to write size");
|
||||
Assertions.assertArrayEquals(expected, result.array());
|
||||
}
|
||||
}
|
||||
|
||||
@ -185,14 +185,14 @@ static void verifySeek(FileSystem fs, Path srcPath, int fileLength,
|
||||
if (!(in.getWrappedStream() instanceof WebHdfsInputStream)) {
|
||||
try {
|
||||
in.seek(-1);
|
||||
Assert.fail("Should be failed if seek to negative offset");
|
||||
Assertions.fail("Should be failed if seek to negative offset");
|
||||
} catch (EOFException e) {
|
||||
// expected
|
||||
}
|
||||
|
||||
try {
|
||||
in.seek(fileLength + 1);
|
||||
Assert.fail("Should be failed if seek after EOF");
|
||||
Assertions.fail("Should be failed if seek after EOF");
|
||||
} catch (EOFException e) {
|
||||
// expected
|
||||
}
|
||||
@ -206,8 +206,8 @@ static void assertSeekAndRead(FSDataInputStream fsdis, int pos,
|
||||
byte[] buf = new byte[writeBytes - pos];
|
||||
IOUtils.readFully(fsdis, buf, 0, buf.length);
|
||||
for (int i = 0; i < buf.length; i++) {
|
||||
assertEquals("Byte at " + i + " should be the same",
|
||||
StripedFileTestUtil.getByte(pos + i), buf[i]);
|
||||
assertEquals(
|
||||
StripedFileTestUtil.getByte(pos + i), buf[i], "Byte at " + i + " should be the same");
|
||||
}
|
||||
}
|
||||
|
||||
@ -225,7 +225,7 @@ static DatanodeInfo getDatanodes(StripedDataStreamer streamer) {
|
||||
final DatanodeInfo[] datanodes = streamer.getNodes();
|
||||
if (datanodes != null) {
|
||||
assertEquals(1, datanodes.length);
|
||||
Assert.assertNotNull(datanodes[0]);
|
||||
Assertions.assertNotNull(datanodes[0]);
|
||||
return datanodes[0];
|
||||
}
|
||||
try {
|
||||
@ -377,13 +377,13 @@ static void checkData(DistributedFileSystem dfs, Path srcPath, int length,
|
||||
final int parityBlkNum = ecPolicy.getNumParityUnits();
|
||||
int index = 0;
|
||||
for (LocatedBlock firstBlock : lbs.getLocatedBlocks()) {
|
||||
Assert.assertTrue(firstBlock instanceof LocatedStripedBlock);
|
||||
Assertions.assertTrue(firstBlock instanceof LocatedStripedBlock);
|
||||
|
||||
final long gs = firstBlock.getBlock().getGenerationStamp();
|
||||
final long oldGS = oldGSList != null ? oldGSList.get(index++) : -1L;
|
||||
final String s = "gs=" + gs + ", oldGS=" + oldGS;
|
||||
LOG.info(s);
|
||||
Assert.assertTrue(s, gs >= oldGS);
|
||||
Assertions.assertTrue(gs >= oldGS, s);
|
||||
|
||||
LocatedBlock[] blocks = StripedBlockUtil.parseStripedBlockGroup(
|
||||
(LocatedStripedBlock) firstBlock, cellSize,
|
||||
@ -456,7 +456,7 @@ static void checkData(DistributedFileSystem dfs, Path srcPath, int length,
|
||||
for (int posInBlk = 0; posInBlk < actual.length; posInBlk++) {
|
||||
final long posInFile = StripedBlockUtil.offsetInBlkToOffsetInBG(
|
||||
cellSize, dataBlkNum, posInBlk, i) + groupPosInFile;
|
||||
Assert.assertTrue(posInFile < length);
|
||||
Assertions.assertTrue(posInFile < length);
|
||||
final byte expected = getByte(posInFile);
|
||||
|
||||
if (killed) {
|
||||
@ -466,7 +466,7 @@ static void checkData(DistributedFileSystem dfs, Path srcPath, int length,
|
||||
String s = "expected=" + expected + " but actual=" + actual[posInBlk]
|
||||
+ ", posInFile=" + posInFile + ", posInBlk=" + posInBlk
|
||||
+ ". group=" + group + ", i=" + i;
|
||||
Assert.fail(s);
|
||||
Assertions.fail(s);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -507,12 +507,12 @@ static void verifyParityBlocks(Configuration conf, final long size,
|
||||
try {
|
||||
encoder.encode(dataBytes, expectedParityBytes);
|
||||
} catch (IOException e) {
|
||||
Assert.fail("Unexpected IOException: " + e.getMessage());
|
||||
Assertions.fail("Unexpected IOException: " + e.getMessage());
|
||||
}
|
||||
for (int i = 0; i < parityBytes.length; i++) {
|
||||
if (checkSet.contains(i + dataBytes.length)){
|
||||
Assert.assertArrayEquals("i=" + i, expectedParityBytes[i],
|
||||
parityBytes[i]);
|
||||
Assertions.assertArrayEquals(expectedParityBytes[i],
|
||||
parityBytes[i], "i=" + i);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -30,10 +30,10 @@
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
||||
import org.apache.hadoop.hdfs.protocol.QuotaExceededException;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* Test abandoning blocks, which clients do on pipeline creation failure.
|
||||
@ -48,14 +48,14 @@ public class TestAbandonBlock {
|
||||
private MiniDFSCluster cluster;
|
||||
private DistributedFileSystem fs;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws Exception {
|
||||
cluster = new MiniDFSCluster.Builder(CONF).numDataNodes(2).build();
|
||||
fs = cluster.getFileSystem();
|
||||
cluster.waitActive();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
if (fs != null) {
|
||||
fs.close();
|
||||
@ -100,8 +100,8 @@ public void testAbandonBlock() throws IOException {
|
||||
cluster.restartNameNode();
|
||||
blocks = dfsclient.getNamenode().getBlockLocations(src, 0,
|
||||
Integer.MAX_VALUE);
|
||||
Assert.assertEquals("Blocks " + b + " has not been abandoned.",
|
||||
orginalNumBlocks, blocks.locatedBlockCount() + 1);
|
||||
Assertions.assertEquals(
|
||||
orginalNumBlocks, blocks.locatedBlockCount() + 1, "Blocks " + b + " has not been abandoned.");
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -43,12 +43,11 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.security.authorize.ProxyUsers;
|
||||
import org.junit.Assert;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This class tests the ACLs system through the full code path. It overlaps
|
||||
@ -89,7 +88,7 @@ public class TestAclsEndToEnd {
|
||||
private MiniDFSCluster cluster;
|
||||
private DistributedFileSystem fs;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void captureUser() throws IOException {
|
||||
realUgi = UserGroupInformation.getCurrentUser();
|
||||
realUser = System.getProperty("user.name");
|
||||
@ -174,7 +173,7 @@ private void setup(Configuration conf, boolean resetKms, boolean resetDfs)
|
||||
|
||||
kmsDir = new File(fsHelper.getTestRootDir()).getAbsoluteFile();
|
||||
|
||||
Assert.assertTrue(kmsDir.mkdirs());
|
||||
Assertions.assertTrue(kmsDir.mkdirs());
|
||||
}
|
||||
|
||||
writeConf(kmsDir, conf);
|
||||
@ -411,66 +410,66 @@ private void doFullAclTest(final Configuration conf,
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
// Create a test key
|
||||
assertTrue("Exception during creation of key " + KEY1 + " by "
|
||||
+ keyadminUgi.getUserName(), createKey(keyadminUgi, KEY1, conf));
|
||||
// Create a test key
|
||||
assertTrue(createKey(keyadminUgi, KEY1, conf), "Exception during creation of key " + KEY1 + " by "
|
||||
+ keyadminUgi.getUserName());
|
||||
|
||||
// Fail to create a test key
|
||||
assertFalse("Allowed creation of key " + KEY2 + " by "
|
||||
+ hdfsUgi.getUserName(), createKey(hdfsUgi, KEY2, conf));
|
||||
assertFalse("Allowed creation of key " + KEY2 + " by "
|
||||
+ userUgi.getUserName(), createKey(userUgi, KEY2, conf));
|
||||
// Fail to create a test key
|
||||
assertFalse(createKey(hdfsUgi, KEY2, conf), "Allowed creation of key " + KEY2 + " by "
|
||||
+ hdfsUgi.getUserName());
|
||||
assertFalse(createKey(userUgi, KEY2, conf), "Allowed creation of key " + KEY2 + " by "
|
||||
+ userUgi.getUserName());
|
||||
|
||||
// Create a directory and chown it to the normal user.
|
||||
fs.mkdirs(ZONE1);
|
||||
fs.setOwner(ZONE1, userUgi.getUserName(),
|
||||
userUgi.getPrimaryGroupName());
|
||||
|
||||
// Create an EZ
|
||||
assertTrue("Exception during creation of EZ " + ZONE1 + " by "
|
||||
+ hdfsUgi.getUserName() + " using key " + KEY1,
|
||||
createEncryptionZone(hdfsUgi, KEY1, ZONE1));
|
||||
// Create an EZ
|
||||
assertTrue(
|
||||
createEncryptionZone(hdfsUgi, KEY1, ZONE1), "Exception during creation of EZ " + ZONE1 + " by "
|
||||
+ hdfsUgi.getUserName() + " using key " + KEY1);
|
||||
|
||||
// Fail to create an EZ
|
||||
assertFalse("Allowed creation of EZ " + ZONE2 + " by "
|
||||
+ keyadminUgi.getUserName() + " using key " + KEY1,
|
||||
createEncryptionZone(keyadminUgi, KEY1, ZONE2));
|
||||
assertFalse("Allowed creation of EZ " + ZONE2 + " by "
|
||||
+ userUgi.getUserName() + " using key " + KEY1,
|
||||
createEncryptionZone(userUgi, KEY1, ZONE2));
|
||||
// Fail to create an EZ
|
||||
assertFalse(
|
||||
createEncryptionZone(keyadminUgi, KEY1, ZONE2), "Allowed creation of EZ " + ZONE2 + " by "
|
||||
+ keyadminUgi.getUserName() + " using key " + KEY1);
|
||||
assertFalse(
|
||||
createEncryptionZone(userUgi, KEY1, ZONE2), "Allowed creation of EZ " + ZONE2 + " by "
|
||||
+ userUgi.getUserName() + " using key " + KEY1);
|
||||
|
||||
// Create a file in the zone
|
||||
assertTrue("Exception during creation of file " + FILE1 + " by "
|
||||
+ userUgi.getUserName(), createFile(userUgi, FILE1, TEXT));
|
||||
// Create a file in the zone
|
||||
assertTrue(createFile(userUgi, FILE1, TEXT), "Exception during creation of file " + FILE1 + " by "
|
||||
+ userUgi.getUserName());
|
||||
|
||||
// Fail to create a file in the zone
|
||||
assertFalse("Allowed creation of file " + FILE1A + " by "
|
||||
+ hdfsUgi.getUserName(), createFile(hdfsUgi, FILE1A, TEXT));
|
||||
assertFalse("Allowed creation of file " + FILE1A + " by "
|
||||
+ keyadminUgi.getUserName(), createFile(keyadminUgi, FILE1A, TEXT));
|
||||
// Fail to create a file in the zone
|
||||
assertFalse(createFile(hdfsUgi, FILE1A, TEXT), "Allowed creation of file " + FILE1A + " by "
|
||||
+ hdfsUgi.getUserName());
|
||||
assertFalse(createFile(keyadminUgi, FILE1A, TEXT), "Allowed creation of file " + FILE1A + " by "
|
||||
+ keyadminUgi.getUserName());
|
||||
|
||||
// Read a file in the zone
|
||||
assertTrue("Exception while reading file " + FILE1 + " by "
|
||||
+ userUgi.getUserName(), compareFile(userUgi, FILE1, TEXT));
|
||||
// Read a file in the zone
|
||||
assertTrue(compareFile(userUgi, FILE1, TEXT), "Exception while reading file " + FILE1 + " by "
|
||||
+ userUgi.getUserName());
|
||||
|
||||
// Fail to read a file in the zone
|
||||
assertFalse("Allowed reading of file " + FILE1 + " by "
|
||||
+ hdfsUgi.getUserName(), compareFile(hdfsUgi, FILE1, TEXT));
|
||||
assertFalse("Allowed reading of file " + FILE1 + " by "
|
||||
+ keyadminUgi.getUserName(), compareFile(keyadminUgi, FILE1, TEXT));
|
||||
// Fail to read a file in the zone
|
||||
assertFalse(compareFile(hdfsUgi, FILE1, TEXT), "Allowed reading of file " + FILE1 + " by "
|
||||
+ hdfsUgi.getUserName());
|
||||
assertFalse(compareFile(keyadminUgi, FILE1, TEXT), "Allowed reading of file " + FILE1 + " by "
|
||||
+ keyadminUgi.getUserName());
|
||||
|
||||
// Remove the zone
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
// Fail to remove the key
|
||||
assertFalse("Allowed deletion of file " + FILE1 + " by "
|
||||
+ hdfsUgi.getUserName(), deleteKey(hdfsUgi, KEY1));
|
||||
assertFalse("Allowed deletion of file " + FILE1 + " by "
|
||||
+ userUgi.getUserName(), deleteKey(userUgi, KEY1));
|
||||
// Fail to remove the key
|
||||
assertFalse(deleteKey(hdfsUgi, KEY1), "Allowed deletion of file " + FILE1 + " by "
|
||||
+ hdfsUgi.getUserName());
|
||||
assertFalse(deleteKey(userUgi, KEY1), "Allowed deletion of file " + FILE1 + " by "
|
||||
+ userUgi.getUserName());
|
||||
|
||||
// Remove
|
||||
assertTrue("Exception during deletion of file " + FILE1 + " by "
|
||||
+ keyadminUgi.getUserName(), deleteKey(keyadminUgi, KEY1));
|
||||
// Remove
|
||||
assertTrue(deleteKey(keyadminUgi, KEY1), "Exception during deletion of file " + FILE1 + " by "
|
||||
+ keyadminUgi.getUserName());
|
||||
} finally {
|
||||
fs.delete(ZONE1, true);
|
||||
fs.delete(ZONE2, true);
|
||||
@ -495,8 +494,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation with correct config"
|
||||
+ " using whitelist key ACLs", createKey(realUgi, KEY1, conf));
|
||||
assertTrue(createKey(realUgi, KEY1, conf), "Exception during key creation with correct config"
|
||||
+ " using whitelist key ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -512,8 +511,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation with correct config"
|
||||
+ " using default key ACLs", createKey(realUgi, KEY2, conf));
|
||||
assertTrue(createKey(realUgi, KEY2, conf), "Exception during key creation with correct config"
|
||||
+ " using default key ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -531,8 +530,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertFalse("Allowed key creation with blacklist for CREATE",
|
||||
createKey(realUgi, KEY3, conf));
|
||||
assertFalse(
|
||||
createKey(realUgi, KEY3, conf), "Allowed key creation with blacklist for CREATE");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -547,8 +546,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertFalse("Allowed key creation without CREATE KMS ACL",
|
||||
createKey(realUgi, KEY3, conf));
|
||||
assertFalse(
|
||||
createKey(realUgi, KEY3, conf), "Allowed key creation without CREATE KMS ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -562,8 +561,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertFalse("Allowed key creation without MANAGMENT key ACL",
|
||||
createKey(realUgi, KEY3, conf));
|
||||
assertFalse(
|
||||
createKey(realUgi, KEY3, conf), "Allowed key creation without MANAGMENT key ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -581,8 +580,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertFalse("Allowed key creation when default key ACL should have been"
|
||||
+ " overridden by key ACL", createKey(realUgi, KEY3, conf));
|
||||
assertFalse(createKey(realUgi, KEY3, conf), "Allowed key creation when default key ACL should have been"
|
||||
+ " overridden by key ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -596,8 +595,8 @@ public void testCreateKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation with default KMS ACLs",
|
||||
createKey(realUgi, KEY3, conf));
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY3, conf), "Exception during key creation with default KMS ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -620,8 +619,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY1, conf));
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY1, conf), "Exception during key creation");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -647,8 +646,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE1);
|
||||
|
||||
assertTrue("Exception during zone creation with correct config using"
|
||||
+ " whitelist key ACLs", createEncryptionZone(realUgi, KEY1, ZONE1));
|
||||
assertTrue(createEncryptionZone(realUgi, KEY1, ZONE1), "Exception during zone creation with correct config using"
|
||||
+ " whitelist key ACLs");
|
||||
} finally {
|
||||
fs.delete(ZONE1, true);
|
||||
teardown();
|
||||
@ -671,8 +670,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE2);
|
||||
|
||||
assertTrue("Exception during zone creation with correct config using"
|
||||
+ " default key ACLs", createEncryptionZone(realUgi, KEY1, ZONE2));
|
||||
assertTrue(createEncryptionZone(realUgi, KEY1, ZONE2), "Exception during zone creation with correct config using"
|
||||
+ " default key ACLs");
|
||||
} finally {
|
||||
fs.delete(ZONE2, true);
|
||||
teardown();
|
||||
@ -697,9 +696,9 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE3);
|
||||
|
||||
assertFalse("Allowed creation of zone when default key ACLs should have"
|
||||
+ " been overridden by key ACL",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3), "Allowed creation of zone when default key ACLs should have"
|
||||
+ " been overridden by key ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE3, true);
|
||||
teardown();
|
||||
@ -724,8 +723,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE3);
|
||||
|
||||
assertFalse("Allowed zone creation of zone with blacklisted GET_METADATA",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3), "Allowed zone creation of zone with blacklisted GET_METADATA");
|
||||
} finally {
|
||||
fs.delete(ZONE3, true);
|
||||
teardown();
|
||||
@ -750,8 +749,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE3);
|
||||
|
||||
assertFalse("Allowed zone creation of zone with blacklisted GENERATE_EEK",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3), "Allowed zone creation of zone with blacklisted GENERATE_EEK");
|
||||
} finally {
|
||||
fs.delete(ZONE3, true);
|
||||
teardown();
|
||||
@ -771,8 +770,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE3);
|
||||
|
||||
assertTrue("Exception during zone creation with default KMS ACLs",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3), "Exception during zone creation with default KMS ACLs");
|
||||
} finally {
|
||||
fs.delete(ZONE3, true);
|
||||
teardown();
|
||||
@ -794,8 +793,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE4);
|
||||
|
||||
assertFalse("Allowed zone creation without GET_METADATA KMS ACL",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4), "Allowed zone creation without GET_METADATA KMS ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE4, true);
|
||||
teardown();
|
||||
@ -817,8 +816,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE4);
|
||||
|
||||
assertFalse("Allowed zone creation without GENERATE_EEK KMS ACL",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4), "Allowed zone creation without GENERATE_EEK KMS ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE4, true);
|
||||
teardown();
|
||||
@ -839,8 +838,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE4);
|
||||
|
||||
assertFalse("Allowed zone creation without READ ACL",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4), "Allowed zone creation without READ ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE4, true);
|
||||
teardown();
|
||||
@ -861,8 +860,8 @@ public void testCreateEncryptionZone() throws Exception {
|
||||
|
||||
fs.mkdirs(ZONE4);
|
||||
|
||||
assertFalse("Allowed zone creation without GENERATE_EEK ACL",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4));
|
||||
assertFalse(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4), "Allowed zone creation without GENERATE_EEK ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE4, true);
|
||||
teardown();
|
||||
@ -896,20 +895,20 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY1, conf));
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY1, conf), "Exception during key creation");
|
||||
fs.mkdirs(ZONE1);
|
||||
assertTrue("Exception during zone creation",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE1));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE1), "Exception during zone creation");
|
||||
fs.mkdirs(ZONE2);
|
||||
assertTrue("Exception during zone creation",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE2));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE2), "Exception during zone creation");
|
||||
fs.mkdirs(ZONE3);
|
||||
assertTrue("Exception during zone creation",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE3), "Exception during zone creation");
|
||||
fs.mkdirs(ZONE4);
|
||||
assertTrue("Exception during zone creation",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE4), "Exception during zone creation");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
fs.delete(ZONE2, true);
|
||||
@ -941,8 +940,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception during file creation with correct config"
|
||||
+ " using whitelist ACL", createFile(realUgi, FILE1, TEXT));
|
||||
assertTrue(createFile(realUgi, FILE1, TEXT), "Exception during file creation with correct config"
|
||||
+ " using whitelist ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE1, true);
|
||||
teardown();
|
||||
@ -963,8 +962,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception during file creation with correct config"
|
||||
+ " using whitelist ACL", createFile(realUgi, FILE2, TEXT));
|
||||
assertTrue(createFile(realUgi, FILE2, TEXT), "Exception during file creation with correct config"
|
||||
+ " using whitelist ACL");
|
||||
} finally {
|
||||
fs.delete(ZONE2, true);
|
||||
teardown();
|
||||
@ -987,8 +986,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation when default key ACLs should have been"
|
||||
+ " overridden by key ACL", createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(createFile(realUgi, FILE3, TEXT), "Allowed file creation when default key ACLs should have been"
|
||||
+ " overridden by key ACL");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1014,8 +1013,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation with blacklist for GENERATE_EEK",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE3, TEXT), "Allowed file creation with blacklist for GENERATE_EEK");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1041,8 +1040,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation with blacklist for DECRYPT_EEK",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE3, TEXT), "Allowed file creation with blacklist for DECRYPT_EEK");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1062,8 +1061,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception during file creation with default KMS ACLs",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertTrue(
|
||||
createFile(realUgi, FILE3, TEXT), "Exception during file creation with default KMS ACLs");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1086,8 +1085,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation without GENERATE_EEK KMS ACL",
|
||||
createFile(realUgi, FILE4, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE4, TEXT), "Allowed file creation without GENERATE_EEK KMS ACL");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1110,8 +1109,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation without DECRYPT_EEK KMS ACL",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE3, TEXT), "Allowed file creation without DECRYPT_EEK KMS ACL");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1133,8 +1132,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation without GENERATE_EEK key ACL",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE3, TEXT), "Allowed file creation without GENERATE_EEK key ACL");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1156,8 +1155,8 @@ public void testCreateFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file creation without DECRYPT_EEK key ACL",
|
||||
createFile(realUgi, FILE3, TEXT));
|
||||
assertFalse(
|
||||
createFile(realUgi, FILE3, TEXT), "Allowed file creation without DECRYPT_EEK key ACL");
|
||||
} catch (Exception ex) {
|
||||
fs.delete(ZONE3, true);
|
||||
|
||||
@ -1198,13 +1197,13 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY1, conf));
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY1, conf), "Exception during key creation");
|
||||
fs.mkdirs(ZONE1);
|
||||
assertTrue("Exception during zone creation",
|
||||
createEncryptionZone(realUgi, KEY1, ZONE1));
|
||||
assertTrue("Exception during file creation",
|
||||
createFile(realUgi, FILE1, TEXT));
|
||||
assertTrue(
|
||||
createEncryptionZone(realUgi, KEY1, ZONE1), "Exception during zone creation");
|
||||
assertTrue(
|
||||
createFile(realUgi, FILE1, TEXT), "Exception during file creation");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1229,8 +1228,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception while reading file with correct config with"
|
||||
+ " whitelist ACLs", compareFile(realUgi, FILE1, TEXT));
|
||||
assertTrue(compareFile(realUgi, FILE1, TEXT), "Exception while reading file with correct config with"
|
||||
+ " whitelist ACLs");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1250,8 +1249,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception while reading file with correct config"
|
||||
+ " with default ACLs", compareFile(realUgi, FILE1, TEXT));
|
||||
assertTrue(compareFile(realUgi, FILE1, TEXT), "Exception while reading file with correct config"
|
||||
+ " with default ACLs");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1273,8 +1272,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file read when default key ACLs should have been"
|
||||
+ " overridden by key ACL", compareFile(realUgi, FILE1, TEXT));
|
||||
assertFalse(compareFile(realUgi, FILE1, TEXT), "Allowed file read when default key ACLs should have been"
|
||||
+ " overridden by key ACL");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1296,8 +1295,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file read with blacklist for DECRYPT_EEK",
|
||||
compareFile(realUgi, FILE1, TEXT));
|
||||
assertFalse(
|
||||
compareFile(realUgi, FILE1, TEXT), "Allowed file read with blacklist for DECRYPT_EEK");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1315,8 +1314,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertTrue("Exception while reading file with default KMS ACLs",
|
||||
compareFile(realUgi, FILE1, TEXT));
|
||||
assertTrue(
|
||||
compareFile(realUgi, FILE1, TEXT), "Exception while reading file with default KMS ACLs");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1335,8 +1334,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file read without DECRYPT_EEK KMS ACL",
|
||||
compareFile(realUgi, FILE1, TEXT));
|
||||
assertFalse(
|
||||
compareFile(realUgi, FILE1, TEXT), "Allowed file read without DECRYPT_EEK KMS ACL");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1351,8 +1350,8 @@ public void testReadFileInEncryptionZone() throws Exception {
|
||||
try {
|
||||
setup(conf, false, false);
|
||||
|
||||
assertFalse("Allowed file read without DECRYPT_EEK key ACL",
|
||||
compareFile(realUgi, FILE1, TEXT));
|
||||
assertFalse(
|
||||
compareFile(realUgi, FILE1, TEXT), "Allowed file read without DECRYPT_EEK key ACL");
|
||||
} catch (Throwable ex) {
|
||||
fs.delete(ZONE1, true);
|
||||
|
||||
@ -1379,12 +1378,12 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf);
|
||||
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY1, conf));
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY2, conf));
|
||||
assertTrue("Exception during key creation",
|
||||
createKey(realUgi, KEY3, conf));
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY1, conf), "Exception during key creation");
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY2, conf), "Exception during key creation");
|
||||
assertTrue(
|
||||
createKey(realUgi, KEY3, conf), "Exception during key creation");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1405,8 +1404,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertTrue("Exception during key deletion with correct config"
|
||||
+ " using whitelist key ACLs", deleteKey(realUgi, KEY1));
|
||||
assertTrue(deleteKey(realUgi, KEY1), "Exception during key deletion with correct config"
|
||||
+ " using whitelist key ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1422,8 +1421,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertTrue("Exception during key deletion with correct config"
|
||||
+ " using default key ACLs", deleteKey(realUgi, KEY2));
|
||||
assertTrue(deleteKey(realUgi, KEY2), "Exception during key deletion with correct config"
|
||||
+ " using default key ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1441,8 +1440,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertFalse("Allowed key deletion with blacklist for DELETE",
|
||||
deleteKey(realUgi, KEY3));
|
||||
assertFalse(
|
||||
deleteKey(realUgi, KEY3), "Allowed key deletion with blacklist for DELETE");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1457,8 +1456,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertFalse("Allowed key deletion without DELETE KMS ACL",
|
||||
deleteKey(realUgi, KEY3));
|
||||
assertFalse(
|
||||
deleteKey(realUgi, KEY3), "Allowed key deletion without DELETE KMS ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1473,8 +1472,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertFalse("Allowed key deletion without MANAGMENT key ACL",
|
||||
deleteKey(realUgi, KEY3));
|
||||
assertFalse(
|
||||
deleteKey(realUgi, KEY3), "Allowed key deletion without MANAGMENT key ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1492,8 +1491,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertFalse("Allowed key deletion when default key ACL should have been"
|
||||
+ " overridden by key ACL", deleteKey(realUgi, KEY3));
|
||||
assertFalse(deleteKey(realUgi, KEY3), "Allowed key deletion when default key ACL should have been"
|
||||
+ " overridden by key ACL");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1507,8 +1506,8 @@ public void testDeleteKey() throws Exception {
|
||||
try {
|
||||
setup(conf, false);
|
||||
|
||||
assertTrue("Exception during key deletion with default KMS ACLs",
|
||||
deleteKey(realUgi, KEY3));
|
||||
assertTrue(
|
||||
deleteKey(realUgi, KEY3), "Exception during key deletion with default KMS ACLs");
|
||||
} finally {
|
||||
teardown();
|
||||
}
|
||||
@ -1596,8 +1595,8 @@ public void execute() throws IOException {
|
||||
FSDataInputStream din = cluster.getFileSystem().open(file);
|
||||
BufferedReader in = new BufferedReader(new InputStreamReader(din));
|
||||
|
||||
assertEquals("The text read does not match the text written",
|
||||
text, in.readLine());
|
||||
assertEquals(
|
||||
text, in.readLine(), "The text read does not match the text written");
|
||||
}
|
||||
});
|
||||
}
|
||||
|
@ -26,10 +26,10 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
|
||||
/**
|
||||
* Test cases for trying to append to a file with a different
|
||||
@ -44,7 +44,7 @@ public class TestAppendDifferentChecksum {
|
||||
private static FileSystem fs;
|
||||
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setupCluster() throws IOException {
|
||||
Configuration conf = new HdfsConfiguration();
|
||||
conf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 4096);
|
||||
@ -55,7 +55,7 @@ public static void setupCluster() throws IOException {
|
||||
fs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void teardown() throws IOException {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -68,7 +68,7 @@ public static void teardown() throws IOException {
|
||||
* difficulties in doing so.
|
||||
*/
|
||||
@Test
|
||||
@Ignore("this is not implemented! See HDFS-2130")
|
||||
@Disabled("this is not implemented! See HDFS-2130")
|
||||
public void testSwitchChunkSize() throws IOException {
|
||||
FileSystem fsWithSmallChunk = createFsWithChecksum("CRC32", 512);
|
||||
FileSystem fsWithBigChunk = createFsWithChecksum("CRC32", 1024);
|
||||
|
@ -45,11 +45,10 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.NameNode;
|
||||
import org.apache.hadoop.hdfs.server.namenode.TestFileTruncate;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.slf4j.event.Level;
|
||||
|
||||
@ -78,7 +77,7 @@ public class TestAppendSnapshotTruncate {
|
||||
static MiniDFSCluster cluster;
|
||||
static DistributedFileSystem dfs;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void startUp() throws IOException {
|
||||
conf = new HdfsConfiguration();
|
||||
conf.setLong(DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, BLOCK_SIZE);
|
||||
@ -95,7 +94,7 @@ public static void startUp() throws IOException {
|
||||
dfs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() throws IOException {
|
||||
if(dfs != null) {
|
||||
dfs.close();
|
||||
@ -167,7 +166,7 @@ String createSnapshot(String snapshot) throws IOException {
|
||||
{
|
||||
//copy all local files to a sub dir to simulate snapshot.
|
||||
final File subDir = new File(localDir, snapshot);
|
||||
Assert.assertFalse(subDir.exists());
|
||||
Assertions.assertFalse(subDir.exists());
|
||||
subDir.mkdir();
|
||||
|
||||
for(File f : localDir.listFiles(FILE_ONLY)) {
|
||||
@ -185,12 +184,12 @@ String checkSnapshot(String snapshot) throws IOException {
|
||||
.append(snapshot);
|
||||
|
||||
final File subDir = new File(localDir, snapshot);
|
||||
Assert.assertTrue(subDir.exists());
|
||||
Assertions.assertTrue(subDir.exists());
|
||||
|
||||
final File[] localFiles = subDir.listFiles(FILE_ONLY);
|
||||
final Path p = snapshotPaths.get(snapshot);
|
||||
final FileStatus[] statuses = dfs.listStatus(p);
|
||||
Assert.assertEquals(localFiles.length, statuses.length);
|
||||
Assertions.assertEquals(localFiles.length, statuses.length);
|
||||
b.append(p).append(" vs ").append(subDir).append(", ")
|
||||
.append(statuses.length).append(" entries");
|
||||
|
||||
@ -374,8 +373,8 @@ int checkLength() throws IOException {
|
||||
|
||||
static int checkLength(Path file, File localFile) throws IOException {
|
||||
final long length = dfs.getFileStatus(file).getLen();
|
||||
Assert.assertEquals(localFile.length(), length);
|
||||
Assert.assertTrue(length <= Integer.MAX_VALUE);
|
||||
Assertions.assertEquals(localFile.length(), length);
|
||||
Assertions.assertTrue(length <= Integer.MAX_VALUE);
|
||||
return (int)length;
|
||||
}
|
||||
|
||||
|
@ -17,8 +17,8 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
import java.io.IOException;
|
||||
@ -27,9 +27,9 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.protocol.BlockStoragePolicy;
|
||||
import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestApplyingStoragePolicy {
|
||||
private static final short REPL = 1;
|
||||
@ -39,7 +39,7 @@ public class TestApplyingStoragePolicy {
|
||||
private static MiniDFSCluster cluster;
|
||||
private static DistributedFileSystem fs;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void clusterSetUp() throws IOException {
|
||||
conf = new HdfsConfiguration();
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL).build();
|
||||
@ -47,7 +47,7 @@ public void clusterSetUp() throws IOException {
|
||||
fs = cluster.getFileSystem();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void clusterShutdown() throws IOException{
|
||||
if(fs != null) {
|
||||
fs.close();
|
||||
|
@ -17,8 +17,8 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.PrintStream;
|
||||
@ -34,7 +34,7 @@
|
||||
import org.apache.hadoop.hdfs.tools.DFSAdmin;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This test ensures that the balancer bandwidth is dynamically adjusted
|
||||
@ -131,11 +131,11 @@ private void runGetBalancerBandwidthCmd(DFSAdmin admin, String[] args,
|
||||
try {
|
||||
System.setOut(outStream);
|
||||
int exitCode = admin.run(args);
|
||||
assertEquals("DFSAdmin should return 0", 0, exitCode);
|
||||
assertEquals(0, exitCode, "DFSAdmin should return 0");
|
||||
String bandwidthOutMsg = "Balancer bandwidth is " + expectedBandwidth
|
||||
+ " bytes per second.";
|
||||
String strOut = new String(outContent.toByteArray(), UTF8);
|
||||
assertTrue("Wrong balancer bandwidth!", strOut.contains(bandwidthOutMsg));
|
||||
assertTrue(strOut.contains(bandwidthOutMsg), "Wrong balancer bandwidth!");
|
||||
} finally {
|
||||
System.setOut(initialStdOut);
|
||||
}
|
||||
|
@ -33,10 +33,10 @@
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.util.Lists;
|
||||
import org.hamcrest.core.StringContains;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Rule;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.rules.ExpectedException;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
@ -46,9 +46,7 @@
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
/**
|
||||
* Tests for the batched listing API.
|
||||
@ -85,15 +83,15 @@ private static Path getFileName(int i, int j, int k) {
|
||||
private static void assertSubDirEquals(int i, int j, Path p) {
|
||||
assertTrue(p.toString().startsWith("hdfs://"));
|
||||
Path expected = getSubDirName(i, j);
|
||||
assertEquals("Unexpected subdir name",
|
||||
expected.toString(), p.toUri().getPath());
|
||||
assertEquals(
|
||||
expected.toString(), p.toUri().getPath(), "Unexpected subdir name");
|
||||
}
|
||||
|
||||
private static void assertFileEquals(int i, int j, int k, Path p) {
|
||||
assertTrue(p.toString().startsWith("hdfs://"));
|
||||
Path expected = getFileName(i, j, k);
|
||||
assertEquals("Unexpected file name",
|
||||
expected.toString(), p.toUri().getPath());
|
||||
assertEquals(
|
||||
expected.toString(), p.toUri().getPath(), "Unexpected file name");
|
||||
}
|
||||
|
||||
private static void loadData() throws Exception {
|
||||
@ -119,7 +117,7 @@ private static void loadData() throws Exception {
|
||||
dfs.setPermission(INACCESSIBLE_DIR_PATH, new FsPermission(0000));
|
||||
}
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void beforeClass() throws Exception {
|
||||
conf = new HdfsConfiguration();
|
||||
conf.setInt(DFSConfigKeys.DFS_LIST_LIMIT, 7);
|
||||
@ -132,7 +130,7 @@ public static void beforeClass() throws Exception {
|
||||
loadData();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void afterClass() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -233,8 +231,8 @@ public void listDirRelative() throws Exception {
|
||||
dfs.setWorkingDirectory(new Path("/dir0"));
|
||||
List<Path> paths = Lists.newArrayList(new Path("."));
|
||||
List<FileStatus> statuses = getStatuses(paths);
|
||||
assertEquals("Wrong number of items",
|
||||
SECOND_LEVEL_DIRS, statuses.size());
|
||||
assertEquals(
|
||||
SECOND_LEVEL_DIRS, statuses.size(), "Wrong number of items");
|
||||
for (int i = 0; i < SECOND_LEVEL_DIRS; i++) {
|
||||
FileStatus stat = statuses.get(i);
|
||||
assertSubDirEquals(0, i, stat.getPath());
|
||||
@ -246,8 +244,8 @@ public void listFilesRelative() throws Exception {
|
||||
dfs.setWorkingDirectory(new Path("/dir0"));
|
||||
List<Path> paths = Lists.newArrayList(new Path("subdir0"));
|
||||
List<FileStatus> statuses = getStatuses(paths);
|
||||
assertEquals("Wrong number of items",
|
||||
FILES_PER_DIR, statuses.size());
|
||||
assertEquals(
|
||||
FILES_PER_DIR, statuses.size(), "Wrong number of items");
|
||||
for (int i = 0; i < FILES_PER_DIR; i++) {
|
||||
FileStatus stat = statuses.get(i);
|
||||
assertFileEquals(0, 0, i, stat.getPath());
|
||||
@ -256,9 +254,9 @@ public void listFilesRelative() throws Exception {
|
||||
|
||||
@Test
|
||||
public void testDFSHasCapability() throws Throwable {
|
||||
assertTrue("FS does not declare PathCapability support",
|
||||
dfs.hasPathCapability(new Path("/"),
|
||||
CommonPathCapabilities.FS_EXPERIMENTAL_BATCH_LISTING));
|
||||
assertTrue(
|
||||
dfs.hasPathCapability(new Path("/"),
|
||||
CommonPathCapabilities.FS_EXPERIMENTAL_BATCH_LISTING), "FS does not declare PathCapability support");
|
||||
}
|
||||
|
||||
private void listFilesInternal(int numFiles) throws Exception {
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
@ -31,7 +31,7 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestBlockMissingException {
|
||||
final static Logger LOG =
|
||||
@ -116,6 +116,6 @@ private void validateFile(FileSystem fileSys, Path name)
|
||||
gotException = true;
|
||||
}
|
||||
stm.close();
|
||||
assertTrue("Expected BlockMissingException ", gotException);
|
||||
assertTrue(gotException, "Expected BlockMissingException ");
|
||||
}
|
||||
}
|
||||
|
@ -18,6 +18,7 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.hdfs.protocol.HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.FileNotFoundException;
|
||||
@ -47,9 +48,8 @@
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
import org.apache.hadoop.util.Lists;
|
||||
import org.apache.hadoop.util.Sets;
|
||||
import org.junit.Assert;
|
||||
import static org.junit.Assert.fail;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
|
||||
/** Test {@link BlockStoragePolicy} */
|
||||
public class TestBlockStoragePolicy {
|
||||
@ -160,10 +160,10 @@ public void testDefaultPolicies() {
|
||||
final BlockStoragePolicy policy = POLICY_SUITE.getPolicy(i);
|
||||
if (policy != null) {
|
||||
final String s = policy.toString();
|
||||
Assert.assertEquals(expectedPolicyStrings.get(i), s);
|
||||
Assertions.assertEquals(expectedPolicyStrings.get(i), s);
|
||||
}
|
||||
}
|
||||
Assert.assertEquals(POLICY_SUITE.getPolicy(HOT), POLICY_SUITE.getDefaultPolicy());
|
||||
Assertions.assertEquals(POLICY_SUITE.getPolicy(HOT), POLICY_SUITE.getDefaultPolicy());
|
||||
|
||||
// check Cold policy
|
||||
final BlockStoragePolicy cold = POLICY_SUITE.getPolicy(COLD);
|
||||
@ -260,11 +260,11 @@ static List<StorageType> asList(int nDisk, int nArchive) {
|
||||
|
||||
static void assertStorageType(List<StorageType> computed, short replication,
|
||||
StorageType... answers) {
|
||||
Assert.assertEquals(replication, computed.size());
|
||||
Assertions.assertEquals(replication, computed.size());
|
||||
final StorageType last = answers[answers.length - 1];
|
||||
for(int i = 0; i < computed.size(); i++) {
|
||||
final StorageType expected = i < answers.length? answers[i]: last;
|
||||
Assert.assertEquals(expected, computed.get(i));
|
||||
Assertions.assertEquals(expected, computed.get(i));
|
||||
}
|
||||
}
|
||||
|
||||
@ -272,27 +272,27 @@ static void assertCreationFallback(BlockStoragePolicy policy,
|
||||
StorageType noneExpected, StorageType archiveExpected,
|
||||
StorageType diskExpected, StorageType ssdExpected,
|
||||
StorageType disk_archiveExpected, StorageType nvdimmExpected) {
|
||||
Assert.assertEquals(noneExpected, policy.getCreationFallback(none));
|
||||
Assert.assertEquals(archiveExpected, policy.getCreationFallback(archive));
|
||||
Assert.assertEquals(diskExpected, policy.getCreationFallback(disk));
|
||||
Assert.assertEquals(ssdExpected, policy.getCreationFallback(ssd));
|
||||
Assert.assertEquals(nvdimmExpected, policy.getCreationFallback(nvdimm));
|
||||
Assert.assertEquals(disk_archiveExpected,
|
||||
Assertions.assertEquals(noneExpected, policy.getCreationFallback(none));
|
||||
Assertions.assertEquals(archiveExpected, policy.getCreationFallback(archive));
|
||||
Assertions.assertEquals(diskExpected, policy.getCreationFallback(disk));
|
||||
Assertions.assertEquals(ssdExpected, policy.getCreationFallback(ssd));
|
||||
Assertions.assertEquals(nvdimmExpected, policy.getCreationFallback(nvdimm));
|
||||
Assertions.assertEquals(disk_archiveExpected,
|
||||
policy.getCreationFallback(disk_archive));
|
||||
Assert.assertEquals(null, policy.getCreationFallback(all));
|
||||
Assertions.assertEquals(null, policy.getCreationFallback(all));
|
||||
}
|
||||
|
||||
static void assertReplicationFallback(BlockStoragePolicy policy,
|
||||
StorageType noneExpected, StorageType archiveExpected,
|
||||
StorageType diskExpected, StorageType ssdExpected,
|
||||
StorageType nvdimmExpected) {
|
||||
Assert.assertEquals(noneExpected, policy.getReplicationFallback(none));
|
||||
Assert
|
||||
Assertions.assertEquals(noneExpected, policy.getReplicationFallback(none));
|
||||
Assertions
|
||||
.assertEquals(archiveExpected, policy.getReplicationFallback(archive));
|
||||
Assert.assertEquals(diskExpected, policy.getReplicationFallback(disk));
|
||||
Assert.assertEquals(ssdExpected, policy.getReplicationFallback(ssd));
|
||||
Assert.assertEquals(nvdimmExpected, policy.getReplicationFallback(nvdimm));
|
||||
Assert.assertEquals(null, policy.getReplicationFallback(all));
|
||||
Assertions.assertEquals(diskExpected, policy.getReplicationFallback(disk));
|
||||
Assertions.assertEquals(ssdExpected, policy.getReplicationFallback(ssd));
|
||||
Assertions.assertEquals(nvdimmExpected, policy.getReplicationFallback(nvdimm));
|
||||
Assertions.assertEquals(null, policy.getReplicationFallback(all));
|
||||
}
|
||||
|
||||
private static interface CheckChooseStorageTypes {
|
||||
@ -879,7 +879,7 @@ static void assertStorageTypes(List<StorageType> computed, StorageType... expect
|
||||
static void assertStorageTypes(StorageType[] computed, StorageType... expected) {
|
||||
Arrays.sort(expected);
|
||||
Arrays.sort(computed);
|
||||
Assert.assertArrayEquals(expected, computed);
|
||||
Assertions.assertArrayEquals(expected, computed);
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -924,9 +924,9 @@ static void checkChooseExcess(BlockStoragePolicy p, short replication,
|
||||
}
|
||||
|
||||
private void checkDirectoryListing(HdfsFileStatus[] stats, byte... policies) {
|
||||
Assert.assertEquals(stats.length, policies.length);
|
||||
Assertions.assertEquals(stats.length, policies.length);
|
||||
for (int i = 0; i < stats.length; i++) {
|
||||
Assert.assertEquals(stats[i].getStoragePolicy(), policies[i]);
|
||||
Assertions.assertEquals(stats[i].getStoragePolicy(), policies[i]);
|
||||
}
|
||||
}
|
||||
|
||||
@ -949,7 +949,7 @@ public void testSetStoragePolicy() throws Exception {
|
||||
final String invalidPolicyName = "INVALID-POLICY";
|
||||
try {
|
||||
fs.setStoragePolicy(fooFile, invalidPolicyName);
|
||||
Assert.fail("Should throw a HadoopIllegalArgumentException");
|
||||
Assertions.fail("Should throw a HadoopIllegalArgumentException");
|
||||
} catch (RemoteException e) {
|
||||
GenericTestUtils.assertExceptionContains(invalidPolicyName, e);
|
||||
}
|
||||
@ -967,14 +967,14 @@ public void testSetStoragePolicy() throws Exception {
|
||||
final Path invalidPath = new Path("/invalidPath");
|
||||
try {
|
||||
fs.setStoragePolicy(invalidPath, HdfsConstants.WARM_STORAGE_POLICY_NAME);
|
||||
Assert.fail("Should throw a FileNotFoundException");
|
||||
Assertions.fail("Should throw a FileNotFoundException");
|
||||
} catch (FileNotFoundException e) {
|
||||
GenericTestUtils.assertExceptionContains(invalidPath.toString(), e);
|
||||
}
|
||||
|
||||
try {
|
||||
fs.getStoragePolicy(invalidPath);
|
||||
Assert.fail("Should throw a FileNotFoundException");
|
||||
Assertions.fail("Should throw a FileNotFoundException");
|
||||
} catch (FileNotFoundException e) {
|
||||
GenericTestUtils.assertExceptionContains(invalidPath.toString(), e);
|
||||
}
|
||||
@ -982,15 +982,15 @@ public void testSetStoragePolicy() throws Exception {
|
||||
fs.setStoragePolicy(fooFile, HdfsConstants.COLD_STORAGE_POLICY_NAME);
|
||||
fs.setStoragePolicy(barDir, HdfsConstants.WARM_STORAGE_POLICY_NAME);
|
||||
fs.setStoragePolicy(barFile2, HdfsConstants.HOT_STORAGE_POLICY_NAME);
|
||||
Assert.assertEquals("File storage policy should be COLD",
|
||||
HdfsConstants.COLD_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(fooFile).getName());
|
||||
Assert.assertEquals("File storage policy should be WARM",
|
||||
HdfsConstants.WARM_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(barDir).getName());
|
||||
Assert.assertEquals("File storage policy should be HOT",
|
||||
HdfsConstants.HOT_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(barFile2).getName());
|
||||
Assertions.assertEquals(
|
||||
HdfsConstants.COLD_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(fooFile).getName(), "File storage policy should be COLD");
|
||||
Assertions.assertEquals(
|
||||
HdfsConstants.WARM_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(barDir).getName(), "File storage policy should be WARM");
|
||||
Assertions.assertEquals(
|
||||
HdfsConstants.HOT_STORAGE_POLICY_NAME,
|
||||
fs.getStoragePolicy(barFile2).getName(), "File storage policy should be HOT");
|
||||
|
||||
dirList = fs.getClient().listPaths(dir.toString(),
|
||||
HdfsFileStatus.EMPTY_NAME).getPartialListing();
|
||||
@ -1040,8 +1040,8 @@ public void testGetStoragePolicy() throws Exception {
|
||||
HdfsConstants.COLD_STORAGE_POLICY_NAME);
|
||||
String policyName = client.getStoragePolicy("/testGetStoragePolicy/foo")
|
||||
.getName();
|
||||
Assert.assertEquals("File storage policy should be COLD",
|
||||
HdfsConstants.COLD_STORAGE_POLICY_NAME, policyName);
|
||||
Assertions.assertEquals(
|
||||
HdfsConstants.COLD_STORAGE_POLICY_NAME, policyName, "File storage policy should be COLD");
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
@ -1140,14 +1140,14 @@ private void checkLocatedBlocks(HdfsLocatedFileStatus status, int blockNum,
|
||||
List<StorageType> typeList = Lists.newArrayList();
|
||||
Collections.addAll(typeList, types);
|
||||
LocatedBlocks lbs = status.getLocatedBlocks();
|
||||
Assert.assertEquals(blockNum, lbs.getLocatedBlocks().size());
|
||||
Assertions.assertEquals(blockNum, lbs.getLocatedBlocks().size());
|
||||
for (LocatedBlock lb : lbs.getLocatedBlocks()) {
|
||||
Assert.assertEquals(replicaNum, lb.getStorageTypes().length);
|
||||
Assertions.assertEquals(replicaNum, lb.getStorageTypes().length);
|
||||
for (StorageType type : lb.getStorageTypes()) {
|
||||
Assert.assertTrue(typeList.remove(type));
|
||||
Assertions.assertTrue(typeList.remove(type));
|
||||
}
|
||||
}
|
||||
Assert.assertTrue(typeList.isEmpty());
|
||||
Assertions.assertTrue(typeList.isEmpty());
|
||||
}
|
||||
|
||||
private void testChangeFileRep(String policyName, byte policyId,
|
||||
@ -1285,12 +1285,12 @@ public void testChooseTargetWithTopology() throws Exception {
|
||||
dataNodes[0], Collections.<DatanodeStorageInfo>emptyList(), false,
|
||||
new HashSet<Node>(), 0, policy1, null);
|
||||
System.out.println(Arrays.asList(targets));
|
||||
Assert.assertEquals(3, targets.length);
|
||||
Assertions.assertEquals(3, targets.length);
|
||||
targets = replicator.chooseTarget("/foo", 3,
|
||||
dataNodes[0], Collections.<DatanodeStorageInfo>emptyList(), false,
|
||||
new HashSet<Node>(), 0, policy2, null);
|
||||
System.out.println(Arrays.asList(targets));
|
||||
Assert.assertEquals(3, targets.length);
|
||||
Assertions.assertEquals(3, targets.length);
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -1332,9 +1332,9 @@ public void testChooseSsdOverDisk() throws Exception {
|
||||
dataNodes[0], Collections.<DatanodeStorageInfo>emptyList(), false,
|
||||
new HashSet<Node>(), 0, policy, null);
|
||||
System.out.println(policy.getName() + ": " + Arrays.asList(targets));
|
||||
Assert.assertEquals(2, targets.length);
|
||||
Assert.assertEquals(StorageType.SSD, targets[0].getStorageType());
|
||||
Assert.assertEquals(StorageType.DISK, targets[1].getStorageType());
|
||||
Assertions.assertEquals(2, targets.length);
|
||||
Assertions.assertEquals(StorageType.SSD, targets[0].getStorageType());
|
||||
Assertions.assertEquals(StorageType.DISK, targets[1].getStorageType());
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -1360,17 +1360,17 @@ public void testGetFileStoragePolicyAfterRestartNN() throws Exception {
|
||||
// 4. Set Dir policy
|
||||
fs.setStoragePolicy(dir, "HOT");
|
||||
HdfsFileStatus status = fs.getClient().getFileInfo(file);
|
||||
// 5. get file policy, it should be parent policy.
|
||||
Assert
|
||||
.assertTrue("File storage policy should be HOT",
|
||||
status.getStoragePolicy() == HOT);
|
||||
// 5. get file policy, it should be parent policy.
|
||||
Assertions
|
||||
.assertTrue(
|
||||
status.getStoragePolicy() == HOT, "File storage policy should be HOT");
|
||||
// 6. restart NameNode for reloading edits logs.
|
||||
cluster.restartNameNode(true);
|
||||
// 7. get file policy, it should be parent policy.
|
||||
status = fs.getClient().getFileInfo(file);
|
||||
Assert
|
||||
.assertTrue("File storage policy should be HOT",
|
||||
status.getStoragePolicy() == HOT);
|
||||
Assertions
|
||||
.assertTrue(
|
||||
status.getStoragePolicy() == HOT, "File storage policy should be HOT");
|
||||
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
@ -1408,8 +1408,8 @@ public void testGetAllStoragePoliciesFromFs() throws IOException {
|
||||
}
|
||||
|
||||
// Ensure that we got the same set of policies in both cases.
|
||||
Assert.assertTrue(Sets.difference(policyNamesSet1, policyNamesSet2).isEmpty());
|
||||
Assert.assertTrue(Sets.difference(policyNamesSet2, policyNamesSet1).isEmpty());
|
||||
Assertions.assertTrue(Sets.difference(policyNamesSet1, policyNamesSet2).isEmpty());
|
||||
Assertions.assertTrue(Sets.difference(policyNamesSet2, policyNamesSet1).isEmpty());
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
@ -1428,21 +1428,21 @@ public void testStorageType() {
|
||||
|
||||
{
|
||||
final Iterator<StorageType> i = map.keySet().iterator();
|
||||
Assert.assertEquals(StorageType.RAM_DISK, i.next());
|
||||
Assert.assertEquals(StorageType.SSD, i.next());
|
||||
Assert.assertEquals(StorageType.DISK, i.next());
|
||||
Assert.assertEquals(StorageType.ARCHIVE, i.next());
|
||||
Assert.assertEquals(StorageType.NVDIMM, i.next());
|
||||
Assertions.assertEquals(StorageType.RAM_DISK, i.next());
|
||||
Assertions.assertEquals(StorageType.SSD, i.next());
|
||||
Assertions.assertEquals(StorageType.DISK, i.next());
|
||||
Assertions.assertEquals(StorageType.ARCHIVE, i.next());
|
||||
Assertions.assertEquals(StorageType.NVDIMM, i.next());
|
||||
}
|
||||
|
||||
{
|
||||
final Iterator<Map.Entry<StorageType, Integer>> i
|
||||
= map.entrySet().iterator();
|
||||
Assert.assertEquals(StorageType.RAM_DISK, i.next().getKey());
|
||||
Assert.assertEquals(StorageType.SSD, i.next().getKey());
|
||||
Assert.assertEquals(StorageType.DISK, i.next().getKey());
|
||||
Assert.assertEquals(StorageType.ARCHIVE, i.next().getKey());
|
||||
Assert.assertEquals(StorageType.NVDIMM, i.next().getKey());
|
||||
Assertions.assertEquals(StorageType.RAM_DISK, i.next().getKey());
|
||||
Assertions.assertEquals(StorageType.SSD, i.next().getKey());
|
||||
Assertions.assertEquals(StorageType.DISK, i.next().getKey());
|
||||
Assertions.assertEquals(StorageType.ARCHIVE, i.next().getKey());
|
||||
Assertions.assertEquals(StorageType.NVDIMM, i.next().getKey());
|
||||
}
|
||||
}
|
||||
|
||||
@ -1600,7 +1600,7 @@ private void testStorageIDCheckAccessResult(String[] requested,
|
||||
public void testCreateDefaultPoliciesFromConf() {
|
||||
BlockStoragePolicySuite suite =
|
||||
BlockStoragePolicySuite.createDefaultSuite();
|
||||
Assert.assertEquals(HdfsConstants.StoragePolicy.HOT.value(),
|
||||
Assertions.assertEquals(HdfsConstants.StoragePolicy.HOT.value(),
|
||||
suite.getDefaultPolicy().getId());
|
||||
|
||||
Configuration newConf = new Configuration();
|
||||
@ -1608,7 +1608,7 @@ public void testCreateDefaultPoliciesFromConf() {
|
||||
HdfsConstants.StoragePolicy.ONE_SSD);
|
||||
BlockStoragePolicySuite suiteConf =
|
||||
BlockStoragePolicySuite.createDefaultSuite(newConf);
|
||||
Assert.assertEquals(HdfsConstants.StoragePolicy.ONE_SSD.value(),
|
||||
Assertions.assertEquals(HdfsConstants.StoragePolicy.ONE_SSD.value(),
|
||||
suiteConf.getDefaultPolicy().getId());
|
||||
}
|
||||
|
||||
@ -1627,7 +1627,7 @@ public void testCreateFileWithConfiguredDefaultPolicies()
|
||||
DFSTestUtil.createFile(newfs, fooFile, 0, REPLICATION, 0L);
|
||||
|
||||
String policy = newfs.getStoragePolicy(fooFile).getName();
|
||||
Assert.assertEquals(HdfsConstants.StoragePolicy.WARM.name(), policy);
|
||||
Assertions.assertEquals(HdfsConstants.StoragePolicy.WARM.name(), policy);
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
|
@ -36,14 +36,15 @@
|
||||
import org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferTestCase;
|
||||
import org.apache.hadoop.io.EnumSetWritable;
|
||||
import org.apache.hadoop.security.TestPermission;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.junit.runners.Parameterized;
|
||||
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.*;
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
|
||||
/**
|
||||
@ -77,7 +78,7 @@ public TestBlockTokenWrappingQOP(String configKey, String qopValue) {
|
||||
this.qopValue = qopValue;
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setup() throws Exception {
|
||||
conf = createSecureConfig(this.configKey);
|
||||
conf.set(DFS_NAMENODE_RPC_ADDRESS_AUXILIARY_KEY, "12000");
|
||||
@ -109,7 +110,7 @@ public void setup() throws Exception {
|
||||
dfs = (DistributedFileSystem) FileSystem.get(uriAuxiliary, conf);
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
@ -35,8 +35,8 @@
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils;
|
||||
import org.apache.hadoop.hdfs.server.namenode.NameNodeAdapter;
|
||||
import org.junit.After;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This class tests DatanodeDescriptor.getBlocksScheduled() at the
|
||||
@ -47,7 +47,7 @@ public class TestBlocksScheduledCounter {
|
||||
MiniDFSCluster cluster = null;
|
||||
FileSystem fs = null;
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws IOException {
|
||||
if (fs != null) {
|
||||
fs.close();
|
||||
@ -104,8 +104,8 @@ public void testScheduledBlocksCounterShouldDecrementOnAbandonBlock()
|
||||
ArrayList<DatanodeDescriptor> dnList = new ArrayList<DatanodeDescriptor>();
|
||||
datanodeManager.fetchDatanodes(dnList, dnList, false);
|
||||
for (DatanodeDescriptor descriptor : dnList) {
|
||||
assertEquals("Blocks scheduled should be 0 for " + descriptor.getName(),
|
||||
0, descriptor.getBlocksScheduled());
|
||||
assertEquals(
|
||||
0, descriptor.getBlocksScheduled(), "Blocks scheduled should be 0 for " + descriptor.getName());
|
||||
}
|
||||
|
||||
cluster.getDataNodes().get(0).shutdown();
|
||||
@ -120,21 +120,21 @@ public void testScheduledBlocksCounterShouldDecrementOnAbandonBlock()
|
||||
|
||||
DatanodeDescriptor abandonedDn = datanodeManager.getDatanode(cluster
|
||||
.getDataNodes().get(0).getDatanodeId());
|
||||
assertEquals("for the abandoned dn scheduled counts should be 0", 0,
|
||||
abandonedDn.getBlocksScheduled());
|
||||
assertEquals(0,
|
||||
abandonedDn.getBlocksScheduled(), "for the abandoned dn scheduled counts should be 0");
|
||||
|
||||
for (DatanodeDescriptor descriptor : dnList) {
|
||||
if (descriptor.equals(abandonedDn)) {
|
||||
continue;
|
||||
}
|
||||
assertEquals("Blocks scheduled should be 1 for " + descriptor.getName(),
|
||||
1, descriptor.getBlocksScheduled());
|
||||
assertEquals(
|
||||
1, descriptor.getBlocksScheduled(), "Blocks scheduled should be 1 for " + descriptor.getName());
|
||||
}
|
||||
// close the file and the counter should go to zero.
|
||||
out.close();
|
||||
for (DatanodeDescriptor descriptor : dnList) {
|
||||
assertEquals("Blocks scheduled should be 0 for " + descriptor.getName(),
|
||||
0, descriptor.getBlocksScheduled());
|
||||
assertEquals(
|
||||
0, descriptor.getBlocksScheduled(), "Blocks scheduled should be 0 for " + descriptor.getName());
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -28,14 +28,11 @@
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
/**
|
||||
* This class tests the DFS positional read functionality on a single node
|
||||
@ -55,7 +52,7 @@ public class TestByteBufferPread {
|
||||
private static final int BLOCK_SIZE = 4096;
|
||||
private static final int FILE_SIZE = 12 * BLOCK_SIZE;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setup() throws IOException {
|
||||
// Setup the cluster with a small block size so we can create small files
|
||||
// that span multiple blocks
|
||||
@ -278,7 +275,7 @@ private void testPreadFullyWithByteBuffer(ByteBuffer buffer)
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void shutdown() throws IOException {
|
||||
try {
|
||||
fs.delete(testFile, false);
|
||||
|
@ -17,9 +17,8 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
@ -31,6 +30,8 @@
|
||||
|
||||
import java.util.function.Supplier;
|
||||
|
||||
import static org.junit.jupiter.api.Assertions.assertFalse;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FSDataInputStream;
|
||||
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||
@ -54,8 +55,8 @@
|
||||
import org.apache.hadoop.hdfs.tools.DFSAdmin;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.mockito.Mockito;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
@ -84,9 +85,9 @@ public class TestClientProtocolForPipelineRecovery {
|
||||
// test getNewStampAndToken on a finalized block
|
||||
try {
|
||||
namenode.updateBlockForPipeline(firstBlock, "");
|
||||
Assert.fail("Can not get a new GS from a finalized block");
|
||||
Assertions.fail("Can not get a new GS from a finalized block");
|
||||
} catch (IOException e) {
|
||||
Assert.assertTrue(e.getMessage().contains(
|
||||
Assertions.assertTrue(e.getMessage().contains(
|
||||
"not " + BlockUCState.UNDER_CONSTRUCTION));
|
||||
}
|
||||
|
||||
@ -96,9 +97,9 @@ public class TestClientProtocolForPipelineRecovery {
|
||||
ExtendedBlock newBlock = new ExtendedBlock(firstBlock.getBlockPoolId(),
|
||||
newBlockId, 0, firstBlock.getGenerationStamp());
|
||||
namenode.updateBlockForPipeline(newBlock, "");
|
||||
Assert.fail("Cannot get a new GS from a non-existent block");
|
||||
Assertions.fail("Cannot get a new GS from a non-existent block");
|
||||
} catch (IOException e) {
|
||||
Assert.assertTrue(e.getMessage().contains("does not exist"));
|
||||
Assertions.assertTrue(e.getMessage().contains("does not exist"));
|
||||
}
|
||||
|
||||
|
||||
@ -122,17 +123,17 @@ public class TestClientProtocolForPipelineRecovery {
|
||||
DFSClient dfs = ((DistributedFileSystem)fileSys).dfs;
|
||||
try {
|
||||
namenode.updateBlockForPipeline(firstBlock, "test" + dfs.clientName);
|
||||
Assert.fail("Cannot get a new GS for a non lease holder");
|
||||
Assertions.fail("Cannot get a new GS for a non lease holder");
|
||||
} catch (LeaseExpiredException e) {
|
||||
Assert.assertTrue(e.getMessage().startsWith("Lease mismatch"));
|
||||
Assertions.assertTrue(e.getMessage().startsWith("Lease mismatch"));
|
||||
}
|
||||
|
||||
// test null lease holder
|
||||
try {
|
||||
namenode.updateBlockForPipeline(firstBlock, null);
|
||||
Assert.fail("Cannot get a new GS for a null lease holder");
|
||||
Assertions.fail("Cannot get a new GS for a null lease holder");
|
||||
} catch (LeaseExpiredException e) {
|
||||
Assert.assertTrue(e.getMessage().startsWith("Lease mismatch"));
|
||||
Assertions.assertTrue(e.getMessage().startsWith("Lease mismatch"));
|
||||
}
|
||||
|
||||
// test getNewStampAndToken on a rbw block
|
||||
@ -177,7 +178,7 @@ public void testPipelineRecoveryForLastBlock() throws IOException {
|
||||
// Test will fail with BlockMissingException if NN does not update the
|
||||
// replica state based on the latest report.
|
||||
} catch (org.apache.hadoop.hdfs.BlockMissingException bme) {
|
||||
Assert.fail("Block is missing because the file was closed with"
|
||||
Assertions.fail("Block is missing because the file was closed with"
|
||||
+ " corrupt replicas.");
|
||||
}
|
||||
} finally {
|
||||
@ -239,7 +240,7 @@ public boolean dropHeartbeatPacket() {
|
||||
contains = true;
|
||||
}
|
||||
}
|
||||
Assert.assertTrue(contains);
|
||||
Assertions.assertTrue(contains);
|
||||
} finally {
|
||||
DataNodeFaultInjector.set(oldDnInjector);
|
||||
if (cluster != null) {
|
||||
@ -322,7 +323,7 @@ public void testPipelineRecoveryOnOOB() throws Exception {
|
||||
final String dnAddr = dn.getDatanodeId().getIpcAddr(false);
|
||||
// issue shutdown to the datanode.
|
||||
final String[] args1 = {"-shutdownDatanode", dnAddr, "upgrade" };
|
||||
Assert.assertEquals(0, dfsadmin.run(args1));
|
||||
Assertions.assertEquals(0, dfsadmin.run(args1));
|
||||
// Wait long enough to receive an OOB ack before closing the file.
|
||||
GenericTestUtils.waitForThreadTermination(
|
||||
"Async datanode shutdown thread", 100, 10000);
|
||||
@ -358,23 +359,23 @@ public void testEvictWriter() throws Exception {
|
||||
// get nodes in the pipeline
|
||||
DFSOutputStream dfsOut = (DFSOutputStream)out.getWrappedStream();
|
||||
DatanodeInfo[] nodes = dfsOut.getPipeline();
|
||||
Assert.assertEquals(2, nodes.length);
|
||||
Assertions.assertEquals(2, nodes.length);
|
||||
String dnAddr = nodes[1].getIpcAddr(false);
|
||||
|
||||
// evict the writer from the second datanode and wait until
|
||||
// the pipeline is rebuilt.
|
||||
DFSAdmin dfsadmin = new DFSAdmin(conf);
|
||||
final String[] args1 = {"-evictWriters", dnAddr };
|
||||
Assert.assertEquals(0, dfsadmin.run(args1));
|
||||
Assertions.assertEquals(0, dfsadmin.run(args1));
|
||||
out.write(0x31);
|
||||
out.hflush();
|
||||
|
||||
// get the new pipline and check the node is not in there.
|
||||
nodes = dfsOut.getPipeline();
|
||||
try {
|
||||
Assert.assertTrue(nodes.length > 0 );
|
||||
Assertions.assertTrue(nodes.length > 0 );
|
||||
for (int i = 0; i < nodes.length; i++) {
|
||||
Assert.assertFalse(dnAddr.equals(nodes[i].getIpcAddr(false)));
|
||||
Assertions.assertFalse(dnAddr.equals(nodes[i].getIpcAddr(false)));
|
||||
}
|
||||
} finally {
|
||||
out.close();
|
||||
@ -410,7 +411,7 @@ public void testPipelineRecoveryOnRestartFailure() throws Exception {
|
||||
final String dnAddr1 = dn.getDatanodeId().getIpcAddr(false);
|
||||
// issue shutdown to the datanode.
|
||||
final String[] args1 = {"-shutdownDatanode", dnAddr1, "upgrade" };
|
||||
Assert.assertEquals(0, dfsadmin.run(args1));
|
||||
Assertions.assertEquals(0, dfsadmin.run(args1));
|
||||
GenericTestUtils.waitForThreadTermination(
|
||||
"Async datanode shutdown thread", 100, 10000);
|
||||
// This should succeed without restarting the node. The restart will
|
||||
@ -427,7 +428,7 @@ public void testPipelineRecoveryOnRestartFailure() throws Exception {
|
||||
final String dnAddr2 = dn.getDatanodeId().getIpcAddr(false);
|
||||
// issue shutdown to the datanode.
|
||||
final String[] args2 = {"-shutdownDatanode", dnAddr2, "upgrade" };
|
||||
Assert.assertEquals(0, dfsadmin.run(args2));
|
||||
Assertions.assertEquals(0, dfsadmin.run(args2));
|
||||
GenericTestUtils.waitForThreadTermination(
|
||||
"Async datanode shutdown thread", 100, 10000);
|
||||
try {
|
||||
@ -480,8 +481,8 @@ public Boolean get() {
|
||||
return out.getBlock().getGenerationStamp() > oldGs;
|
||||
}
|
||||
}, 100, 10000);
|
||||
Assert.assertEquals("The pipeline recovery count shouldn't increase",
|
||||
0, out.getStreamer().getPipelineRecoveryCount());
|
||||
Assertions.assertEquals(
|
||||
0, out.getStreamer().getPipelineRecoveryCount(), "The pipeline recovery count shouldn't increase");
|
||||
out.write(1);
|
||||
out.close();
|
||||
// Ensure that subsequent closes are idempotent and do not throw errors
|
||||
@ -539,7 +540,7 @@ public void run() {
|
||||
Thread.sleep(1000);
|
||||
DatanodeInfo[] pipeline = out.getPipeline();
|
||||
for (DatanodeInfo node : pipeline) {
|
||||
assertFalse("Write should be going on", failed.get());
|
||||
assertFalse(failed.get(), "Write should be going on");
|
||||
ArrayList<DataNode> dataNodes = cluster.getDataNodes();
|
||||
int indexToShutdown = 0;
|
||||
for (int i = 0; i < dataNodes.size(); i++) {
|
||||
@ -564,15 +565,15 @@ public Boolean get() {
|
||||
return out.getBlock().getGenerationStamp() > oldGs;
|
||||
}
|
||||
}, 100, 10000);
|
||||
Assert.assertEquals("The pipeline recovery count shouldn't increase", 0,
|
||||
out.getStreamer().getPipelineRecoveryCount());
|
||||
Assertions.assertEquals(0,
|
||||
out.getStreamer().getPipelineRecoveryCount(), "The pipeline recovery count shouldn't increase");
|
||||
}
|
||||
assertFalse("Write should be going on", failed.get());
|
||||
assertFalse(failed.get(), "Write should be going on");
|
||||
running.set(false);
|
||||
t.join();
|
||||
out.write("testagain".getBytes());
|
||||
assertTrue("There should be atleast 2 nodes in pipeline still", out
|
||||
.getPipeline().length >= 2);
|
||||
assertTrue(out
|
||||
.getPipeline().length >= 2, "There should be atleast 2 nodes in pipeline still");
|
||||
out.close();
|
||||
} finally {
|
||||
DFSClientFaultInjector.set(old);
|
||||
@ -723,7 +724,7 @@ public void failPipeline(ReplicaInPipeline replicaInfo,
|
||||
o.hflush();
|
||||
}
|
||||
|
||||
assertTrue("Expected a failure in the pipeline", failed.get());
|
||||
assertTrue(failed.get(), "Expected a failure in the pipeline");
|
||||
DatanodeInfo[] newNodes = dfsO.getStreamer().getNodes();
|
||||
o.close();
|
||||
// Trigger block report to NN
|
||||
|
@ -41,10 +41,10 @@
|
||||
import org.apache.hadoop.hdfs.tools.DFSck;
|
||||
import org.apache.hadoop.security.AccessControlException;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* Class is used to test client reporting corrupted block replica to name node.
|
||||
@ -67,7 +67,7 @@ public class TestClientReportBadBlock {
|
||||
|
||||
Random rand = new Random();
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void startUpCluster() throws IOException {
|
||||
// disable block scanner
|
||||
conf.setInt(DFSConfigKeys.DFS_DATANODE_SCAN_PERIOD_HOURS_KEY, -1);
|
||||
@ -80,7 +80,7 @@ public void startUpCluster() throws IOException {
|
||||
buffersize = conf.getInt(CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096);
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void shutDownCluster() throws IOException {
|
||||
if (dfs != null) {
|
||||
dfs.close();
|
||||
@ -211,7 +211,7 @@ private void createAFileWithCorruptedBlockReplicas(Path filePath, short repl,
|
||||
// Locate the file blocks by asking name node
|
||||
final LocatedBlocks locatedblocks = dfs.dfs.getNamenode()
|
||||
.getBlockLocations(filePath.toString(), 0L, BLOCK_SIZE);
|
||||
Assert.assertEquals(repl, locatedblocks.get(0).getLocations().length);
|
||||
Assertions.assertEquals(repl, locatedblocks.get(0).getLocations().length);
|
||||
// The file only has one block
|
||||
LocatedBlock lblock = locatedblocks.get(0);
|
||||
DatanodeInfo[] datanodeinfos = lblock.getLocations();
|
||||
@ -236,7 +236,7 @@ private void verifyFirstBlockCorrupted(Path filePath, boolean isCorrupted)
|
||||
final LocatedBlocks locatedBlocks = dfs.dfs.getNamenode()
|
||||
.getBlockLocations(filePath.toUri().getPath(), 0, Long.MAX_VALUE);
|
||||
final LocatedBlock firstLocatedBlock = locatedBlocks.get(0);
|
||||
Assert.assertEquals(isCorrupted, firstLocatedBlock.isCorrupt());
|
||||
Assertions.assertEquals(isCorrupted, firstLocatedBlock.isCorrupt());
|
||||
}
|
||||
|
||||
/**
|
||||
@ -250,7 +250,7 @@ private void verifyCorruptedBlockCount(Path filePath, int expectedReplicas)
|
||||
filePath.toUri().getPath(), 0, Long.MAX_VALUE);
|
||||
// we expect only the first block of the file is used for this test
|
||||
LocatedBlock firstLocatedBlock = lBlocks.get(0);
|
||||
Assert.assertEquals(expectedReplicas,
|
||||
Assertions.assertEquals(expectedReplicas,
|
||||
firstLocatedBlock.getLocations().length);
|
||||
}
|
||||
|
||||
@ -300,23 +300,23 @@ private static void verifyFsckHealth(String expected) throws Exception {
|
||||
// Make sure filesystem is in healthy state
|
||||
String outStr = runFsck(conf, 0, true, "/");
|
||||
LOG.info(outStr);
|
||||
Assert.assertTrue(outStr.contains(NamenodeFsck.HEALTHY_STATUS));
|
||||
Assertions.assertTrue(outStr.contains(NamenodeFsck.HEALTHY_STATUS));
|
||||
if (!expected.equals("")) {
|
||||
Assert.assertTrue(outStr.contains(expected));
|
||||
Assertions.assertTrue(outStr.contains(expected));
|
||||
}
|
||||
}
|
||||
|
||||
private static void verifyFsckBlockCorrupted() throws Exception {
|
||||
String outStr = runFsck(conf, 1, true, "/");
|
||||
LOG.info(outStr);
|
||||
Assert.assertTrue(outStr.contains(NamenodeFsck.CORRUPT_STATUS));
|
||||
Assertions.assertTrue(outStr.contains(NamenodeFsck.CORRUPT_STATUS));
|
||||
}
|
||||
|
||||
private static void testFsckListCorruptFilesBlocks(Path filePath, int errorCode) throws Exception{
|
||||
String outStr = runFsck(conf, errorCode, true, filePath.toString(), "-list-corruptfileblocks");
|
||||
LOG.info("fsck -list-corruptfileblocks out: " + outStr);
|
||||
if (errorCode != 0) {
|
||||
Assert.assertTrue(outStr.contains("CORRUPT blocks"));
|
||||
Assertions.assertTrue(outStr.contains("CORRUPT blocks"));
|
||||
}
|
||||
}
|
||||
|
||||
@ -326,7 +326,7 @@ static String runFsck(Configuration conf, int expectedErrCode,
|
||||
PrintStream out = new PrintStream(bStream, true);
|
||||
int errCode = ToolRunner.run(new DFSck(conf, out), path);
|
||||
if (checkErrorCode)
|
||||
Assert.assertEquals(expectedErrCode, errCode);
|
||||
Assertions.assertEquals(expectedErrCode, errCode);
|
||||
return bStream.toString();
|
||||
}
|
||||
}
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
@ -26,7 +26,7 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
public class TestClose {
|
||||
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.InetSocketAddress;
|
||||
@ -27,8 +27,8 @@
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
import org.apache.hadoop.hdfs.client.impl.BlockReaderTestUtil;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This class tests the client connection caching in a single node
|
||||
@ -52,7 +52,7 @@ private void pread(DFSInputStream in,
|
||||
int length,
|
||||
byte[] authenticData)
|
||||
throws IOException {
|
||||
Assert.assertTrue("Test buffer too small", buffer.length >= offset + length);
|
||||
Assertions.assertTrue(buffer.length >= offset + length, "Test buffer too small");
|
||||
|
||||
if (pos >= 0)
|
||||
in.seek(pos);
|
||||
@ -62,7 +62,7 @@ private void pread(DFSInputStream in,
|
||||
|
||||
while (length > 0) {
|
||||
int cnt = in.read(buffer, offset, length);
|
||||
Assert.assertTrue("Error in read", cnt > 0);
|
||||
Assertions.assertTrue(cnt > 0, "Error in read");
|
||||
offset += cnt;
|
||||
length -= cnt;
|
||||
}
|
||||
@ -71,9 +71,9 @@ private void pread(DFSInputStream in,
|
||||
for (int i = 0; i < length; ++i) {
|
||||
byte actual = buffer[i];
|
||||
byte expect = authenticData[(int)pos + i];
|
||||
assertEquals("Read data mismatch at file offset " + (pos + i) +
|
||||
". Expects " + expect + "; got " + actual,
|
||||
actual, expect);
|
||||
assertEquals(
|
||||
actual, expect, "Read data mismatch at file offset " + (pos + i) +
|
||||
". Expects " + expect + "; got " + actual);
|
||||
}
|
||||
}
|
||||
|
||||
@ -116,7 +116,7 @@ public void testReadFromOneDN() throws Exception {
|
||||
|
||||
in.close();
|
||||
client.close();
|
||||
Assert.assertEquals(1,
|
||||
Assertions.assertEquals(1,
|
||||
ClientContext.getFromConf(configuration).getPeerCache().size());
|
||||
}
|
||||
}
|
||||
|
@ -18,9 +18,7 @@
|
||||
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
@ -36,8 +34,8 @@
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||
import org.apache.hadoop.hdfs.server.datanode.ReplicaInfo;
|
||||
import org.apache.hadoop.io.IOUtils;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.mockito.Mockito;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
@ -75,7 +73,7 @@ public class TestCrcCorruption {
|
||||
|
||||
private DFSClientFaultInjector faultInjector;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() throws IOException {
|
||||
faultInjector = Mockito.mock(DFSClientFaultInjector.class);
|
||||
DFSClientFaultInjector.set(faultInjector);
|
||||
@ -174,7 +172,7 @@ private void thistest(Configuration conf, DFSTestUtil util) throws Exception {
|
||||
final String bpid = cluster.getNamesystem().getBlockPoolId();
|
||||
List<ReplicaInfo> replicas =
|
||||
dn.getFSDataset().getFinalizedBlocks(bpid);
|
||||
assertTrue("Replicas do not exist", !replicas.isEmpty());
|
||||
assertTrue(!replicas.isEmpty(), "Replicas do not exist");
|
||||
|
||||
for (int idx = 0; idx < replicas.size(); idx++) {
|
||||
ReplicaInfo replica = replicas.get(idx);
|
||||
@ -192,12 +190,12 @@ private void thistest(Configuration conf, DFSTestUtil util) throws Exception {
|
||||
}
|
||||
}
|
||||
|
||||
//
|
||||
// Only one replica is possibly corrupted. The other replica should still
|
||||
// be good. Verify.
|
||||
//
|
||||
assertTrue("Corrupted replicas not handled properly.",
|
||||
util.checkFiles(fs, "/srcdat"));
|
||||
//
|
||||
// Only one replica is possibly corrupted. The other replica should still
|
||||
// be good. Verify.
|
||||
//
|
||||
assertTrue(
|
||||
util.checkFiles(fs, "/srcdat"), "Corrupted replicas not handled properly.");
|
||||
LOG.info("All File still have a valid replica");
|
||||
|
||||
//
|
||||
@ -287,7 +285,7 @@ private void doTestEntirelyCorruptFile(int numDataNodes) throws Exception {
|
||||
|
||||
ExtendedBlock block = DFSTestUtil.getFirstBlock(fs, file);
|
||||
int blockFilesCorrupted = cluster.corruptBlockOnDataNodes(block);
|
||||
assertEquals("All replicas not corrupted", replFactor, blockFilesCorrupted);
|
||||
assertEquals(replFactor, blockFilesCorrupted, "All replicas not corrupted");
|
||||
|
||||
try {
|
||||
IOUtils.copyBytes(fs.open(file), new IOUtils.NullOutputStream(), conf,
|
||||
|
@ -27,8 +27,8 @@
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_ADDRESS_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_HTTP_ADDRESS_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_IPC_ADDRESS_KEY;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertNotNull;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
@ -37,7 +37,7 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster.DataNodeProperties;
|
||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
|
||||
public class TestDFSAddressConfig {
|
||||
@ -67,7 +67,7 @@ public void testDFSAddressConfig() throws IOException {
|
||||
*------------------------------------------------------------------------*/
|
||||
for (int i = 0; i < dns.size(); i++) {
|
||||
DataNodeProperties dnp = cluster.stopDataNode(i);
|
||||
assertNotNull("Should have been able to stop simulated datanode", dnp);
|
||||
assertNotNull(dnp, "Should have been able to stop simulated datanode");
|
||||
}
|
||||
|
||||
conf.unset(DFS_DATANODE_ADDRESS_KEY);
|
||||
@ -92,7 +92,7 @@ public void testDFSAddressConfig() throws IOException {
|
||||
*------------------------------------------------------------------------*/
|
||||
for (int i = 0; i < dns.size(); i++) {
|
||||
DataNodeProperties dnp = cluster.stopDataNode(i);
|
||||
assertNotNull("Should have been able to stop simulated datanode", dnp);
|
||||
assertNotNull(dnp, "Should have been able to stop simulated datanode");
|
||||
}
|
||||
|
||||
conf.set(DFS_DATANODE_ADDRESS_KEY, "0.0.0.0:0");
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.fail;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
@ -29,10 +29,10 @@
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster.DataNodeProperties;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||
import org.apache.hadoop.util.ThreadUtil;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
|
||||
|
||||
/**
|
||||
@ -44,13 +44,13 @@ public class TestDFSClientExcludedNodes {
|
||||
private MiniDFSCluster cluster;
|
||||
private Configuration conf;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUp() {
|
||||
cluster = null;
|
||||
conf = new HdfsConfiguration();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -122,8 +122,8 @@ public void testExcludedNodesForgiveness() throws IOException {
|
||||
|
||||
// Bring back the older DNs, since they are gonna be forgiven only
|
||||
// afterwards of this previous block write.
|
||||
Assert.assertEquals(true, cluster.restartDataNode(one, true));
|
||||
Assert.assertEquals(true, cluster.restartDataNode(two, true));
|
||||
Assertions.assertEquals(true, cluster.restartDataNode(one, true));
|
||||
Assertions.assertEquals(true, cluster.restartDataNode(two, true));
|
||||
cluster.waitActive();
|
||||
|
||||
// Sleep for 5s, to let the excluded nodes be expired
|
||||
|
@ -17,10 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.lang.reflect.Field;
|
||||
@ -55,10 +52,10 @@
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.apache.hadoop.util.StringUtils;
|
||||
import org.junit.After;
|
||||
import org.junit.Assume;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assumptions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.mockito.ArgumentMatcher;
|
||||
import org.mockito.Mockito;
|
||||
|
||||
@ -74,7 +71,7 @@ public class TestDFSClientFailover {
|
||||
private final Configuration conf = new Configuration();
|
||||
private MiniDFSCluster cluster;
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setUpCluster() throws IOException {
|
||||
cluster = new MiniDFSCluster.Builder(conf)
|
||||
.nnTopology(MiniDFSNNTopology.simpleHATopology())
|
||||
@ -83,7 +80,7 @@ public void setUpCluster() throws IOException {
|
||||
cluster.waitActive();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDownCluster() throws IOException {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
@ -91,7 +88,7 @@ public void tearDownCluster() throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void clearConfig() {
|
||||
SecurityUtil.setTokenServiceUseIp(true);
|
||||
}
|
||||
@ -217,9 +214,9 @@ public void testFailureWithMisconfiguredHaNNs() throws Exception {
|
||||
fail("Successfully got proxy provider for misconfigured FS");
|
||||
} catch (IOException ioe) {
|
||||
LOG.info("got expected exception", ioe);
|
||||
assertTrue("expected exception did not contain helpful message",
|
||||
StringUtils.stringifyException(ioe).contains(
|
||||
"Could not find any configured addresses for URI " + uri));
|
||||
assertTrue(
|
||||
StringUtils.stringifyException(ioe).contains(
|
||||
"Could not find any configured addresses for URI " + uri), "expected exception did not contain helpful message");
|
||||
}
|
||||
}
|
||||
|
||||
@ -233,7 +230,7 @@ private NameService spyOnNameService() {
|
||||
try {
|
||||
Field f = InetAddress.class.getDeclaredField("nameServices");
|
||||
f.setAccessible(true);
|
||||
Assume.assumeNotNull(f);
|
||||
Assumptions.assumeNotNull(f);
|
||||
@SuppressWarnings("unchecked")
|
||||
List<NameService> nsList = (List<NameService>) f.get(null);
|
||||
|
||||
@ -248,7 +245,7 @@ private NameService spyOnNameService() {
|
||||
LOG.info("Unable to spy on DNS. Skipping test.", t);
|
||||
// In case the JDK we're testing on doesn't work like Sun's, just
|
||||
// skip the test.
|
||||
Assume.assumeNoException(t);
|
||||
Assumptions.assumeNoException(t);
|
||||
throw new RuntimeException(t);
|
||||
}
|
||||
}
|
||||
@ -377,9 +374,9 @@ public void testWrappedFailoverProxyProvider() throws Exception {
|
||||
// not to use IP address for token service
|
||||
SecurityUtil.setTokenServiceUseIp(false);
|
||||
|
||||
// Logical URI should be used.
|
||||
assertTrue("Legacy proxy providers should use logical URI.",
|
||||
HAUtil.useLogicalUri(config, p.toUri()));
|
||||
// Logical URI should be used.
|
||||
assertTrue(
|
||||
HAUtil.useLogicalUri(config, p.toUri()), "Legacy proxy providers should use logical URI.");
|
||||
}
|
||||
|
||||
/**
|
||||
@ -394,8 +391,8 @@ public void testIPFailoverProxyProviderLogicalUri() throws Exception {
|
||||
nnUri.getHost(),
|
||||
IPFailoverProxyProvider.class.getName());
|
||||
|
||||
assertFalse("IPFailoverProxyProvider should not use logical URI.",
|
||||
HAUtil.useLogicalUri(config, nnUri));
|
||||
assertFalse(
|
||||
HAUtil.useLogicalUri(config, nnUri), "IPFailoverProxyProvider should not use logical URI.");
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -18,10 +18,7 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
import static org.mockito.ArgumentMatchers.any;
|
||||
import static org.mockito.ArgumentMatchers.anyBoolean;
|
||||
import static org.mockito.ArgumentMatchers.anyLong;
|
||||
@ -90,9 +87,9 @@
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.apache.hadoop.util.StringUtils;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.mockito.Mockito;
|
||||
import org.mockito.internal.stubbing.answers.ThrowsException;
|
||||
import org.mockito.invocation.InvocationOnMock;
|
||||
@ -160,7 +157,7 @@ private static void writeData(OutputStream out, int len) throws IOException {
|
||||
}
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setupConf(){
|
||||
conf = new HdfsConfiguration();
|
||||
}
|
||||
@ -285,8 +282,8 @@ public Object answer(InvocationOnMock invocation)
|
||||
try {
|
||||
os.close();
|
||||
} catch (Exception e) {
|
||||
assertTrue("Retries are not being stopped correctly: " + e.getMessage(),
|
||||
e.getMessage().equals(exceptionMsg));
|
||||
assertTrue(
|
||||
e.getMessage().equals(exceptionMsg), "Retries are not being stopped correctly: " + e.getMessage());
|
||||
}
|
||||
}
|
||||
|
||||
@ -632,7 +629,7 @@ public void testDFSClientRetriesOnBusyBlocks() throws IOException {
|
||||
timestamp = Time.now();
|
||||
pass = busyTest(xcievers, threads, fileLen, timeWin, retries);
|
||||
timestamp2 = Time.now();
|
||||
assertTrue("Something wrong! Test 2 got Exception with maxmum retries!", pass);
|
||||
assertTrue(pass, "Something wrong! Test 2 got Exception with maxmum retries!");
|
||||
LOG.info("Test 2 succeeded! Time spent: " + (timestamp2-timestamp)/1000.0 + " sec.");
|
||||
|
||||
//
|
||||
@ -657,7 +654,7 @@ public void testDFSClientRetriesOnBusyBlocks() throws IOException {
|
||||
timestamp = Time.now();
|
||||
pass = busyTest(xcievers, threads, fileLen, timeWin, retries);
|
||||
timestamp2 = Time.now();
|
||||
assertTrue("Something wrong! Test 4 got Exception with maxmum retries!", pass);
|
||||
assertTrue(pass, "Something wrong! Test 4 got Exception with maxmum retries!");
|
||||
LOG.info("Test 4 succeeded! Time spent: " + (timestamp2-timestamp)/1000.0 + " sec.");
|
||||
}
|
||||
|
||||
@ -692,10 +689,10 @@ private boolean busyTest(int xcievers, int threads, int fileLen, int timeWin, in
|
||||
bufferSize,
|
||||
replicationFactor,
|
||||
blockSize);
|
||||
|
||||
// verify that file exists in FS namespace
|
||||
assertTrue(file1 + " should be a file",
|
||||
fs.getFileStatus(file1).isFile());
|
||||
|
||||
// verify that file exists in FS namespace
|
||||
assertTrue(
|
||||
fs.getFileStatus(file1).isFile(), file1 + " should be a file");
|
||||
System.out.println("Path : \"" + file1 + "\"");
|
||||
LOG.info("Path : \"" + file1 + "\"");
|
||||
|
||||
@ -706,10 +703,10 @@ private boolean busyTest(int xcievers, int threads, int fileLen, int timeWin, in
|
||||
|
||||
// verify that file size has changed to the full size
|
||||
long len = fs.getFileStatus(file1).getLen();
|
||||
|
||||
assertTrue(file1 + " should be of size " + fileLen +
|
||||
" but found to be of size " + len,
|
||||
len == fileLen);
|
||||
|
||||
assertTrue(
|
||||
len == fileLen, file1 + " should be of size " + fileLen +
|
||||
" but found to be of size " + len);
|
||||
|
||||
// read back and check data integrigy
|
||||
byte[] read_buf = new byte[fileLen];
|
||||
@ -809,11 +806,11 @@ public void run() {
|
||||
in.close();
|
||||
fs.close();
|
||||
|
||||
assertTrue("hashed keys are not the same size",
|
||||
hash_sha.length == expected_sha.length);
|
||||
assertTrue(
|
||||
hash_sha.length == expected_sha.length, "hashed keys are not the same size");
|
||||
|
||||
assertTrue("hashed keys are not equal",
|
||||
Arrays.equals(hash_sha, expected_sha));
|
||||
assertTrue(
|
||||
Arrays.equals(hash_sha, expected_sha), "hashed keys are not equal");
|
||||
|
||||
counter.inc(); // count this thread as successful
|
||||
|
||||
@ -928,8 +925,8 @@ public void testRetryOnChecksumFailure() throws Exception {
|
||||
|
||||
ExtendedBlock block = DFSTestUtil.getFirstBlock(fs, path);
|
||||
int blockFilesCorrupted = cluster.corruptBlockOnDataNodes(block);
|
||||
assertEquals("All replicas not corrupted", REPL_FACTOR,
|
||||
blockFilesCorrupted);
|
||||
assertEquals(REPL_FACTOR,
|
||||
blockFilesCorrupted, "All replicas not corrupted");
|
||||
|
||||
InetSocketAddress nnAddr =
|
||||
new InetSocketAddress("localhost", cluster.getNameNodePort());
|
||||
@ -1107,13 +1104,13 @@ public void run() {
|
||||
final FSDataInputStream in = fs.open(file4);
|
||||
int count = 0;
|
||||
for(int r; (r = in.read()) != -1; count++) {
|
||||
Assert.assertEquals(String.format("count=%d", count),
|
||||
bytes[count % bytes.length], (byte)r);
|
||||
Assertions.assertEquals(
|
||||
bytes[count % bytes.length], (byte) r, String.format("count=%d", count));
|
||||
}
|
||||
if (!isWebHDFS) {
|
||||
Assert.assertEquals(5 * bytes.length, count);
|
||||
Assertions.assertEquals(5 * bytes.length, count);
|
||||
} else {
|
||||
Assert.assertEquals(2 * bytes.length, count);
|
||||
Assertions.assertEquals(2 * bytes.length, count);
|
||||
}
|
||||
in.close();
|
||||
}
|
||||
|
@ -20,9 +20,7 @@
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hdfs.protocol.DatanodeInfo.DatanodeInfoBuilder;
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
|
||||
import org.junit.Test;
|
||||
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.slf4j.event.Level;
|
||||
@ -31,7 +29,7 @@
|
||||
import java.net.Socket;
|
||||
|
||||
import static org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.DFS_CLIENT_SOCKET_SEND_BUFFER_SIZE_KEY;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.assertTrue;
|
||||
|
||||
public class TestDFSClientSocketSize {
|
||||
private static final Logger LOG = LoggerFactory.getLogger(
|
||||
@ -49,8 +47,8 @@ public void testDefaultSendBufferSize() throws IOException {
|
||||
final int sendBufferSize = getSendBufferSize(new Configuration());
|
||||
LOG.info("If not specified, the auto tuned send buffer size is: {}",
|
||||
sendBufferSize);
|
||||
assertTrue("Send buffer size should be non-negative value which is " +
|
||||
"determined by system (kernel).", sendBufferSize > 0);
|
||||
assertTrue(sendBufferSize > 0, "Send buffer size should be non-negative value which is " +
|
||||
"determined by system (kernel).");
|
||||
}
|
||||
|
||||
/**
|
||||
@ -69,8 +67,8 @@ public void testSpecifiedSendBufferSize() throws IOException {
|
||||
|
||||
LOG.info("Large buf size is {}, small is {}",
|
||||
sendBufferSize1, sendBufferSize2);
|
||||
assertTrue("Larger specified send buffer should have effect",
|
||||
sendBufferSize1 > sendBufferSize2);
|
||||
assertTrue(
|
||||
sendBufferSize1 > sendBufferSize2, "Larger specified send buffer should have effect");
|
||||
}
|
||||
|
||||
/**
|
||||
@ -83,8 +81,8 @@ public void testAutoTuningSendBufferSize() throws IOException {
|
||||
conf.setInt(DFS_CLIENT_SOCKET_SEND_BUFFER_SIZE_KEY, 0);
|
||||
final int sendBufferSize = getSendBufferSize(conf);
|
||||
LOG.info("The auto tuned send buffer size is: {}", sendBufferSize);
|
||||
assertTrue("Send buffer size should be non-negative value which is " +
|
||||
"determined by system (kernel).", sendBufferSize > 0);
|
||||
assertTrue(sendBufferSize > 0, "Send buffer size should be non-negative value which is " +
|
||||
"determined by system (kernel).");
|
||||
}
|
||||
|
||||
private int getSendBufferSize(Configuration conf) throws IOException {
|
||||
|
@ -18,8 +18,8 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType.DATA_NODE;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
import static org.junit.jupiter.api.Assertions.assertFalse;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.Collections;
|
||||
@ -33,8 +33,8 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSImageTestUtil;
|
||||
import org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceStorage;
|
||||
import org.apache.hadoop.hdfs.server.datanode.DataStorage;
|
||||
import org.junit.After;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This test ensures the appropriate response from the system when
|
||||
@ -184,7 +184,7 @@ public void testFinalize() throws Exception {
|
||||
} // end numDir loop
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void tearDown() throws Exception {
|
||||
LOG.info("Shutting down MiniDFSCluster");
|
||||
if (cluster != null) {
|
||||
|
@ -35,8 +35,8 @@
|
||||
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOpCodes;
|
||||
import org.apache.hadoop.hdfs.server.namenode.ha.HATestUtil;
|
||||
import org.apache.hadoop.util.ExitUtil;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.OutputStream;
|
||||
@ -60,8 +60,8 @@ public static EventBatch waitForNextEvents(DFSInotifyEventInputStream eis)
|
||||
}
|
||||
|
||||
private static long checkTxid(EventBatch batch, long prevTxid){
|
||||
Assert.assertTrue("Previous txid " + prevTxid + " was not less than " +
|
||||
"new txid " + batch.getTxid(), prevTxid < batch.getTxid());
|
||||
Assertions.assertTrue(prevTxid < batch.getTxid(), "Previous txid " + prevTxid + " was not less than " +
|
||||
"new txid " + batch.getTxid());
|
||||
return batch.getTxid();
|
||||
}
|
||||
|
||||
@ -73,7 +73,7 @@ private static long checkTxid(EventBatch batch, long prevTxid){
|
||||
*/
|
||||
@Test
|
||||
public void testOpcodeCount() {
|
||||
Assert.assertEquals(54, FSEditLogOpCodes.values().length);
|
||||
Assertions.assertEquals(54, FSEditLogOpCodes.values().length);
|
||||
}
|
||||
|
||||
|
||||
@ -146,287 +146,287 @@ public void testBasic() throws IOException, URISyntaxException,
|
||||
|
||||
// RenameOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
long txid = batch.getTxid();
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Event.RenameEvent re = (Event.RenameEvent) batch.getEvents()[0];
|
||||
Assert.assertEquals("/file4", re.getDstPath());
|
||||
Assert.assertEquals("/file", re.getSrcPath());
|
||||
Assert.assertTrue(re.getTimestamp() > 0);
|
||||
Assertions.assertEquals("/file4", re.getDstPath());
|
||||
Assertions.assertEquals("/file", re.getSrcPath());
|
||||
Assertions.assertTrue(re.getTimestamp() > 0);
|
||||
LOG.info(re.toString());
|
||||
Assert.assertTrue(re.toString().startsWith("RenameEvent [srcPath="));
|
||||
Assertions.assertTrue(re.toString().startsWith("RenameEvent [srcPath="));
|
||||
|
||||
long eventsBehind = eis.getTxidsBehindEstimate();
|
||||
|
||||
// RenameOldOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Event.RenameEvent re2 = (Event.RenameEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(re2.getDstPath().equals("/file2"));
|
||||
Assert.assertTrue(re2.getSrcPath().equals("/file4"));
|
||||
Assert.assertTrue(re2.getTimestamp() > 0);
|
||||
Assertions.assertTrue(re2.getDstPath().equals("/file2"));
|
||||
Assertions.assertTrue(re2.getSrcPath().equals("/file4"));
|
||||
Assertions.assertTrue(re2.getTimestamp() > 0);
|
||||
LOG.info(re2.toString());
|
||||
|
||||
// AddOp with overwrite
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Event.CreateEvent ce = (Event.CreateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assert.assertTrue(ce.getPath().equals("/file2"));
|
||||
Assert.assertTrue(ce.getCtime() > 0);
|
||||
Assert.assertTrue(ce.getReplication() > 0);
|
||||
Assert.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assert.assertTrue(ce.getOverwrite());
|
||||
Assert.assertEquals(BLOCK_SIZE, ce.getDefaultBlockSize());
|
||||
Assert.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assert.assertFalse(ce.isErasureCoded().get());
|
||||
Assertions.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assertions.assertTrue(ce.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(ce.getCtime() > 0);
|
||||
Assertions.assertTrue(ce.getReplication() > 0);
|
||||
Assertions.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assertions.assertTrue(ce.getOverwrite());
|
||||
Assertions.assertEquals(BLOCK_SIZE, ce.getDefaultBlockSize());
|
||||
Assertions.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assertions.assertFalse(ce.isErasureCoded().get());
|
||||
LOG.info(ce.toString());
|
||||
Assert.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
Assertions.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
|
||||
// CloseOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CLOSE);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CLOSE);
|
||||
Event.CloseEvent ce2 = (Event.CloseEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce2.getPath().equals("/file2"));
|
||||
Assert.assertTrue(ce2.getFileSize() > 0);
|
||||
Assert.assertTrue(ce2.getTimestamp() > 0);
|
||||
Assertions.assertTrue(ce2.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(ce2.getFileSize() > 0);
|
||||
Assertions.assertTrue(ce2.getTimestamp() > 0);
|
||||
LOG.info(ce2.toString());
|
||||
Assert.assertTrue(ce2.toString().startsWith("CloseEvent [path="));
|
||||
Assertions.assertTrue(ce2.toString().startsWith("CloseEvent [path="));
|
||||
|
||||
// AppendOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.APPEND);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.APPEND);
|
||||
Event.AppendEvent append2 = (Event.AppendEvent)batch.getEvents()[0];
|
||||
Assert.assertEquals("/file2", append2.getPath());
|
||||
Assert.assertFalse(append2.toNewBlock());
|
||||
Assertions.assertEquals("/file2", append2.getPath());
|
||||
Assertions.assertFalse(append2.toNewBlock());
|
||||
LOG.info(append2.toString());
|
||||
Assert.assertTrue(append2.toString().startsWith("AppendEvent [path="));
|
||||
Assertions.assertTrue(append2.toString().startsWith("AppendEvent [path="));
|
||||
|
||||
// CloseOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CLOSE);
|
||||
Assert.assertTrue(((Event.CloseEvent) batch.getEvents()[0]).getPath().equals("/file2"));
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CLOSE);
|
||||
Assertions.assertTrue(((Event.CloseEvent) batch.getEvents()[0]).getPath().equals("/file2"));
|
||||
|
||||
// TimesOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue.getPath().equals("/file2"));
|
||||
Assert.assertTrue(mue.getMetadataType() ==
|
||||
Assertions.assertTrue(mue.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(mue.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.TIMES);
|
||||
LOG.info(mue.toString());
|
||||
Assert.assertTrue(mue.toString().startsWith("MetadataUpdateEvent [path="));
|
||||
Assertions.assertTrue(mue.toString().startsWith("MetadataUpdateEvent [path="));
|
||||
|
||||
// SetReplicationOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue2 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue2.getPath().equals("/file2"));
|
||||
Assert.assertTrue(mue2.getMetadataType() ==
|
||||
Assertions.assertTrue(mue2.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(mue2.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.REPLICATION);
|
||||
Assert.assertTrue(mue2.getReplication() == 1);
|
||||
Assertions.assertTrue(mue2.getReplication() == 1);
|
||||
LOG.info(mue2.toString());
|
||||
|
||||
// ConcatDeleteOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(3, batch.getEvents().length);
|
||||
Assertions.assertEquals(3, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.APPEND);
|
||||
Assert.assertTrue(((Event.AppendEvent) batch.getEvents()[0]).getPath().equals("/file2"));
|
||||
Assert.assertTrue(batch.getEvents()[1].getEventType() == Event.EventType.UNLINK);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.APPEND);
|
||||
Assertions.assertTrue(((Event.AppendEvent) batch.getEvents()[0]).getPath().equals("/file2"));
|
||||
Assertions.assertTrue(batch.getEvents()[1].getEventType() == Event.EventType.UNLINK);
|
||||
Event.UnlinkEvent ue2 = (Event.UnlinkEvent) batch.getEvents()[1];
|
||||
Assert.assertTrue(ue2.getPath().equals("/file3"));
|
||||
Assert.assertTrue(ue2.getTimestamp() > 0);
|
||||
Assertions.assertTrue(ue2.getPath().equals("/file3"));
|
||||
Assertions.assertTrue(ue2.getTimestamp() > 0);
|
||||
LOG.info(ue2.toString());
|
||||
Assert.assertTrue(ue2.toString().startsWith("UnlinkEvent [path="));
|
||||
Assert.assertTrue(batch.getEvents()[2].getEventType() == Event.EventType.CLOSE);
|
||||
Assertions.assertTrue(ue2.toString().startsWith("UnlinkEvent [path="));
|
||||
Assertions.assertTrue(batch.getEvents()[2].getEventType() == Event.EventType.CLOSE);
|
||||
Event.CloseEvent ce3 = (Event.CloseEvent) batch.getEvents()[2];
|
||||
Assert.assertTrue(ce3.getPath().equals("/file2"));
|
||||
Assert.assertTrue(ce3.getTimestamp() > 0);
|
||||
Assertions.assertTrue(ce3.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(ce3.getTimestamp() > 0);
|
||||
|
||||
// DeleteOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.UNLINK);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.UNLINK);
|
||||
Event.UnlinkEvent ue = (Event.UnlinkEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ue.getPath().equals("/file2"));
|
||||
Assert.assertTrue(ue.getTimestamp() > 0);
|
||||
Assertions.assertTrue(ue.getPath().equals("/file2"));
|
||||
Assertions.assertTrue(ue.getTimestamp() > 0);
|
||||
LOG.info(ue.toString());
|
||||
|
||||
// MkdirOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Event.CreateEvent ce4 = (Event.CreateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce4.getiNodeType() ==
|
||||
Assertions.assertTrue(ce4.getiNodeType() ==
|
||||
Event.CreateEvent.INodeType.DIRECTORY);
|
||||
Assert.assertTrue(ce4.getPath().equals("/dir"));
|
||||
Assert.assertTrue(ce4.getCtime() > 0);
|
||||
Assert.assertTrue(ce4.getReplication() == 0);
|
||||
Assert.assertTrue(ce4.getSymlinkTarget() == null);
|
||||
Assertions.assertTrue(ce4.getPath().equals("/dir"));
|
||||
Assertions.assertTrue(ce4.getCtime() > 0);
|
||||
Assertions.assertTrue(ce4.getReplication() == 0);
|
||||
Assertions.assertTrue(ce4.getSymlinkTarget() == null);
|
||||
LOG.info(ce4.toString());
|
||||
|
||||
// SetPermissionsOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue3 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue3.getPath().equals("/dir"));
|
||||
Assert.assertTrue(mue3.getMetadataType() ==
|
||||
Assertions.assertTrue(mue3.getPath().equals("/dir"));
|
||||
Assertions.assertTrue(mue3.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.PERMS);
|
||||
Assert.assertTrue(mue3.getPerms().toString().contains("rw-rw-rw-"));
|
||||
Assertions.assertTrue(mue3.getPerms().toString().contains("rw-rw-rw-"));
|
||||
LOG.info(mue3.toString());
|
||||
|
||||
// SetOwnerOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue4 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue4.getPath().equals("/dir"));
|
||||
Assert.assertTrue(mue4.getMetadataType() ==
|
||||
Assertions.assertTrue(mue4.getPath().equals("/dir"));
|
||||
Assertions.assertTrue(mue4.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.OWNER);
|
||||
Assert.assertTrue(mue4.getOwnerName().equals("username"));
|
||||
Assert.assertTrue(mue4.getGroupName().equals("groupname"));
|
||||
Assertions.assertTrue(mue4.getOwnerName().equals("username"));
|
||||
Assertions.assertTrue(mue4.getGroupName().equals("groupname"));
|
||||
LOG.info(mue4.toString());
|
||||
|
||||
// SymlinkOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Event.CreateEvent ce5 = (Event.CreateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce5.getiNodeType() ==
|
||||
Assertions.assertTrue(ce5.getiNodeType() ==
|
||||
Event.CreateEvent.INodeType.SYMLINK);
|
||||
Assert.assertTrue(ce5.getPath().equals("/dir2"));
|
||||
Assert.assertTrue(ce5.getCtime() > 0);
|
||||
Assert.assertTrue(ce5.getReplication() == 0);
|
||||
Assert.assertTrue(ce5.getSymlinkTarget().equals("/dir"));
|
||||
Assertions.assertTrue(ce5.getPath().equals("/dir2"));
|
||||
Assertions.assertTrue(ce5.getCtime() > 0);
|
||||
Assertions.assertTrue(ce5.getReplication() == 0);
|
||||
Assertions.assertTrue(ce5.getSymlinkTarget().equals("/dir"));
|
||||
LOG.info(ce5.toString());
|
||||
|
||||
// SetXAttrOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue5 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue5.getPath().equals("/file5"));
|
||||
Assert.assertTrue(mue5.getMetadataType() ==
|
||||
Assertions.assertTrue(mue5.getPath().equals("/file5"));
|
||||
Assertions.assertTrue(mue5.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.XATTRS);
|
||||
Assert.assertTrue(mue5.getxAttrs().size() == 1);
|
||||
Assert.assertTrue(mue5.getxAttrs().get(0).getName().contains("field"));
|
||||
Assert.assertTrue(!mue5.isxAttrsRemoved());
|
||||
Assertions.assertTrue(mue5.getxAttrs().size() == 1);
|
||||
Assertions.assertTrue(mue5.getxAttrs().get(0).getName().contains("field"));
|
||||
Assertions.assertTrue(!mue5.isxAttrsRemoved());
|
||||
LOG.info(mue5.toString());
|
||||
|
||||
// RemoveXAttrOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue6 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue6.getPath().equals("/file5"));
|
||||
Assert.assertTrue(mue6.getMetadataType() ==
|
||||
Assertions.assertTrue(mue6.getPath().equals("/file5"));
|
||||
Assertions.assertTrue(mue6.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.XATTRS);
|
||||
Assert.assertTrue(mue6.getxAttrs().size() == 1);
|
||||
Assert.assertTrue(mue6.getxAttrs().get(0).getName().contains("field"));
|
||||
Assert.assertTrue(mue6.isxAttrsRemoved());
|
||||
Assertions.assertTrue(mue6.getxAttrs().size() == 1);
|
||||
Assertions.assertTrue(mue6.getxAttrs().get(0).getName().contains("field"));
|
||||
Assertions.assertTrue(mue6.isxAttrsRemoved());
|
||||
LOG.info(mue6.toString());
|
||||
|
||||
// SetAclOp (1)
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue7 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue7.getPath().equals("/file5"));
|
||||
Assert.assertTrue(mue7.getMetadataType() ==
|
||||
Assertions.assertTrue(mue7.getPath().equals("/file5"));
|
||||
Assertions.assertTrue(mue7.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.ACLS);
|
||||
Assert.assertTrue(mue7.getAcls().contains(
|
||||
Assertions.assertTrue(mue7.getAcls().contains(
|
||||
AclEntry.parseAclEntry("user::rwx", true)));
|
||||
LOG.info(mue7.toString());
|
||||
|
||||
// SetAclOp (2)
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.METADATA);
|
||||
Event.MetadataUpdateEvent mue8 = (Event.MetadataUpdateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(mue8.getPath().equals("/file5"));
|
||||
Assert.assertTrue(mue8.getMetadataType() ==
|
||||
Assertions.assertTrue(mue8.getPath().equals("/file5"));
|
||||
Assertions.assertTrue(mue8.getMetadataType() ==
|
||||
Event.MetadataUpdateEvent.MetadataType.ACLS);
|
||||
Assert.assertTrue(mue8.getAcls() == null);
|
||||
Assertions.assertTrue(mue8.getAcls() == null);
|
||||
LOG.info(mue8.toString());
|
||||
|
||||
// RenameOp (2)
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.RENAME);
|
||||
Event.RenameEvent re3 = (Event.RenameEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(re3.getDstPath().equals("/dir/file5"));
|
||||
Assert.assertTrue(re3.getSrcPath().equals("/file5"));
|
||||
Assert.assertTrue(re3.getTimestamp() > 0);
|
||||
Assertions.assertTrue(re3.getDstPath().equals("/dir/file5"));
|
||||
Assertions.assertTrue(re3.getSrcPath().equals("/file5"));
|
||||
Assertions.assertTrue(re3.getTimestamp() > 0);
|
||||
LOG.info(re3.toString());
|
||||
|
||||
// TruncateOp
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert
|
||||
Assertions
|
||||
.assertTrue(batch.getEvents()[0].getEventType() ==
|
||||
Event.EventType.TRUNCATE);
|
||||
Event.TruncateEvent et = ((Event.TruncateEvent) batch.getEvents()[0]);
|
||||
Assert.assertTrue(et.getPath().equals("/truncate_file"));
|
||||
Assert.assertTrue(et.getFileSize() == BLOCK_SIZE);
|
||||
Assert.assertTrue(et.getTimestamp() > 0);
|
||||
Assertions.assertTrue(et.getPath().equals("/truncate_file"));
|
||||
Assertions.assertTrue(et.getFileSize() == BLOCK_SIZE);
|
||||
Assertions.assertTrue(et.getTimestamp() > 0);
|
||||
LOG.info(et.toString());
|
||||
Assert.assertTrue(et.toString().startsWith("TruncateEvent [path="));
|
||||
Assertions.assertTrue(et.toString().startsWith("TruncateEvent [path="));
|
||||
|
||||
// CreateEvent without overwrite
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType()
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType()
|
||||
== Event.EventType.CREATE);
|
||||
ce = (Event.CreateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assert.assertTrue(ce.getPath().equals("/file_ec_test1"));
|
||||
Assert.assertTrue(ce.getCtime() > 0);
|
||||
Assert.assertTrue(ce.getReplication() > 0);
|
||||
Assert.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assert.assertFalse(ce.getOverwrite());
|
||||
Assert.assertEquals(BLOCK_SIZE, ce.getDefaultBlockSize());
|
||||
Assert.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assert.assertFalse(ce.isErasureCoded().get());
|
||||
Assertions.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assertions.assertTrue(ce.getPath().equals("/file_ec_test1"));
|
||||
Assertions.assertTrue(ce.getCtime() > 0);
|
||||
Assertions.assertTrue(ce.getReplication() > 0);
|
||||
Assertions.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assertions.assertFalse(ce.getOverwrite());
|
||||
Assertions.assertEquals(BLOCK_SIZE, ce.getDefaultBlockSize());
|
||||
Assertions.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assertions.assertFalse(ce.isErasureCoded().get());
|
||||
LOG.info(ce.toString());
|
||||
Assert.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
Assertions.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
|
||||
// Returns null when there are no further events
|
||||
Assert.assertTrue(eis.poll() == null);
|
||||
Assertions.assertTrue(eis.poll() == null);
|
||||
|
||||
// make sure the estimate hasn't changed since the above assertion
|
||||
// tells us that we are fully caught up to the current namesystem state
|
||||
// and we should not have been behind at all when eventsBehind was set
|
||||
// either, since there were few enough events that they should have all
|
||||
// been read to the client during the first poll() call
|
||||
Assert.assertTrue(eis.getTxidsBehindEstimate() == eventsBehind);
|
||||
Assertions.assertTrue(eis.getTxidsBehindEstimate() == eventsBehind);
|
||||
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
@ -470,41 +470,41 @@ public void testErasureCodedFiles() throws Exception {
|
||||
EventBatch batch = null;
|
||||
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
long txid = batch.getTxid();
|
||||
long eventsBehind = eis.getTxidsBehindEstimate();
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType()
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType()
|
||||
== Event.EventType.CREATE);
|
||||
Event.CreateEvent ce = (Event.CreateEvent) batch.getEvents()[0];
|
||||
Assert.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assert.assertTrue(ce.getPath().equals("/ecdir/file_ec_test2"));
|
||||
Assert.assertTrue(ce.getCtime() > 0);
|
||||
Assert.assertEquals(1, ce.getReplication());
|
||||
Assert.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assert.assertTrue(ce.getOverwrite());
|
||||
Assert.assertEquals(ecPolicy.getCellSize(), ce.getDefaultBlockSize());
|
||||
Assert.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assert.assertTrue(ce.isErasureCoded().get());
|
||||
Assertions.assertTrue(ce.getiNodeType() == Event.CreateEvent.INodeType.FILE);
|
||||
Assertions.assertTrue(ce.getPath().equals("/ecdir/file_ec_test2"));
|
||||
Assertions.assertTrue(ce.getCtime() > 0);
|
||||
Assertions.assertEquals(1, ce.getReplication());
|
||||
Assertions.assertTrue(ce.getSymlinkTarget() == null);
|
||||
Assertions.assertTrue(ce.getOverwrite());
|
||||
Assertions.assertEquals(ecPolicy.getCellSize(), ce.getDefaultBlockSize());
|
||||
Assertions.assertTrue(ce.isErasureCoded().isPresent());
|
||||
Assertions.assertTrue(ce.isErasureCoded().get());
|
||||
LOG.info(ce.toString());
|
||||
Assert.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
Assertions.assertTrue(ce.toString().startsWith("CreateEvent [INodeType="));
|
||||
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
txid = checkTxid(batch, txid);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType()
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType()
|
||||
== Event.EventType.CLOSE);
|
||||
Assert.assertTrue(((Event.CloseEvent) batch.getEvents()[0]).getPath()
|
||||
Assertions.assertTrue(((Event.CloseEvent) batch.getEvents()[0]).getPath()
|
||||
.equals("/ecdir/file_ec_test2"));
|
||||
|
||||
// Returns null when there are no further events
|
||||
Assert.assertTrue(eis.poll() == null);
|
||||
Assertions.assertTrue(eis.poll() == null);
|
||||
|
||||
// make sure the estimate hasn't changed since the above assertion
|
||||
// tells us that we are fully caught up to the current namesystem state
|
||||
// and we should not have been behind at all when eventsBehind was set
|
||||
// either, since there were few enough events that they should have all
|
||||
// been read to the client during the first poll() call
|
||||
Assert.assertTrue(eis.getTxidsBehindEstimate() == eventsBehind);
|
||||
Assertions.assertTrue(eis.getTxidsBehindEstimate() == eventsBehind);
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
@ -532,12 +532,12 @@ public void testNNFailover() throws IOException, URISyntaxException,
|
||||
// active
|
||||
for (int i = 0; i < 10; i++) {
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assert.assertTrue(((Event.CreateEvent) batch.getEvents()[0]).getPath().equals("/dir" +
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertTrue(((Event.CreateEvent) batch.getEvents()[0]).getPath().equals("/dir" +
|
||||
i));
|
||||
}
|
||||
Assert.assertTrue(eis.poll() == null);
|
||||
Assertions.assertTrue(eis.poll() == null);
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
@ -571,12 +571,12 @@ public void testTwoActiveNNs() throws IOException, MissingEventsException {
|
||||
EventBatch batch = null;
|
||||
for (int i = 0; i < 10; i++) {
|
||||
batch = waitForNextEvents(eis);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assert.assertTrue(((Event.CreateEvent) batch.getEvents()[0]).getPath().equals("/dir" +
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertTrue(((Event.CreateEvent) batch.getEvents()[0]).getPath().equals("/dir" +
|
||||
i));
|
||||
}
|
||||
Assert.assertTrue(eis.poll() == null);
|
||||
Assertions.assertTrue(eis.poll() == null);
|
||||
} finally {
|
||||
try {
|
||||
cluster.shutdown();
|
||||
@ -615,10 +615,10 @@ public void run() {
|
||||
// a very generous wait period -- the edit will definitely have been
|
||||
// processed by the time this is up
|
||||
EventBatch batch = eis.poll(5, TimeUnit.SECONDS);
|
||||
Assert.assertNotNull(batch);
|
||||
Assert.assertEquals(1, batch.getEvents().length);
|
||||
Assert.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assert.assertEquals("/dir", ((Event.CreateEvent) batch.getEvents()[0]).getPath());
|
||||
Assertions.assertNotNull(batch);
|
||||
Assertions.assertEquals(1, batch.getEvents().length);
|
||||
Assertions.assertTrue(batch.getEvents()[0].getEventType() == Event.EventType.CREATE);
|
||||
Assertions.assertEquals("/dir", ((Event.CreateEvent) batch.getEvents()[0]).getPath());
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
|
@ -33,10 +33,10 @@
|
||||
import org.apache.hadoop.security.SecurityUtil;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.security.ssl.KeyStoreTestUtil;
|
||||
import org.junit.After;
|
||||
import org.junit.Before;
|
||||
import org.junit.Rule;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.rules.Timeout;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
@ -64,10 +64,7 @@
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_KEYTAB_FILE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_SERVER_HTTPS_KEYSTORE_RESOURCE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_WEB_AUTHENTICATION_KERBEROS_PRINCIPAL_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
/**
|
||||
* Class for Kerberized test cases for {@link DFSInotifyEventInputStream}.
|
||||
@ -131,7 +128,7 @@ public Void run() throws Exception {
|
||||
while ((batch = eis.poll()) != null) {
|
||||
LOG.info("txid: " + batch.getTxid());
|
||||
}
|
||||
assertNull("poll should not return anything", eis.poll());
|
||||
assertNull(eis.poll(), "poll should not return anything");
|
||||
|
||||
Thread.sleep(6000);
|
||||
LOG.info("Slept 6 seconds to make sure the TGT has expired.");
|
||||
@ -143,16 +140,16 @@ public Void run() throws Exception {
|
||||
|
||||
// verify we can poll after a tgt expiration interval
|
||||
batch = eis.poll();
|
||||
assertNotNull("poll should return something", batch);
|
||||
assertNotNull(batch, "poll should return something");
|
||||
assertEquals(1, batch.getEvents().length);
|
||||
assertNull("poll should not return anything", eis.poll());
|
||||
assertNull(eis.poll(), "poll should not return anything");
|
||||
return null;
|
||||
}
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void initKerberizedCluster() throws Exception {
|
||||
baseDir = new File(System.getProperty("test.build.dir", "target/test-dir"),
|
||||
TestDFSInotifyEventInputStreamKerberized.class.getSimpleName());
|
||||
@ -169,8 +166,8 @@ public void initKerberizedCluster() throws Exception {
|
||||
SecurityUtil.setAuthenticationMethod(
|
||||
UserGroupInformation.AuthenticationMethod.KERBEROS, baseConf);
|
||||
UserGroupInformation.setConfiguration(baseConf);
|
||||
assertTrue("Expected configuration to enable security",
|
||||
UserGroupInformation.isSecurityEnabled());
|
||||
assertTrue(
|
||||
UserGroupInformation.isSecurityEnabled(), "Expected configuration to enable security");
|
||||
|
||||
final String userName = "hdfs";
|
||||
nnKeytabFile = new File(baseDir, userName + ".keytab");
|
||||
@ -218,7 +215,7 @@ public void initKerberizedCluster() throws Exception {
|
||||
KeyStoreTestUtil.getServerSSLConfigFileName());
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void shutdownCluster() throws Exception {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
@ -18,10 +18,7 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.DFS_CLIENT_READ_USE_CACHE_PRIORITY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
import static org.hamcrest.CoreMatchers.equalTo;
|
||||
import static org.mockito.Mockito.mock;
|
||||
import static org.mockito.Mockito.when;
|
||||
@ -47,8 +44,8 @@
|
||||
import org.apache.hadoop.hdfs.client.impl.DfsClientConf;
|
||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.Retry;
|
||||
|
||||
import org.junit.Assume;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Assumptions;
|
||||
|
||||
public class TestDFSInputStream {
|
||||
private void testSkipInner(MiniDFSCluster cluster) throws IOException {
|
||||
@ -108,7 +105,7 @@ public void testSkipWithRemoteBlockReader2() throws IOException {
|
||||
|
||||
@Test(timeout=60000)
|
||||
public void testSkipWithLocalBlockReader() throws IOException {
|
||||
Assume.assumeThat(DomainSocket.getLoadingFailureReason(), equalTo(null));
|
||||
Assumptions.assumeThat(DomainSocket.getLoadingFailureReason(), equalTo(null));
|
||||
TemporarySocketDirectory sockDir = new TemporarySocketDirectory();
|
||||
DomainSocket.disableBindPathValidation();
|
||||
Configuration conf = new Configuration();
|
||||
@ -218,10 +215,10 @@ public void testNullCheckSumWhenDNRestarted()
|
||||
final List<DatanodeDescriptor> live = new ArrayList<DatanodeDescriptor>();
|
||||
cluster.getNameNode().getNamesystem().getBlockManager()
|
||||
.getDatanodeManager().fetchDatanodes(live, null, false);
|
||||
assertTrue("DN start should be success and live dn should be 2",
|
||||
live.size() == 2);
|
||||
assertTrue("File size should be " + chunkSize,
|
||||
fs.getFileStatus(file).getLen() == chunkSize);
|
||||
assertTrue(
|
||||
live.size() == 2, "DN start should be success and live dn should be 2");
|
||||
assertTrue(
|
||||
fs.getFileStatus(file).getLen() == chunkSize, "File size should be " + chunkSize);
|
||||
} finally {
|
||||
cluster.shutdown();
|
||||
}
|
||||
|
@ -19,11 +19,7 @@
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_SIZE_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotEquals;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.InetSocketAddress;
|
||||
@ -41,10 +37,10 @@
|
||||
import org.apache.hadoop.hdfs.server.datanode.InternalDataNodeTestUtils;
|
||||
import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.junit.After;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.junit.runners.Parameterized;
|
||||
|
||||
@ -84,7 +80,7 @@ public TestDFSInputStreamBlockLocations(Boolean enableExpiration) {
|
||||
enableBlkExpiration = enableExpiration;
|
||||
}
|
||||
|
||||
@Before
|
||||
@BeforeEach
|
||||
public void setup() throws IOException {
|
||||
conf = new HdfsConfiguration();
|
||||
conf.setBoolean(
|
||||
@ -119,7 +115,7 @@ public void setup() throws IOException {
|
||||
fs = dfsCluster.getFileSystem();
|
||||
}
|
||||
|
||||
@After
|
||||
@AfterEach
|
||||
public void teardown() throws IOException {
|
||||
if (dfsClient != null) {
|
||||
dfsClient.close();
|
||||
@ -172,21 +168,21 @@ public void testRead() throws Exception {
|
||||
DatanodeInfo[] firstBlkDNInfos = firstLocatedBlk.getLocations();
|
||||
while (fin.getPos() < firstBlockMark) {
|
||||
bytesRead = fin.read(readBuffer);
|
||||
Assert.assertTrue("Unexpected number of read bytes",
|
||||
chunkReadSize >= bytesRead);
|
||||
Assertions.assertTrue(
|
||||
chunkReadSize >= bytesRead, "Unexpected number of read bytes");
|
||||
if (currDNInfo == null) {
|
||||
currDNInfo = fin.getCurrentDatanode();
|
||||
assertNotNull("current FIS datanode is null", currDNInfo);
|
||||
assertNotNull(currDNInfo, "current FIS datanode is null");
|
||||
continue;
|
||||
}
|
||||
prevDNInfo = currDNInfo;
|
||||
currDNInfo = fin.getCurrentDatanode();
|
||||
assertEquals("the DFSInput stream does not read from same node",
|
||||
prevDNInfo, currDNInfo);
|
||||
assertEquals(
|
||||
prevDNInfo, currDNInfo, "the DFSInput stream does not read from same node");
|
||||
}
|
||||
|
||||
assertEquals("InputStream exceeds expected position",
|
||||
firstBlockMark, fin.getPos());
|
||||
assertEquals(
|
||||
firstBlockMark, fin.getPos(), "InputStream exceeds expected position");
|
||||
// get the second block locations
|
||||
LocatedBlock secondLocatedBlk =
|
||||
fin.locatedBlocks.getLocatedBlocks().get(1);
|
||||
@ -216,23 +212,23 @@ public void testRead() throws Exception {
|
||||
}
|
||||
while (fin.getPos() < secondBlockMark) {
|
||||
bytesRead = fin.read(readBuffer);
|
||||
assertTrue("dead node used to read at position: " + fin.getPos(),
|
||||
fin.deadNodesContain(deadNodeInfo));
|
||||
Assert.assertTrue("Unexpected number of read bytes",
|
||||
chunkReadSize >= bytesRead);
|
||||
assertTrue(
|
||||
fin.deadNodesContain(deadNodeInfo), "dead node used to read at position: " + fin.getPos());
|
||||
Assertions.assertTrue(
|
||||
chunkReadSize >= bytesRead, "Unexpected number of read bytes");
|
||||
prevDNInfo = currDNInfo;
|
||||
currDNInfo = fin.getCurrentDatanode();
|
||||
assertNotEquals(deadNodeInfo, currDNInfo);
|
||||
if (firstIteration) {
|
||||
// currDNInfo has to be different unless first block locs is different
|
||||
assertFalse("FSInputStream should pick a different DN",
|
||||
firstBlkDNInfos[0].equals(deadNodeInfo)
|
||||
&& prevDNInfo.equals(currDNInfo));
|
||||
// currDNInfo has to be different unless first block locs is different
|
||||
assertFalse(
|
||||
firstBlkDNInfos[0].equals(deadNodeInfo)
|
||||
&& prevDNInfo.equals(currDNInfo), "FSInputStream should pick a different DN");
|
||||
firstIteration = false;
|
||||
}
|
||||
}
|
||||
assertEquals("InputStream exceeds expected position",
|
||||
secondBlockMark, fin.getPos());
|
||||
assertEquals(
|
||||
secondBlockMark, fin.getPos(), "InputStream exceeds expected position");
|
||||
// restart the dead node with the same port
|
||||
assertTrue(dfsCluster.restartDataNode(stoppedDNProps, true));
|
||||
dfsCluster.waitActive();
|
||||
@ -244,13 +240,13 @@ public void testRead() throws Exception {
|
||||
while (fin.getPos() < thirdBlockMark) {
|
||||
bytesRead = fin.read(readBuffer);
|
||||
if (this.enableBlkExpiration) {
|
||||
assertEquals("node is removed from deadNodes after 1st iteration",
|
||||
firstIteration, fin.deadNodesContain(deadNodeInfo));
|
||||
assertEquals(
|
||||
firstIteration, fin.deadNodesContain(deadNodeInfo), "node is removed from deadNodes after 1st iteration");
|
||||
} else {
|
||||
assertTrue(fin.deadNodesContain(deadNodeInfo));
|
||||
}
|
||||
Assert.assertTrue("Unexpected number of read bytes",
|
||||
chunkReadSize >= bytesRead);
|
||||
Assertions.assertTrue(
|
||||
chunkReadSize >= bytesRead, "Unexpected number of read bytes");
|
||||
prevDNInfo = currDNInfo;
|
||||
currDNInfo = fin.getCurrentDatanode();
|
||||
if (!this.enableBlkExpiration) {
|
||||
@ -266,8 +262,8 @@ public void testRead() throws Exception {
|
||||
}
|
||||
}
|
||||
}
|
||||
assertEquals("InputStream exceeds expected position",
|
||||
thirdBlockMark, fin.getPos());
|
||||
assertEquals(
|
||||
thirdBlockMark, fin.getPos(), "InputStream exceeds expected position");
|
||||
} finally {
|
||||
if (fout != null) {
|
||||
fout.close();
|
||||
|
@ -17,7 +17,7 @@
|
||||
*/
|
||||
package org.apache.hadoop.hdfs;
|
||||
|
||||
import static org.junit.Assert.*;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
import java.io.IOException;
|
||||
@ -30,7 +30,7 @@
|
||||
import org.apache.hadoop.fs.permission.FsPermission;
|
||||
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
/**
|
||||
* This class tests that the DFS command mkdirs only creates valid
|
||||
@ -106,10 +106,10 @@ public void testMkdir() throws IOException {
|
||||
} catch (IOException e) {
|
||||
expectedException = e;
|
||||
}
|
||||
assertTrue("Create a directory when parent dir exists as file using"
|
||||
+ " mkdir() should throw ParentNotDirectoryException ",
|
||||
expectedException != null
|
||||
&& expectedException instanceof ParentNotDirectoryException);
|
||||
assertTrue(
|
||||
expectedException != null
|
||||
&& expectedException instanceof ParentNotDirectoryException, "Create a directory when parent dir exists as file using"
|
||||
+ " mkdir() should throw ParentNotDirectoryException ");
|
||||
// Create a dir in a non-exist directory, should fail
|
||||
expectedException = null;
|
||||
try {
|
||||
@ -118,10 +118,10 @@ public void testMkdir() throws IOException {
|
||||
} catch (IOException e) {
|
||||
expectedException = e;
|
||||
}
|
||||
assertTrue("Create a directory in a non-exist parent dir using"
|
||||
+ " mkdir() should throw FileNotFoundException ",
|
||||
expectedException != null
|
||||
&& expectedException instanceof FileNotFoundException);
|
||||
assertTrue(
|
||||
expectedException != null
|
||||
&& expectedException instanceof FileNotFoundException, "Create a directory in a non-exist parent dir using"
|
||||
+ " mkdir() should throw FileNotFoundException ");
|
||||
} finally {
|
||||
dfs.close();
|
||||
cluster.shutdown();
|
||||
|
@ -57,22 +57,19 @@
|
||||
import org.apache.hadoop.test.GenericTestUtils;
|
||||
import org.apache.hadoop.test.PathUtils;
|
||||
import org.apache.hadoop.test.Whitebox;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.jupiter.api.AfterAll;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.BeforeAll;
|
||||
|
||||
import static org.apache.hadoop.hdfs.client.HdfsClientConfigKeys.Write.RECOVER_LEASE_ON_CLOSE_EXCEPTION_KEY;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.junit.jupiter.api.Assertions.*;
|
||||
import static org.mockito.ArgumentMatchers.anyBoolean;
|
||||
import static org.mockito.ArgumentMatchers.anyLong;
|
||||
import org.mockito.Mockito;
|
||||
import static org.mockito.Mockito.times;
|
||||
import static org.mockito.Mockito.verify;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.mockito.Mockito.doReturn;
|
||||
import static org.mockito.Mockito.mock;
|
||||
import static org.mockito.Mockito.doThrow;
|
||||
@ -85,7 +82,7 @@
|
||||
public class TestDFSOutputStream {
|
||||
static MiniDFSCluster cluster;
|
||||
|
||||
@BeforeClass
|
||||
@BeforeAll
|
||||
public static void setup() throws IOException {
|
||||
Configuration conf = new Configuration();
|
||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
|
||||
@ -107,7 +104,7 @@ public void testCloseTwice() throws IOException {
|
||||
LastExceptionInStreamer ex = (LastExceptionInStreamer) Whitebox
|
||||
.getInternalState(streamer, "lastException");
|
||||
Throwable thrown = (Throwable) Whitebox.getInternalState(ex, "thrown");
|
||||
Assert.assertNull(thrown);
|
||||
Assertions.assertNull(thrown);
|
||||
|
||||
dos.close();
|
||||
|
||||
@ -119,7 +116,7 @@ public void testCloseTwice() throws IOException {
|
||||
assertEquals(e, dummy);
|
||||
}
|
||||
thrown = (Throwable) Whitebox.getInternalState(ex, "thrown");
|
||||
Assert.assertNull(thrown);
|
||||
Assertions.assertNull(thrown);
|
||||
dos.close();
|
||||
}
|
||||
|
||||
@ -145,10 +142,10 @@ public void testComputePacketChunkSize() throws Exception {
|
||||
Field field = dos.getClass().getDeclaredField("packetSize");
|
||||
field.setAccessible(true);
|
||||
|
||||
Assert.assertTrue((Integer) field.get(dos) + 33 < packetSize);
|
||||
Assertions.assertTrue((Integer) field.get(dos) + 33 < packetSize);
|
||||
// If PKT_MAX_HEADER_LEN is 257, actual packet size come to over 64KB
|
||||
// without a fix on HDFS-7308.
|
||||
Assert.assertTrue((Integer) field.get(dos) + 257 < packetSize);
|
||||
Assertions.assertTrue((Integer) field.get(dos) + 257 < packetSize);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -246,21 +243,21 @@ private void runAdjustChunkBoundary(
|
||||
final Field writePacketSizeField = dos.getClass()
|
||||
.getDeclaredField("writePacketSize");
|
||||
writePacketSizeField.setAccessible(true);
|
||||
Assert.assertEquals(writePacketSizeField.getInt(dos),
|
||||
Assertions.assertEquals(writePacketSizeField.getInt(dos),
|
||||
finalWritePacketSize);
|
||||
|
||||
/* get and verify chunksPerPacket */
|
||||
final Field chunksPerPacketField = dos.getClass()
|
||||
.getDeclaredField("chunksPerPacket");
|
||||
chunksPerPacketField.setAccessible(true);
|
||||
Assert.assertEquals(chunksPerPacketField.getInt(dos),
|
||||
Assertions.assertEquals(chunksPerPacketField.getInt(dos),
|
||||
(finalWritePacketSize - packateMaxHeaderLength) / chunkSize);
|
||||
|
||||
/* get and verify packetSize */
|
||||
final Field packetSizeField = dos.getClass()
|
||||
.getDeclaredField("packetSize");
|
||||
packetSizeField.setAccessible(true);
|
||||
Assert.assertEquals(packetSizeField.getInt(dos),
|
||||
Assertions.assertEquals(packetSizeField.getInt(dos),
|
||||
chunksPerPacketField.getInt(dos) * chunkSize);
|
||||
} finally {
|
||||
if (dfsCluster != null) {
|
||||
@ -297,7 +294,7 @@ public void testCongestionBackoff() throws IOException {
|
||||
DFSPacket packet = mock(DFSPacket.class);
|
||||
dataQueue.add(packet);
|
||||
stream.run();
|
||||
Assert.assertTrue(congestedNodes.isEmpty());
|
||||
Assertions.assertTrue(congestedNodes.isEmpty());
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -359,11 +356,11 @@ public void testEndLeaseCall() throws Exception {
|
||||
public void testStreamFlush() throws Exception {
|
||||
FileSystem fs = cluster.getFileSystem();
|
||||
FSDataOutputStream os = fs.create(new Path("/normal-file"));
|
||||
// Verify output stream supports hsync() and hflush().
|
||||
assertTrue("DFSOutputStream should support hflush()!",
|
||||
os.hasCapability(StreamCapability.HFLUSH.getValue()));
|
||||
assertTrue("DFSOutputStream should support hsync()!",
|
||||
os.hasCapability(StreamCapability.HSYNC.getValue()));
|
||||
// Verify output stream supports hsync() and hflush().
|
||||
assertTrue(
|
||||
os.hasCapability(StreamCapability.HFLUSH.getValue()), "DFSOutputStream should support hflush()!");
|
||||
assertTrue(
|
||||
os.hasCapability(StreamCapability.HSYNC.getValue()), "DFSOutputStream should support hsync()!");
|
||||
byte[] bytes = new byte[1024];
|
||||
InputStream is = new ByteArrayInputStream(bytes);
|
||||
IOUtils.copyBytes(is, os, bytes.length);
|
||||
@ -422,7 +419,7 @@ public void testExceptionInCloseWithoutRecoverLease() throws Exception {
|
||||
}
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
@AfterAll
|
||||
public static void tearDown() {
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
x
Reference in New Issue
Block a user