HADOOP-8770. NN should not RPC to self to find trash defaults. Contributed by Eli Collins
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1381321 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
36a45e3995
commit
1c71954df8
|
@ -68,8 +68,26 @@ public class Trash extends Configured {
|
|||
public static boolean moveToAppropriateTrash(FileSystem fs, Path p,
|
||||
Configuration conf) throws IOException {
|
||||
Path fullyResolvedPath = fs.resolvePath(p);
|
||||
Trash trash = new Trash(FileSystem.get(fullyResolvedPath.toUri(), conf), conf);
|
||||
boolean success = trash.moveToTrash(fullyResolvedPath);
|
||||
FileSystem fullyResolvedFs =
|
||||
FileSystem.get(fullyResolvedPath.toUri(), conf);
|
||||
// If the trash interval is configured server side then clobber this
|
||||
// configuration so that we always respect the server configuration.
|
||||
try {
|
||||
long trashInterval = fullyResolvedFs.getServerDefaults(
|
||||
fullyResolvedPath).getTrashInterval();
|
||||
if (0 != trashInterval) {
|
||||
Configuration confCopy = new Configuration(conf);
|
||||
confCopy.setLong(CommonConfigurationKeysPublic.FS_TRASH_INTERVAL_KEY,
|
||||
trashInterval);
|
||||
conf = confCopy;
|
||||
}
|
||||
} catch (Exception e) {
|
||||
// If we can not determine that trash is enabled server side then
|
||||
// bail rather than potentially deleting a file when trash is enabled.
|
||||
throw new IOException("Failed to get server trash configuration", e);
|
||||
}
|
||||
Trash trash = new Trash(fullyResolvedFs, conf);
|
||||
boolean success = trash.moveToTrash(fullyResolvedPath);
|
||||
if (success) {
|
||||
System.out.println("Moved: '" + p + "' to trash at: " +
|
||||
trash.getCurrentTrashDir() );
|
||||
|
|
|
@ -79,24 +79,9 @@ public class TrashPolicyDefault extends TrashPolicy {
|
|||
this.trash = new Path(home, TRASH);
|
||||
this.homesParent = home.getParent();
|
||||
this.current = new Path(trash, CURRENT);
|
||||
long trashInterval = 0;
|
||||
try {
|
||||
trashInterval = fs.getServerDefaults(home).getTrashInterval();
|
||||
} catch (IOException ioe) {
|
||||
LOG.warn("Unable to get server defaults", ioe);
|
||||
}
|
||||
// If the trash interval is not configured or is disabled on the
|
||||
// server side then check the config which may be client side.
|
||||
if (0 == trashInterval) {
|
||||
this.deletionInterval = (long)(conf.getFloat(
|
||||
FS_TRASH_INTERVAL_KEY, FS_TRASH_INTERVAL_DEFAULT)
|
||||
* MSECS_PER_MINUTE);
|
||||
} else {
|
||||
this.deletionInterval = trashInterval * MSECS_PER_MINUTE;
|
||||
}
|
||||
// For the checkpoint interval use the given config instead of
|
||||
// checking the server as it's OK if a client starts an emptier
|
||||
// with a different interval than the server.
|
||||
this.deletionInterval = (long)(conf.getFloat(
|
||||
FS_TRASH_INTERVAL_KEY, FS_TRASH_INTERVAL_DEFAULT)
|
||||
* MSECS_PER_MINUTE);
|
||||
this.emptierInterval = (long)(conf.getFloat(
|
||||
FS_TRASH_CHECKPOINT_INTERVAL_KEY, FS_TRASH_CHECKPOINT_INTERVAL_DEFAULT)
|
||||
* MSECS_PER_MINUTE);
|
||||
|
|
|
@ -99,7 +99,6 @@ public class TestTrash extends TestCase {
|
|||
}
|
||||
|
||||
/**
|
||||
*
|
||||
* Test trash for the shell's delete command for the default file system
|
||||
* specified in the paramter conf
|
||||
* @param conf
|
||||
|
@ -429,8 +428,10 @@ public class TestTrash extends TestCase {
|
|||
String output = byteStream.toString();
|
||||
System.setOut(stdout);
|
||||
System.setErr(stderr);
|
||||
assertTrue("skipTrash wasn't suggested as remedy to failed rm command",
|
||||
output.indexOf(("Consider using -skipTrash option")) != -1 );
|
||||
assertTrue("skipTrash wasn't suggested as remedy to failed rm command" +
|
||||
" or we deleted / even though we could not get server defaults",
|
||||
output.indexOf("Consider using -skipTrash option") != -1 ||
|
||||
output.indexOf("Failed to determine server trash configuration") != -1);
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -57,6 +57,8 @@ import org.apache.hadoop.util.StringUtils;
|
|||
import org.apache.hadoop.util.ToolRunner;
|
||||
import org.junit.Test;
|
||||
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.FS_TRASH_INTERVAL_KEY;
|
||||
|
||||
/**
|
||||
* This class tests commands from DFSShell.
|
||||
*/
|
||||
|
@ -1480,4 +1482,95 @@ public class TestDFSShell {
|
|||
|
||||
}
|
||||
|
||||
/**
|
||||
* Delete a file optionally configuring trash on the server and client.
|
||||
*/
|
||||
private void deleteFileUsingTrash(
|
||||
boolean serverTrash, boolean clientTrash) throws Exception {
|
||||
// Run a cluster, optionally with trash enabled on the server
|
||||
Configuration serverConf = new HdfsConfiguration();
|
||||
if (serverTrash) {
|
||||
serverConf.setLong(FS_TRASH_INTERVAL_KEY, 1);
|
||||
}
|
||||
|
||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(serverConf)
|
||||
.numDataNodes(1).format(true).build();
|
||||
Configuration clientConf = new Configuration(serverConf);
|
||||
|
||||
// Create a client, optionally with trash enabled
|
||||
if (clientTrash) {
|
||||
clientConf.setLong(FS_TRASH_INTERVAL_KEY, 1);
|
||||
} else {
|
||||
clientConf.setLong(FS_TRASH_INTERVAL_KEY, 0);
|
||||
}
|
||||
|
||||
FsShell shell = new FsShell(clientConf);
|
||||
FileSystem fs = null;
|
||||
|
||||
try {
|
||||
// Create and delete a file
|
||||
fs = cluster.getFileSystem();
|
||||
writeFile(fs, new Path(TEST_ROOT_DIR, "foo"));
|
||||
final String testFile = TEST_ROOT_DIR + "/foo";
|
||||
final String trashFile = shell.getCurrentTrashDir() + "/" + testFile;
|
||||
String[] argv = new String[] { "-rm", testFile };
|
||||
int res = ToolRunner.run(shell, argv);
|
||||
assertEquals("rm failed", 0, res);
|
||||
|
||||
if (serverTrash) {
|
||||
// If the server config was set we should use it unconditionally
|
||||
assertTrue("File not in trash", fs.exists(new Path(trashFile)));
|
||||
} else if (clientTrash) {
|
||||
// If the server config was not set but the client config was
|
||||
// set then we should use it
|
||||
assertTrue("File not in trashed", fs.exists(new Path(trashFile)));
|
||||
} else {
|
||||
// If neither was set then we should not have trashed the file
|
||||
assertFalse("File was not removed", fs.exists(new Path(testFile)));
|
||||
assertFalse("File was trashed", fs.exists(new Path(trashFile)));
|
||||
}
|
||||
} finally {
|
||||
if (fs != null) {
|
||||
fs.close();
|
||||
}
|
||||
if (cluster != null) {
|
||||
cluster.shutdown();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that the server trash configuration is respected when
|
||||
* the client configuration is not set.
|
||||
*/
|
||||
@Test
|
||||
public void testServerConfigRespected() throws Exception {
|
||||
deleteFileUsingTrash(true, false);
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that server trash configuration is respected even when the
|
||||
* client configuration is set.
|
||||
*/
|
||||
@Test
|
||||
public void testServerConfigRespectedWithClient() throws Exception {
|
||||
deleteFileUsingTrash(true, true);
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that the client trash configuration is respected when
|
||||
* the server configuration is not set.
|
||||
*/
|
||||
@Test
|
||||
public void testClientConfigRespected() throws Exception {
|
||||
deleteFileUsingTrash(false, true);
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that trash is disabled by default.
|
||||
*/
|
||||
@Test
|
||||
public void testNoTrashConfig() throws Exception {
|
||||
deleteFileUsingTrash(false, false);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -23,11 +23,6 @@ import org.apache.hadoop.conf.Configuration;
|
|||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.fs.TestTrash;
|
||||
import org.apache.hadoop.fs.Trash;
|
||||
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.FS_TRASH_INTERVAL_KEY;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
|
@ -62,53 +57,4 @@ public class TestHDFSTrash {
|
|||
conf.set(DFSConfigKeys.FS_DEFAULT_NAME_KEY, fs.getUri().toString());
|
||||
TestTrash.trashNonDefaultFS(conf);
|
||||
}
|
||||
|
||||
/** Clients should always use trash if enabled server side */
|
||||
@Test
|
||||
public void testTrashEnabledServerSide() throws IOException {
|
||||
Configuration serverConf = new HdfsConfiguration();
|
||||
Configuration clientConf = new Configuration();
|
||||
|
||||
// Enable trash on the server and client
|
||||
serverConf.setLong(FS_TRASH_INTERVAL_KEY, 1);
|
||||
clientConf.setLong(FS_TRASH_INTERVAL_KEY, 1);
|
||||
|
||||
MiniDFSCluster cluster2 = null;
|
||||
try {
|
||||
cluster2 = new MiniDFSCluster.Builder(serverConf).numDataNodes(1).build();
|
||||
FileSystem fs = cluster2.getFileSystem();
|
||||
assertTrue(new Trash(fs, clientConf).isEnabled());
|
||||
|
||||
// Disabling trash on the client is ignored
|
||||
clientConf.setLong(FS_TRASH_INTERVAL_KEY, 0);
|
||||
assertTrue(new Trash(fs, clientConf).isEnabled());
|
||||
} finally {
|
||||
if (cluster2 != null) cluster2.shutdown();
|
||||
}
|
||||
}
|
||||
|
||||
/** Clients should always use trash if enabled client side */
|
||||
@Test
|
||||
public void testTrashEnabledClientSide() throws IOException {
|
||||
Configuration serverConf = new HdfsConfiguration();
|
||||
Configuration clientConf = new Configuration();
|
||||
|
||||
// Disable server side
|
||||
serverConf.setLong(FS_TRASH_INTERVAL_KEY, 0);
|
||||
|
||||
MiniDFSCluster cluster2 = null;
|
||||
try {
|
||||
cluster2 = new MiniDFSCluster.Builder(serverConf).numDataNodes(1).build();
|
||||
|
||||
// Client side is disabled by default
|
||||
FileSystem fs = cluster2.getFileSystem();
|
||||
assertFalse(new Trash(fs, clientConf).isEnabled());
|
||||
|
||||
// Enabling on the client works even though its disabled on the server
|
||||
clientConf.setLong(FS_TRASH_INTERVAL_KEY, 1);
|
||||
assertTrue(new Trash(fs, clientConf).isEnabled());
|
||||
} finally {
|
||||
if (cluster2 != null) cluster2.shutdown();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue