diff --git a/.gitignore b/.gitignore index 8b132cb5fd1..15c040cc300 100644 --- a/.gitignore +++ b/.gitignore @@ -21,3 +21,4 @@ hadoop-common-project/hadoop-common/src/test/resources/contract-test-options.xml hadoop-tools/hadoop-openstack/src/test/resources/contract-test-options.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-registry/src/main/tla/yarnregistry.toolbox yarnregistry.pdf +hadoop-tools/hadoop-aws/src/test/resources/contract-test-options.xml diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index be69d805ef9..563ed8463af 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -63,6 +63,9 @@ Release 2.7.0 - UNRELEASED HADOOP-11267. TestSecurityUtil fails when run with JDK8 because of empty principal names. (Stephen Chu via wheat9) + HADOOP-10714. AmazonS3Client.deleteObjects() need to be limited to 1000 + entries per call. (Juan Yu via atm) + Release 2.6.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-common-project/hadoop-common/src/site/markdown/filesystem/testing.md b/hadoop-common-project/hadoop-common/src/site/markdown/filesystem/testing.md index bc66e670468..444fb609850 100644 --- a/hadoop-common-project/hadoop-common/src/site/markdown/filesystem/testing.md +++ b/hadoop-common-project/hadoop-common/src/site/markdown/filesystem/testing.md @@ -28,53 +28,6 @@ These filesystem bindings must be defined in an XML configuration file, usually `hadoop-common-project/hadoop-common/src/test/resources/contract-test-options.xml`. This file is excluded should not be checked in. -### s3:// - -In `contract-test-options.xml`, the filesystem name must be defined in the property `fs.contract.test.fs.s3`. The standard configuration options to define the S3 authentication details must also be provided. - -Example: - - - - fs.contract.test.fs.s3 - s3://tests3hdfs/ - - - - fs.s3.awsAccessKeyId - DONOTPCOMMITTHISKEYTOSCM - - - - fs.s3.awsSecretAccessKey - DONOTEVERSHARETHISSECRETKEY! - - - -### s3n:// - - -In `contract-test-options.xml`, the filesystem name must be defined in the property `fs.contract.test.fs.s3n`. The standard configuration options to define the S3N authentication details muse also be provided. - -Example: - - - - - fs.contract.test.fs.s3n - s3n://tests3contract - - - - fs.s3n.awsAccessKeyId - DONOTPCOMMITTHISKEYTOSCM - - - - fs.s3n.awsSecretAccessKey - DONOTEVERSHARETHISSECRETKEY! - - ### ftp:// diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContractBaseTest.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContractBaseTest.java index 36f2fdbc4dc..84e4e43c032 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContractBaseTest.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContractBaseTest.java @@ -464,11 +464,11 @@ protected void createFile(Path path) throws IOException { out.close(); } - private void rename(Path src, Path dst, boolean renameSucceeded, + protected void rename(Path src, Path dst, boolean renameSucceeded, boolean srcExists, boolean dstExists) throws IOException { assertEquals("Rename result", renameSucceeded, fs.rename(src, dst)); assertEquals("Source exists", srcExists, fs.exists(src)); - assertEquals("Destination exists", dstExists, fs.exists(dst)); + assertEquals("Destination exists" + dst, dstExists, fs.exists(dst)); } /** diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractDeleteTest.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractDeleteTest.java index c90efd19386..2bd60ca3731 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractDeleteTest.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractDeleteTest.java @@ -19,6 +19,7 @@ package org.apache.hadoop.fs.contract; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.FileSystem; import org.junit.Test; import java.io.IOException; @@ -94,4 +95,30 @@ public void testDeleteNonEmptyDirRecursive() throws Throwable { ContractTestUtils.assertPathDoesNotExist(getFileSystem(), "not deleted", file); } + @Test + public void testDeleteDeepEmptyDir() throws Throwable { + mkdirs(path("testDeleteDeepEmptyDir/d1/d2/d3/d4")); + assertDeleted(path("testDeleteDeepEmptyDir/d1/d2/d3"), true); + + FileSystem fs = getFileSystem(); + ContractTestUtils.assertPathDoesNotExist(fs, + "not deleted", path("testDeleteDeepEmptyDir/d1/d2/d3/d4")); + ContractTestUtils.assertPathDoesNotExist(fs, + "not deleted", path("testDeleteDeepEmptyDir/d1/d2/d3")); + ContractTestUtils.assertPathExists(fs, "parent dir is deleted", + path("testDeleteDeepEmptyDir/d1/d2")); + } + + @Test + public void testDeleteSingleFile() throws Throwable { + // Test delete of just a file + Path path = path("testDeleteSingleFile/d1/d2"); + mkdirs(path); + Path file = new Path(path, "childfile"); + ContractTestUtils.writeTextFile(getFileSystem(), file, + "single file to be deleted.", true); + ContractTestUtils.assertPathExists(getFileSystem(), + "single file not created", file); + assertDeleted(file, false); + } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractMkdirTest.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractMkdirTest.java index dad3b7f2c46..86fd61f72b2 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractMkdirTest.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractMkdirTest.java @@ -112,4 +112,23 @@ public void testMkdirOverParentFile() throws Throwable { assertPathExists("mkdir failed", path); assertDeleted(path, true); } + + @Test + public void testMkdirSlashHandling() throws Throwable { + describe("verify mkdir slash handling"); + FileSystem fs = getFileSystem(); + + // No trailing slash + assertTrue(fs.mkdirs(path("testmkdir/a"))); + assertPathExists("mkdir without trailing slash failed", + path("testmkdir/a")); + + // With trailing slash + assertTrue(fs.mkdirs(path("testmkdir/b/"))); + assertPathExists("mkdir with trailing slash failed", path("testmkdir/b/")); + + // Mismatched slashes + assertPathExists("check path existence without trailing slash failed", + path("testmkdir/b")); + } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractRenameTest.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractRenameTest.java index 32f27a713fb..04c444de8d8 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractRenameTest.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractRenameTest.java @@ -182,4 +182,45 @@ public void testRenameFileNonexistentDir() throws Throwable { assertFalse(renameCreatesDestDirs); } } + + @Test + public void testRenameWithNonEmptySubDir() throws Throwable { + final Path renameTestDir = path("testRenameWithNonEmptySubDir"); + final Path srcDir = new Path(renameTestDir, "src1"); + final Path srcSubDir = new Path(srcDir, "sub"); + final Path finalDir = new Path(renameTestDir, "dest"); + FileSystem fs = getFileSystem(); + boolean renameRemoveEmptyDest = isSupported(RENAME_REMOVE_DEST_IF_EMPTY_DIR); + ContractTestUtils.rm(fs, renameTestDir, true, false); + + fs.mkdirs(srcDir); + fs.mkdirs(finalDir); + ContractTestUtils.writeTextFile(fs, new Path(srcDir, "source.txt"), + "this is the file in src dir", false); + ContractTestUtils.writeTextFile(fs, new Path(srcSubDir, "subfile.txt"), + "this is the file in src/sub dir", false); + + ContractTestUtils.assertPathExists(fs, "not created in src dir", + new Path(srcDir, "source.txt")); + ContractTestUtils.assertPathExists(fs, "not created in src/sub dir", + new Path(srcSubDir, "subfile.txt")); + + fs.rename(srcDir, finalDir); + // Accept both POSIX rename behavior and CLI rename behavior + if (renameRemoveEmptyDest) { + // POSIX rename behavior + ContractTestUtils.assertPathExists(fs, "not renamed into dest dir", + new Path(finalDir, "source.txt")); + ContractTestUtils.assertPathExists(fs, "not renamed into dest/sub dir", + new Path(finalDir, "sub/subfile.txt")); + } else { + // CLI rename behavior + ContractTestUtils.assertPathExists(fs, "not renamed into dest dir", + new Path(finalDir, "src1/source.txt")); + ContractTestUtils.assertPathExists(fs, "not renamed into dest/sub dir", + new Path(finalDir, "src1/sub/subfile.txt")); + } + ContractTestUtils.assertPathDoesNotExist(fs, "not deleted", + new Path(srcDir, "source.txt")); + } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractOptions.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractOptions.java index 61279b02ee8..d9427c6c9d0 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractOptions.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractOptions.java @@ -79,6 +79,13 @@ public interface ContractOptions { String RENAME_RETURNS_FALSE_IF_SOURCE_MISSING = "rename-returns-false-if-source-missing"; + /** + * Flag to indicate that the FS remove dest first if it is an empty directory + * mean the FS honors POSIX rename behavior. + * @{value} + */ + String RENAME_REMOVE_DEST_IF_EMPTY_DIR = "rename-remove-dest-if-empty-dir"; + /** * Flag to indicate that append is supported * @{value} diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java index cd9cc1ba154..3f16724ec26 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java @@ -31,8 +31,11 @@ import java.io.EOFException; import java.io.FileNotFoundException; import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; import java.util.Arrays; import java.util.Properties; +import java.util.UUID; /** * Utilities used across test cases @@ -44,6 +47,13 @@ public class ContractTestUtils extends Assert { public static final String IO_FILE_BUFFER_SIZE = "io.file.buffer.size"; + // For scale testing, we can repeatedly write small chunk data to generate + // a large file. + public static final String IO_CHUNK_BUFFER_SIZE = "io.chunk.buffer.size"; + public static final int DEFAULT_IO_CHUNK_BUFFER_SIZE = 128; + public static final String IO_CHUNK_MODULUS_SIZE = "io.chunk.modulus.size"; + public static final int DEFAULT_IO_CHUNK_MODULUS_SIZE = 128; + /** * Assert that a property in the property set matches the expected value * @param props property set @@ -755,5 +765,134 @@ public static void validateFileContent(byte[] concat, byte[][] bytes) { mismatch); } + /** + * Receives test data from the given input file and checks the size of the + * data as well as the pattern inside the received data. + * + * @param fs FileSystem + * @param path Input file to be checked + * @param expectedSize the expected size of the data to be read from the + * input file in bytes + * @param bufferLen Pattern length + * @param modulus Pattern modulus + * @throws IOException + * thrown if an error occurs while reading the data + */ + public static void verifyReceivedData(FileSystem fs, Path path, + final long expectedSize, + final int bufferLen, + final int modulus) throws IOException { + final byte[] testBuffer = new byte[bufferLen]; + long totalBytesRead = 0; + int nextExpectedNumber = 0; + final InputStream inputStream = fs.open(path); + try { + while (true) { + final int bytesRead = inputStream.read(testBuffer); + if (bytesRead < 0) { + break; + } + + totalBytesRead += bytesRead; + + for (int i = 0; i < bytesRead; ++i) { + if (testBuffer[i] != nextExpectedNumber) { + throw new IOException("Read number " + testBuffer[i] + + " but expected " + nextExpectedNumber); + } + + ++nextExpectedNumber; + + if (nextExpectedNumber == modulus) { + nextExpectedNumber = 0; + } + } + } + + if (totalBytesRead != expectedSize) { + throw new IOException("Expected to read " + expectedSize + + " bytes but only received " + totalBytesRead); + } + } finally { + inputStream.close(); + } + } + + /** + * Generates test data of the given size according to some specific pattern + * and writes it to the provided output file. + * + * @param fs FileSystem + * @param path Test file to be generated + * @param size The size of the test data to be generated in bytes + * @param bufferLen Pattern length + * @param modulus Pattern modulus + * @throws IOException + * thrown if an error occurs while writing the data + */ + public static long generateTestFile(FileSystem fs, Path path, + final long size, + final int bufferLen, + final int modulus) throws IOException { + final byte[] testBuffer = new byte[bufferLen]; + for (int i = 0; i < testBuffer.length; ++i) { + testBuffer[i] = (byte) (i % modulus); + } + + final OutputStream outputStream = fs.create(path, false); + long bytesWritten = 0; + try { + while (bytesWritten < size) { + final long diff = size - bytesWritten; + if (diff < testBuffer.length) { + outputStream.write(testBuffer, 0, (int) diff); + bytesWritten += diff; + } else { + outputStream.write(testBuffer); + bytesWritten += testBuffer.length; + } + } + + return bytesWritten; + } finally { + outputStream.close(); + } + } + + /** + * Creates and reads a file with the given size. The test file is generated + * according to a specific pattern so it can be easily verified even if it's + * a multi-GB one. + * During the read phase the incoming data stream is also checked against + * this pattern. + * + * @param fs FileSystem + * @param parent Test file parent dir path + * @throws IOException + * thrown if an I/O error occurs while writing or reading the test file + */ + public static void createAndVerifyFile(FileSystem fs, Path parent, final long fileSize) + throws IOException { + int testBufferSize = fs.getConf() + .getInt(IO_CHUNK_BUFFER_SIZE, DEFAULT_IO_CHUNK_BUFFER_SIZE); + int modulus = fs.getConf() + .getInt(IO_CHUNK_MODULUS_SIZE, DEFAULT_IO_CHUNK_MODULUS_SIZE); + + final String objectName = UUID.randomUUID().toString(); + final Path objectPath = new Path(parent, objectName); + + // Write test file in a specific pattern + assertEquals(fileSize, + generateTestFile(fs, objectPath, fileSize, testBufferSize, modulus)); + assertPathExists(fs, "not created successful", objectPath); + + // Now read the same file back and verify its content + try { + verifyReceivedData(fs, objectPath, fileSize, testBufferSize, modulus); + } finally { + // Delete test file + fs.delete(objectPath, false); + } + } } diff --git a/hadoop-common-project/hadoop-common/src/test/resources/contract/localfs.xml b/hadoop-common-project/hadoop-common/src/test/resources/contract/localfs.xml index b8857eb730f..38d68b33474 100644 --- a/hadoop-common-project/hadoop-common/src/test/resources/contract/localfs.xml +++ b/hadoop-common-project/hadoop-common/src/test/resources/contract/localfs.xml @@ -57,6 +57,10 @@ case sensitivity and permission options are determined at run time from OS type true + + fs.contract.rename-remove-dest-if-empty-dir + true + + +# Hadoop-AWS module: Integration with Amazon Web Services + +The `hadoop-aws` module provides support for AWS integration. The generated +JAR file, `hadoop-aws.jar` also declares a transitive dependency on all +external artifacts which are needed for this support —enabling downstream +applications to easily use this support. + +Features + +1. The "classic" `s3:` filesystem for storing objects in Amazon S3 Storage +1. The second-generation, `s3n:` filesystem, making it easy to share +data between hadoop and other applications via the S3 object store +1. The third generation, `s3a:` filesystem. Designed to be a switch in +replacement for `s3n:`, this filesystem binding supports larger files and promises +higher performance. + +The specifics of using these filesystems are documented below. + +## Warning: Object Stores are not filesystems. + +Amazon S3 is an example of "an object store". In order to achieve scalalablity +and especially high availability, S3 has —as many other cloud object stores have +done— relaxed some of the constraints which classic "POSIX" filesystems promise. + +Specifically + +1. Files that are newly created from the Hadoop Filesystem APIs may not be +immediately visible. +2. File delete and update operations may not immediately propagate. Old +copies of the file may exist for an indeterminate time period. +3. Directory operations: `delete()` and `rename()` are implemented by +recursive file-by-file operations. They take time at least proportional to +the number of files, during which time partial updates may be visible. If +the operations are interrupted, the filesystem is left in an intermediate state. + +For further discussion on these topics, please consult +[/filesystem](The Hadoop FileSystem API Definition). + +## Warning #2: your AWS credentials are valuable + +Your AWS credentials not only pay for services, they offer read and write +access to the data. Anyone with the credentials can not only read your datasets +—they can delete them. + +Do not inadvertently share these credentials through means such as +1. Checking in Hadoop configuration files containing the credentials. +1. Logging them to a console, as they invariably end up being seen. + +If you do any of these: change your credentials immediately! + + +## S3 + +### Authentication properties + + + fs.s3.awsAccessKeyId + AWS access key ID + + + + fs.s3.awsSecretAccessKey + AWS secret key + + + +## S3N + +### Authentication properties + + + fs.s3n.awsAccessKeyId + AWS access key ID + + + + fs.s3n.awsSecretAccessKey + AWS secret key + + +### Other properties + + + + fs.s3n.block.size + 67108864 + Block size to use when reading files using the native S3 + filesystem (s3n: URIs). + + + + fs.s3n.multipart.uploads.enabled + false + Setting this property to true enables multiple uploads to + native S3 filesystem. When uploading a file, it is split into blocks + if the size is larger than fs.s3n.multipart.uploads.block.size. + + + + + fs.s3n.multipart.uploads.block.size + 67108864 + The block size for multipart uploads to native S3 filesystem. + Default size is 64MB. + + + + + fs.s3n.multipart.copy.block.size + 5368709120 + The block size for multipart copy in native S3 filesystem. + Default size is 5GB. + + + + + fs.s3n.server-side-encryption-algorithm + + Specify a server-side encryption algorithm for S3. + The default is NULL, and the only other currently allowable value is AES256. + + + +## S3A + + +### Authentication properties + + + fs.s3a.awsAccessKeyId + AWS access key ID. Omit for Role-based authentication. + + + + fs.s3a.awsSecretAccessKey + AWS secret key. Omit for Role-based authentication. + + +### Other properties + + + fs.s3a.connection.maximum + 15 + Controls the maximum number of simultaneous connections to S3. + + + + fs.s3a.connection.ssl.enabled + true + Enables or disables SSL connections to S3. + + + + fs.s3a.attempts.maximum + 10 + How many times we should retry commands on transient errors. + + + + fs.s3a.connection.timeout + 5000 + Socket connection timeout in seconds. + + + + fs.s3a.paging.maximum + 5000 + How many keys to request from S3 when doing + directory listings at a time. + + + + fs.s3a.multipart.size + 104857600 + How big (in bytes) to split upload or copy operations up into. + + + + fs.s3a.multipart.threshold + 2147483647 + Threshold before uploads or copies use parallel multipart operations. + + + + fs.s3a.acl.default + Set a canned ACL for newly created and copied objects. Value may be private, + public-read, public-read-write, authenticated-read, log-delivery-write, + bucket-owner-read, or bucket-owner-full-control. + + + + fs.s3a.multipart.purge + false + True if you want to purge existing multipart uploads that may not have been + completed/aborted correctly + + + + fs.s3a.multipart.purge.age + 86400 + Minimum age in seconds of multipart uploads to purge + + + + fs.s3a.buffer.dir + ${hadoop.tmp.dir}/s3a + Comma separated list of directories that will be used to buffer file + uploads to. + + + + fs.s3a.impl + org.apache.hadoop.fs.s3a.S3AFileSystem + The implementation class of the S3A Filesystem + + + +## Testing the S3 filesystem clients + +To test the S3* filesystem clients, you need to provide two files +which pass in authentication details to the test runner + +1. `auth-keys.xml` +1. `core-site.xml` + +These are both Hadoop XML configuration files, which must be placed into +`hadoop-tools/hadoop-aws/src/test/resources`. + + +### `auth-keys.xml` + +The presence of this file triggers the testing of the S3 classes. + +Without this file, *none of the tests in this module will be executed* + +The XML file must contain all the ID/key information needed to connect +each of the filesystem clients to the object stores, and a URL for +each filesystem for its testing. + +1. `test.fs.s3n.name` : the URL of the bucket for S3n tests +1. `test.fs.s3a.name` : the URL of the bucket for S3a tests +2. `test.fs.s3.name` : the URL of the bucket for "S3" tests + +The contents of each bucket will be destroyed during the test process: +do not use the bucket for any purpose other than testing. + +Example: + + + + + test.fs.s3n.name + s3n://test-aws-s3n/ + + + + test.fs.s3a.name + s3a://test-aws-s3a/ + + + + test.fs.s3.name + s3a://test-aws-s3/ + + + + fs.s3.awsAccessKeyId + DONOTPCOMMITTHISKEYTOSCM + + + + fs.s3.awsSecretAccessKey + DONOTEVERSHARETHISSECRETKEY! + + + + fs.s3n.awsAccessKeyId + DONOTPCOMMITTHISKEYTOSCM + + + + fs.s3n.awsSecretAccessKey + DONOTEVERSHARETHISSECRETKEY! + + + + fs.s3a.awsAccessKeyId + AWS access key ID. Omit for Role-based authentication. + DONOTPCOMMITTHISKEYTOSCM + + + + fs.s3a.awsSecretAccessKey + AWS secret key. Omit for Role-based authentication. + DONOTEVERSHARETHISSECRETKEY! + + + +## File `contract-test-options.xml` + +The file `hadoop-tools/hadoop-aws/src/test/resources/contract-test-options.xml` +must be created and configured for the test fileystems. + +If a specific file `fs.contract.test.fs.*` test path is not defined for +any of the filesystems, those tests will be skipped. + +The standard S3 authentication details must also be provided. This can be +through copy-and-paste of the `auth-keys.xml` credentials, or it can be +through direct XInclude inclustion. + +#### s3:// + +The filesystem name must be defined in the property `fs.contract.test.fs.s3`. + + +Example: + + + fs.contract.test.fs.s3 + s3://test-aws-s3/ + + +### s3n:// + + +In the file `src/test/resources/contract-test-options.xml`, the filesystem +name must be defined in the property `fs.contract.test.fs.s3n`. +The standard configuration options to define the S3N authentication details +must also be provided. + +Example: + + + fs.contract.test.fs.s3n + s3n://test-aws-s3n/ + + +### s3a:// + + +In the file `src/test/resources/contract-test-options.xml`, the filesystem +name must be defined in the property `fs.contract.test.fs.s3a`. +The standard configuration options to define the S3N authentication details +must also be provided. + +Example: + + + fs.contract.test.fs.s3a + s3a://test-aws-s3a/ + + +### Complete example of `contract-test-options.xml` + + + + + + + + + + + + + fs.contract.test.fs.s3 + s3://test-aws-s3/ + + + + + fs.contract.test.fs.s3a + s3a://test-aws-s3a/ + + + + fs.contract.test.fs.s3n + s3n://test-aws-s3n/ + + + + +This example pulls in the `auth-keys.xml` file for the credentials. +This provides one single place to keep the keys up to date —and means +that the file `contract-test-options.xml` does not contain any +secret credentials itself. \ No newline at end of file diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/TestS3AContractRename.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/TestS3AContractRename.java index 88ed6d6a7be..af1ed377ef7 100644 --- a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/TestS3AContractRename.java +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/TestS3AContractRename.java @@ -21,10 +21,10 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.contract.AbstractContractRenameTest; import org.apache.hadoop.fs.contract.AbstractFSContract; -import org.apache.hadoop.fs.contract.AbstractFSContractTestBase; -import org.apache.hadoop.fs.contract.ContractTestUtils; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.contract.ContractTestUtils; +import org.junit.Test; import static org.apache.hadoop.fs.contract.ContractTestUtils.dataset; import static org.apache.hadoop.fs.contract.ContractTestUtils.writeDataset; @@ -51,14 +51,11 @@ public void testRenameDirIntoExistingDir() throws Throwable { Path destFilePath = new Path(destDir, "dest-512.txt"); byte[] destDateset = dataset(512, 'A', 'Z'); - writeDataset(fs, destFilePath, destDateset, destDateset.length, 1024, false); + writeDataset(fs, destFilePath, destDateset, destDateset.length, 1024, + false); assertIsFile(destFilePath); boolean rename = fs.rename(srcDir, destDir); - Path renamedSrcFilePath = new Path(destDir, "source-256.txt"); - assertIsFile(destFilePath); - assertIsFile(renamedSrcFilePath); - ContractTestUtils.verifyFileContents(fs, destFilePath, destDateset); - assertTrue("rename returned false though the contents were copied", rename); + assertFalse("s3a doesn't support rename to non-empty directory", rename); } } diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3/S3FileSystemContractBaseTest.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3/S3FileSystemContractBaseTest.java index 28b0507f0fa..de106f8eb8d 100644 --- a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3/S3FileSystemContractBaseTest.java +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3/S3FileSystemContractBaseTest.java @@ -21,13 +21,15 @@ import java.io.IOException; import java.net.URI; +import org.apache.commons.lang.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystemContractBaseTest; -import org.apache.hadoop.fs.Path; +import org.junit.internal.AssumptionViolatedException; public abstract class S3FileSystemContractBaseTest extends FileSystemContractBaseTest { + public static final String KEY_TEST_FS = "test.fs.s3.name"; private FileSystemStore store; abstract FileSystemStore getFileSystemStore() throws IOException; @@ -37,7 +39,12 @@ protected void setUp() throws Exception { Configuration conf = new Configuration(); store = getFileSystemStore(); fs = new S3FileSystem(store); - fs.initialize(URI.create(conf.get("test.fs.s3.name")), conf); + String fsname = conf.get(KEY_TEST_FS); + if (StringUtils.isEmpty(fsname)) { + throw new AssumptionViolatedException( + "No test FS defined in :" + KEY_TEST_FS); + } + fs.initialize(URI.create(fsname), conf); } @Override diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3AFileSystemContractBaseTest.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3AFileSystemContractBaseTest.java deleted file mode 100644 index 8455233466c..00000000000 --- a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3AFileSystemContractBaseTest.java +++ /dev/null @@ -1,327 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.fs.s3a; - -import static org.junit.Assume.*; - -import org.junit.Test; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FSDataInputStream; -import org.apache.hadoop.fs.FileSystemContractBaseTest; -import org.apache.hadoop.fs.Path; - -import java.io.IOException; -import java.io.InputStream; -import java.io.OutputStream; -import java.net.URI; -import java.util.UUID; - -/** - * Tests a live S3 system. If you keys and bucket aren't specified, all tests - * are marked as passed - * - * This uses BlockJUnit4ClassRunner because FileSystemContractBaseTest from - * TestCase which uses the old Junit3 runner that doesn't ignore assumptions - * properly making it impossible to skip the tests if we don't have a valid - * bucket. - **/ -public class S3AFileSystemContractBaseTest extends FileSystemContractBaseTest { - private static final int TEST_BUFFER_SIZE = 128; - private static final int MODULUS = 128; - - protected static final Logger LOG = LoggerFactory.getLogger(S3AFileSystemContractBaseTest.class); - - @Override - public void setUp() throws Exception { - Configuration conf = new Configuration(); - - URI testURI = URI.create(conf.get("test.fs.s3a.name")); - - boolean liveTest = testURI != null && !testURI.equals("s3a:///"); - - // This doesn't work with our JUnit 3 style test cases, so instead we'll - // make this whole class not run by default - assumeTrue(liveTest); - - fs = new S3AFileSystem(); - fs.initialize(testURI, conf); - super.setUp(); - } - - @Override - protected void tearDown() throws Exception { - if (fs != null) { - fs.delete(path("/tests3a"), true); - } - super.tearDown(); - } - - @Test(timeout = 10000) - public void testMkdirs() throws IOException { - // No trailing slash - assertTrue(fs.mkdirs(path("/tests3a/a"))); - assertTrue(fs.exists(path("/tests3a/a"))); - - // With trailing slash - assertTrue(fs.mkdirs(path("/tests3a/b/"))); - assertTrue(fs.exists(path("/tests3a/b/"))); - - // Two levels deep - assertTrue(fs.mkdirs(path("/tests3a/c/a/"))); - assertTrue(fs.exists(path("/tests3a/c/a/"))); - - // Mismatched slashes - assertTrue(fs.exists(path("/tests3a/c/a"))); - } - - - @Test(timeout=20000) - public void testDelete() throws IOException { - // Test deleting an empty directory - assertTrue(fs.mkdirs(path("/tests3a/d"))); - assertTrue(fs.delete(path("/tests3a/d"), true)); - assertFalse(fs.exists(path("/tests3a/d"))); - - // Test deleting a deep empty directory - assertTrue(fs.mkdirs(path("/tests3a/e/f/g/h"))); - assertTrue(fs.delete(path("/tests3a/e/f/g"), true)); - assertFalse(fs.exists(path("/tests3a/e/f/g/h"))); - assertFalse(fs.exists(path("/tests3a/e/f/g"))); - assertTrue(fs.exists(path("/tests3a/e/f"))); - - // Test delete of just a file - writeFile(path("/tests3a/f/f/file"), 1000); - assertTrue(fs.exists(path("/tests3a/f/f/file"))); - assertTrue(fs.delete(path("/tests3a/f/f/file"), false)); - assertFalse(fs.exists(path("/tests3a/f/f/file"))); - - - // Test delete of a path with files in various directories - writeFile(path("/tests3a/g/h/i/file"), 1000); - assertTrue(fs.exists(path("/tests3a/g/h/i/file"))); - writeFile(path("/tests3a/g/h/j/file"), 1000); - assertTrue(fs.exists(path("/tests3a/g/h/j/file"))); - try { - assertFalse(fs.delete(path("/tests3a/g/h"), false)); - fail("Expected delete to fail with recursion turned off"); - } catch (IOException e) {} - assertTrue(fs.exists(path("/tests3a/g/h/j/file"))); - assertTrue(fs.delete(path("/tests3a/g/h"), true)); - assertFalse(fs.exists(path("/tests3a/g/h/j"))); - } - - - @Test(timeout = 3600000) - public void testOpenCreate() throws IOException { - try { - createAndReadFileTest(1024); - } catch (IOException e) { - fail(e.getMessage()); - } - - try { - createAndReadFileTest(5 * 1024 * 1024); - } catch (IOException e) { - fail(e.getMessage()); - } - - try { - createAndReadFileTest(20 * 1024 * 1024); - } catch (IOException e) { - fail(e.getMessage()); - } - - /* - Enable to test the multipart upload - try { - createAndReadFileTest((long)6 * 1024 * 1024 * 1024); - } catch (IOException e) { - fail(e.getMessage()); - } - */ - } - - @Test(timeout = 1200000) - public void testRenameFile() throws IOException { - Path srcPath = path("/tests3a/a/srcfile"); - - final OutputStream outputStream = fs.create(srcPath, false); - generateTestData(outputStream, 11 * 1024 * 1024); - outputStream.close(); - - assertTrue(fs.exists(srcPath)); - - Path dstPath = path("/tests3a/b/dstfile"); - - assertFalse(fs.rename(srcPath, dstPath)); - assertTrue(fs.mkdirs(dstPath.getParent())); - assertTrue(fs.rename(srcPath, dstPath)); - assertTrue(fs.exists(dstPath)); - assertFalse(fs.exists(srcPath)); - assertTrue(fs.exists(srcPath.getParent())); - } - - - @Test(timeout = 10000) - public void testRenameDirectory() throws IOException { - Path srcPath = path("/tests3a/a"); - - assertTrue(fs.mkdirs(srcPath)); - writeFile(new Path(srcPath, "b/testfile"), 1024); - - Path nonEmptyPath = path("/tests3a/nonempty"); - writeFile(new Path(nonEmptyPath, "b/testfile"), 1024); - - assertFalse(fs.rename(srcPath, nonEmptyPath)); - - Path dstPath = path("/tests3a/b"); - assertTrue(fs.rename(srcPath, dstPath)); - assertFalse(fs.exists(srcPath)); - assertTrue(fs.exists(new Path(dstPath, "b/testfile"))); - } - - - @Test(timeout=10000) - public void testSeek() throws IOException { - Path path = path("/tests3a/testfile.seek"); - writeFile(path, TEST_BUFFER_SIZE * 10); - - - FSDataInputStream inputStream = fs.open(path, TEST_BUFFER_SIZE); - inputStream.seek(inputStream.getPos() + MODULUS); - - testReceivedData(inputStream, TEST_BUFFER_SIZE * 10 - MODULUS); - } - - /** - * Creates and reads a file with the given size in S3. The test file is - * generated according to a specific pattern. - * During the read phase the incoming data stream is also checked against this pattern. - * - * @param fileSize - * the size of the file to be generated in bytes - * @throws IOException - * thrown if an I/O error occurs while writing or reading the test file - */ - private void createAndReadFileTest(final long fileSize) throws IOException { - final String objectName = UUID.randomUUID().toString(); - final Path objectPath = new Path("/tests3a/", objectName); - - // Write test file to S3 - final OutputStream outputStream = fs.create(objectPath, false); - generateTestData(outputStream, fileSize); - outputStream.close(); - - // Now read the same file back from S3 - final InputStream inputStream = fs.open(objectPath); - testReceivedData(inputStream, fileSize); - inputStream.close(); - - // Delete test file - fs.delete(objectPath, false); - } - - - /** - * Receives test data from the given input stream and checks the size of the - * data as well as the pattern inside the received data. - * - * @param inputStream - * the input stream to read the test data from - * @param expectedSize - * the expected size of the data to be read from the input stream in bytes - * @throws IOException - * thrown if an error occurs while reading the data - */ - private void testReceivedData(final InputStream inputStream, - final long expectedSize) throws IOException { - final byte[] testBuffer = new byte[TEST_BUFFER_SIZE]; - - long totalBytesRead = 0; - int nextExpectedNumber = 0; - while (true) { - final int bytesRead = inputStream.read(testBuffer); - if (bytesRead < 0) { - break; - } - - totalBytesRead += bytesRead; - - for (int i = 0; i < bytesRead; ++i) { - if (testBuffer[i] != nextExpectedNumber) { - throw new IOException("Read number " + testBuffer[i] + " but expected " - + nextExpectedNumber); - } - - ++nextExpectedNumber; - - if (nextExpectedNumber == MODULUS) { - nextExpectedNumber = 0; - } - } - } - - if (totalBytesRead != expectedSize) { - throw new IOException("Expected to read " + expectedSize + - " bytes but only received " + totalBytesRead); - } - } - - - /** - * Generates test data of the given size according to some specific pattern - * and writes it to the provided output stream. - * - * @param outputStream - * the output stream to write the data to - * @param size - * the size of the test data to be generated in bytes - * @throws IOException - * thrown if an error occurs while writing the data - */ - private void generateTestData(final OutputStream outputStream, - final long size) throws IOException { - - final byte[] testBuffer = new byte[TEST_BUFFER_SIZE]; - for (int i = 0; i < testBuffer.length; ++i) { - testBuffer[i] = (byte) (i % MODULUS); - } - - long bytesWritten = 0; - while (bytesWritten < size) { - - final long diff = size - bytesWritten; - if (diff < testBuffer.length) { - outputStream.write(testBuffer, 0, (int)diff); - bytesWritten += diff; - } else { - outputStream.write(testBuffer); - bytesWritten += testBuffer.length; - } - } - } - - private void writeFile(Path name, int fileSize) throws IOException { - final OutputStream outputStream = fs.create(name, false); - generateTestData(outputStream, fileSize); - outputStream.close(); - } -} diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3ATestUtils.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3ATestUtils.java new file mode 100644 index 00000000000..514647c3743 --- /dev/null +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/S3ATestUtils.java @@ -0,0 +1,51 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.s3a; + +import org.apache.commons.lang.StringUtils; +import org.apache.hadoop.conf.Configuration; +import org.junit.internal.AssumptionViolatedException; + +import java.io.IOException; +import java.net.URI; + +public class S3ATestUtils { + + public static S3AFileSystem createTestFileSystem(Configuration conf) throws + IOException { + String fsname = conf.getTrimmed(TestS3AFileSystemContract.TEST_FS_S3A_NAME, ""); + + + boolean liveTest = !StringUtils.isEmpty(fsname); + URI testURI = null; + if (liveTest) { + testURI = URI.create(fsname); + liveTest = testURI.getScheme().equals(Constants.FS_S3A); + } + if (!liveTest) { + // This doesn't work with our JUnit 3 style test cases, so instead we'll + // make this whole class not run by default + throw new AssumptionViolatedException( + "No test filesystem in " + TestS3AFileSystemContract.TEST_FS_S3A_NAME); + } + S3AFileSystem fs1 = new S3AFileSystem(); + fs1.initialize(testURI, conf); + return fs1; + } +} diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AFileSystemContract.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AFileSystemContract.java new file mode 100644 index 00000000000..5c88358e5e2 --- /dev/null +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AFileSystemContract.java @@ -0,0 +1,105 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.s3a; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.FileSystemContractBaseTest; +import org.apache.hadoop.fs.Path; + +/** + * Tests a live S3 system. If your keys and bucket aren't specified, all tests + * are marked as passed. + * + * This uses BlockJUnit4ClassRunner because FileSystemContractBaseTest from + * TestCase which uses the old Junit3 runner that doesn't ignore assumptions + * properly making it impossible to skip the tests if we don't have a valid + * bucket. + **/ +public class TestS3AFileSystemContract extends FileSystemContractBaseTest { + + protected static final Logger LOG = + LoggerFactory.getLogger(TestS3AFileSystemContract.class); + public static final String TEST_FS_S3A_NAME = "test.fs.s3a.name"; + + @Override + public void setUp() throws Exception { + Configuration conf = new Configuration(); + + fs = S3ATestUtils.createTestFileSystem(conf); + super.setUp(); + } + + @Override + protected void tearDown() throws Exception { + if (fs != null) { + fs.delete(path("test"), true); + } + super.tearDown(); + } + + @Override + public void testMkdirsWithUmask() throws Exception { + // not supported + } + + @Override + public void testRenameFileAsExistingFile() throws Exception { + if (!renameSupported()) return; + + Path src = path("/test/hadoop/file"); + createFile(src); + Path dst = path("/test/new/newfile"); + createFile(dst); + // s3 doesn't support rename option + // rename-overwrites-dest is always allowed. + rename(src, dst, true, false, true); + } + + @Override + public void testRenameDirectoryAsExistingDirectory() throws Exception { + if (!renameSupported()) { + return; + } + + Path src = path("/test/hadoop/dir"); + fs.mkdirs(src); + createFile(path("/test/hadoop/dir/file1")); + createFile(path("/test/hadoop/dir/subdir/file2")); + + Path dst = path("/test/new/newdir"); + fs.mkdirs(dst); + rename(src, dst, true, false, true); + assertFalse("Nested file1 exists", + fs.exists(path("/test/hadoop/dir/file1"))); + assertFalse("Nested file2 exists", + fs.exists(path("/test/hadoop/dir/subdir/file2"))); + assertTrue("Renamed nested file1 exists", + fs.exists(path("/test/new/newdir/file1"))); + assertTrue("Renamed nested exists", + fs.exists(path("/test/new/newdir/subdir/file2"))); + } + +// @Override + public void testMoveDirUnderParent() throws Throwable { + // not support because + // Fails if dst is a directory that is not empty. + } +} diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/S3AScaleTestBase.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/S3AScaleTestBase.java new file mode 100644 index 00000000000..e0cbc92f5c9 --- /dev/null +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/S3AScaleTestBase.java @@ -0,0 +1,89 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.s3a.scale; + +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.contract.ContractTestUtils; +import org.apache.hadoop.fs.s3a.S3AFileSystem; +import org.apache.hadoop.fs.Path; + +import org.apache.hadoop.fs.s3a.S3ATestUtils; +import org.junit.After; +import org.junit.Before; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import java.io.IOException; +import java.net.URI; + +import static org.junit.Assume.assumeTrue; + +/** + * Base class for scale tests; here is where the common scale configuration + * keys are defined + */ +public class S3AScaleTestBase { + + public static final String SCALE_TEST = "scale.test."; + public static final String KEY_OPERATION_COUNT = + SCALE_TEST + "operation.count"; + public static final long DEFAULT_OPERATION_COUNT = 2005; + + protected S3AFileSystem fs; + private static final Logger LOG = + LoggerFactory.getLogger(S3AScaleTestBase.class); + + private Configuration conf; + + /** + * Configuration generator. May be overridden to inject + * some custom options + * @return a configuration with which to create FS instances + */ + protected Configuration createConfiguration() { + return new Configuration(); + } + + /** + * Get the configuration used to set up the FS + * @return the configuration + */ + public Configuration getConf() { + return conf; + } + + @Before + public void setUp() throws Exception { + conf = createConfiguration(); + fs = S3ATestUtils.createTestFileSystem(conf); + } + + @After + public void tearDown() throws Exception { + ContractTestUtils.rm(fs, getTestPath(), true, true); + } + + protected Path getTestPath() { + return new Path("/tests3a"); + } + + protected long getOperationCount() { + return getConf().getLong(KEY_OPERATION_COUNT, DEFAULT_OPERATION_COUNT); + } +} diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/TestS3ADeleteManyFiles.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/TestS3ADeleteManyFiles.java new file mode 100644 index 00000000000..c913a67a4a9 --- /dev/null +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/TestS3ADeleteManyFiles.java @@ -0,0 +1,131 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.s3a.scale; + +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.contract.ContractTestUtils; +import org.junit.Rule; +import org.junit.Test; +import org.junit.rules.Timeout; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import java.io.IOException; +import java.util.concurrent.Callable; +import java.util.concurrent.ExecutionException; +import java.util.concurrent.ExecutorCompletionService; +import java.util.concurrent.ExecutorService; +import java.util.concurrent.Executors; +import java.util.concurrent.Future; + +import static org.junit.Assert.assertEquals; + +public class TestS3ADeleteManyFiles extends S3AScaleTestBase { + private static final Logger LOG = + LoggerFactory.getLogger(TestS3ADeleteManyFiles.class); + + + @Rule + public Timeout testTimeout = new Timeout(30 * 60 * 1000); + + @Test + public void testBulkRenameAndDelete() throws Throwable { + final Path scaleTestDir = getTestPath(); + final Path srcDir = new Path(scaleTestDir, "src"); + final Path finalDir = new Path(scaleTestDir, "final"); + final long count = getOperationCount(); + ContractTestUtils.rm(fs, scaleTestDir, true, false); + + fs.mkdirs(srcDir); + fs.mkdirs(finalDir); + + int testBufferSize = fs.getConf() + .getInt(ContractTestUtils.IO_CHUNK_BUFFER_SIZE, + ContractTestUtils.DEFAULT_IO_CHUNK_BUFFER_SIZE); + // use Executor to speed up file creation + ExecutorService exec = Executors.newFixedThreadPool(16); + final ExecutorCompletionService completionService = + new ExecutorCompletionService(exec); + try { + final byte[] data = ContractTestUtils.dataset(testBufferSize, 'a', 'z'); + + for (int i = 0; i < count; ++i) { + final String fileName = "foo-" + i; + completionService.submit(new Callable() { + @Override + public Boolean call() throws IOException { + ContractTestUtils.createFile(fs, new Path(srcDir, fileName), + false, data); + return fs.exists(new Path(srcDir, fileName)); + } + }); + } + for (int i = 0; i < count; ++i) { + final Future future = completionService.take(); + try { + if (!future.get()) { + LOG.warn("cannot create file"); + } + } catch (ExecutionException e) { + LOG.warn("Error while uploading file", e.getCause()); + throw e; + } + } + } finally { + exec.shutdown(); + } + + int nSrcFiles = fs.listStatus(srcDir).length; + fs.rename(srcDir, finalDir); + assertEquals(nSrcFiles, fs.listStatus(finalDir).length); + ContractTestUtils.assertPathDoesNotExist(fs, "not deleted after rename", + new Path(srcDir, "foo-" + 0)); + ContractTestUtils.assertPathDoesNotExist(fs, "not deleted after rename", + new Path(srcDir, "foo-" + count / 2)); + ContractTestUtils.assertPathDoesNotExist(fs, "not deleted after rename", + new Path(srcDir, "foo-" + (count - 1))); + ContractTestUtils.assertPathExists(fs, "not renamed to dest dir", + new Path(finalDir, "foo-" + 0)); + ContractTestUtils.assertPathExists(fs, "not renamed to dest dir", + new Path(finalDir, "foo-" + count/2)); + ContractTestUtils.assertPathExists(fs, "not renamed to dest dir", + new Path(finalDir, "foo-" + (count-1))); + + ContractTestUtils.assertDeleted(fs, finalDir, true, false); + } + + @Test + public void testOpenCreate() throws IOException { + Path dir = new Path("/tests3a"); + ContractTestUtils.createAndVerifyFile(fs, dir, 1024); + ContractTestUtils.createAndVerifyFile(fs, dir, 5 * 1024 * 1024); + ContractTestUtils.createAndVerifyFile(fs, dir, 20 * 1024 * 1024); + + + /* + Enable to test the multipart upload + try { + ContractTestUtils.createAndVerifyFile(fs, dir, + (long)6 * 1024 * 1024 * 1024); + } catch (IOException e) { + fail(e.getMessage()); + } + */ + } +} diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/NativeS3FileSystemContractBaseTest.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/NativeS3FileSystemContractBaseTest.java index ac6b9ec3251..f215219aee9 100644 --- a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/NativeS3FileSystemContractBaseTest.java +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/NativeS3FileSystemContractBaseTest.java @@ -22,15 +22,17 @@ import java.io.InputStream; import java.net.URI; +import org.apache.commons.lang.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystemContractBaseTest; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.s3native.NativeS3FileSystem.NativeS3FsInputStream; +import org.junit.internal.AssumptionViolatedException; public abstract class NativeS3FileSystemContractBaseTest extends FileSystemContractBaseTest { - + public static final String KEY_TEST_FS = "test.fs.s3n.name"; private NativeFileSystemStore store; abstract NativeFileSystemStore getNativeFileSystemStore() throws IOException; @@ -40,7 +42,12 @@ protected void setUp() throws Exception { Configuration conf = new Configuration(); store = getNativeFileSystemStore(); fs = new NativeS3FileSystem(store); - fs.initialize(URI.create(conf.get("test.fs.s3n.name")), conf); + String fsname = conf.get(KEY_TEST_FS); + if (StringUtils.isEmpty(fsname)) { + throw new AssumptionViolatedException( + "No test FS defined in :" + KEY_TEST_FS); + } + fs.initialize(URI.create(fsname), conf); } @Override diff --git a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/TestJets3tNativeFileSystemStore.java b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/TestJets3tNativeFileSystemStore.java index b1078a45144..dbd476e4ecf 100644 --- a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/TestJets3tNativeFileSystemStore.java +++ b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3native/TestJets3tNativeFileSystemStore.java @@ -117,10 +117,13 @@ public void testMediumUpload() throws IOException, NoSuchAlgorithmException { writeRenameReadCompare(new Path("/test/medium"), 33554432); // 100 MB } + /* + Enable Multipart upload to run this test @Test public void testExtraLargeUpload() throws IOException, NoSuchAlgorithmException { // Multipart upload, multipart copy writeRenameReadCompare(new Path("/test/xlarge"), 5368709121L); // 5GB+1byte } + */ } diff --git a/hadoop-tools/hadoop-aws/src/test/resources/contract/s3a.xml b/hadoop-tools/hadoop-aws/src/test/resources/contract/s3a.xml index 4142471d17a..4f9c0818ffa 100644 --- a/hadoop-tools/hadoop-aws/src/test/resources/contract/s3a.xml +++ b/hadoop-tools/hadoop-aws/src/test/resources/contract/s3a.xml @@ -47,6 +47,11 @@ true + + fs.contract.rename-remove-dest-if-empty-dir + true + + fs.contract.supports-append false diff --git a/hadoop-tools/hadoop-aws/src/test/resources/core-site.xml b/hadoop-tools/hadoop-aws/src/test/resources/core-site.xml new file mode 100644 index 00000000000..3397769d3ac --- /dev/null +++ b/hadoop-tools/hadoop-aws/src/test/resources/core-site.xml @@ -0,0 +1,51 @@ + + + + + + + + + + hadoop.tmp.dir + target/build/test + A base for other temporary directories. + true + + + + + hadoop.security.authentication + simple + + + + + + + +