HADOOP-18482. ITestS3APrefetchingInputStream to skip if CSV test file unavailable (#4983)
Contributed by Danny Jones
This commit is contained in:
parent
caf58937a0
commit
5a90d98972
@ -92,11 +92,12 @@ public void teardown() throws Exception {
|
|||||||
|
|
||||||
private void openFS() throws Exception {
|
private void openFS() throws Exception {
|
||||||
Configuration conf = getConfiguration();
|
Configuration conf = getConfiguration();
|
||||||
|
String largeFileUri = S3ATestUtils.getCSVTestFile(conf);
|
||||||
|
|
||||||
largeFile = new Path(DEFAULT_CSVTEST_FILE);
|
largeFile = new Path(largeFileUri);
|
||||||
blockSize = conf.getInt(PREFETCH_BLOCK_SIZE_KEY, PREFETCH_BLOCK_DEFAULT_SIZE);
|
blockSize = conf.getInt(PREFETCH_BLOCK_SIZE_KEY, PREFETCH_BLOCK_DEFAULT_SIZE);
|
||||||
largeFileFS = new S3AFileSystem();
|
largeFileFS = new S3AFileSystem();
|
||||||
largeFileFS.initialize(new URI(DEFAULT_CSVTEST_FILE), getConfiguration());
|
largeFileFS.initialize(new URI(largeFileUri), getConfiguration());
|
||||||
FileStatus fileStatus = largeFileFS.getFileStatus(largeFile);
|
FileStatus fileStatus = largeFileFS.getFileStatus(largeFile);
|
||||||
largeFileSize = fileStatus.getLen();
|
largeFileSize = fileStatus.getLen();
|
||||||
numBlocks = calculateNumBlocks(largeFileSize, blockSize);
|
numBlocks = calculateNumBlocks(largeFileSize, blockSize);
|
||||||
|
Loading…
x
Reference in New Issue
Block a user