HADOOP-16269. ABFS: add listFileStatus with StartFrom.
Author: Da Zhou
This commit is contained in:
parent
006ae258b3
commit
dd636127e9
|
@ -31,6 +31,7 @@ import java.nio.charset.CharacterCodingException;
|
||||||
import java.nio.charset.Charset;
|
import java.nio.charset.Charset;
|
||||||
import java.nio.charset.CharsetDecoder;
|
import java.nio.charset.CharsetDecoder;
|
||||||
import java.nio.charset.CharsetEncoder;
|
import java.nio.charset.CharsetEncoder;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
import java.text.ParseException;
|
import java.text.ParseException;
|
||||||
import java.text.SimpleDateFormat;
|
import java.text.SimpleDateFormat;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
|
@ -46,6 +47,7 @@ import java.util.Set;
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
|
import com.google.common.base.Strings;
|
||||||
|
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.classification.InterfaceStability;
|
import org.apache.hadoop.classification.InterfaceStability;
|
||||||
|
@ -79,6 +81,7 @@ import org.apache.hadoop.fs.azurebfs.services.AuthType;
|
||||||
import org.apache.hadoop.fs.azurebfs.services.ExponentialRetryPolicy;
|
import org.apache.hadoop.fs.azurebfs.services.ExponentialRetryPolicy;
|
||||||
import org.apache.hadoop.fs.azurebfs.services.SharedKeyCredentials;
|
import org.apache.hadoop.fs.azurebfs.services.SharedKeyCredentials;
|
||||||
import org.apache.hadoop.fs.azurebfs.utils.Base64;
|
import org.apache.hadoop.fs.azurebfs.utils.Base64;
|
||||||
|
import org.apache.hadoop.fs.azurebfs.utils.CRC64;
|
||||||
import org.apache.hadoop.fs.azurebfs.utils.UriUtils;
|
import org.apache.hadoop.fs.azurebfs.utils.UriUtils;
|
||||||
import org.apache.hadoop.fs.permission.AclEntry;
|
import org.apache.hadoop.fs.permission.AclEntry;
|
||||||
import org.apache.hadoop.fs.permission.AclStatus;
|
import org.apache.hadoop.fs.permission.AclStatus;
|
||||||
|
@ -89,7 +92,17 @@ import org.apache.http.client.utils.URIBuilder;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_EQUALS;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_FORWARD_SLASH;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_HYPHEN;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_PLUS;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_STAR;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_UNDERSCORE;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.ROOT_PATH;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.SINGLE_WHITE_SPACE;
|
||||||
|
import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.TOKEN_VERSION;
|
||||||
import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.AZURE_ABFS_ENDPOINT;
|
import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.AZURE_ABFS_ENDPOINT;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Provides the bridging logic between Hadoop's abstract filesystem and Azure Storage.
|
* Provides the bridging logic between Hadoop's abstract filesystem and Azure Storage.
|
||||||
*/
|
*/
|
||||||
|
@ -103,6 +116,7 @@ public class AzureBlobFileSystemStore {
|
||||||
private String userName;
|
private String userName;
|
||||||
private String primaryUserGroup;
|
private String primaryUserGroup;
|
||||||
private static final String DATE_TIME_PATTERN = "E, dd MMM yyyy HH:mm:ss 'GMT'";
|
private static final String DATE_TIME_PATTERN = "E, dd MMM yyyy HH:mm:ss 'GMT'";
|
||||||
|
private static final String TOKEN_DATE_PATTERN = "yyyy-MM-dd'T'HH:mm:ss.SSSSSSS'Z'";
|
||||||
private static final String XMS_PROPERTIES_ENCODING = "ISO-8859-1";
|
private static final String XMS_PROPERTIES_ENCODING = "ISO-8859-1";
|
||||||
private static final int LIST_MAX_RESULTS = 500;
|
private static final int LIST_MAX_RESULTS = 500;
|
||||||
|
|
||||||
|
@ -514,15 +528,43 @@ public class AzureBlobFileSystemStore {
|
||||||
eTag);
|
eTag);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param path The list path.
|
||||||
|
* @return the entries in the path.
|
||||||
|
* */
|
||||||
public FileStatus[] listStatus(final Path path) throws IOException {
|
public FileStatus[] listStatus(final Path path) throws IOException {
|
||||||
LOG.debug("listStatus filesystem: {} path: {}",
|
return listStatus(path, null);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param path Path the list path.
|
||||||
|
* @param startFrom the entry name that list results should start with.
|
||||||
|
* For example, if folder "/folder" contains four files: "afile", "bfile", "hfile", "ifile".
|
||||||
|
* Then listStatus(Path("/folder"), "hfile") will return "/folder/hfile" and "folder/ifile"
|
||||||
|
* Notice that if startFrom is a non-existent entry name, then the list response contains
|
||||||
|
* all entries after this non-existent entry in lexical order:
|
||||||
|
* listStatus(Path("/folder"), "cfile") will return "/folder/hfile" and "/folder/ifile".
|
||||||
|
*
|
||||||
|
* @return the entries in the path start from "startFrom" in lexical order.
|
||||||
|
* */
|
||||||
|
@InterfaceStability.Unstable
|
||||||
|
public FileStatus[] listStatus(final Path path, final String startFrom) throws IOException {
|
||||||
|
LOG.debug("listStatus filesystem: {} path: {}, startFrom: {}",
|
||||||
client.getFileSystem(),
|
client.getFileSystem(),
|
||||||
path);
|
path,
|
||||||
|
startFrom);
|
||||||
|
|
||||||
String relativePath = path.isRoot() ? AbfsHttpConstants.EMPTY_STRING : getRelativePath(path);
|
final String relativePath = path.isRoot() ? AbfsHttpConstants.EMPTY_STRING : getRelativePath(path);
|
||||||
String continuation = null;
|
String continuation = null;
|
||||||
ArrayList<FileStatus> fileStatuses = new ArrayList<>();
|
|
||||||
|
|
||||||
|
// generate continuation token if a valid startFrom is provided.
|
||||||
|
if (startFrom != null && !startFrom.isEmpty()) {
|
||||||
|
continuation = getIsNamespaceEnabled()
|
||||||
|
? generateContinuationTokenForXns(startFrom)
|
||||||
|
: generateContinuationTokenForNonXns(path.isRoot() ? ROOT_PATH : relativePath, startFrom);
|
||||||
|
}
|
||||||
|
|
||||||
|
ArrayList<FileStatus> fileStatuses = new ArrayList<>();
|
||||||
do {
|
do {
|
||||||
AbfsRestOperation op = client.listPath(relativePath, false, LIST_MAX_RESULTS, continuation);
|
AbfsRestOperation op = client.listPath(relativePath, false, LIST_MAX_RESULTS, continuation);
|
||||||
continuation = op.getResult().getResponseHeader(HttpHeaderConfigurations.X_MS_CONTINUATION);
|
continuation = op.getResult().getResponseHeader(HttpHeaderConfigurations.X_MS_CONTINUATION);
|
||||||
|
@ -575,6 +617,61 @@ public class AzureBlobFileSystemStore {
|
||||||
return fileStatuses.toArray(new FileStatus[fileStatuses.size()]);
|
return fileStatuses.toArray(new FileStatus[fileStatuses.size()]);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// generate continuation token for xns account
|
||||||
|
private String generateContinuationTokenForXns(final String firstEntryName) {
|
||||||
|
Preconditions.checkArgument(!Strings.isNullOrEmpty(firstEntryName)
|
||||||
|
&& !firstEntryName.startsWith(AbfsHttpConstants.ROOT_PATH),
|
||||||
|
"startFrom must be a dir/file name and it can not be a full path");
|
||||||
|
|
||||||
|
StringBuilder sb = new StringBuilder();
|
||||||
|
sb.append(firstEntryName).append("#$").append("0");
|
||||||
|
|
||||||
|
CRC64 crc64 = new CRC64();
|
||||||
|
StringBuilder token = new StringBuilder();
|
||||||
|
token.append(crc64.compute(sb.toString().getBytes(StandardCharsets.UTF_8)))
|
||||||
|
.append(SINGLE_WHITE_SPACE)
|
||||||
|
.append("0")
|
||||||
|
.append(SINGLE_WHITE_SPACE)
|
||||||
|
.append(firstEntryName);
|
||||||
|
|
||||||
|
return Base64.encode(token.toString().getBytes(StandardCharsets.UTF_8));
|
||||||
|
}
|
||||||
|
|
||||||
|
// generate continuation token for non-xns account
|
||||||
|
private String generateContinuationTokenForNonXns(final String path, final String firstEntryName) {
|
||||||
|
Preconditions.checkArgument(!Strings.isNullOrEmpty(firstEntryName)
|
||||||
|
&& !firstEntryName.startsWith(AbfsHttpConstants.ROOT_PATH),
|
||||||
|
"startFrom must be a dir/file name and it can not be a full path");
|
||||||
|
|
||||||
|
// Notice: non-xns continuation token requires full path (first "/" is not included) for startFrom
|
||||||
|
final String startFrom = (path.isEmpty() || path.equals(ROOT_PATH))
|
||||||
|
? firstEntryName
|
||||||
|
: path + ROOT_PATH + firstEntryName;
|
||||||
|
|
||||||
|
SimpleDateFormat simpleDateFormat = new SimpleDateFormat(TOKEN_DATE_PATTERN, Locale.US);
|
||||||
|
String date = simpleDateFormat.format(new Date());
|
||||||
|
String token = String.format("%06d!%s!%06d!%s!%06d!%s!",
|
||||||
|
path.length(), path, startFrom.length(), startFrom, date.length(), date);
|
||||||
|
String base64EncodedToken = Base64.encode(token.getBytes(StandardCharsets.UTF_8));
|
||||||
|
|
||||||
|
StringBuilder encodedTokenBuilder = new StringBuilder(base64EncodedToken.length() + 5);
|
||||||
|
encodedTokenBuilder.append(String.format("%s!%d!", TOKEN_VERSION, base64EncodedToken.length()));
|
||||||
|
|
||||||
|
for (int i = 0; i < base64EncodedToken.length(); i++) {
|
||||||
|
char current = base64EncodedToken.charAt(i);
|
||||||
|
if (CHAR_FORWARD_SLASH == current) {
|
||||||
|
current = CHAR_UNDERSCORE;
|
||||||
|
} else if (CHAR_PLUS == current) {
|
||||||
|
current = CHAR_STAR;
|
||||||
|
} else if (CHAR_EQUALS == current) {
|
||||||
|
current = CHAR_HYPHEN;
|
||||||
|
}
|
||||||
|
encodedTokenBuilder.append(current);
|
||||||
|
}
|
||||||
|
|
||||||
|
return encodedTokenBuilder.toString();
|
||||||
|
}
|
||||||
|
|
||||||
public void setOwner(final Path path, final String owner, final String group) throws
|
public void setOwner(final Path path, final String owner, final String group) throws
|
||||||
AzureBlobFileSystemException {
|
AzureBlobFileSystemException {
|
||||||
if (!getIsNamespaceEnabled()) {
|
if (!getIsNamespaceEnabled()) {
|
||||||
|
@ -992,7 +1089,7 @@ public class AzureBlobFileSystemStore {
|
||||||
|
|
||||||
FileStatus other = (FileStatus) obj;
|
FileStatus other = (FileStatus) obj;
|
||||||
|
|
||||||
if (!other.equals(this)) {// compare the path
|
if (!this.getPath().equals(other.getPath())) {// compare the path
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -39,6 +39,7 @@ public final class AbfsHttpConstants {
|
||||||
public static final String GET_ACCESS_CONTROL = "getAccessControl";
|
public static final String GET_ACCESS_CONTROL = "getAccessControl";
|
||||||
public static final String GET_STATUS = "getStatus";
|
public static final String GET_STATUS = "getStatus";
|
||||||
public static final String DEFAULT_TIMEOUT = "90";
|
public static final String DEFAULT_TIMEOUT = "90";
|
||||||
|
public static final String TOKEN_VERSION = "2";
|
||||||
|
|
||||||
public static final String JAVA_VERSION = "java.version";
|
public static final String JAVA_VERSION = "java.version";
|
||||||
public static final String OS_NAME = "os.name";
|
public static final String OS_NAME = "os.name";
|
||||||
|
@ -91,5 +92,13 @@ public final class AbfsHttpConstants {
|
||||||
public static final String PERMISSION_FORMAT = "%04d";
|
public static final String PERMISSION_FORMAT = "%04d";
|
||||||
public static final String SUPER_USER = "$superuser";
|
public static final String SUPER_USER = "$superuser";
|
||||||
|
|
||||||
|
public static final char CHAR_FORWARD_SLASH = '/';
|
||||||
|
public static final char CHAR_EXCLAMATION_POINT = '!';
|
||||||
|
public static final char CHAR_UNDERSCORE = '_';
|
||||||
|
public static final char CHAR_HYPHEN = '-';
|
||||||
|
public static final char CHAR_EQUALS = '=';
|
||||||
|
public static final char CHAR_STAR = '*';
|
||||||
|
public static final char CHAR_PLUS = '+';
|
||||||
|
|
||||||
private AbfsHttpConstants() {}
|
private AbfsHttpConstants() {}
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,60 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.fs.azurebfs.utils;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* CRC64 implementation for AzureBlobFileSystem.
|
||||||
|
*/
|
||||||
|
public class CRC64 {
|
||||||
|
|
||||||
|
private static final long POLY = 0x9a6c9329ac4bc9b5L;
|
||||||
|
private static final int TABLE_LENGTH = 256;
|
||||||
|
private static final long[] TABLE = new long[TABLE_LENGTH];
|
||||||
|
|
||||||
|
private long value = -1;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param input byte arrays.
|
||||||
|
* @return long value of the CRC-64 checksum of the data.
|
||||||
|
* */
|
||||||
|
public long compute(byte[] input) {
|
||||||
|
init();
|
||||||
|
for (int i = 0; i < input.length; i++) {
|
||||||
|
value = TABLE[(input[i] ^ (int) value) & 0xFF] ^ (value >>> 8);
|
||||||
|
}
|
||||||
|
return ~value;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Initialize a table constructed from POLY (0x9a6c9329ac4bc9b5L).
|
||||||
|
* */
|
||||||
|
private void init() {
|
||||||
|
value = -1;
|
||||||
|
for (int n = 0; n < TABLE_LENGTH; ++n) {
|
||||||
|
long crc = n;
|
||||||
|
for (int i = 0; i < 8; ++i) {
|
||||||
|
if ((crc & 1) == 1) {
|
||||||
|
crc = (crc >>> 1) ^ POLY;
|
||||||
|
} else {
|
||||||
|
crc >>>= 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
TABLE[n] = crc;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -24,7 +24,6 @@ import java.util.Hashtable;
|
||||||
import java.util.UUID;
|
import java.util.UUID;
|
||||||
import java.util.concurrent.Callable;
|
import java.util.concurrent.Callable;
|
||||||
|
|
||||||
import com.google.common.base.Preconditions;
|
|
||||||
import org.junit.After;
|
import org.junit.After;
|
||||||
import org.junit.Before;
|
import org.junit.Before;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
|
@ -210,9 +209,9 @@ public abstract class AbstractAbfsIntegrationTest extends
|
||||||
* @throws IOException failure during create/init.
|
* @throws IOException failure during create/init.
|
||||||
*/
|
*/
|
||||||
public AzureBlobFileSystem createFileSystem() throws IOException {
|
public AzureBlobFileSystem createFileSystem() throws IOException {
|
||||||
Preconditions.checkState(abfs == null,
|
if (abfs == null) {
|
||||||
"existing ABFS instance exists: %s", abfs);
|
|
||||||
abfs = (AzureBlobFileSystem) FileSystem.newInstance(rawConfig);
|
abfs = (AzureBlobFileSystem) FileSystem.newInstance(rawConfig);
|
||||||
|
}
|
||||||
return abfs;
|
return abfs;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,151 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.fs.azurebfs;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.Arrays;
|
||||||
|
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.junit.runners.Parameterized;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.contract.ContractTestUtils;
|
||||||
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test AzureBlobFileSystemStore listStatus with startFrom.
|
||||||
|
* */
|
||||||
|
@RunWith(Parameterized.class)
|
||||||
|
public class ITestAzureBlobFileSystemStoreListStatusWithRange extends
|
||||||
|
AbstractAbfsIntegrationTest {
|
||||||
|
private static final boolean SUCCEED = true;
|
||||||
|
private static final boolean FAIL = false;
|
||||||
|
private static final String[] SORTED_ENTRY_NAMES = {"1_folder", "A0", "D01", "a+", "c0", "name5"};
|
||||||
|
|
||||||
|
private AzureBlobFileSystemStore store;
|
||||||
|
private AzureBlobFileSystem fs;
|
||||||
|
|
||||||
|
@Parameterized.Parameter
|
||||||
|
public String path;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A valid startFrom for listFileStatus with range is a non-fully qualified dir/file name
|
||||||
|
* */
|
||||||
|
@Parameterized.Parameter(1)
|
||||||
|
public String startFrom;
|
||||||
|
|
||||||
|
@Parameterized.Parameter(2)
|
||||||
|
public int expectedStartIndexInArray;
|
||||||
|
|
||||||
|
@Parameterized.Parameter(3)
|
||||||
|
public boolean expectedResult;
|
||||||
|
|
||||||
|
@Parameterized.Parameters(name = "Testing path \"{0}\", startFrom: \"{1}\", Expecting result : {3}") // Test path
|
||||||
|
public static Iterable<Object[]> params() {
|
||||||
|
return Arrays.asList(
|
||||||
|
new Object[][]{
|
||||||
|
// case 0: list in root, without range
|
||||||
|
{"/", null, 0, SUCCEED},
|
||||||
|
|
||||||
|
// case 1: list in the root, start from the second file
|
||||||
|
{"/", SORTED_ENTRY_NAMES[1], 1, SUCCEED},
|
||||||
|
|
||||||
|
// case 2: list in the root, invalid startFrom
|
||||||
|
{"/", "/", -1, FAIL},
|
||||||
|
|
||||||
|
// case 3: list in non-root level, valid startFrom : dir name
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], SORTED_ENTRY_NAMES[1], 1, SUCCEED},
|
||||||
|
|
||||||
|
// case 4: list in non-root level, valid startFrom : file name
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], SORTED_ENTRY_NAMES[2], 2, SUCCEED},
|
||||||
|
|
||||||
|
// case 5: list in non root level, invalid startFrom
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], "/" + SORTED_ENTRY_NAMES[3], -1, FAIL},
|
||||||
|
|
||||||
|
// case 6: list using non existent startFrom, startFrom is smaller than the entries in lexical order
|
||||||
|
// expecting return all entries
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], "0-non-existent", 0, SUCCEED},
|
||||||
|
|
||||||
|
// case 7: list using non existent startFrom, startFrom is larger than the entries in lexical order
|
||||||
|
// expecting return 0 entries
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], "z-non-existent", -1, SUCCEED},
|
||||||
|
|
||||||
|
// case 8: list using non existent startFrom, startFrom is in the range
|
||||||
|
{"/" + SORTED_ENTRY_NAMES[2], "A1", 2, SUCCEED}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
public ITestAzureBlobFileSystemStoreListStatusWithRange() throws Exception {
|
||||||
|
super();
|
||||||
|
if (this.getFileSystem() == null) {
|
||||||
|
super.createFileSystem();
|
||||||
|
}
|
||||||
|
fs = this.getFileSystem();
|
||||||
|
store = fs.getAbfsStore();
|
||||||
|
prepareTestFiles();
|
||||||
|
// Sort the names for verification, ABFS service should return the results in order.
|
||||||
|
Arrays.sort(SORTED_ENTRY_NAMES);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testListWithRange() throws IOException {
|
||||||
|
try {
|
||||||
|
FileStatus[] listResult = store.listStatus(new Path(path), startFrom);
|
||||||
|
if (!expectedResult) {
|
||||||
|
Assert.fail("Excepting failure with IllegalArgumentException");
|
||||||
|
}
|
||||||
|
verifyFileStatus(listResult, new Path(path), expectedStartIndexInArray);
|
||||||
|
} catch (IllegalArgumentException ex) {
|
||||||
|
if (expectedResult) {
|
||||||
|
Assert.fail("Excepting success");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// compare the file status
|
||||||
|
private void verifyFileStatus(FileStatus[] listResult, Path parentPath, int startIndexInSortedName) throws IOException {
|
||||||
|
if (startIndexInSortedName == -1) {
|
||||||
|
Assert.assertEquals("Expected empty FileStatus array", 0, listResult.length);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
FileStatus[] allFileStatuses = fs.listStatus(parentPath);
|
||||||
|
Assert.assertEquals("number of dir/file doesn't match",
|
||||||
|
SORTED_ENTRY_NAMES.length, allFileStatuses.length);
|
||||||
|
int indexInResult = 0;
|
||||||
|
for (int index = startIndexInSortedName; index < SORTED_ENTRY_NAMES.length; index++) {
|
||||||
|
Assert.assertEquals("fileStatus doesn't match", allFileStatuses[index], listResult[indexInResult++]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void prepareTestFiles() throws IOException {
|
||||||
|
final AzureBlobFileSystem fs = getFileSystem();
|
||||||
|
// created 2 level file structures
|
||||||
|
for (String levelOneFolder : SORTED_ENTRY_NAMES) {
|
||||||
|
Path levelOnePath = new Path("/" + levelOneFolder);
|
||||||
|
Assert.assertTrue(fs.mkdirs(levelOnePath));
|
||||||
|
for (String fileName : SORTED_ENTRY_NAMES) {
|
||||||
|
Path filePath = new Path(levelOnePath, fileName);
|
||||||
|
ContractTestUtils.touch(fs, filePath);
|
||||||
|
ContractTestUtils.assertIsFile(fs, filePath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,38 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.fs.azurebfs;
|
||||||
|
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.azurebfs.utils.CRC64;
|
||||||
|
/**
|
||||||
|
* Test for Crc64 in AzureBlobFileSystem, notice that ABFS CRC64 has its own polynomial.
|
||||||
|
* */
|
||||||
|
public class TestAbfsCrc64 {
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void tesCrc64Compute() {
|
||||||
|
CRC64 crc64 = new CRC64();
|
||||||
|
final String[] testStr = {"#$", "dir_2_ac83abee", "dir_42_976df1f5"};
|
||||||
|
final String[] expected = {"f91f7e6a837dbfa8", "203f9fefc38ae97b", "cc0d56eafe58a855"};
|
||||||
|
for (int i = 0; i < testStr.length; i++) {
|
||||||
|
Assert.assertEquals(expected[i], Long.toHexString(crc64.compute(testStr[i].getBytes())));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue