HDFS-6125. Merge 1581242 from trunk
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1581243 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
45d6277ec2
commit
590df94253
|
@ -11,6 +11,8 @@ Release 2.5.0 - UNRELEASED
|
||||||
HDFS-6007. Update documentation about short-circuit local reads (iwasakims
|
HDFS-6007. Update documentation about short-circuit local reads (iwasakims
|
||||||
via cmccabe)
|
via cmccabe)
|
||||||
|
|
||||||
|
HDFS-6125. Cleanup unnecessary cast in HDFS code base. (suresh)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
|
@ -203,7 +203,7 @@ public class EditLogLedgerMetadata {
|
||||||
hash = hash * 31 + (int) ledgerId;
|
hash = hash * 31 + (int) ledgerId;
|
||||||
hash = hash * 31 + (int) firstTxId;
|
hash = hash * 31 + (int) firstTxId;
|
||||||
hash = hash * 31 + (int) lastTxId;
|
hash = hash * 31 + (int) lastTxId;
|
||||||
hash = hash * 31 + (int) dataLayoutVersion;
|
hash = hash * 31 + dataLayoutVersion;
|
||||||
return hash;
|
return hash;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -143,7 +143,7 @@ public class DataStorage extends Storage {
|
||||||
*/
|
*/
|
||||||
public String getTrashDirectoryForBlockFile(String bpid, File blockFile) {
|
public String getTrashDirectoryForBlockFile(String bpid, File blockFile) {
|
||||||
if (trashEnabledBpids.contains(bpid)) {
|
if (trashEnabledBpids.contains(bpid)) {
|
||||||
return ((BlockPoolSliceStorage) getBPStorage(bpid)).getTrashDirectory(blockFile);
|
return getBPStorage(bpid).getTrashDirectory(blockFile);
|
||||||
}
|
}
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
|
@ -308,7 +308,7 @@ class ClusterJspHelper {
|
||||||
// Inner map of attribute name to value
|
// Inner map of attribute name to value
|
||||||
Map<String, Object> innerMap = entry.getValue();
|
Map<String, Object> innerMap = entry.getValue();
|
||||||
if (innerMap != null) {
|
if (innerMap != null) {
|
||||||
if (((String) innerMap.get("adminState"))
|
if (innerMap.get("adminState")
|
||||||
.equals(AdminStates.DECOMMISSIONED.toString())) {
|
.equals(AdminStates.DECOMMISSIONED.toString())) {
|
||||||
nn.liveDecomCount++;
|
nn.liveDecomCount++;
|
||||||
}
|
}
|
||||||
|
|
|
@ -2396,7 +2396,7 @@ public class FSDirectory implements Closeable {
|
||||||
*/
|
*/
|
||||||
public final void addToInodeMap(INode inode) {
|
public final void addToInodeMap(INode inode) {
|
||||||
if (inode instanceof INodeWithAdditionalFields) {
|
if (inode instanceof INodeWithAdditionalFields) {
|
||||||
inodeMap.put((INodeWithAdditionalFields)inode);
|
inodeMap.put(inode);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -3507,7 +3507,7 @@ public abstract class FSEditLogOp {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
void readFields(DataInputStream in, int logVersion) throws IOException {
|
void readFields(DataInputStream in, int logVersion) throws IOException {
|
||||||
AclEditLogProto p = AclEditLogProto.parseDelimitedFrom((DataInputStream)in);
|
AclEditLogProto p = AclEditLogProto.parseDelimitedFrom(in);
|
||||||
if (p == null) {
|
if (p == null) {
|
||||||
throw new IOException("Failed to read fields from SetAclOp");
|
throw new IOException("Failed to read fields from SetAclOp");
|
||||||
}
|
}
|
||||||
|
|
|
@ -3911,7 +3911,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
|
||||||
assert hasReadLock();
|
assert hasReadLock();
|
||||||
final BlockCollection bc = blockUC.getBlockCollection();
|
final BlockCollection bc = blockUC.getBlockCollection();
|
||||||
if (bc == null || !(bc instanceof INodeFile)
|
if (bc == null || !(bc instanceof INodeFile)
|
||||||
|| !((INodeFile) bc).isUnderConstruction()) {
|
|| !bc.isUnderConstruction()) {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -55,7 +55,7 @@ public class FileChecksumServlets {
|
||||||
HttpServletRequest request, NameNode nn)
|
HttpServletRequest request, NameNode nn)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
final String hostname = host instanceof DatanodeInfo
|
final String hostname = host instanceof DatanodeInfo
|
||||||
? ((DatanodeInfo)host).getHostName() : host.getIpAddr();
|
? host.getHostName() : host.getIpAddr();
|
||||||
final String scheme = request.getScheme();
|
final String scheme = request.getScheme();
|
||||||
int port = host.getInfoPort();
|
int port = host.getInfoPort();
|
||||||
if ("https".equals(scheme)) {
|
if ("https".equals(scheme)) {
|
||||||
|
|
|
@ -56,7 +56,7 @@ public class FileDataServlet extends DfsServlet {
|
||||||
final DatanodeID host = pickSrcDatanode(blks, status, conf);
|
final DatanodeID host = pickSrcDatanode(blks, status, conf);
|
||||||
final String hostname;
|
final String hostname;
|
||||||
if (host instanceof DatanodeInfo) {
|
if (host instanceof DatanodeInfo) {
|
||||||
hostname = ((DatanodeInfo)host).getHostName();
|
hostname = host.getHostName();
|
||||||
} else {
|
} else {
|
||||||
hostname = host.getIpAddr();
|
hostname = host.getIpAddr();
|
||||||
}
|
}
|
||||||
|
|
|
@ -128,7 +128,7 @@ public class ConfiguredFailoverProxyProvider<T> implements
|
||||||
throw new RuntimeException(e);
|
throw new RuntimeException(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return new ProxyInfo<T>((T)current.namenode, current.address.toString());
|
return new ProxyInfo<T>(current.namenode, current.address.toString());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
@ -414,7 +414,7 @@ public class FSImageFormatPBSnapshot {
|
||||||
rb.setLastSnapshotId(((WithName) ref).getLastSnapshotId()).setName(
|
rb.setLastSnapshotId(((WithName) ref).getLastSnapshotId()).setName(
|
||||||
ByteString.copyFrom(ref.getLocalNameBytes()));
|
ByteString.copyFrom(ref.getLocalNameBytes()));
|
||||||
} else if (ref instanceof DstReference) {
|
} else if (ref instanceof DstReference) {
|
||||||
rb.setDstSnapshotId(((DstReference) ref).getDstSnapshotId());
|
rb.setDstSnapshotId(ref.getDstSnapshotId());
|
||||||
}
|
}
|
||||||
return rb;
|
return rb;
|
||||||
}
|
}
|
||||||
|
|
|
@ -346,8 +346,7 @@ public class DelegationTokenFetcher {
|
||||||
Class.forName(exceptionClass).asSubclass(Exception.class);
|
Class.forName(exceptionClass).asSubclass(Exception.class);
|
||||||
// we are interested in constructor with String arguments
|
// we are interested in constructor with String arguments
|
||||||
java.lang.reflect.Constructor<? extends Exception> constructor =
|
java.lang.reflect.Constructor<? extends Exception> constructor =
|
||||||
(java.lang.reflect.Constructor<? extends Exception>)
|
ec.getConstructor (new Class[] {String.class});
|
||||||
ec.getConstructor (new Class[] {String.class});
|
|
||||||
|
|
||||||
// create an instance
|
// create an instance
|
||||||
e = (IOException) constructor.newInstance (exceptionMsg);
|
e = (IOException) constructor.newInstance (exceptionMsg);
|
||||||
|
|
|
@ -317,7 +317,7 @@ public class JsonUtil {
|
||||||
if (value == null) {
|
if (value == null) {
|
||||||
return defaultValue;
|
return defaultValue;
|
||||||
}
|
}
|
||||||
return (long) (Long) value;
|
return (Long) value;
|
||||||
}
|
}
|
||||||
|
|
||||||
private static String getString(Map<?, ?> m, String key,
|
private static String getString(Map<?, ?> m, String key,
|
||||||
|
|
|
@ -62,7 +62,7 @@ public class FileAppendTest4 {
|
||||||
conf = new HdfsConfiguration();
|
conf = new HdfsConfiguration();
|
||||||
init(conf);
|
init(conf);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
||||||
fs = (DistributedFileSystem)cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@AfterClass
|
@AfterClass
|
||||||
|
|
|
@ -50,7 +50,7 @@ public class TestBalancerBandwidth {
|
||||||
try {
|
try {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
|
|
||||||
DistributedFileSystem fs = (DistributedFileSystem) cluster.getFileSystem();
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
|
||||||
ArrayList<DataNode> datanodes = cluster.getDataNodes();
|
ArrayList<DataNode> datanodes = cluster.getDataNodes();
|
||||||
// Ensure value from the configuration is reflected in the datanodes.
|
// Ensure value from the configuration is reflected in the datanodes.
|
||||||
|
|
|
@ -54,7 +54,7 @@ public class TestBlockMissingException {
|
||||||
try {
|
try {
|
||||||
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(NUM_DATANODES).build();
|
dfs = new MiniDFSCluster.Builder(conf).numDataNodes(NUM_DATANODES).build();
|
||||||
dfs.waitActive();
|
dfs.waitActive();
|
||||||
fileSys = (DistributedFileSystem)dfs.getFileSystem();
|
fileSys = dfs.getFileSystem();
|
||||||
Path file1 = new Path("/user/dhruba/raidtest/file1");
|
Path file1 = new Path("/user/dhruba/raidtest/file1");
|
||||||
createOldFile(fileSys, file1, 1, numBlocks, blockSize);
|
createOldFile(fileSys, file1, 1, numBlocks, blockSize);
|
||||||
|
|
||||||
|
|
|
@ -77,7 +77,7 @@ public class TestClientReportBadBlock {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDataNodes)
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDataNodes)
|
||||||
.build();
|
.build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
dfs = (DistributedFileSystem) cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
buffersize = conf.getInt(CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096);
|
buffersize = conf.getInt(CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -92,7 +92,7 @@ public class TestDFSMkdirs {
|
||||||
public void testMkdir() throws IOException {
|
public void testMkdir() throws IOException {
|
||||||
Configuration conf = new HdfsConfiguration();
|
Configuration conf = new HdfsConfiguration();
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem) cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
try {
|
try {
|
||||||
// Create a dir in root dir, should succeed
|
// Create a dir in root dir, should succeed
|
||||||
assertTrue(dfs.mkdir(new Path("/mkdir-" + Time.now()),
|
assertTrue(dfs.mkdir(new Path("/mkdir-" + Time.now()),
|
||||||
|
|
|
@ -657,8 +657,7 @@ public class TestDFSShell {
|
||||||
|
|
||||||
// Test a simple codec. Extension based detection. We use
|
// Test a simple codec. Extension based detection. We use
|
||||||
// Bzip2 cause its non-native.
|
// Bzip2 cause its non-native.
|
||||||
CompressionCodec codec = (CompressionCodec)
|
CompressionCodec codec = ReflectionUtils.newInstance(BZip2Codec.class, conf);
|
||||||
ReflectionUtils.newInstance(BZip2Codec.class, conf);
|
|
||||||
String extension = codec.getDefaultExtension();
|
String extension = codec.getDefaultExtension();
|
||||||
Path p = new Path(root, "file." + extension);
|
Path p = new Path(root, "file." + extension);
|
||||||
OutputStream fout = new DataOutputStream(codec.createOutputStream(
|
OutputStream fout = new DataOutputStream(codec.createOutputStream(
|
||||||
|
@ -785,7 +784,7 @@ public class TestDFSShell {
|
||||||
public void testCount() throws Exception {
|
public void testCount() throws Exception {
|
||||||
Configuration conf = new HdfsConfiguration();
|
Configuration conf = new HdfsConfiguration();
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
FsShell shell = new FsShell();
|
FsShell shell = new FsShell();
|
||||||
shell.setConf(conf);
|
shell.setConf(conf);
|
||||||
|
|
||||||
|
@ -1488,7 +1487,7 @@ public class TestDFSShell {
|
||||||
try {
|
try {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).format(true)
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).format(true)
|
||||||
.build();
|
.build();
|
||||||
dfs = (DistributedFileSystem)cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
|
|
||||||
mkdir(dfs, root);
|
mkdir(dfs, root);
|
||||||
dfs.copyFromLocalFile(false, false, new Path(localf.getPath()), remotef);
|
dfs.copyFromLocalFile(false, false, new Path(localf.getPath()), remotef);
|
||||||
|
@ -1517,7 +1516,7 @@ public class TestDFSShell {
|
||||||
// Start the cluster again, but do not reformat, so prior files remain.
|
// Start the cluster again, but do not reformat, so prior files remain.
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).format(false)
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).format(false)
|
||||||
.build();
|
.build();
|
||||||
dfs = (DistributedFileSystem)cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
|
|
||||||
assertEquals(null, runner.run(1));
|
assertEquals(null, runner.run(1));
|
||||||
String corruptedcontent = runner.run(0, "-ignoreCrc");
|
String corruptedcontent = runner.run(0, "-ignoreCrc");
|
||||||
|
@ -1541,7 +1540,7 @@ public class TestDFSShell {
|
||||||
public void testLsr() throws Exception {
|
public void testLsr() throws Exception {
|
||||||
final Configuration conf = new HdfsConfiguration();
|
final Configuration conf = new HdfsConfiguration();
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
|
|
||||||
try {
|
try {
|
||||||
final String root = createTree(dfs, "lsr");
|
final String root = createTree(dfs, "lsr");
|
||||||
|
|
|
@ -443,7 +443,7 @@ public class TestDFSUpgradeFromImage {
|
||||||
.clusterId("testClusterId");
|
.clusterId("testClusterId");
|
||||||
cluster = bld.build();
|
cluster = bld.build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
DFSClient dfsClient = dfs.dfs;
|
DFSClient dfsClient = dfs.dfs;
|
||||||
//Safemode will be off only after upgrade is complete. Wait for it.
|
//Safemode will be off only after upgrade is complete. Wait for it.
|
||||||
while ( dfsClient.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_GET) ) {
|
while ( dfsClient.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_GET) ) {
|
||||||
|
|
|
@ -156,7 +156,7 @@ public class TestDecommission {
|
||||||
assertTrue("Not HDFS:"+fileSys.getUri(),
|
assertTrue("Not HDFS:"+fileSys.getUri(),
|
||||||
fileSys instanceof DistributedFileSystem);
|
fileSys instanceof DistributedFileSystem);
|
||||||
HdfsDataInputStream dis = (HdfsDataInputStream)
|
HdfsDataInputStream dis = (HdfsDataInputStream)
|
||||||
((DistributedFileSystem)fileSys).open(name);
|
fileSys.open(name);
|
||||||
Collection<LocatedBlock> dinfo = dis.getAllBlocks();
|
Collection<LocatedBlock> dinfo = dis.getAllBlocks();
|
||||||
for (LocatedBlock blk : dinfo) { // for each block
|
for (LocatedBlock blk : dinfo) { // for each block
|
||||||
int hasdown = 0;
|
int hasdown = 0;
|
||||||
|
|
|
@ -479,7 +479,7 @@ public class TestDistributedFileSystem {
|
||||||
current.getShortUserName() + "x", new String[]{"user"});
|
current.getShortUserName() + "x", new String[]{"user"});
|
||||||
|
|
||||||
try {
|
try {
|
||||||
((DistributedFileSystem) hdfs).getFileChecksum(new Path(
|
hdfs.getFileChecksum(new Path(
|
||||||
"/test/TestNonExistingFile"));
|
"/test/TestNonExistingFile"));
|
||||||
fail("Expecting FileNotFoundException");
|
fail("Expecting FileNotFoundException");
|
||||||
} catch (FileNotFoundException e) {
|
} catch (FileNotFoundException e) {
|
||||||
|
@ -490,7 +490,7 @@ public class TestDistributedFileSystem {
|
||||||
try {
|
try {
|
||||||
Path path = new Path("/test/TestExistingDir/");
|
Path path = new Path("/test/TestExistingDir/");
|
||||||
hdfs.mkdirs(path);
|
hdfs.mkdirs(path);
|
||||||
((DistributedFileSystem) hdfs).getFileChecksum(path);
|
hdfs.getFileChecksum(path);
|
||||||
fail("Expecting FileNotFoundException");
|
fail("Expecting FileNotFoundException");
|
||||||
} catch (FileNotFoundException e) {
|
} catch (FileNotFoundException e) {
|
||||||
assertTrue("Not throwing the intended exception message", e.getMessage()
|
assertTrue("Not throwing the intended exception message", e.getMessage()
|
||||||
|
|
|
@ -74,7 +74,7 @@ public class TestFileAppend3 {
|
||||||
conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, 512);
|
conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, 512);
|
||||||
buffersize = conf.getInt(CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096);
|
buffersize = conf.getInt(CommonConfigurationKeys.IO_FILE_BUFFER_SIZE_KEY, 4096);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
||||||
fs = (DistributedFileSystem)cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@AfterClass
|
@AfterClass
|
||||||
|
|
|
@ -62,7 +62,7 @@ public class TestFileCreationEmpty {
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
||||||
try {
|
try {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
|
|
||||||
// create a new file.
|
// create a new file.
|
||||||
TestFileCreation.createFile(dfs, new Path("/foo"), DATANODE_NUM);
|
TestFileCreation.createFile(dfs, new Path("/foo"), DATANODE_NUM);
|
||||||
|
|
|
@ -44,7 +44,7 @@ public class TestFileLengthOnClusterRestart {
|
||||||
HdfsDataInputStream in = null;
|
HdfsDataInputStream in = null;
|
||||||
try {
|
try {
|
||||||
Path path = new Path("/tmp/TestFileLengthOnClusterRestart", "test");
|
Path path = new Path("/tmp/TestFileLengthOnClusterRestart", "test");
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem) cluster
|
DistributedFileSystem dfs = cluster
|
||||||
.getFileSystem();
|
.getFileSystem();
|
||||||
FSDataOutputStream out = dfs.create(path);
|
FSDataOutputStream out = dfs.create(path);
|
||||||
int fileLength = 1030;
|
int fileLength = 1030;
|
||||||
|
|
|
@ -104,7 +104,7 @@ public class TestHFlush {
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(
|
||||||
2).build();
|
2).build();
|
||||||
DistributedFileSystem fileSystem =
|
DistributedFileSystem fileSystem =
|
||||||
(DistributedFileSystem)cluster.getFileSystem();
|
cluster.getFileSystem();
|
||||||
|
|
||||||
try {
|
try {
|
||||||
Path path = new Path(fName);
|
Path path = new Path(fName);
|
||||||
|
@ -198,7 +198,7 @@ public class TestHFlush {
|
||||||
.numDataNodes(replicas).build();
|
.numDataNodes(replicas).build();
|
||||||
// Make sure we work with DFS in order to utilize all its functionality
|
// Make sure we work with DFS in order to utilize all its functionality
|
||||||
DistributedFileSystem fileSystem =
|
DistributedFileSystem fileSystem =
|
||||||
(DistributedFileSystem)cluster.getFileSystem();
|
cluster.getFileSystem();
|
||||||
|
|
||||||
FSDataInputStream is;
|
FSDataInputStream is;
|
||||||
try {
|
try {
|
||||||
|
@ -278,7 +278,7 @@ public class TestHFlush {
|
||||||
|
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
||||||
try {
|
try {
|
||||||
DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
|
||||||
byte[] fileContents = AppendTestUtil.initBuffer(fileLen);
|
byte[] fileContents = AppendTestUtil.initBuffer(fileLen);
|
||||||
|
|
||||||
|
@ -328,7 +328,7 @@ public class TestHFlush {
|
||||||
|
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(DATANODE_NUM).build();
|
||||||
try {
|
try {
|
||||||
DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
|
||||||
// create a new file.
|
// create a new file.
|
||||||
FSDataOutputStream stm = AppendTestUtil.createFile(fs, p, DATANODE_NUM);
|
FSDataOutputStream stm = AppendTestUtil.createFile(fs, p, DATANODE_NUM);
|
||||||
|
|
|
@ -89,7 +89,7 @@ public class TestLeaseRecovery {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
|
|
||||||
//create a file
|
//create a file
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem)cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
String filestr = "/foo";
|
String filestr = "/foo";
|
||||||
Path filepath = new Path(filestr);
|
Path filepath = new Path(filestr);
|
||||||
DFSTestUtil.createFile(dfs, filepath, ORG_FILE_SIZE, REPLICATION_NUM, 0L);
|
DFSTestUtil.createFile(dfs, filepath, ORG_FILE_SIZE, REPLICATION_NUM, 0L);
|
||||||
|
|
|
@ -109,7 +109,7 @@ public class TestMiniDFSCluster {
|
||||||
conf.set(MiniDFSCluster.HDFS_MINIDFS_BASEDIR, c4Path);
|
conf.set(MiniDFSCluster.HDFS_MINIDFS_BASEDIR, c4Path);
|
||||||
MiniDFSCluster cluster4 = new MiniDFSCluster.Builder(conf).build();
|
MiniDFSCluster cluster4 = new MiniDFSCluster.Builder(conf).build();
|
||||||
try {
|
try {
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem) cluster4.getFileSystem();
|
DistributedFileSystem dfs = cluster4.getFileSystem();
|
||||||
dfs.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_ENTER);
|
dfs.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_ENTER);
|
||||||
cluster4.shutdown();
|
cluster4.shutdown();
|
||||||
} finally {
|
} finally {
|
||||||
|
|
|
@ -62,8 +62,8 @@ public class TestMissingBlocksAlert {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
|
|
||||||
final BlockManager bm = cluster.getNamesystem().getBlockManager();
|
final BlockManager bm = cluster.getNamesystem().getBlockManager();
|
||||||
DistributedFileSystem dfs =
|
DistributedFileSystem dfs =
|
||||||
(DistributedFileSystem) cluster.getFileSystem();
|
cluster.getFileSystem();
|
||||||
|
|
||||||
// create a normal file
|
// create a normal file
|
||||||
DFSTestUtil.createFile(dfs, new Path("/testMissingBlocksAlert/file1"),
|
DFSTestUtil.createFile(dfs, new Path("/testMissingBlocksAlert/file1"),
|
||||||
|
|
|
@ -63,7 +63,7 @@ public class TestPipelines {
|
||||||
@Before
|
@Before
|
||||||
public void startUpCluster() throws IOException {
|
public void startUpCluster() throws IOException {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL_FACTOR).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL_FACTOR).build();
|
||||||
fs = (DistributedFileSystem) cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@After
|
@After
|
||||||
|
|
|
@ -122,7 +122,7 @@ public class TestReplaceDatanodeOnFailure {
|
||||||
).racks(racks).numDataNodes(REPLICATION).build();
|
).racks(racks).numDataNodes(REPLICATION).build();
|
||||||
|
|
||||||
try {
|
try {
|
||||||
final DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();
|
final DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
final Path dir = new Path(DIR);
|
final Path dir = new Path(DIR);
|
||||||
|
|
||||||
final SlowWriter[] slowwriters = new SlowWriter[10];
|
final SlowWriter[] slowwriters = new SlowWriter[10];
|
||||||
|
@ -246,7 +246,7 @@ public class TestReplaceDatanodeOnFailure {
|
||||||
).numDataNodes(1).build();
|
).numDataNodes(1).build();
|
||||||
|
|
||||||
try {
|
try {
|
||||||
final DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();
|
final DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
final Path f = new Path(DIR, "testAppend");
|
final Path f = new Path(DIR, "testAppend");
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
|
@ -102,7 +102,7 @@ public class TestClientProtocolWithDelegationToken {
|
||||||
public Object run() throws Exception {
|
public Object run() throws Exception {
|
||||||
ClientProtocol proxy = null;
|
ClientProtocol proxy = null;
|
||||||
try {
|
try {
|
||||||
proxy = (ClientProtocol) RPC.getProxy(ClientProtocol.class,
|
proxy = RPC.getProxy(ClientProtocol.class,
|
||||||
ClientProtocol.versionID, addr, conf);
|
ClientProtocol.versionID, addr, conf);
|
||||||
proxy.getServerDefaults();
|
proxy.getServerDefaults();
|
||||||
} finally {
|
} finally {
|
||||||
|
|
|
@ -184,9 +184,9 @@ public class TestHeartbeatHandling {
|
||||||
DatanodeInfo[] recoveringNodes = recoveryCommand.getRecoveringBlocks()
|
DatanodeInfo[] recoveringNodes = recoveryCommand.getRecoveringBlocks()
|
||||||
.toArray(new BlockRecoveryCommand.RecoveringBlock[0])[0].getLocations();
|
.toArray(new BlockRecoveryCommand.RecoveringBlock[0])[0].getLocations();
|
||||||
assertEquals(3, recoveringNodes.length);
|
assertEquals(3, recoveringNodes.length);
|
||||||
assertEquals(recoveringNodes[0], (DatanodeInfo)dd1);
|
assertEquals(recoveringNodes[0], dd1);
|
||||||
assertEquals(recoveringNodes[1], (DatanodeInfo)dd2);
|
assertEquals(recoveringNodes[1], dd2);
|
||||||
assertEquals(recoveringNodes[2], (DatanodeInfo)dd3);
|
assertEquals(recoveringNodes[2], dd3);
|
||||||
|
|
||||||
// Test with one stale node.
|
// Test with one stale node.
|
||||||
dd1.setLastUpdate(System.currentTimeMillis());
|
dd1.setLastUpdate(System.currentTimeMillis());
|
||||||
|
@ -206,8 +206,8 @@ public class TestHeartbeatHandling {
|
||||||
.toArray(new BlockRecoveryCommand.RecoveringBlock[0])[0].getLocations();
|
.toArray(new BlockRecoveryCommand.RecoveringBlock[0])[0].getLocations();
|
||||||
assertEquals(2, recoveringNodes.length);
|
assertEquals(2, recoveringNodes.length);
|
||||||
// dd2 is skipped.
|
// dd2 is skipped.
|
||||||
assertEquals(recoveringNodes[0], (DatanodeInfo)dd1);
|
assertEquals(recoveringNodes[0], dd1);
|
||||||
assertEquals(recoveringNodes[1], (DatanodeInfo)dd3);
|
assertEquals(recoveringNodes[1], dd3);
|
||||||
|
|
||||||
// Test with all stale node.
|
// Test with all stale node.
|
||||||
dd1.setLastUpdate(System.currentTimeMillis() - 60 * 1000);
|
dd1.setLastUpdate(System.currentTimeMillis() - 60 * 1000);
|
||||||
|
@ -228,9 +228,9 @@ public class TestHeartbeatHandling {
|
||||||
// Only dd1 is included since it heart beated and hence its not stale
|
// Only dd1 is included since it heart beated and hence its not stale
|
||||||
// when the list of recovery blocks is constructed.
|
// when the list of recovery blocks is constructed.
|
||||||
assertEquals(3, recoveringNodes.length);
|
assertEquals(3, recoveringNodes.length);
|
||||||
assertEquals(recoveringNodes[0], (DatanodeInfo)dd1);
|
assertEquals(recoveringNodes[0], dd1);
|
||||||
assertEquals(recoveringNodes[1], (DatanodeInfo)dd2);
|
assertEquals(recoveringNodes[1], dd2);
|
||||||
assertEquals(recoveringNodes[2], (DatanodeInfo)dd3);
|
assertEquals(recoveringNodes[2], dd3);
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
namesystem.writeUnlock();
|
namesystem.writeUnlock();
|
||||||
|
|
|
@ -79,8 +79,8 @@ public class TestCachingStrategy {
|
||||||
LOG.debug("got fadvise(offset=" + offset + ", len=" + len +
|
LOG.debug("got fadvise(offset=" + offset + ", len=" + len +
|
||||||
",flags=" + flags + ")");
|
",flags=" + flags + ")");
|
||||||
if (flags == NativeIO.POSIX.POSIX_FADV_DONTNEED) {
|
if (flags == NativeIO.POSIX.POSIX_FADV_DONTNEED) {
|
||||||
for (int i = 0; i < (int)len; i++) {
|
for (int i = 0; i < len; i++) {
|
||||||
dropped[(int)(offset + i)] = true;
|
dropped[(offset + i)] = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -165,7 +165,7 @@ public class TestCachingStrategy {
|
||||||
}
|
}
|
||||||
byte buf[] = new byte[8196];
|
byte buf[] = new byte[8196];
|
||||||
while (length > 0) {
|
while (length > 0) {
|
||||||
int amt = (length > buf.length) ? (int)buf.length : (int)length;
|
int amt = (length > buf.length) ? buf.length : (int)length;
|
||||||
fos.write(buf, 0, amt);
|
fos.write(buf, 0, amt);
|
||||||
length -= amt;
|
length -= amt;
|
||||||
}
|
}
|
||||||
|
@ -189,7 +189,7 @@ public class TestCachingStrategy {
|
||||||
}
|
}
|
||||||
byte buf[] = new byte[8196];
|
byte buf[] = new byte[8196];
|
||||||
while (length > 0) {
|
while (length > 0) {
|
||||||
int amt = (length > buf.length) ? (int)buf.length : (int)length;
|
int amt = (length > buf.length) ? buf.length : (int)length;
|
||||||
int ret = fis.read(buf, 0, amt);
|
int ret = fis.read(buf, 0, amt);
|
||||||
if (ret == -1) {
|
if (ret == -1) {
|
||||||
return totalRead;
|
return totalRead;
|
||||||
|
|
|
@ -103,7 +103,7 @@ public class TestDataNodeMetrics {
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
|
||||||
try {
|
try {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
DistributedFileSystem fs = (DistributedFileSystem) cluster.getFileSystem();
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
|
||||||
Path testFile = new Path("/testFlushNanosMetric.txt");
|
Path testFile = new Path("/testFlushNanosMetric.txt");
|
||||||
FSDataOutputStream fout = fs.create(testFile);
|
FSDataOutputStream fout = fs.create(testFile);
|
||||||
|
|
|
@ -78,7 +78,7 @@ public class TestTransferRbw {
|
||||||
).numDataNodes(REPLICATION).build();
|
).numDataNodes(REPLICATION).build();
|
||||||
try {
|
try {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
final DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();
|
final DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
|
|
||||||
//create a file, write some data and leave it open.
|
//create a file, write some data and leave it open.
|
||||||
final Path p = new Path("/foo");
|
final Path p = new Path("/foo");
|
||||||
|
|
|
@ -126,7 +126,7 @@ public class OfflineEditsViewerHelper {
|
||||||
private CheckpointSignature runOperations() throws IOException {
|
private CheckpointSignature runOperations() throws IOException {
|
||||||
LOG.info("Creating edits by performing fs operations");
|
LOG.info("Creating edits by performing fs operations");
|
||||||
// no check, if it's not it throws an exception which is what we want
|
// no check, if it's not it throws an exception which is what we want
|
||||||
DistributedFileSystem dfs = (DistributedFileSystem) cluster.getFileSystem();
|
DistributedFileSystem dfs = cluster.getFileSystem();
|
||||||
DFSTestUtil.runOperations(cluster, dfs, cluster.getConfiguration(0),
|
DFSTestUtil.runOperations(cluster, dfs, cluster.getConfiguration(0),
|
||||||
dfs.getDefaultBlockSize(), 0);
|
dfs.getDefaultBlockSize(), 0);
|
||||||
|
|
||||||
|
|
|
@ -56,7 +56,7 @@ public class TestBlockUnderConstruction {
|
||||||
Configuration conf = new HdfsConfiguration();
|
Configuration conf = new HdfsConfiguration();
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
hdfs = (DistributedFileSystem)cluster.getFileSystem();
|
hdfs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@AfterClass
|
@AfterClass
|
||||||
|
|
|
@ -71,7 +71,7 @@ public class TestCheckPointForSecurityTokens {
|
||||||
DFSConfigKeys.DFS_NAMENODE_DELEGATION_TOKEN_ALWAYS_USE_KEY, true);
|
DFSConfigKeys.DFS_NAMENODE_DELEGATION_TOKEN_ALWAYS_USE_KEY, true);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDatanodes).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDatanodes).build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
fs = (DistributedFileSystem)(cluster.getFileSystem());
|
fs = cluster.getFileSystem();
|
||||||
FSNamesystem namesystem = cluster.getNamesystem();
|
FSNamesystem namesystem = cluster.getNamesystem();
|
||||||
String renewer = UserGroupInformation.getLoginUser().getUserName();
|
String renewer = UserGroupInformation.getLoginUser().getUserName();
|
||||||
Token<DelegationTokenIdentifier> token1 = namesystem
|
Token<DelegationTokenIdentifier> token1 = namesystem
|
||||||
|
|
|
@ -372,7 +372,7 @@ public class TestFSPermissionChecker {
|
||||||
|
|
||||||
private void addAcl(INodeWithAdditionalFields inode, AclEntry... acl)
|
private void addAcl(INodeWithAdditionalFields inode, AclEntry... acl)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
AclStorage.updateINodeAcl((INodeWithAdditionalFields)inode,
|
AclStorage.updateINodeAcl(inode,
|
||||||
Arrays.asList(acl), Snapshot.CURRENT_STATE_ID);
|
Arrays.asList(acl), Snapshot.CURRENT_STATE_ID);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -464,7 +464,7 @@ public class TestFsck {
|
||||||
len = (int)(in.getFileLength() % blockSize);
|
len = (int)(in.getFileLength() % blockSize);
|
||||||
if (len == 0) len = blockBuffer.length;
|
if (len == 0) len = blockBuffer.length;
|
||||||
}
|
}
|
||||||
IOUtils.readFully(in, blockBuffer, 0, (int)len);
|
IOUtils.readFully(in, blockBuffer, 0, len);
|
||||||
int startIdx = blockIdx * blockSize;
|
int startIdx = blockIdx * blockSize;
|
||||||
for (int i = 0; i < len; i++) {
|
for (int i = 0; i < len; i++) {
|
||||||
if (initialContents[startIdx + i] != blockBuffer[i]) {
|
if (initialContents[startIdx + i] != blockBuffer[i]) {
|
||||||
|
@ -851,7 +851,7 @@ public class TestFsck {
|
||||||
new MiniDFSCluster.Builder(conf).numDataNodes(NUM_REPLICAS).build();
|
new MiniDFSCluster.Builder(conf).numDataNodes(NUM_REPLICAS).build();
|
||||||
assertNotNull("Failed Cluster Creation", cluster);
|
assertNotNull("Failed Cluster Creation", cluster);
|
||||||
cluster.waitClusterUp();
|
cluster.waitClusterUp();
|
||||||
dfs = (DistributedFileSystem) cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
assertNotNull("Failed to get FileSystem", dfs);
|
assertNotNull("Failed to get FileSystem", dfs);
|
||||||
|
|
||||||
// Create a file that will be intentionally under-replicated
|
// Create a file that will be intentionally under-replicated
|
||||||
|
@ -923,7 +923,7 @@ public class TestFsck {
|
||||||
.racks(racks).build();
|
.racks(racks).build();
|
||||||
assertNotNull("Failed Cluster Creation", cluster);
|
assertNotNull("Failed Cluster Creation", cluster);
|
||||||
cluster.waitClusterUp();
|
cluster.waitClusterUp();
|
||||||
dfs = (DistributedFileSystem) cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
assertNotNull("Failed to get FileSystem", dfs);
|
assertNotNull("Failed to get FileSystem", dfs);
|
||||||
|
|
||||||
// Create a file that will be intentionally under-replicated
|
// Create a file that will be intentionally under-replicated
|
||||||
|
@ -946,7 +946,7 @@ public class TestFsck {
|
||||||
PrintWriter out = new PrintWriter(result, true);
|
PrintWriter out = new PrintWriter(result, true);
|
||||||
InetAddress remoteAddress = InetAddress.getLocalHost();
|
InetAddress remoteAddress = InetAddress.getLocalHost();
|
||||||
NamenodeFsck fsck = new NamenodeFsck(conf, namenode, nettop, pmap, out,
|
NamenodeFsck fsck = new NamenodeFsck(conf, namenode, nettop, pmap, out,
|
||||||
NUM_DN, (short)REPL_FACTOR, remoteAddress);
|
NUM_DN, REPL_FACTOR, remoteAddress);
|
||||||
|
|
||||||
// Run the fsck and check the Result
|
// Run the fsck and check the Result
|
||||||
final HdfsFileStatus file =
|
final HdfsFileStatus file =
|
||||||
|
|
|
@ -70,7 +70,7 @@ public class TestHDFSConcat {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL_FACTOR).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL_FACTOR).build();
|
||||||
assertNotNull("Failed Cluster Creation", cluster);
|
assertNotNull("Failed Cluster Creation", cluster);
|
||||||
cluster.waitClusterUp();
|
cluster.waitClusterUp();
|
||||||
dfs = (DistributedFileSystem) cluster.getFileSystem();
|
dfs = cluster.getFileSystem();
|
||||||
assertNotNull("Failed to get FileSystem", dfs);
|
assertNotNull("Failed to get FileSystem", dfs);
|
||||||
nn = cluster.getNameNodeRpc();
|
nn = cluster.getNameNodeRpc();
|
||||||
assertNotNull("Failed to get NameNode", nn);
|
assertNotNull("Failed to get NameNode", nn);
|
||||||
|
|
|
@ -361,7 +361,7 @@ public class TestNNStorageRetentionManager {
|
||||||
public Void answer(InvocationOnMock invocation) throws Throwable {
|
public Void answer(InvocationOnMock invocation) throws Throwable {
|
||||||
Object[] args = invocation.getArguments();
|
Object[] args = invocation.getArguments();
|
||||||
journalSet.selectInputStreams((Collection<EditLogInputStream>)args[0],
|
journalSet.selectInputStreams((Collection<EditLogInputStream>)args[0],
|
||||||
(long)((Long)args[1]), (boolean)((Boolean)args[2]));
|
(Long)args[1], (Boolean)args[2]);
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
}).when(mockLog).selectInputStreams(Mockito.anyCollection(),
|
}).when(mockLog).selectInputStreams(Mockito.anyCollection(),
|
||||||
|
|
|
@ -609,7 +609,7 @@ public class TestSaveNamespace {
|
||||||
MiniDFSCluster cluster = new MiniDFSCluster.Builder(new Configuration())
|
MiniDFSCluster cluster = new MiniDFSCluster.Builder(new Configuration())
|
||||||
.numDataNodes(1).build();
|
.numDataNodes(1).build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
DistributedFileSystem fs = (DistributedFileSystem) cluster.getFileSystem();
|
DistributedFileSystem fs = cluster.getFileSystem();
|
||||||
try {
|
try {
|
||||||
cluster.getNamesystem().leaseManager.addLease("me", "/non-existent");
|
cluster.getNamesystem().leaseManager.addLease("me", "/non-existent");
|
||||||
fs.setSafeMode(SafeModeAction.SAFEMODE_ENTER);
|
fs.setSafeMode(SafeModeAction.SAFEMODE_ENTER);
|
||||||
|
|
|
@ -191,9 +191,9 @@ public class TestSecurityTokenEditLog {
|
||||||
Token<DelegationTokenIdentifier> token1 = fsn.getDelegationToken(renewer);
|
Token<DelegationTokenIdentifier> token1 = fsn.getDelegationToken(renewer);
|
||||||
Token<DelegationTokenIdentifier> token2 = fsn.getDelegationToken(renewer);
|
Token<DelegationTokenIdentifier> token2 = fsn.getDelegationToken(renewer);
|
||||||
DelegationTokenIdentifier ident1 =
|
DelegationTokenIdentifier ident1 =
|
||||||
(DelegationTokenIdentifier)token1.decodeIdentifier();
|
token1.decodeIdentifier();
|
||||||
DelegationTokenIdentifier ident2 =
|
DelegationTokenIdentifier ident2 =
|
||||||
(DelegationTokenIdentifier)token2.decodeIdentifier();
|
token2.decodeIdentifier();
|
||||||
|
|
||||||
// verify we got the tokens
|
// verify we got the tokens
|
||||||
verify(log, times(1)).logGetDelegationToken(eq(ident1), anyLong());
|
verify(log, times(1)).logGetDelegationToken(eq(ident1), anyLong());
|
||||||
|
|
|
@ -212,7 +212,7 @@ public class TestStartup {
|
||||||
private void corruptNameNodeFiles() throws IOException {
|
private void corruptNameNodeFiles() throws IOException {
|
||||||
// now corrupt/delete the directrory
|
// now corrupt/delete the directrory
|
||||||
List<URI> nameDirs = (List<URI>)FSNamesystem.getNamespaceDirs(config);
|
List<URI> nameDirs = (List<URI>)FSNamesystem.getNamespaceDirs(config);
|
||||||
List<URI> nameEditsDirs = (List<URI>)FSNamesystem.getNamespaceEditsDirs(config);
|
List<URI> nameEditsDirs = FSNamesystem.getNamespaceEditsDirs(config);
|
||||||
|
|
||||||
// get name dir and its length, then delete and recreate the directory
|
// get name dir and its length, then delete and recreate the directory
|
||||||
File dir = new File(nameDirs.get(0).getPath()); // has only one
|
File dir = new File(nameDirs.get(0).getPath()); // has only one
|
||||||
|
|
|
@ -954,7 +954,7 @@ public class TestRetryCacheWithHA {
|
||||||
boolean checkNamenodeBeforeReturn() throws Exception {
|
boolean checkNamenodeBeforeReturn() throws Exception {
|
||||||
for (int i = 0; i < CHECKTIMES; i++) {
|
for (int i = 0; i < CHECKTIMES; i++) {
|
||||||
RemoteIterator<CachePoolEntry> iter = dfs.listCachePools();
|
RemoteIterator<CachePoolEntry> iter = dfs.listCachePools();
|
||||||
if (iter.hasNext() && (long)iter.next().getInfo().getLimit() == 99) {
|
if (iter.hasNext() && iter.next().getInfo().getLimit() == 99) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
Thread.sleep(1000);
|
Thread.sleep(1000);
|
||||||
|
|
|
@ -57,7 +57,7 @@ public class TestNNMetricFilesInGetListingOps {
|
||||||
cluster = new MiniDFSCluster.Builder(CONF).build();
|
cluster = new MiniDFSCluster.Builder(CONF).build();
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
cluster.getNameNode();
|
cluster.getNameNode();
|
||||||
fs = (DistributedFileSystem) cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@After
|
@After
|
||||||
|
|
|
@ -105,7 +105,7 @@ public class TestNameNodeMetrics {
|
||||||
cluster.waitActive();
|
cluster.waitActive();
|
||||||
namesystem = cluster.getNamesystem();
|
namesystem = cluster.getNamesystem();
|
||||||
bm = namesystem.getBlockManager();
|
bm = namesystem.getBlockManager();
|
||||||
fs = (DistributedFileSystem) cluster.getFileSystem();
|
fs = cluster.getFileSystem();
|
||||||
}
|
}
|
||||||
|
|
||||||
@After
|
@After
|
||||||
|
|
|
@ -1592,7 +1592,7 @@ public class TestRenameWithSnapshots {
|
||||||
INode barNode = fsdir.getINode4Write(bar.toString());
|
INode barNode = fsdir.getINode4Write(bar.toString());
|
||||||
assertTrue(barNode.getClass() == INodeFile.class);
|
assertTrue(barNode.getClass() == INodeFile.class);
|
||||||
assertSame(fooNode, barNode.getParent());
|
assertSame(fooNode, barNode.getParent());
|
||||||
List<DirectoryDiff> diffList = ((INodeDirectorySnapshottable) dir1Node)
|
List<DirectoryDiff> diffList = dir1Node
|
||||||
.getDiffs().asList();
|
.getDiffs().asList();
|
||||||
assertEquals(1, diffList.size());
|
assertEquals(1, diffList.size());
|
||||||
DirectoryDiff diff = diffList.get(0);
|
DirectoryDiff diff = diffList.get(0);
|
||||||
|
@ -1666,7 +1666,7 @@ public class TestRenameWithSnapshots {
|
||||||
INode fooNode = childrenList.get(0);
|
INode fooNode = childrenList.get(0);
|
||||||
assertTrue(fooNode.asDirectory().isWithSnapshot());
|
assertTrue(fooNode.asDirectory().isWithSnapshot());
|
||||||
assertSame(dir1Node, fooNode.getParent());
|
assertSame(dir1Node, fooNode.getParent());
|
||||||
List<DirectoryDiff> diffList = ((INodeDirectorySnapshottable) dir1Node)
|
List<DirectoryDiff> diffList = dir1Node
|
||||||
.getDiffs().asList();
|
.getDiffs().asList();
|
||||||
assertEquals(1, diffList.size());
|
assertEquals(1, diffList.size());
|
||||||
DirectoryDiff diff = diffList.get(0);
|
DirectoryDiff diff = diffList.get(0);
|
||||||
|
|
|
@ -490,7 +490,7 @@ public class TestSnapshot {
|
||||||
// ...
|
// ...
|
||||||
//
|
//
|
||||||
Modification create = new FileCreation(
|
Modification create = new FileCreation(
|
||||||
node.fileList.get(node.nullFileIndex), hdfs, (int) BLOCKSIZE);
|
node.fileList.get(node.nullFileIndex), hdfs, BLOCKSIZE);
|
||||||
Modification delete = new FileDeletion(
|
Modification delete = new FileDeletion(
|
||||||
node.fileList.get((node.nullFileIndex + 1) % node.fileList.size()),
|
node.fileList.get((node.nullFileIndex + 1) % node.fileList.size()),
|
||||||
hdfs);
|
hdfs);
|
||||||
|
|
|
@ -421,7 +421,7 @@ public class TestLightWeightHashSet{
|
||||||
assertEquals(NUM - 10, array2.length);
|
assertEquals(NUM - 10, array2.length);
|
||||||
|
|
||||||
for (int i = 0; i < array2.length; i++) {
|
for (int i = 0; i < array2.length; i++) {
|
||||||
assertTrue(sub2.contains((Integer) array2[i]));
|
assertTrue(sub2.contains(array2[i]));
|
||||||
}
|
}
|
||||||
|
|
||||||
LOG.info("Test other - DONE");
|
LOG.info("Test other - DONE");
|
||||||
|
|
|
@ -358,7 +358,7 @@ public class TestLightWeightLinkedSet {
|
||||||
Object[] array2 = set.toArray();
|
Object[] array2 = set.toArray();
|
||||||
assertEquals(NUM, array2.length);
|
assertEquals(NUM, array2.length);
|
||||||
for (int i = 0; i < array2.length; i++) {
|
for (int i = 0; i < array2.length; i++) {
|
||||||
assertTrue(list.contains((Integer) array2[i]));
|
assertTrue(list.contains(array2[i]));
|
||||||
}
|
}
|
||||||
LOG.info("Test capacity - DONE");
|
LOG.info("Test capacity - DONE");
|
||||||
}
|
}
|
||||||
|
|
|
@ -103,7 +103,7 @@ public static class MockHttpURLConnection extends HttpURLConnection {
|
||||||
doReturn(new MockHttpURLConnection(ospy.getURL())).when(ospy)
|
doReturn(new MockHttpURLConnection(ospy.getURL())).when(ospy)
|
||||||
.openConnection();
|
.openConnection();
|
||||||
HftpFileSystem.RangeHeaderUrlOpener rspy = spy(
|
HftpFileSystem.RangeHeaderUrlOpener rspy = spy(
|
||||||
new HftpFileSystem.RangeHeaderUrlOpener(factory, (URL) null));
|
new HftpFileSystem.RangeHeaderUrlOpener(factory, null));
|
||||||
doReturn(new MockHttpURLConnection(rspy.getURL())).when(rspy)
|
doReturn(new MockHttpURLConnection(rspy.getURL())).when(rspy)
|
||||||
.openConnection();
|
.openConnection();
|
||||||
ByteRangeInputStream is = new HftpFileSystem.RangeHeaderInputStream(ospy, rspy);
|
ByteRangeInputStream is = new HftpFileSystem.RangeHeaderInputStream(ospy, rspy);
|
||||||
|
|
|
@ -124,8 +124,8 @@ public class TestJsonUtil {
|
||||||
|
|
||||||
// The encoded result should contain name, ipAddr and xferPort.
|
// The encoded result should contain name, ipAddr and xferPort.
|
||||||
Map<String, Object> r = JsonUtil.toJsonMap(di);
|
Map<String, Object> r = JsonUtil.toJsonMap(di);
|
||||||
Assert.assertEquals(name, (String)r.get("name"));
|
Assert.assertEquals(name, r.get("name"));
|
||||||
Assert.assertEquals("127.0.0.1", (String)r.get("ipAddr"));
|
Assert.assertEquals("127.0.0.1", r.get("ipAddr"));
|
||||||
// In this test, it is Integer instead of Long since json was not actually
|
// In this test, it is Integer instead of Long since json was not actually
|
||||||
// involved in constructing the map.
|
// involved in constructing the map.
|
||||||
Assert.assertEquals(1004, (int)(Integer)r.get("xferPort"));
|
Assert.assertEquals(1004, (int)(Integer)r.get("xferPort"));
|
||||||
|
|
Loading…
Reference in New Issue