HDFS-3659. Add missing @Override to methods across the hadoop-hdfs project. Contributed by Brandon Li. (harsh)

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1361894 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Harsh J 2012-07-16 02:58:03 +00:00
parent ae6cc14611
commit 0e8e499ff4
142 changed files with 347 additions and 1 deletions

View File

@ -330,6 +330,9 @@ Branch-2 ( Unreleased changes )
HDFS-3663. MiniDFSCluster should capture the code path that led to
the first ExitException. (eli)
HDFS-3659. Add missing @Override to methods across the hadoop-hdfs
project. (Brandon Li via harsh)
OPTIMIZATIONS
HDFS-2982. Startup performance suffers when there are many edit log

View File

@ -597,6 +597,7 @@ public class DFSClient implements java.io.Closeable {
* Close the file system, abandoning all of the leases and files being
* created and close connections to the namenode.
*/
@Override
public synchronized void close() throws IOException {
if(clientRunning) {
closeAllFilesBeingWritten(false);

View File

@ -268,6 +268,7 @@ public class DFSOutputStream extends FSOutputSummer implements Syncable {
return seqno == HEART_BEAT_SEQNO;
}
@Override
public String toString() {
return "packet seqno:" + this.seqno +
" offsetInBlock:" + this.offsetInBlock +
@ -396,6 +397,7 @@ public class DFSOutputStream extends FSOutputSummer implements Syncable {
* streamer thread is the only thread that opens streams to datanode,
* and closes them. Any error recovery is also done by this thread.
*/
@Override
public void run() {
long lastPacket = Time.now();
while (!streamerClosed && dfsClient.clientRunning) {
@ -654,6 +656,7 @@ public class DFSOutputStream extends FSOutputSummer implements Syncable {
this.targets = targets;
}
@Override
public void run() {
setName("ResponseProcessor for block " + block);

View File

@ -110,6 +110,7 @@ public class DFSUtil {
* Address matcher for matching an address to local address
*/
static final AddressMatcher LOCAL_ADDRESS_MATCHER = new AddressMatcher() {
@Override
public boolean match(InetSocketAddress s) {
return NetUtils.isLocalAddress(s.getAddress());
};

View File

@ -113,6 +113,7 @@ public class HftpFileSystem extends FileSystem
protected static final ThreadLocal<SimpleDateFormat> df =
new ThreadLocal<SimpleDateFormat>() {
@Override
protected SimpleDateFormat initialValue() {
return getDateFormat();
}
@ -240,6 +241,7 @@ public class HftpFileSystem extends FileSystem
//Renew TGT if needed
ugi.reloginFromKeytab();
return ugi.doAs(new PrivilegedExceptionAction<Token<?>>() {
@Override
public Token<?> run() throws IOException {
final String nnHttpUrl = nnSecureUri.toString();
Credentials c;
@ -402,6 +404,7 @@ public class HftpFileSystem extends FileSystem
ArrayList<FileStatus> fslist = new ArrayList<FileStatus>();
@Override
public void startElement(String ns, String localname, String qname,
Attributes attrs) throws SAXException {
if ("listing".equals(qname)) return;
@ -541,6 +544,7 @@ public class HftpFileSystem extends FileSystem
public void setWorkingDirectory(Path f) { }
/** This optional operation is not yet supported. */
@Override
public FSDataOutputStream append(Path f, int bufferSize,
Progressable progress) throws IOException {
throw new IOException("Not supported");

View File

@ -189,6 +189,7 @@ public class HsftpFileSystem extends HftpFileSystem {
* Dummy hostname verifier that is used to bypass hostname checking
*/
protected static class DummyHostnameVerifier implements HostnameVerifier {
@Override
public boolean verify(String hostname, SSLSession session) {
return true;
}
@ -198,12 +199,15 @@ public class HsftpFileSystem extends HftpFileSystem {
* Dummy trustmanager that is used to trust all server certificates
*/
protected static class DummyTrustManager implements X509TrustManager {
@Override
public void checkClientTrusted(X509Certificate[] chain, String authType) {
}
@Override
public void checkServerTrusted(X509Certificate[] chain, String authType) {
}
@Override
public X509Certificate[] getAcceptedIssuers() {
return null;
}

View File

@ -134,6 +134,7 @@ class SocketCache {
multimap.clear();
}
@Override
protected void finalize() {
clear();
}

View File

@ -40,6 +40,7 @@ public class Block implements Writable, Comparable<Block> {
WritableFactories.setFactory
(Block.class,
new WritableFactory() {
@Override
public Writable newInstance() { return new Block(); }
});
}
@ -146,6 +147,7 @@ public class Block implements Writable, Comparable<Block> {
/**
*/
@Override
public String toString() {
return getBlockName() + "_" + getGenerationStamp();
}

View File

@ -148,10 +148,12 @@ public class BlockListAsLongs implements Iterable<Block> {
this.currentReplicaState = null;
}
@Override
public boolean hasNext() {
return currentBlockIndex < getNumberOfBlocks();
}
@Override
public Block next() {
block.set(blockId(currentBlockIndex),
blockLength(currentBlockIndex),
@ -161,6 +163,7 @@ public class BlockListAsLongs implements Iterable<Block> {
return block;
}
@Override
public void remove() {
throw new UnsupportedOperationException("Sorry. can't remove.");
}
@ -178,6 +181,7 @@ public class BlockListAsLongs implements Iterable<Block> {
/**
* Returns an iterator over blocks in the block report.
*/
@Override
public Iterator<Block> iterator() {
return getBlockReportIterator();
}

View File

@ -37,6 +37,7 @@ public class DSQuotaExceededException extends QuotaExceededException {
super(quota, count);
}
@Override
public String getMessage() {
String msg = super.getMessage();
if (msg == null) {

View File

@ -150,6 +150,7 @@ public class DatanodeID implements Comparable<DatanodeID> {
return ipcPort;
}
@Override
public boolean equals(Object to) {
if (this == to) {
return true;
@ -161,10 +162,12 @@ public class DatanodeID implements Comparable<DatanodeID> {
storageID.equals(((DatanodeID)to).getStorageID()));
}
@Override
public int hashCode() {
return getXferAddr().hashCode()^ storageID.hashCode();
}
@Override
public String toString() {
return getXferAddr();
}
@ -187,6 +190,7 @@ public class DatanodeID implements Comparable<DatanodeID> {
* @param that
* @return as specified by Comparable
*/
@Override
public int compareTo(DatanodeID that) {
return getXferAddr().compareTo(that.getXferAddr());
}

View File

@ -56,6 +56,7 @@ public class DatanodeInfo extends DatanodeID implements Node {
this.value = v;
}
@Override
public String toString() {
return value;
}
@ -126,6 +127,7 @@ public class DatanodeInfo extends DatanodeID implements Node {
}
/** Network location name */
@Override
public String getName() {
return getXferAddr();
}
@ -200,9 +202,11 @@ public class DatanodeInfo extends DatanodeID implements Node {
}
/** network location */
@Override
public synchronized String getNetworkLocation() {return location;}
/** Sets the network location */
@Override
public synchronized void setNetworkLocation(String location) {
this.location = NodeBase.normalize(location);
}
@ -334,13 +338,17 @@ public class DatanodeInfo extends DatanodeID implements Node {
private transient Node parent; //its parent
/** Return this node's parent */
@Override
public Node getParent() { return parent; }
@Override
public void setParent(Node parent) {this.parent = parent;}
/** Return this node's level in the tree.
* E.g. the root of a tree returns 0 and its children return 1
*/
@Override
public int getLevel() { return level; }
@Override
public void setLevel(int level) {this.level = level;}
@Override

View File

@ -113,6 +113,7 @@ public class LocatedBlocks {
Comparator<LocatedBlock> comp =
new Comparator<LocatedBlock>() {
// Returns 0 iff a is inside b or b is inside a
@Override
public int compare(LocatedBlock a, LocatedBlock b) {
long aBeg = a.getStartOffset();
long bBeg = b.getStartOffset();

View File

@ -36,6 +36,7 @@ public final class NSQuotaExceededException extends QuotaExceededException {
super(quota, count);
}
@Override
public String getMessage() {
String msg = super.getMessage();
if (msg == null) {

View File

@ -58,6 +58,7 @@ public class QuotaExceededException extends IOException {
this.pathName = path;
}
@Override
public String getMessage() {
return super.getMessage();
}

View File

@ -131,6 +131,7 @@ public class ClientNamenodeProtocolTranslatorPB implements
rpcProxy = proxy;
}
@Override
public void close() {
RPC.stopProxy(rpcProxy);
}

View File

@ -82,6 +82,7 @@ public class NamenodeProtocolTranslatorPB implements NamenodeProtocol,
this.rpcProxy = rpcProxy;
}
@Override
public void close() {
RPC.stopProxy(rpcProxy);
}

View File

@ -119,6 +119,7 @@ public class BlockTokenIdentifier extends TokenIdentifier {
}
/** {@inheritDoc} */
@Override
public boolean equals(Object obj) {
if (obj == this) {
return true;
@ -135,12 +136,14 @@ public class BlockTokenIdentifier extends TokenIdentifier {
}
/** {@inheritDoc} */
@Override
public int hashCode() {
return (int) expiryDate ^ keyId ^ (int) blockId ^ modes.hashCode()
^ (userId == null ? 0 : userId.hashCode())
^ (blockPoolId == null ? 0 : blockPoolId.hashCode());
}
@Override
public void readFields(DataInput in) throws IOException {
this.cache = null;
expiryDate = WritableUtils.readVLong(in);
@ -155,6 +158,7 @@ public class BlockTokenIdentifier extends TokenIdentifier {
}
}
@Override
public void write(DataOutput out) throws IOException {
WritableUtils.writeVLong(out, expiryDate);
WritableUtils.writeVInt(out, keyId);

View File

@ -31,6 +31,7 @@ import org.apache.hadoop.security.token.TokenSelector;
@InterfaceAudience.Private
public class BlockTokenSelector implements TokenSelector<BlockTokenIdentifier> {
@Override
@SuppressWarnings("unchecked")
public Token<BlockTokenIdentifier> selectToken(Text service,
Collection<Token<? extends TokenIdentifier>> tokens) {

View File

@ -78,6 +78,7 @@ public class ExportedBlockKeys implements Writable {
static { // register a ctor
WritableFactories.setFactory(ExportedBlockKeys.class,
new WritableFactory() {
@Override
public Writable newInstance() {
return new ExportedBlockKeys();
}
@ -86,6 +87,7 @@ public class ExportedBlockKeys implements Writable {
/**
*/
@Override
public void write(DataOutput out) throws IOException {
out.writeBoolean(isBlockTokenEnabled);
out.writeLong(keyUpdateInterval);
@ -99,6 +101,7 @@ public class ExportedBlockKeys implements Writable {
/**
*/
@Override
public void readFields(DataInput in) throws IOException {
isBlockTokenEnabled = in.readBoolean();
keyUpdateInterval = in.readLong();

View File

@ -378,6 +378,7 @@ public class Balancer {
/* start a thread to dispatch the block move */
private void scheduleBlockMove() {
moverExecutor.execute(new Runnable() {
@Override
public void run() {
if (LOG.isDebugEnabled()) {
LOG.debug("Starting moving "+ block.getBlockId() +
@ -570,6 +571,7 @@ public class Balancer {
/* A thread that initiates a block move
* and waits for block move to complete */
private class BlockMoveDispatcher implements Runnable {
@Override
public void run() {
dispatchBlocks();
}

View File

@ -189,6 +189,7 @@ class NameNodeConnector {
* Periodically updates access keys.
*/
class BlockKeyUpdater implements Runnable {
@Override
public void run() {
try {
while (shouldRun) {

View File

@ -52,6 +52,7 @@ public class BlockPlacementPolicyWithNodeGroup extends BlockPlacementPolicyDefau
BlockPlacementPolicyWithNodeGroup() {
}
@Override
public void initialize(Configuration conf, FSClusterStats stats,
NetworkTopology clusterMap) {
super.initialize(conf, stats, clusterMap);

View File

@ -37,15 +37,18 @@ class BlocksMap {
this.blockInfo = blkInfo;
}
@Override
public boolean hasNext() {
return blockInfo != null && nextIdx < blockInfo.getCapacity()
&& blockInfo.getDatanode(nextIdx) != null;
}
@Override
public DatanodeDescriptor next() {
return blockInfo.getDatanode(nextIdx++);
}
@Override
public void remove() {
throw new UnsupportedOperationException("Sorry. can't remove.");
}

View File

@ -326,16 +326,19 @@ public class DatanodeDescriptor extends DatanodeInfo {
this.node = dn;
}
@Override
public boolean hasNext() {
return current != null;
}
@Override
public BlockInfo next() {
BlockInfo res = current;
current = current.getNext(current.findDatanode(node));
return res;
}
@Override
public void remove() {
throw new UnsupportedOperationException("Sorry. can't remove.");
}
@ -542,6 +545,7 @@ public class DatanodeDescriptor extends DatanodeInfo {
/**
* @param nodeReg DatanodeID to update registration for.
*/
@Override
public void updateRegInfo(DatanodeID nodeReg) {
super.updateRegInfo(nodeReg);
}

View File

@ -192,6 +192,7 @@ class PendingReplicationBlocks {
* their replication request.
*/
class PendingReplicationMonitor implements Runnable {
@Override
public void run() {
while (fsRunning) {
long period = Math.min(defaultRecheckInterval, timeout);

View File

@ -141,6 +141,7 @@ public final class HdfsServerConstants {
private String description = null;
private NamenodeRole(String arg) {this.description = arg;}
@Override
public String toString() {
return description;
}

View File

@ -109,6 +109,7 @@ public class JspHelper {
// compare two records based on their frequency
private static class NodeRecordComparator implements Comparator<NodeRecord> {
@Override
public int compare(NodeRecord o1, NodeRecord o2) {
if (o1.frequency < o2.frequency) {
return -1;
@ -312,6 +313,7 @@ public class JspHelper {
}
}
@Override
public int compare(DatanodeDescriptor d1,
DatanodeDescriptor d2) {
int ret = 0;

View File

@ -123,6 +123,7 @@ public abstract class Storage extends StorageInfo {
this.prevIndex = 0;
}
@Override
public boolean hasNext() {
if (storageDirs.isEmpty() || nextIndex >= storageDirs.size())
return false;
@ -138,6 +139,7 @@ public abstract class Storage extends StorageInfo {
return true;
}
@Override
public StorageDirectory next() {
StorageDirectory sd = getStorageDir(nextIndex);
prevIndex = nextIndex;
@ -152,6 +154,7 @@ public abstract class Storage extends StorageInfo {
return sd;
}
@Override
public void remove() {
nextIndex = prevIndex; // restore previous state
storageDirs.remove(prevIndex); // remove last returned element

View File

@ -78,6 +78,7 @@ public class StorageInfo {
cTime = from.cTime;
}
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
sb.append("lv=").append(layoutVersion).append(";cid=").append(clusterID)

View File

@ -32,19 +32,23 @@ import org.apache.hadoop.hdfs.server.common.UpgradeObjectCollection.UOSignature;
public abstract class UpgradeObject implements Upgradeable {
protected short status;
@Override
public short getUpgradeStatus() {
return status;
}
@Override
public String getDescription() {
return "Upgrade object for " + getType() + " layout version " + getVersion();
}
@Override
public UpgradeStatusReport getUpgradeStatusReport(boolean details)
throws IOException {
return new UpgradeStatusReport(getVersion(), getUpgradeStatus(), false);
}
@Override
public int compareTo(Upgradeable o) {
if(this.getVersion() != o.getVersion())
return (getVersion() > o.getVersion() ? -1 : 1);
@ -55,6 +59,7 @@ public abstract class UpgradeObject implements Upgradeable {
o.getClass().getCanonicalName());
}
@Override
public boolean equals(Object o) {
if (!(o instanceof UpgradeObject)) {
return false;
@ -62,6 +67,7 @@ public abstract class UpgradeObject implements Upgradeable {
return this.compareTo((UpgradeObject)o) == 0;
}
@Override
public int hashCode() {
return new UOSignature(this).hashCode();
}

View File

@ -73,6 +73,7 @@ public class UpgradeObjectCollection {
}
}
@Override
public int compareTo(UOSignature o) {
if(this.version != o.version)
return (version < o.version ? -1 : 1);
@ -82,6 +83,7 @@ public class UpgradeObjectCollection {
return className.compareTo(o.className);
}
@Override
public boolean equals(Object o) {
if (!(o instanceof UOSignature)) {
return false;
@ -89,6 +91,7 @@ public class UpgradeObjectCollection {
return this.compareTo((UOSignature)o) == 0;
}
@Override
public int hashCode() {
return version ^ ((type==null)?0:type.hashCode())
^ ((className==null)?0:className.hashCode());

View File

@ -82,6 +82,7 @@ public class UpgradeStatusReport {
/**
* Print basic upgradeStatus details.
*/
@Override
public String toString() {
return getStatusText(false);
}

View File

@ -122,6 +122,7 @@ class BlockPoolManager {
try {
UserGroupInformation.getLoginUser().doAs(
new PrivilegedExceptionAction<Object>() {
@Override
public Object run() throws Exception {
for (BPOfferService bpos : offerServices) {
bpos.start();

View File

@ -115,10 +115,12 @@ class BlockPoolSliceScanner {
this.block = block;
}
@Override
public int hashCode() {
return block.hashCode();
}
@Override
public boolean equals(Object other) {
return other instanceof BlockScanInfo &&
compareTo((BlockScanInfo)other) == 0;
@ -128,6 +130,7 @@ class BlockPoolSliceScanner {
return (lastScanType == ScanType.NONE) ? 0 : lastScanTime;
}
@Override
public int compareTo(BlockScanInfo other) {
long t1 = lastScanTime;
long t2 = other.lastScanTime;

View File

@ -440,6 +440,7 @@ public class BlockPoolSliceStorage extends Storage {
// delete finalized.tmp dir in a separate thread
new Daemon(new Runnable() {
@Override
public void run() {
try {
deleteDir(tmpDir);
@ -449,6 +450,7 @@ public class BlockPoolSliceStorage extends Storage {
LOG.info("Finalize upgrade for " + dataDirPath + " is complete.");
}
@Override
public String toString() {
return "Finalize " + dataDirPath;
}

View File

@ -246,6 +246,7 @@ class BlockReceiver implements Closeable {
/**
* close files.
*/
@Override
public void close() throws IOException {
IOException ioe = null;
if (syncOnClose && (out != null || checksumOut != null)) {
@ -1033,6 +1034,7 @@ class BlockReceiver implements Closeable {
* Thread to process incoming acks.
* @see java.lang.Runnable#run()
*/
@Override
public void run() {
boolean lastPacketInBlock = false;
final long startTime = ClientTraceLog.isInfoEnabled() ? System.nanoTime() : 0;

View File

@ -330,6 +330,7 @@ class BlockSender implements java.io.Closeable {
/**
* close opened files.
*/
@Override
public void close() throws IOException {
if (blockInFd != null && shouldDropCacheBehindRead && isLongRead()) {
// drop the last few MB of the file from cache

View File

@ -63,6 +63,7 @@ public class DataBlockScanner implements Runnable {
this.conf = conf;
}
@Override
public void run() {
String currentBpId = "";
boolean firstRun = true;
@ -273,6 +274,7 @@ public class DataBlockScanner implements Runnable {
public static class Servlet extends HttpServlet {
private static final long serialVersionUID = 1L;
@Override
public void doGet(HttpServletRequest request,
HttpServletResponse response) throws IOException {
response.setContentType("text/plain");

View File

@ -928,6 +928,7 @@ public class DataNode extends Configured
try {
return loginUgi
.doAs(new PrivilegedExceptionAction<InterDatanodeProtocol>() {
@Override
public InterDatanodeProtocol run() throws IOException {
return new InterDatanodeProtocolTranslatorPB(addr, loginUgi,
conf, NetUtils.getDefaultSocketFactory(conf), socketTimeout);
@ -1367,6 +1368,7 @@ public class DataNode extends Configured
/**
* Do the deed, write the bytes
*/
@Override
public void run() {
xmitsInProgress.getAndIncrement();
Socket sock = null;
@ -1723,6 +1725,7 @@ public class DataNode extends Configured
Daemon d = new Daemon(threadGroup, new Runnable() {
/** Recover a list of blocks. It is run by the primary datanode. */
@Override
public void run() {
for(RecoveringBlock b : blocks) {
try {

View File

@ -587,6 +587,7 @@ public class DataStorage extends Storage {
// 2. delete finalized.tmp dir in a separate thread
new Daemon(new Runnable() {
@Override
public void run() {
try {
deleteDir(tmpDir);
@ -595,6 +596,7 @@ public class DataStorage extends Storage {
}
LOG.info("Finalize upgrade for " + dataDirPath + " is complete.");
}
@Override
public String toString() { return "Finalize " + dataDirPath; }
}).start();
}
@ -677,6 +679,7 @@ public class DataStorage extends Storage {
throw new IOException("Cannot create directory " + to);
String[] blockNames = from.list(new java.io.FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return name.startsWith(BLOCK_FILE_PREFIX);
}
@ -694,6 +697,7 @@ public class DataStorage extends Storage {
// Now take care of the rest of the files and subdirectories
String[] otherNames = from.list(new java.io.FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return name.startsWith(BLOCK_SUBDIR_PREFIX)
|| name.startsWith(COPY_FILE_PREFIX);

View File

@ -145,6 +145,7 @@ class DataXceiver extends Receiver implements Runnable {
/**
* Read/write data from/to the DataXceiverServer.
*/
@Override
public void run() {
int opsProcessed = 0;
Op op = null;

View File

@ -60,6 +60,7 @@ public class DatanodeJspHelper {
InterruptedException {
return
user.doAs(new PrivilegedExceptionAction<DFSClient>() {
@Override
public DFSClient run() throws IOException {
return new DFSClient(NetUtils.createSocketAddr(addr), conf);
}

View File

@ -88,6 +88,7 @@ public class DirectoryScanner implements Runnable {
this.bpid = bpid;
}
@Override
public String toString() {
return "BlockPool " + bpid
+ " Total blocks: " + totalBlocks + ", missing metadata files:"

View File

@ -47,6 +47,7 @@ class UpgradeManagerDatanode extends UpgradeManager {
this.bpid = bpid;
}
@Override
public HdfsServerConstants.NodeType getType() {
return HdfsServerConstants.NodeType.DATA_NODE;
}
@ -71,6 +72,7 @@ class UpgradeManagerDatanode extends UpgradeManager {
* @return true if distributed upgrade is required or false otherwise
* @throws IOException
*/
@Override
public synchronized boolean startUpgrade() throws IOException {
if(upgradeState) { // upgrade is already in progress
assert currentUpgrades != null :
@ -134,6 +136,7 @@ class UpgradeManagerDatanode extends UpgradeManager {
+ "The upgrade object is not defined.");
}
@Override
public synchronized void completeUpgrade() throws IOException {
assert currentUpgrades != null :
"UpgradeManagerDatanode.currentUpgrades is null.";

View File

@ -36,6 +36,7 @@ public abstract class UpgradeObjectDatanode extends UpgradeObject implements Run
private DataNode dataNode = null;
private String bpid = null;
@Override
public HdfsServerConstants.NodeType getType() {
return HdfsServerConstants.NodeType.DATA_NODE;
}
@ -96,6 +97,7 @@ public abstract class UpgradeObjectDatanode extends UpgradeObject implements Run
throw new IOException(errorMsg);
}
@Override
public void run() {
assert dataNode != null : "UpgradeObjectDatanode.dataNode is null";
while(dataNode.shouldRun) {
@ -132,6 +134,7 @@ public abstract class UpgradeObjectDatanode extends UpgradeObject implements Run
* The data-node needs to re-confirm with the name-node that the upgrade
* is complete while other nodes are still upgrading.
*/
@Override
public UpgradeCommand completeUpgrade() throws IOException {
return new UpgradeCommand(UpgradeCommand.UC_ACTION_REPORT_STATUS,
getVersion(), (short)100);

View File

@ -264,6 +264,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
/**
* Return the number of failed volumes in the FSDataset.
*/
@Override
public int getNumFailedVolumes() {
return volumes.numberOfFailedVolumes();
}

View File

@ -68,6 +68,7 @@ public class CancelDelegationTokenServlet extends DfsServlet {
try {
ugi.doAs(new PrivilegedExceptionAction<Void>() {
@Override
public Void run() throws Exception {
nn.getRpcServer().cancelDelegationToken(token);
return null;

View File

@ -72,6 +72,7 @@ public class CheckpointSignature extends StorageInfo
* Get the cluster id from CheckpointSignature
* @return the cluster id
*/
@Override
public String getClusterID() {
return clusterID;
}
@ -101,6 +102,7 @@ public class CheckpointSignature extends StorageInfo
this.blockpoolID = blockpoolID;
}
@Override
public String toString() {
return String.valueOf(layoutVersion) + FIELD_SEPARATOR
+ String.valueOf(namespaceID) + FIELD_SEPARATOR
@ -133,6 +135,7 @@ public class CheckpointSignature extends StorageInfo
//
// Comparable interface
//
@Override
public int compareTo(CheckpointSignature o) {
return ComparisonChain.start()
.compare(layoutVersion, o.layoutVersion)
@ -145,6 +148,7 @@ public class CheckpointSignature extends StorageInfo
.result();
}
@Override
public boolean equals(Object o) {
if (!(o instanceof CheckpointSignature)) {
return false;
@ -152,6 +156,7 @@ public class CheckpointSignature extends StorageInfo
return compareTo((CheckpointSignature)o) == 0;
}
@Override
public int hashCode() {
return layoutVersion ^ namespaceID ^
(int)(cTime ^ mostRecentCheckpointTxId ^ curSegmentTxId)

View File

@ -118,6 +118,7 @@ class Checkpointer extends Daemon {
//
// The main work loop
//
@Override
public void run() {
// Check the size of the edit log once every 5 minutes.
long periodMSec = 5 * 60; // 5 minutes

View File

@ -656,6 +656,7 @@ class ClusterJspHelper {
this.value = v;
}
@Override
public String toString() {
return value;
}

View File

@ -57,6 +57,7 @@ public abstract class EditLogInputStream implements Closeable {
* Close the stream.
* @throws IOException if an error occurred while closing
*/
@Override
public abstract void close() throws IOException;
/**

View File

@ -74,6 +74,7 @@ public abstract class EditLogOutputStream implements Closeable {
* @throws IOException if the journal can't be closed,
* or if there are unflushed edits
*/
@Override
abstract public void close() throws IOException;
/**

View File

@ -187,6 +187,7 @@ public class FSDirectory implements Closeable {
/**
* Shutdown the filestore
*/
@Override
public void close() throws IOException {
fsImage.close();
}

View File

@ -173,6 +173,7 @@ public class FSEditLog {
// stores the most current transactionId of this thread.
private static final ThreadLocal<TransactionId> myTransactionId = new ThreadLocal<TransactionId>() {
@Override
protected synchronized TransactionId initialValue() {
return new TransactionId(Long.MAX_VALUE);
}

View File

@ -183,6 +183,7 @@ public abstract class FSEditLogOp {
return (T)this;
}
@Override
public String getPath() {
return path;
}
@ -216,6 +217,7 @@ public abstract class FSEditLogOp {
return (T)this;
}
@Override
public Block[] getBlocks() {
return blocks;
}
@ -409,6 +411,7 @@ public abstract class FSEditLogOp {
return (AddOp)cache.get(OP_ADD);
}
@Override
public boolean shouldCompleteLastBlock() {
return false;
}
@ -431,6 +434,7 @@ public abstract class FSEditLogOp {
return (CloseOp)cache.get(OP_CLOSE);
}
@Override
public boolean shouldCompleteLastBlock() {
return true;
}
@ -462,6 +466,7 @@ public abstract class FSEditLogOp {
return this;
}
@Override
public String getPath() {
return path;
}
@ -471,6 +476,7 @@ public abstract class FSEditLogOp {
return this;
}
@Override
public Block[] getBlocks() {
return blocks;
}
@ -2082,6 +2088,7 @@ public abstract class FSEditLogOp {
return (LogSegmentOp)cache.get(code);
}
@Override
public void readFields(DataInputStream in, int logVersion)
throws IOException {
// no data stored in these ops yet
@ -2174,6 +2181,7 @@ public abstract class FSEditLogOp {
WritableFactories.setFactory
(BlockTwo.class,
new WritableFactory() {
@Override
public Writable newInstance() { return new BlockTwo(); }
});
}
@ -2186,11 +2194,13 @@ public abstract class FSEditLogOp {
/////////////////////////////////////
// Writable
/////////////////////////////////////
@Override
public void write(DataOutput out) throws IOException {
out.writeLong(blkid);
out.writeLong(len);
}
@Override
public void readFields(DataInput in) throws IOException {
this.blkid = in.readLong();
this.len = in.readLong();

View File

@ -785,6 +785,7 @@ public class FSImage implements Closeable {
this.sd = sd;
}
@Override
public void run() {
try {
saveFSImage(context, sd);
@ -798,6 +799,7 @@ public class FSImage implements Closeable {
}
}
@Override
public String toString() {
return "FSImageSaver for " + sd.getRoot() +
" of type " + sd.getStorageDirType();
@ -1081,6 +1083,7 @@ public class FSImage implements Closeable {
}
}
@Override
synchronized public void close() throws IOException {
if (editLog != null) { // 2NN doesn't have any edit log
getEditLog().close();

View File

@ -114,6 +114,7 @@ class FSImageTransactionalStorageInspector extends FSImageStorageInspector {
*
* @throws FileNotFoundException if not images are found.
*/
@Override
FSImageFile getLatestImage() throws IOException {
if (foundImages.isEmpty()) {
throw new FileNotFoundException("No valid image files found");

View File

@ -233,6 +233,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
private static final ThreadLocal<StringBuilder> auditBuffer =
new ThreadLocal<StringBuilder>() {
@Override
protected StringBuilder initialValue() {
return new StringBuilder();
}
@ -4140,6 +4141,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
/**
*/
@Override
public void run() {
while (fsRunning && (safeMode != null && !safeMode.canLeave())) {
try {
@ -4244,6 +4246,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
* @param deltaSafe the change in number of safe blocks
* @param deltaTotal the change i nnumber of total blocks expected
*/
@Override
public void adjustSafeModeBlockTotals(int deltaSafe, int deltaTotal) {
// safeMode is volatile, and may be set to null at any time
SafeModeInfo safeMode = this.safeMode;
@ -4966,6 +4969,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
block = b;
}
@Override
public String toString() {
return block.getBlockName() + "\t" + path;
}
@ -5460,6 +5464,7 @@ public class FSNamesystem implements Namesystem, FSClusterStats,
return blockManager;
}
@Override
public boolean isGenStampInFuture(long genStamp) {
return (genStamp > getGenerationStamp());
}

View File

@ -106,6 +106,7 @@ public class FileDataServlet extends DfsServlet {
* GET http://<nn>:<port>/data[/<path>] HTTP/1.1
* }
*/
@Override
public void doGet(final HttpServletRequest request,
final HttpServletResponse response)
throws IOException {

View File

@ -339,6 +339,7 @@ class FileJournalManager implements JournalManager {
final static Comparator<EditLogFile> COMPARE_BY_START_TXID
= new Comparator<EditLogFile>() {
@Override
public int compare(EditLogFile a, EditLogFile b) {
return ComparisonChain.start()
.compare(a.getFirstTxId(), b.getFirstTxId())

View File

@ -42,6 +42,7 @@ public class FsckServlet extends DfsServlet {
private static final long serialVersionUID = 1L;
/** Handle fsck request */
@Override
public void doGet(HttpServletRequest request, HttpServletResponse response
) throws IOException {
@SuppressWarnings("unchecked")

View File

@ -79,6 +79,7 @@ public class GetImageServlet extends HttpServlet {
private static Set<Long> currentlyDownloadingCheckpoints =
Collections.<Long>synchronizedSet(new HashSet<Long>());
@Override
public void doGet(final HttpServletRequest request,
final HttpServletResponse response
) throws ServletException, IOException {

View File

@ -65,6 +65,7 @@ class INodeDirectory extends INode {
/**
* Check whether it's a directory
*/
@Override
public boolean isDirectory() {
return true;
}
@ -422,6 +423,7 @@ class INodeDirectory extends INode {
return children;
}
@Override
int collectSubtreeBlocksAndClear(List<Block> v) {
int total = 1;
if (children == null) {

View File

@ -71,6 +71,7 @@ class INodeDirectoryWithQuota extends INodeDirectory {
/** Get this directory's namespace quota
* @return this directory's namespace quota
*/
@Override
long getNsQuota() {
return nsQuota;
}
@ -78,6 +79,7 @@ class INodeDirectoryWithQuota extends INodeDirectory {
/** Get this directory's diskspace quota
* @return this directory's diskspace quota
*/
@Override
long getDsQuota() {
return dsQuota;
}

View File

@ -59,10 +59,12 @@ public class INodeFile extends INode implements BlockCollection {
* Since this is a file,
* the {@link FsAction#EXECUTE} action, if any, is ignored.
*/
@Override
void setPermission(FsPermission permission) {
super.setPermission(permission.applyUMask(UMASK));
}
@Override
boolean isDirectory() {
return false;
}
@ -138,6 +140,7 @@ public class INodeFile extends INode implements BlockCollection {
this.blocks[idx] = blk;
}
@Override
int collectSubtreeBlocksAndClear(List<Block> v) {
parent = null;
if(blocks != null && v != null) {

View File

@ -147,6 +147,7 @@ class INodeFileUnderConstruction extends INodeFile implements MutableBlockCollec
* Convert the last block of the file to an under-construction block.
* Set its locations.
*/
@Override
public BlockInfoUnderConstruction setLastBlock(BlockInfo lastBlock,
DatanodeDescriptor[] targets)
throws IOException {

View File

@ -40,6 +40,7 @@ public class INodeSymlink extends INode {
setAccessTime(atime);
}
@Override
public boolean isLink() {
return true;
}

View File

@ -83,6 +83,7 @@ public interface JournalManager extends Closeable {
/**
* Close the journal manager, freeing any resources it may hold.
*/
@Override
void close() throws IOException;
/**

View File

@ -390,6 +390,7 @@ public class LeaseManager {
final String name = getClass().getSimpleName();
/** Check leases periodically. */
@Override
public void run() {
for(; shouldRunMonitor && fsnamesystem.isRunning(); ) {
try {

View File

@ -54,6 +54,7 @@ public class ListPathsServlet extends DfsServlet {
public static final ThreadLocal<SimpleDateFormat> df =
new ThreadLocal<SimpleDateFormat>() {
@Override
protected SimpleDateFormat initialValue() {
return HftpFileSystem.getDateFormat();
}
@ -128,6 +129,7 @@ public class ListPathsServlet extends DfsServlet {
* </listing>
* }
*/
@Override
public void doGet(HttpServletRequest request, HttpServletResponse response)
throws ServletException, IOException {
final PrintWriter out = response.getWriter();

View File

@ -101,10 +101,12 @@ public class NNStorage extends Storage implements Closeable,
EDITS,
IMAGE_AND_EDITS;
@Override
public StorageDirType getStorageDirType() {
return this;
}
@Override
public boolean isOfType(StorageDirType type) {
if ((this == IMAGE_AND_EDITS) && (type == IMAGE || type == EDITS))
return true;

View File

@ -119,6 +119,7 @@ public class NameNodeResourceChecker {
Collection<URI> localEditDirs = Collections2.filter(
FSNamesystem.getNamespaceEditsDirs(conf),
new Predicate<URI>() {
@Override
public boolean apply(URI input) {
if (input.getScheme().equals(NNStorage.LOCAL_URI_SCHEME)) {
return true;

View File

@ -369,6 +369,7 @@ class NamenodeJspHelper {
final UserGroupInformation ugi) throws IOException, InterruptedException {
Token<DelegationTokenIdentifier> token = ugi
.doAs(new PrivilegedExceptionAction<Token<DelegationTokenIdentifier>>() {
@Override
public Token<DelegationTokenIdentifier> run() throws IOException {
return nn.getDelegationToken(new Text(ugi.getUserName()));
}

View File

@ -68,6 +68,7 @@ public class RenewDelegationTokenServlet extends DfsServlet {
try {
long result = ugi.doAs(new PrivilegedExceptionAction<Long>() {
@Override
public Long run() throws Exception {
return nn.getRpcServer().renewDelegationToken(token);
}

View File

@ -282,6 +282,7 @@ public class SecondaryNameNode implements Runnable {
}
}
@Override
public void run() {
SecurityUtil.doAsLoginUserOrFatal(
new PrivilegedAction<Object>() {

View File

@ -59,6 +59,7 @@ public class StreamFile extends DfsServlet {
return DatanodeJspHelper.getDFSClient(request, datanode, conf, ugi);
}
@Override
@SuppressWarnings("unchecked")
public void doGet(HttpServletRequest request, HttpServletResponse response)
throws ServletException, IOException {

View File

@ -38,6 +38,7 @@ import org.apache.hadoop.hdfs.server.protocol.UpgradeCommand;
* and updates its status.
*/
class UpgradeManagerNamenode extends UpgradeManager {
@Override
public HdfsServerConstants.NodeType getType() {
return HdfsServerConstants.NodeType.NAME_NODE;
}
@ -55,6 +56,7 @@ class UpgradeManagerNamenode extends UpgradeManager {
* @return true if distributed upgrade is required or false otherwise
* @throws IOException
*/
@Override
public synchronized boolean startUpgrade() throws IOException {
if(!upgradeState) {
initializeUpgrade();
@ -108,6 +110,7 @@ class UpgradeManagerNamenode extends UpgradeManager {
return reply;
}
@Override
public synchronized void completeUpgrade() throws IOException {
// set and write new upgrade state into disk
setUpgradeState(false, HdfsConstants.LAYOUT_VERSION);

View File

@ -44,12 +44,14 @@ public abstract class UpgradeObjectNamenode extends UpgradeObject {
public abstract UpgradeCommand processUpgradeCommand(UpgradeCommand command
) throws IOException;
@Override
public HdfsServerConstants.NodeType getType() {
return HdfsServerConstants.NodeType.NAME_NODE;
}
/**
*/
@Override
public UpgradeCommand startUpgrade() throws IOException {
// broadcast that data-nodes must start the upgrade
return new UpgradeCommand(UpgradeCommand.UC_ACTION_START_UPGRADE,

View File

@ -86,6 +86,7 @@ public class BootstrapStandby implements Tool, Configurable {
static final int ERR_CODE_ALREADY_FORMATTED = 5;
static final int ERR_CODE_LOGS_UNAVAILABLE = 6;
@Override
public int run(String[] args) throws Exception {
parseArgs(args);
parseConfAndFindOtherNN();

View File

@ -77,6 +77,7 @@ public class NamespaceInfo extends StorageInfo {
return softwareVersion;
}
@Override
public String toString(){
return super.toString() + ";bpid=" + blockPoolID;
}

View File

@ -42,5 +42,6 @@ public interface NodeRegistration {
*/
public int getVersion();
@Override
public String toString();
}

View File

@ -82,6 +82,7 @@ public class ReceivedDeletedBlockInfo {
return status;
}
@Override
public boolean equals(Object o) {
if (!(o instanceof ReceivedDeletedBlockInfo)) {
return false;
@ -93,6 +94,7 @@ public class ReceivedDeletedBlockInfo {
this.delHints != null && this.delHints.equals(other.delHints));
}
@Override
public int hashCode() {
assert false : "hashCode not designed";
return 0;
@ -106,6 +108,7 @@ public class ReceivedDeletedBlockInfo {
return status == BlockStatus.DELETED_BLOCK;
}
@Override
public String toString() {
return block.toString() + ", status: " + status +
", delHint: " + delHints;

View File

@ -115,6 +115,7 @@ public class DFSck extends Configured implements Tool {
/**
* @param args
*/
@Override
public int run(final String[] args) throws IOException {
if (args.length == 0) {
printUsage();

View File

@ -57,6 +57,7 @@ class OfflineEditsBinaryLoader implements OfflineEditsLoader {
/**
* Loads edits file, uses visitor to process all elements
*/
@Override
public void loadEdits() throws IOException {
visitor.start(inputStream.getVersion());
while (true) {

View File

@ -77,6 +77,7 @@ class OfflineEditsXmlLoader
/**
* Loads edits file, uses visitor to process all elements
*/
@Override
public void loadEdits() throws IOException {
try {
XMLReader xr = XMLReaderFactory.createXMLReader();
@ -120,6 +121,7 @@ class OfflineEditsXmlLoader
}
}
@Override
public void startElement (String uri, String name,
String qName, Attributes atts) {
switch (state) {
@ -168,6 +170,7 @@ class OfflineEditsXmlLoader
}
}
@Override
public void endElement (String uri, String name, String qName) {
String str = cbuf.toString().trim();
cbuf = new StringBuffer();
@ -248,6 +251,7 @@ class OfflineEditsXmlLoader
}
}
@Override
public void characters (char ch[], int start, int length) {
cbuf.append(ch, start, length);
}

View File

@ -59,6 +59,7 @@ class IndentedImageVisitor extends TextWriterImageVisitor {
write(element + " = " + value + "\n");
}
@Override
void visit(ImageElement element, long value) throws IOException {
if ((element == ImageElement.DELEGATION_TOKEN_IDENTIFIER_EXPIRY_TIME) ||
(element == ImageElement.DELEGATION_TOKEN_IDENTIFIER_ISSUE_DATE) ||

View File

@ -54,6 +54,7 @@ public class CyclicIteration<K, V> implements Iterable<Map.Entry<K, V>> {
}
/** {@inheritDoc} */
@Override
public Iterator<Map.Entry<K, V>> iterator() {
return new CyclicIterator();
}
@ -89,11 +90,13 @@ public class CyclicIteration<K, V> implements Iterable<Map.Entry<K, V>> {
}
/** {@inheritDoc} */
@Override
public boolean hasNext() {
return hasnext;
}
/** {@inheritDoc} */
@Override
public Map.Entry<K, V> next() {
if (!hasnext) {
throw new NoSuchElementException();
@ -106,6 +109,7 @@ public class CyclicIteration<K, V> implements Iterable<Map.Entry<K, V>> {
}
/** Not supported */
@Override
public void remove() {
throw new UnsupportedOperationException("Not supported");
}

View File

@ -55,6 +55,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
this.hashCode = hash;
}
@Override
public String toString() {
return element.toString();
}
@ -142,6 +143,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
*
* @return true is set empty, false otherwise
*/
@Override
public boolean isEmpty() {
return size == 0;
}
@ -156,6 +158,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
/**
* Return the number of stored elements.
*/
@Override
public int size() {
return size;
}
@ -217,6 +220,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
* @param toAdd - elements to add.
* @return true if the set has changed, false otherwise
*/
@Override
public boolean addAll(Collection<? extends T> toAdd) {
boolean changed = false;
for (T elem : toAdd) {
@ -231,6 +235,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
*
* @return true if the element was not present in the table, false otherwise
*/
@Override
public boolean add(final T element) {
boolean added = addElem(element);
expandIfNecessary();
@ -270,6 +275,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
*
* @return If such element exists, return true. Otherwise, return false.
*/
@Override
@SuppressWarnings("unchecked")
public boolean remove(final Object key) {
// validate key
@ -489,6 +495,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
}
}
@Override
public Iterator<T> iterator() {
return new LinkedSetIterator();
}
@ -560,6 +567,7 @@ public class LightWeightHashSet<T> implements Collection<T> {
/**
* Clear the set. Resize it to the original capacity.
*/
@Override
@SuppressWarnings("unchecked")
public void clear() {
this.capacity = this.initialCapacity;

View File

@ -47,6 +47,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
this.after = null;
}
@Override
public String toString() {
return super.toString();
}
@ -79,6 +80,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
*
* @return true if the element was not present in the table, false otherwise
*/
@Override
protected boolean addElem(final T element) {
// validate element
if (element == null) {
@ -118,6 +120,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
*
* @return Return the entry with the element if exists. Otherwise return null.
*/
@Override
protected DoubleLinkedElement<T> removeElem(final T key) {
DoubleLinkedElement<T> found = (DoubleLinkedElement<T>) (super
.removeElem(key));
@ -162,6 +165,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
*
* @return first element
*/
@Override
public List<T> pollN(int n) {
if (n >= size) {
// if we need to remove all elements then do fast polling
@ -182,6 +186,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
* link list, don't worry about hashtable - faster version of the parent
* method.
*/
@Override
public List<T> pollAll() {
List<T> retList = new ArrayList<T>(size);
while (head != null) {
@ -212,6 +217,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
return a;
}
@Override
public Iterator<T> iterator() {
return new LinkedSetIterator();
}
@ -251,6 +257,7 @@ public class LightWeightLinkedSet<T> extends LightWeightHashSet<T> {
/**
* Clear the set. Resize it to the original capacity.
*/
@Override
public void clear() {
super.clear();
this.head = null;

View File

@ -146,6 +146,7 @@ public class XMLUtils {
/**
* Convert a stanza to a human-readable string.
*/
@Override
public String toString() {
StringBuilder bld = new StringBuilder();
bld.append("{");

View File

@ -49,6 +49,7 @@ public class PostOpParam extends HttpOpParam<PostOpParam.Op> {
}
/** @return a URI query string. */
@Override
public String toQueryString() {
return NAME + "=" + this;
}

View File

@ -55,6 +55,7 @@ public class TestFcHdfsCreateMkdir extends
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
super.setUp();

View File

@ -55,6 +55,7 @@ public class TestFcHdfsPermission extends FileContextPermissionBase {
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
super.setUp();
@ -72,6 +73,7 @@ public class TestFcHdfsPermission extends FileContextPermissionBase {
*/
static final FsPermission FILE_MASK_IGNORE_X_BIT =
new FsPermission((short) ~0666);
@Override
FsPermission getFileMask() {
return FILE_MASK_IGNORE_X_BIT;
}

View File

@ -53,18 +53,22 @@ public class TestFcHdfsSymlink extends FileContextSymlinkBaseTest {
private static WebHdfsFileSystem webhdfs;
@Override
protected String getScheme() {
return "hdfs";
}
@Override
protected String testBaseDir1() throws IOException {
return "/test1";
}
@Override
protected String testBaseDir2() throws IOException {
return "/test2";
}
@Override
protected URI testURI() {
return cluster.getURI(0);
}

View File

@ -35,6 +35,7 @@ public class TestGlobPaths extends TestCase {
this.regex = regex;
}
@Override
public boolean accept(Path path) {
return path.toString().matches(regex);
}
@ -47,6 +48,7 @@ public class TestGlobPaths extends TestCase {
static final String USER_DIR = "/user/"+System.getProperty("user.name");
private Path[] path = new Path[NUM_OF_PATHS];
@Override
protected void setUp() throws Exception {
try {
Configuration conf = new HdfsConfiguration();
@ -57,6 +59,7 @@ public class TestGlobPaths extends TestCase {
}
}
@Override
protected void tearDown() throws Exception {
if(dfsCluster!=null) {
dfsCluster.shutdown();

View File

@ -75,6 +75,7 @@ public class TestHDFSFileContextMainOperations extends
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
super.setUp();

View File

@ -62,6 +62,7 @@ public class TestViewFileSystemAtHdfsRoot extends ViewFileSystemBaseTest {
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
fsTarget = fHdfs;

View File

@ -80,6 +80,7 @@ public class TestViewFileSystemHdfs extends ViewFileSystemBaseTest {
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
// create the test root on local_fs
@ -89,6 +90,7 @@ public class TestViewFileSystemHdfs extends ViewFileSystemBaseTest {
super.setUp();
}
@Override
@After
public void tearDown() throws Exception {
super.tearDown();

View File

@ -61,6 +61,7 @@ public class TestViewFsAtHdfsRoot extends ViewFsBaseTest {
cluster.shutdown();
}
@Override
@Before
public void setUp() throws Exception {
// create the test root on local_fs

Some files were not shown because too many files have changed in this diff Show More