From a626fa04f983623b1e2c00189df6f0b83b806b5f Mon Sep 17 00:00:00 2001 From: Todd Lipcon Date: Fri, 10 Feb 2012 18:17:21 +0000 Subject: [PATCH] Revert HDFS-2922 via svn merge -c -1242572 The patch broke a lot of unit tests in the nightly build. Will recommit after it is fixed. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-1623@1242874 13f79535-47bb-0310-9956-ffa450edef68 --- .../io/retry/RetryInvocationHandler.java | 2 ++ .../hadoop-hdfs/CHANGES.HDFS-1623.txt | 2 -- .../hadoop/hdfs/DistributedFileSystem.java | 4 +-- .../server/namenode/NameNodeRpcServer.java | 26 +++++++++---------- 4 files changed, 17 insertions(+), 17 deletions(-) diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/RetryInvocationHandler.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/RetryInvocationHandler.java index dc65f2c2aba..3d702d9879d 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/RetryInvocationHandler.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/retry/RetryInvocationHandler.java @@ -113,6 +113,8 @@ public Object invoke(Object proxy, Method method, Object[] args) msg += ". Trying to fail over " + formatSleepMessage(action.delayMillis); if (LOG.isDebugEnabled()) { LOG.debug(msg, e); + } else { + LOG.warn(msg); } } else { if(LOG.isDebugEnabled()) { diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-1623.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-1623.txt index 23028003649..3eec7452792 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-1623.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-1623.txt @@ -193,6 +193,4 @@ HDFS-2915. HA: TestFailureOfSharedDir.testFailureOfSharedDir() has race conditio HDFS-2912. Namenode not shutting down when shared edits dir is inaccessible. (Bikas Saha via atm) -HDFS-2922. HA: close out operation categories. (eli) - HDFS-2917. HA: haadmin should not work if run by regular user (eli) diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java index eae4cf1d3ac..8dfced350a2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java @@ -646,7 +646,7 @@ public UpgradeStatusReport distributedUpgradeProgress(UpgradeAction action } /* - * Requests the namenode to dump data structures into specified + * Requests the namenode to dump data strcutures into specified * file. */ public void metaSave(String pathname) throws IOException { @@ -664,7 +664,7 @@ public FsServerDefaults getServerDefaults() throws IOException { * we can consider figuring out exactly which block is corrupt. */ // We do not see a need for user to report block checksum errors and do not - // want to rely on user to report block corruption. + // want to rely on user to report block corruptions. @Deprecated public boolean reportChecksumFailure(Path f, FSDataInputStream in, long inPos, diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java index c6058851086..9f85dacc624 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java @@ -700,7 +700,7 @@ public long[] getStats() throws IOException { @Override // ClientProtocol public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) throws IOException { - namesystem.checkOperation(OperationCategory.READ); + // TODO(HA): decide on OperationCategory for this DatanodeInfo results[] = namesystem.datanodeReport(type); if (results == null ) { throw new IOException("Cannot find datanode report"); @@ -710,32 +710,32 @@ public DatanodeInfo[] getDatanodeReport(DatanodeReportType type) @Override // ClientProtocol public boolean setSafeMode(SafeModeAction action) throws IOException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this return namesystem.setSafeMode(action); } @Override // ClientProtocol public boolean restoreFailedStorage(String arg) throws AccessControlException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this return namesystem.restoreFailedStorage(arg); } @Override // ClientProtocol public void saveNamespace() throws IOException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this namesystem.saveNamespace(); } @Override // ClientProtocol public void refreshNodes() throws IOException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this namesystem.getBlockManager().getDatanodeManager().refreshNodes( new HdfsConfiguration()); } @Override // NamenodeProtocol - public long getTransactionID() throws IOException { - namesystem.checkOperation(OperationCategory.READ); + public long getTransactionID() { + // TODO:HA decide on OperationCategory for this return namesystem.getEditLog().getSyncTxId(); } @@ -747,26 +747,26 @@ public CheckpointSignature rollEditLog() throws IOException { @Override // NamenodeProtocol public RemoteEditLogManifest getEditLogManifest(long sinceTxId) throws IOException { - namesystem.checkOperation(OperationCategory.READ); + // TODO:HA decide on OperationCategory for this return namesystem.getEditLog().getEditLogManifest(sinceTxId); } @Override // ClientProtocol public void finalizeUpgrade() throws IOException { - namesystem.checkOperation(OperationCategory.WRITE); + // TODO:HA decide on OperationCategory for this namesystem.finalizeUpgrade(); } @Override // ClientProtocol public UpgradeStatusReport distributedUpgradeProgress(UpgradeAction action) throws IOException { - namesystem.checkOperation(OperationCategory.READ); + // TODO:HA decide on OperationCategory for this return namesystem.distributedUpgradeProgress(action); } @Override // ClientProtocol public void metaSave(String filename) throws IOException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this namesystem.metaSave(filename); } @Override // ClientProtocol @@ -787,12 +787,12 @@ public CorruptFileBlocks listCorruptFileBlocks(String path, String cookie) /** * Tell all datanodes to use a new, non-persistent bandwidth value for * dfs.datanode.balance.bandwidthPerSec. - * @param bandwidth Balancer bandwidth in bytes per second for all datanodes. + * @param bandwidth Blanacer bandwidth in bytes per second for all datanodes. * @throws IOException */ @Override // ClientProtocol public void setBalancerBandwidth(long bandwidth) throws IOException { - // NB: not checking OperationCategory so this works on a standby + // TODO:HA decide on OperationCategory for this namesystem.getBlockManager().getDatanodeManager().setBalancerBandwidth(bandwidth); }