HDFS-6003. Add the new -rollingUpgrade startup option to the namenode usage message. Contributed by Vinayakumar B
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/HDFS-5535@1571596 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
98db64dd49
commit
2c8a6f5712
|
@ -85,3 +85,6 @@ HDFS-5535 subtasks:
|
||||||
HDFS-5583. Make DN send an OOB Ack on shutdown before restarting. (kihwal)
|
HDFS-5583. Make DN send an OOB Ack on shutdown before restarting. (kihwal)
|
||||||
|
|
||||||
HDFS-5778. Add rolling upgrade user document. (szetszwo)
|
HDFS-5778. Add rolling upgrade user document. (szetszwo)
|
||||||
|
|
||||||
|
HDFS-6003. Add the new -rollingUpgrade startup option to the namenode
|
||||||
|
usage message. (Vinayakumar B via szetszwo)
|
||||||
|
|
|
@ -21,6 +21,7 @@ import com.google.common.annotations.VisibleForTesting;
|
||||||
import com.google.common.base.Joiner;
|
import com.google.common.base.Joiner;
|
||||||
import com.google.common.base.Preconditions;
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.HadoopIllegalArgumentException;
|
import org.apache.hadoop.HadoopIllegalArgumentException;
|
||||||
|
@ -40,6 +41,7 @@ import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||||
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
|
||||||
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption;
|
||||||
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.ha.*;
|
import org.apache.hadoop.hdfs.server.namenode.ha.*;
|
||||||
import org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics;
|
import org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics;
|
||||||
|
@ -64,6 +66,7 @@ import org.apache.hadoop.util.ServicePlugin;
|
||||||
import org.apache.hadoop.util.StringUtils;
|
import org.apache.hadoop.util.StringUtils;
|
||||||
|
|
||||||
import javax.management.ObjectName;
|
import javax.management.ObjectName;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.PrintStream;
|
import java.io.PrintStream;
|
||||||
import java.net.InetSocketAddress;
|
import java.net.InetSocketAddress;
|
||||||
|
@ -205,6 +208,9 @@ public class NameNode implements NameNodeStatusMXBean {
|
||||||
" [" + StartupOption.CLUSTERID.getName() + " cid]" +
|
" [" + StartupOption.CLUSTERID.getName() + " cid]" +
|
||||||
" [" + StartupOption.RENAMERESERVED.getName() + "<k-v pairs>] ] | ["
|
" [" + StartupOption.RENAMERESERVED.getName() + "<k-v pairs>] ] | ["
|
||||||
+ StartupOption.ROLLBACK.getName() + "] | ["
|
+ StartupOption.ROLLBACK.getName() + "] | ["
|
||||||
|
+ StartupOption.ROLLINGUPGRADE.getName() + " <"
|
||||||
|
+ RollingUpgradeStartupOption.DOWNGRADE.name().toLowerCase() + "|"
|
||||||
|
+ RollingUpgradeStartupOption.ROLLBACK.name().toLowerCase() + "> ] | ["
|
||||||
+ StartupOption.FINALIZE.getName() + "] | ["
|
+ StartupOption.FINALIZE.getName() + "] | ["
|
||||||
+ StartupOption.IMPORT.getName() + "] | ["
|
+ StartupOption.IMPORT.getName() + "] | ["
|
||||||
+ StartupOption.INITIALIZESHAREDEDITS.getName() + "] | ["
|
+ StartupOption.INITIALIZESHAREDEDITS.getName() + "] | ["
|
||||||
|
|
Loading…
Reference in New Issue