merge HDFS-2574

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1226212 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Harsh J 2012-01-01 03:16:21 +00:00
parent 01b14ae65f
commit c0f4668bd4
4 changed files with 17 additions and 15 deletions

View File

@ -19,7 +19,7 @@
<!-- file system properties --> <!-- file system properties -->
<property> <property>
<name>dfs.name.dir</name> <name>dfs.namenode.name.dir</name>
<value>${HADOOP_NN_DIR}</value> <value>${HADOOP_NN_DIR}</value>
<description>Determines where on the local filesystem the DFS name node <description>Determines where on the local filesystem the DFS name node
should store the name table. If this is a comma-delimited list should store the name table. If this is a comma-delimited list
@ -29,7 +29,7 @@
</property> </property>
<property> <property>
<name>dfs.data.dir</name> <name>dfs.datanode.data.dir</name>
<value>${HADOOP_DN_DIR}</value> <value>${HADOOP_DN_DIR}</value>
<description>Determines where on the local filesystem an DFS data node <description>Determines where on the local filesystem an DFS data node
should store its blocks. If this is a comma-delimited should store its blocks. If this is a comma-delimited
@ -41,11 +41,11 @@
</property> </property>
<property> <property>
<name>dfs.safemode.threshold.pct</name> <name>dfs.namenode.safemode.threshold-pct</name>
<value>1.0f</value> <value>1.0f</value>
<description> <description>
Specifies the percentage of blocks that should satisfy Specifies the percentage of blocks that should satisfy the minimal
the minimal replication requirement defined by dfs.replication.min. replication requirement defined by dfs.namenode.replication.min.
Values less than or equal to 0 mean not to start in safe mode. Values less than or equal to 0 mean not to start in safe mode.
Values greater than 1 will make safe mode permanent. Values greater than 1 will make safe mode permanent.
</description> </description>
@ -62,7 +62,7 @@
</property> </property>
<property> <property>
<name>dfs.http.address</name> <name>dfs.namenode.http-address</name>
<value>${HADOOP_NN_HOST}:50070</value> <value>${HADOOP_NN_HOST}:50070</value>
<description>The name of the default file system. Either the <description>The name of the default file system. Either the
literal string "local" or a host:port for NDFS. literal string "local" or a host:port for NDFS.
@ -167,7 +167,7 @@
</property> </property>
<property> <property>
<name>dfs.https.address</name> <name>dfs.namenode.https-address</name>
<value>${HADOOP_NN_HOST}:50470</value> <value>${HADOOP_NN_HOST}:50470</value>
<description>The https address where namenode binds</description> <description>The https address where namenode binds</description>
</property> </property>
@ -175,9 +175,9 @@
<property> <property>
<name>dfs.datanode.data.dir.perm</name> <name>dfs.datanode.data.dir.perm</name>
<value>${DFS_DATANODE_DIR_PERM}</value> <value>${DFS_DATANODE_DIR_PERM}</value>
<description>The permissions that should be there on dfs.data.dir <description>The permissions that should be there on dfs.datanode.data.dir
directories. The datanode will not come up if the permissions are directories. The datanode will not come up if the permissions are
different on existing dfs.data.dir directories. If the directories different on existing dfs.datanode.data.dir directories. If the directories
don't exist, they will be created with this permission. don't exist, they will be created with this permission.
</description> </description>
</property> </property>
@ -209,7 +209,7 @@
</property> </property>
<property> <property>
<name>dfs.secondary.http.address</name> <name>dfs.namenode.secondary.http-address</name>
<value>${HADOOP_SNN_HOST}:50090</value> <value>${HADOOP_SNN_HOST}:50090</value>
<description> <description>
The secondary namenode http server address and port. The secondary namenode http server address and port.

View File

@ -61,6 +61,8 @@ Release 0.23.1 - UNRELEASED
HDFS-2335. DataNodeCluster and NNStorage always pull fresh entropy. HDFS-2335. DataNodeCluster and NNStorage always pull fresh entropy.
(Uma Maheswara Rao G via eli) (Uma Maheswara Rao G via eli)
HDFS-2574. Remove references to some deprecated properties in conf templates and defaults files. (Joe Crobak via harsh)
OPTIMIZATIONS OPTIMIZATIONS
HDFS-2130. Switch default checksum to CRC32C. (todd) HDFS-2130. Switch default checksum to CRC32C. (todd)

View File

@ -22,11 +22,11 @@
<value>${HADOOP_REPLICATION}</value> <value>${HADOOP_REPLICATION}</value>
</property> </property>
<property> <property>
<name>dfs.name.dir</name> <name>dfs.namenode.name.dir</name>
<value>${HADOOP_NN_DIR}</value> <value>${HADOOP_NN_DIR}</value>
</property> </property>
<property> <property>
<name>dfs.data.dir</name> <name>dfs.datanode.data.dir</name>
<value>${HADOOP_DN_DIR}</value> <value>${HADOOP_DN_DIR}</value>
</property> </property>
<property> <property>

View File

@ -199,8 +199,8 @@ creations/deletions), or "all".</description>
<name>dfs.namenode.name.dir.restore</name> <name>dfs.namenode.name.dir.restore</name>
<value>false</value> <value>false</value>
<description>Set to true to enable NameNode to attempt recovering a <description>Set to true to enable NameNode to attempt recovering a
previously failed dfs.name.dir. When enabled, a recovery of any failed previously failed dfs.namenode.name.dir. When enabled, a recovery of any
directory is attempted during checkpoint.</description> failed directory is attempted during checkpoint.</description>
</property> </property>
<property> <property>
@ -223,7 +223,7 @@ creations/deletions), or "all".</description>
<description>Determines where on the local filesystem the DFS name node <description>Determines where on the local filesystem the DFS name node
should store the transaction (edits) file. If this is a comma-delimited list should store the transaction (edits) file. If this is a comma-delimited list
of directories then the transaction file is replicated in all of the of directories then the transaction file is replicated in all of the
directories, for redundancy. Default value is same as dfs.name.dir directories, for redundancy. Default value is same as dfs.namenode.name.dir
</description> </description>
</property> </property>
<property> <property>