HADOOP-12733. Remove references to obsolete io.seqfile configuration variables. Contributed by Ray Chiang.
This commit is contained in:
parent
87bb1c49bb
commit
01d31fe938
|
@ -1262,22 +1262,6 @@
|
||||||
</description>
|
</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<property>
|
|
||||||
<name>io.seqfile.lazydecompress</name>
|
|
||||||
<value>true</value>
|
|
||||||
<description>Should values of block-compressed SequenceFiles be decompressed
|
|
||||||
only when necessary.
|
|
||||||
</description>
|
|
||||||
</property>
|
|
||||||
|
|
||||||
<property>
|
|
||||||
<name>io.seqfile.sorter.recordlimit</name>
|
|
||||||
<value>1000000</value>
|
|
||||||
<description>The limit on number of records to be kept in memory in a spill
|
|
||||||
in SequenceFiles.Sorter
|
|
||||||
</description>
|
|
||||||
</property>
|
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>io.mapfile.bloom.size</name>
|
<name>io.mapfile.bloom.size</name>
|
||||||
<value>1048576</value>
|
<value>1048576</value>
|
||||||
|
|
|
@ -124,8 +124,6 @@ public class TestCommonConfigurationFields extends TestConfigurationFieldsBase {
|
||||||
configurationPropsToSkipCompare.add("dr.who");
|
configurationPropsToSkipCompare.add("dr.who");
|
||||||
|
|
||||||
// XML deprecated properties.
|
// XML deprecated properties.
|
||||||
xmlPropsToSkipCompare.add("io.seqfile.lazydecompress");
|
|
||||||
xmlPropsToSkipCompare.add("io.seqfile.sorter.recordlimit");
|
|
||||||
// - org.apache.hadoop.hdfs.client.HdfsClientConfigKeys
|
// - org.apache.hadoop.hdfs.client.HdfsClientConfigKeys
|
||||||
xmlPropsToSkipCompare
|
xmlPropsToSkipCompare
|
||||||
.add("io.bytes.per.checksum");
|
.add("io.bytes.per.checksum");
|
||||||
|
|
|
@ -16,7 +16,6 @@
|
||||||
<property><!--Loaded from job.xml--><name>mapreduce.reduce.shuffle.memory.limit.percent</name><value>0.25</value></property>
|
<property><!--Loaded from job.xml--><name>mapreduce.reduce.shuffle.memory.limit.percent</name><value>0.25</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>hadoop.http.authentication.kerberos.keytab</name><value>${user.home}/hadoop.keytab</value></property>
|
<property><!--Loaded from job.xml--><name>hadoop.http.authentication.kerberos.keytab</name><value>${user.home}/hadoop.keytab</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>yarn.nodemanager.keytab</name><value>/etc/krb5.keytab</value></property>
|
<property><!--Loaded from job.xml--><name>yarn.nodemanager.keytab</name><value>/etc/krb5.keytab</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>io.seqfile.sorter.recordlimit</name><value>1000000</value></property>
|
|
||||||
<property><!--Loaded from job.xml--><name>mapreduce.task.io.sort.factor</name><value>10</value></property>
|
<property><!--Loaded from job.xml--><name>mapreduce.task.io.sort.factor</name><value>10</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>yarn.nodemanager.disk-health-checker.interval-ms</name><value>120000</value></property>
|
<property><!--Loaded from job.xml--><name>yarn.nodemanager.disk-health-checker.interval-ms</name><value>120000</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>mapreduce.job.working.dir</name><value>hdfs://localhost:8021/user/user</value></property>
|
<property><!--Loaded from job.xml--><name>mapreduce.job.working.dir</name><value>hdfs://localhost:8021/user/user</value></property>
|
||||||
|
@ -102,7 +101,6 @@
|
||||||
<property><!--Loaded from job.xml--><name>dfs.client.block.write.retries</name><value>3</value></property>
|
<property><!--Loaded from job.xml--><name>dfs.client.block.write.retries</name><value>3</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>hadoop.proxyuser.user.groups</name><value>users</value></property>
|
<property><!--Loaded from job.xml--><name>hadoop.proxyuser.user.groups</name><value>users</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>dfs.namenode.name.dir.restore</name><value>false</value></property>
|
<property><!--Loaded from job.xml--><name>dfs.namenode.name.dir.restore</name><value>false</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>io.seqfile.lazydecompress</name><value>true</value></property>
|
|
||||||
<property><!--Loaded from job.xml--><name>mapreduce.reduce.merge.inmem.threshold</name><value>1000</value></property>
|
<property><!--Loaded from job.xml--><name>mapreduce.reduce.merge.inmem.threshold</name><value>1000</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>mapreduce.input.fileinputformat.split.minsize</name><value>0</value></property>
|
<property><!--Loaded from job.xml--><name>mapreduce.input.fileinputformat.split.minsize</name><value>0</value></property>
|
||||||
<property><!--Loaded from job.xml--><name>dfs.replication</name><value>3</value></property>
|
<property><!--Loaded from job.xml--><name>dfs.replication</name><value>3</value></property>
|
||||||
|
|
|
@ -4545,7 +4545,6 @@
|
||||||
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
||||||
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
||||||
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
||||||
"io.seqfile.sorter.recordlimit" : "1000000",
|
|
||||||
"mapreduce.task.io.sort.factor" : "10",
|
"mapreduce.task.io.sort.factor" : "10",
|
||||||
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
||||||
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
||||||
|
@ -4662,7 +4661,6 @@
|
||||||
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
||||||
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
||||||
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
||||||
"io.seqfile.lazydecompress" : "true",
|
|
||||||
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
||||||
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
||||||
"dfs.replication" : "3",
|
"dfs.replication" : "3",
|
||||||
|
@ -9627,7 +9625,6 @@
|
||||||
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
||||||
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
||||||
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
||||||
"io.seqfile.sorter.recordlimit" : "1000000",
|
|
||||||
"mapreduce.task.io.sort.factor" : "10",
|
"mapreduce.task.io.sort.factor" : "10",
|
||||||
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
||||||
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
||||||
|
@ -9744,7 +9741,6 @@
|
||||||
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
||||||
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
||||||
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
||||||
"io.seqfile.lazydecompress" : "true",
|
|
||||||
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
||||||
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
||||||
"dfs.replication" : "3",
|
"dfs.replication" : "3",
|
||||||
|
@ -10209,7 +10205,6 @@
|
||||||
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
"hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory",
|
||||||
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
"hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab",
|
||||||
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
"yarn.nodemanager.keytab" : "/etc/krb5.keytab",
|
||||||
"io.seqfile.sorter.recordlimit" : "1000000",
|
|
||||||
"mapreduce.task.io.sort.factor" : "10",
|
"mapreduce.task.io.sort.factor" : "10",
|
||||||
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
"yarn.nodemanager.disk-health-checker.interval-ms" : "120000",
|
||||||
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
"mapreduce.job.working.dir" : "hdfs://a2115.smile.com:9820/user/jenkins",
|
||||||
|
@ -10327,7 +10322,6 @@
|
||||||
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
"rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine",
|
||||||
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
"dfs.datanode.hdfs-blocks-metadata.enabled" : "true",
|
||||||
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
"ha.zookeeper.parent-znode" : "/hadoop-ha",
|
||||||
"io.seqfile.lazydecompress" : "true",
|
|
||||||
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
"mapreduce.reduce.merge.inmem.threshold" : "1000",
|
||||||
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
"mapreduce.input.fileinputformat.split.minsize" : "0",
|
||||||
"dfs.replication" : "3",
|
"dfs.replication" : "3",
|
||||||
|
|
Loading…
Reference in New Issue