From 3e2d26a97f83461d491180bbc6aa0ba926c0681b Mon Sep 17 00:00:00 2001 From: Akira Ajisaka Date: Wed, 4 Jan 2017 14:10:36 +0900 Subject: [PATCH] HADOOP-12733. Remove references to obsolete io.seqfile configuration variables. Contributed by Ray Chiang. (cherry picked from commit 01d31fe9389ccdc153d7f4bf6574bf8e509867c1) --- .../src/main/resources/core-default.xml | 16 ---------------- .../conf/TestCommonConfigurationFields.java | 2 -- .../resources/job_1329348432655_0001_conf.xml | 2 -- .../src/main/data/2jobs2min-rumen-jh.json | 6 ------ 4 files changed, 26 deletions(-) diff --git a/hadoop-common-project/hadoop-common/src/main/resources/core-default.xml b/hadoop-common-project/hadoop-common/src/main/resources/core-default.xml index 4431d603be4..503969d6dcb 100644 --- a/hadoop-common-project/hadoop-common/src/main/resources/core-default.xml +++ b/hadoop-common-project/hadoop-common/src/main/resources/core-default.xml @@ -1259,22 +1259,6 @@ - - io.seqfile.lazydecompress - true - Should values of block-compressed SequenceFiles be decompressed - only when necessary. - - - - - io.seqfile.sorter.recordlimit - 1000000 - The limit on number of records to be kept in memory in a spill - in SequenceFiles.Sorter - - - io.mapfile.bloom.size 1048576 diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestCommonConfigurationFields.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestCommonConfigurationFields.java index 02082e9d97a..7713c19be74 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestCommonConfigurationFields.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestCommonConfigurationFields.java @@ -120,8 +120,6 @@ public void initializeMemberVariables() { configurationPropsToSkipCompare.add("dr.who"); // XML deprecated properties. - xmlPropsToSkipCompare.add("io.seqfile.lazydecompress"); - xmlPropsToSkipCompare.add("io.seqfile.sorter.recordlimit"); // - org.apache.hadoop.hdfs.client.HdfsClientConfigKeys xmlPropsToSkipCompare .add("io.bytes.per.checksum"); diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml index 9d4d3602063..27fed7ba7e8 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml @@ -16,7 +16,6 @@ mapreduce.reduce.shuffle.memory.limit.percent0.25 hadoop.http.authentication.kerberos.keytab${user.home}/hadoop.keytab yarn.nodemanager.keytab/etc/krb5.keytab -io.seqfile.sorter.recordlimit1000000 s3.blocksize67108864 mapreduce.task.io.sort.factor10 yarn.nodemanager.disk-health-checker.interval-ms120000 @@ -111,7 +110,6 @@ dfs.client.block.write.retries3 hadoop.proxyuser.user.groupsusers dfs.namenode.name.dir.restorefalse -io.seqfile.lazydecompresstrue mapreduce.reduce.merge.inmem.threshold1000 mapreduce.input.fileinputformat.split.minsize0 dfs.replication3 diff --git a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json index 7ad31bfd5be..be868aff147 100644 --- a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json +++ b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json @@ -4545,7 +4545,6 @@ "hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory", "hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab", "yarn.nodemanager.keytab" : "/etc/krb5.keytab", - "io.seqfile.sorter.recordlimit" : "1000000", "s3.blocksize" : "67108864", "mapreduce.task.io.sort.factor" : "10", "yarn.nodemanager.disk-health-checker.interval-ms" : "120000", @@ -4671,7 +4670,6 @@ "rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine", "dfs.datanode.hdfs-blocks-metadata.enabled" : "true", "ha.zookeeper.parent-znode" : "/hadoop-ha", - "io.seqfile.lazydecompress" : "true", "mapreduce.reduce.merge.inmem.threshold" : "1000", "mapreduce.input.fileinputformat.split.minsize" : "0", "dfs.replication" : "3", @@ -9654,7 +9652,6 @@ "hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory", "hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab", "yarn.nodemanager.keytab" : "/etc/krb5.keytab", - "io.seqfile.sorter.recordlimit" : "1000000", "s3.blocksize" : "67108864", "mapreduce.task.io.sort.factor" : "10", "yarn.nodemanager.disk-health-checker.interval-ms" : "120000", @@ -9780,7 +9777,6 @@ "rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine", "dfs.datanode.hdfs-blocks-metadata.enabled" : "true", "ha.zookeeper.parent-znode" : "/hadoop-ha", - "io.seqfile.lazydecompress" : "true", "mapreduce.reduce.merge.inmem.threshold" : "1000", "mapreduce.input.fileinputformat.split.minsize" : "0", "dfs.replication" : "3", @@ -10263,7 +10259,6 @@ "hadoop.ssl.keystores.factory.class" : "org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory", "hadoop.http.authentication.kerberos.keytab" : "${user.home}/hadoop.keytab", "yarn.nodemanager.keytab" : "/etc/krb5.keytab", -"io.seqfile.sorter.recordlimit" : "1000000", "s3.blocksize" : "67108864", "mapreduce.task.io.sort.factor" : "10", "yarn.nodemanager.disk-health-checker.interval-ms" : "120000", @@ -10389,7 +10384,6 @@ "rpc.engine.org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB" : "org.apache.hadoop.ipc.ProtobufRpcEngine", "dfs.datanode.hdfs-blocks-metadata.enabled" : "true", "ha.zookeeper.parent-znode" : "/hadoop-ha", -"io.seqfile.lazydecompress" : "true", "mapreduce.reduce.merge.inmem.threshold" : "1000", "mapreduce.input.fileinputformat.split.minsize" : "0", "dfs.replication" : "3",