Merge -r 1244263:1244264 from trunk to branch. FIXES: MAPREDUCE-3736
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1244265 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
c8debd083b
commit
9ea26a7b08
|
@ -216,6 +216,13 @@
|
||||||
determine the host, port, etc. for a filesystem.</description>
|
determine the host, port, etc. for a filesystem.</description>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
|
<property>
|
||||||
|
<name>fs.default.name</name>
|
||||||
|
<value>file:///</value>
|
||||||
|
<description>Deprecated. Use (fs.defaultFS) property
|
||||||
|
instead</description>
|
||||||
|
</property>
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>fs.trash.interval</name>
|
<name>fs.trash.interval</name>
|
||||||
<value>0</value>
|
<value>0</value>
|
||||||
|
|
|
@ -52,6 +52,9 @@ Release 0.23.1 - 2012-02-08
|
||||||
MAPREDUCE-3840. JobEndNotifier doesn't use the proxyToUse during connecting
|
MAPREDUCE-3840. JobEndNotifier doesn't use the proxyToUse during connecting
|
||||||
(Ravi Prakash via bobby)
|
(Ravi Prakash via bobby)
|
||||||
|
|
||||||
|
MAPREDUCE-3736. Variable substitution depth too large for fs.default.name
|
||||||
|
causes jobs to fail (ahmed via tucu).
|
||||||
|
|
||||||
MAPREDUCE-2733. [Gridmix] Gridmix3 cpu emulation system tests.
|
MAPREDUCE-2733. [Gridmix] Gridmix3 cpu emulation system tests.
|
||||||
(Vinay Thota via amarrk)
|
(Vinay Thota via amarrk)
|
||||||
|
|
||||||
|
|
|
@ -159,7 +159,7 @@ public class TestMRWithDistributedCache extends TestCase {
|
||||||
public void testLocalJobRunner() throws Exception {
|
public void testLocalJobRunner() throws Exception {
|
||||||
Configuration c = new Configuration();
|
Configuration c = new Configuration();
|
||||||
c.set(JTConfig.JT_IPC_ADDRESS, "local");
|
c.set(JTConfig.JT_IPC_ADDRESS, "local");
|
||||||
c.set("fs.default.name", "file:///");
|
c.set("fs.defaultFS", "file:///");
|
||||||
testWithConf(c);
|
testWithConf(c);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -59,7 +59,7 @@ public class TestNoDefaultsJobConf extends HadoopTestCase {
|
||||||
|
|
||||||
JobConf conf = new JobConf(false);
|
JobConf conf = new JobConf(false);
|
||||||
|
|
||||||
conf.set("fs.default.name", createJobConf().get("fs.default.name"));
|
conf.set("fs.defaultFS", createJobConf().get("fs.defaultFS"));
|
||||||
|
|
||||||
conf.setJobName("mr");
|
conf.setJobName("mr");
|
||||||
|
|
||||||
|
|
|
@ -1024,7 +1024,7 @@ public class JHLogAnalyzer {
|
||||||
if(testFile != null) {
|
if(testFile != null) {
|
||||||
LOG.info("Start JHLA test ============ ");
|
LOG.info("Start JHLA test ============ ");
|
||||||
LocalFileSystem lfs = FileSystem.getLocal(conf);
|
LocalFileSystem lfs = FileSystem.getLocal(conf);
|
||||||
conf.set("fs.default.name", "file:///");
|
conf.set("fs.defaultFS", "file:///");
|
||||||
JHLAMapper map = new JHLAMapper(conf);
|
JHLAMapper map = new JHLAMapper(conf);
|
||||||
map.parseLogFile(lfs, new Path(testFile), 0L,
|
map.parseLogFile(lfs, new Path(testFile), 0L,
|
||||||
new LoggingCollector(), Reporter.NULL);
|
new LoggingCollector(), Reporter.NULL);
|
||||||
|
|
|
@ -53,7 +53,7 @@ public class FileBench extends Configured implements Tool {
|
||||||
"unless they are also explicitly included, as in \"-pln -zip\"\n" +
|
"unless they are also explicitly included, as in \"-pln -zip\"\n" +
|
||||||
"Note that CompressionType params only apply to SequenceFiles\n\n" +
|
"Note that CompressionType params only apply to SequenceFiles\n\n" +
|
||||||
"Useful options to set:\n" +
|
"Useful options to set:\n" +
|
||||||
"-D fs.default.name=\"file:///\" \\\n" +
|
"-D fs.defaultFS=\"file:///\" \\\n" +
|
||||||
"-D fs.file.impl=org.apache.hadoop.fs.RawLocalFileSystem \\\n" +
|
"-D fs.file.impl=org.apache.hadoop.fs.RawLocalFileSystem \\\n" +
|
||||||
"-D filebench.file.bytes=$((10*1024*1024*1024)) \\\n" +
|
"-D filebench.file.bytes=$((10*1024*1024*1024)) \\\n" +
|
||||||
"-D filebench.key.words=5 \\\n" +
|
"-D filebench.key.words=5 \\\n" +
|
||||||
|
|
|
@ -41,7 +41,7 @@ public class TestCombineFileInputFormat {
|
||||||
private static FileSystem localFs = null;
|
private static FileSystem localFs = null;
|
||||||
static {
|
static {
|
||||||
try {
|
try {
|
||||||
defaultConf.set("fs.default.name", "file:///");
|
defaultConf.set("fs.defaultFS", "file:///");
|
||||||
localFs = FileSystem.getLocal(defaultConf);
|
localFs = FileSystem.getLocal(defaultConf);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException("init failure", e);
|
throw new RuntimeException("init failure", e);
|
||||||
|
|
|
@ -69,7 +69,7 @@ public class TestConcatenatedCompressedInput {
|
||||||
|
|
||||||
static {
|
static {
|
||||||
try {
|
try {
|
||||||
defaultConf.set("fs.default.name", "file:///");
|
defaultConf.set("fs.defaultFS", "file:///");
|
||||||
localFs = FileSystem.getLocal(defaultConf);
|
localFs = FileSystem.getLocal(defaultConf);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException("init failure", e);
|
throw new RuntimeException("init failure", e);
|
||||||
|
|
|
@ -55,7 +55,7 @@ public class TestTextInputFormat {
|
||||||
private static FileSystem localFs = null;
|
private static FileSystem localFs = null;
|
||||||
static {
|
static {
|
||||||
try {
|
try {
|
||||||
defaultConf.set("fs.default.name", "file:///");
|
defaultConf.set("fs.defaultFS", "file:///");
|
||||||
localFs = FileSystem.getLocal(defaultConf);
|
localFs = FileSystem.getLocal(defaultConf);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException("init failure", e);
|
throw new RuntimeException("init failure", e);
|
||||||
|
|
|
@ -314,7 +314,7 @@ public class TestMapCollection {
|
||||||
job.setNumReduceTasks(1);
|
job.setNumReduceTasks(1);
|
||||||
job.getConfiguration().set(MRConfig.FRAMEWORK_NAME, MRConfig.LOCAL_FRAMEWORK_NAME);
|
job.getConfiguration().set(MRConfig.FRAMEWORK_NAME, MRConfig.LOCAL_FRAMEWORK_NAME);
|
||||||
job.getConfiguration().setInt(MRJobConfig.IO_SORT_FACTOR, 1000);
|
job.getConfiguration().setInt(MRJobConfig.IO_SORT_FACTOR, 1000);
|
||||||
job.getConfiguration().set("fs.default.name", "file:///");
|
job.getConfiguration().set("fs.defaultFS", "file:///");
|
||||||
job.getConfiguration().setInt("test.mapcollection.num.maps", 1);
|
job.getConfiguration().setInt("test.mapcollection.num.maps", 1);
|
||||||
job.setInputFormatClass(FakeIF.class);
|
job.setInputFormatClass(FakeIF.class);
|
||||||
job.setOutputFormatClass(NullOutputFormat.class);
|
job.setOutputFormatClass(NullOutputFormat.class);
|
||||||
|
|
|
@ -45,9 +45,9 @@ public class TestFileInputFormat {
|
||||||
@Test
|
@Test
|
||||||
public void testAddInputPath() throws IOException {
|
public void testAddInputPath() throws IOException {
|
||||||
final Configuration conf = new Configuration();
|
final Configuration conf = new Configuration();
|
||||||
conf.set("fs.default.name", "s3://abc:xyz@hostname/");
|
conf.set("fs.defaultFS", "s3://abc:xyz@hostname/");
|
||||||
final Job j = Job.getInstance(conf);
|
final Job j = Job.getInstance(conf);
|
||||||
j.getConfiguration().set("fs.default.name", "s3://abc:xyz@hostname/");
|
j.getConfiguration().set("fs.defaultFS", "s3://abc:xyz@hostname/");
|
||||||
|
|
||||||
//setup default fs
|
//setup default fs
|
||||||
final FileSystem defaultfs = FileSystem.get(conf);
|
final FileSystem defaultfs = FileSystem.get(conf);
|
||||||
|
|
|
@ -57,7 +57,7 @@ public class TestMRKeyValueTextInputFormat {
|
||||||
private static FileSystem localFs = null;
|
private static FileSystem localFs = null;
|
||||||
static {
|
static {
|
||||||
try {
|
try {
|
||||||
defaultConf.set("fs.default.name", "file:///");
|
defaultConf.set("fs.defaultFS", "file:///");
|
||||||
localFs = FileSystem.getLocal(defaultConf);
|
localFs = FileSystem.getLocal(defaultConf);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException("init failure", e);
|
throw new RuntimeException("init failure", e);
|
||||||
|
|
|
@ -457,7 +457,7 @@
|
||||||
|
|
||||||
<property>
|
<property>
|
||||||
<name>mapreduce.job.hdfs-servers</name>
|
<name>mapreduce.job.hdfs-servers</name>
|
||||||
<value>${fs.default.name}</value>
|
<value>${fs.defaultFS}</value>
|
||||||
</property>
|
</property>
|
||||||
|
|
||||||
<!-- WebAppProxy Configuration-->
|
<!-- WebAppProxy Configuration-->
|
||||||
|
|
Loading…
Reference in New Issue