Merge -r 1244263:1244264 from trunk to branch. FIXES: MAPREDUCE-3736
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1244265 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
c8debd083b
commit
9ea26a7b08
|
@ -216,6 +216,13 @@
|
|||
determine the host, port, etc. for a filesystem.</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>fs.default.name</name>
|
||||
<value>file:///</value>
|
||||
<description>Deprecated. Use (fs.defaultFS) property
|
||||
instead</description>
|
||||
</property>
|
||||
|
||||
<property>
|
||||
<name>fs.trash.interval</name>
|
||||
<value>0</value>
|
||||
|
|
|
@ -52,6 +52,9 @@ Release 0.23.1 - 2012-02-08
|
|||
MAPREDUCE-3840. JobEndNotifier doesn't use the proxyToUse during connecting
|
||||
(Ravi Prakash via bobby)
|
||||
|
||||
MAPREDUCE-3736. Variable substitution depth too large for fs.default.name
|
||||
causes jobs to fail (ahmed via tucu).
|
||||
|
||||
MAPREDUCE-2733. [Gridmix] Gridmix3 cpu emulation system tests.
|
||||
(Vinay Thota via amarrk)
|
||||
|
||||
|
|
|
@ -159,7 +159,7 @@ public class TestMRWithDistributedCache extends TestCase {
|
|||
public void testLocalJobRunner() throws Exception {
|
||||
Configuration c = new Configuration();
|
||||
c.set(JTConfig.JT_IPC_ADDRESS, "local");
|
||||
c.set("fs.default.name", "file:///");
|
||||
c.set("fs.defaultFS", "file:///");
|
||||
testWithConf(c);
|
||||
}
|
||||
|
||||
|
|
|
@ -59,7 +59,7 @@ public class TestNoDefaultsJobConf extends HadoopTestCase {
|
|||
|
||||
JobConf conf = new JobConf(false);
|
||||
|
||||
conf.set("fs.default.name", createJobConf().get("fs.default.name"));
|
||||
conf.set("fs.defaultFS", createJobConf().get("fs.defaultFS"));
|
||||
|
||||
conf.setJobName("mr");
|
||||
|
||||
|
|
|
@ -1024,7 +1024,7 @@ public class JHLogAnalyzer {
|
|||
if(testFile != null) {
|
||||
LOG.info("Start JHLA test ============ ");
|
||||
LocalFileSystem lfs = FileSystem.getLocal(conf);
|
||||
conf.set("fs.default.name", "file:///");
|
||||
conf.set("fs.defaultFS", "file:///");
|
||||
JHLAMapper map = new JHLAMapper(conf);
|
||||
map.parseLogFile(lfs, new Path(testFile), 0L,
|
||||
new LoggingCollector(), Reporter.NULL);
|
||||
|
|
|
@ -53,7 +53,7 @@ public class FileBench extends Configured implements Tool {
|
|||
"unless they are also explicitly included, as in \"-pln -zip\"\n" +
|
||||
"Note that CompressionType params only apply to SequenceFiles\n\n" +
|
||||
"Useful options to set:\n" +
|
||||
"-D fs.default.name=\"file:///\" \\\n" +
|
||||
"-D fs.defaultFS=\"file:///\" \\\n" +
|
||||
"-D fs.file.impl=org.apache.hadoop.fs.RawLocalFileSystem \\\n" +
|
||||
"-D filebench.file.bytes=$((10*1024*1024*1024)) \\\n" +
|
||||
"-D filebench.key.words=5 \\\n" +
|
||||
|
|
|
@ -41,7 +41,7 @@ public class TestCombineFileInputFormat {
|
|||
private static FileSystem localFs = null;
|
||||
static {
|
||||
try {
|
||||
defaultConf.set("fs.default.name", "file:///");
|
||||
defaultConf.set("fs.defaultFS", "file:///");
|
||||
localFs = FileSystem.getLocal(defaultConf);
|
||||
} catch (IOException e) {
|
||||
throw new RuntimeException("init failure", e);
|
||||
|
|
|
@ -69,7 +69,7 @@ public class TestConcatenatedCompressedInput {
|
|||
|
||||
static {
|
||||
try {
|
||||
defaultConf.set("fs.default.name", "file:///");
|
||||
defaultConf.set("fs.defaultFS", "file:///");
|
||||
localFs = FileSystem.getLocal(defaultConf);
|
||||
} catch (IOException e) {
|
||||
throw new RuntimeException("init failure", e);
|
||||
|
|
|
@ -55,7 +55,7 @@ public class TestTextInputFormat {
|
|||
private static FileSystem localFs = null;
|
||||
static {
|
||||
try {
|
||||
defaultConf.set("fs.default.name", "file:///");
|
||||
defaultConf.set("fs.defaultFS", "file:///");
|
||||
localFs = FileSystem.getLocal(defaultConf);
|
||||
} catch (IOException e) {
|
||||
throw new RuntimeException("init failure", e);
|
||||
|
|
|
@ -314,7 +314,7 @@ public class TestMapCollection {
|
|||
job.setNumReduceTasks(1);
|
||||
job.getConfiguration().set(MRConfig.FRAMEWORK_NAME, MRConfig.LOCAL_FRAMEWORK_NAME);
|
||||
job.getConfiguration().setInt(MRJobConfig.IO_SORT_FACTOR, 1000);
|
||||
job.getConfiguration().set("fs.default.name", "file:///");
|
||||
job.getConfiguration().set("fs.defaultFS", "file:///");
|
||||
job.getConfiguration().setInt("test.mapcollection.num.maps", 1);
|
||||
job.setInputFormatClass(FakeIF.class);
|
||||
job.setOutputFormatClass(NullOutputFormat.class);
|
||||
|
|
|
@ -45,9 +45,9 @@ public class TestFileInputFormat {
|
|||
@Test
|
||||
public void testAddInputPath() throws IOException {
|
||||
final Configuration conf = new Configuration();
|
||||
conf.set("fs.default.name", "s3://abc:xyz@hostname/");
|
||||
conf.set("fs.defaultFS", "s3://abc:xyz@hostname/");
|
||||
final Job j = Job.getInstance(conf);
|
||||
j.getConfiguration().set("fs.default.name", "s3://abc:xyz@hostname/");
|
||||
j.getConfiguration().set("fs.defaultFS", "s3://abc:xyz@hostname/");
|
||||
|
||||
//setup default fs
|
||||
final FileSystem defaultfs = FileSystem.get(conf);
|
||||
|
|
|
@ -57,7 +57,7 @@ public class TestMRKeyValueTextInputFormat {
|
|||
private static FileSystem localFs = null;
|
||||
static {
|
||||
try {
|
||||
defaultConf.set("fs.default.name", "file:///");
|
||||
defaultConf.set("fs.defaultFS", "file:///");
|
||||
localFs = FileSystem.getLocal(defaultConf);
|
||||
} catch (IOException e) {
|
||||
throw new RuntimeException("init failure", e);
|
||||
|
|
|
@ -457,7 +457,7 @@
|
|||
|
||||
<property>
|
||||
<name>mapreduce.job.hdfs-servers</name>
|
||||
<value>${fs.default.name}</value>
|
||||
<value>${fs.defaultFS}</value>
|
||||
</property>
|
||||
|
||||
<!-- WebAppProxy Configuration-->
|
||||
|
|
Loading…
Reference in New Issue