From 95e8912bc0fdbc76a9dd29bd99c15c910e0042a8 Mon Sep 17 00:00:00 2001
From: Harsh J
Date: Fri, 9 Mar 2012 21:21:30 +0000
Subject: [PATCH] svn merge -c 1299045. Merge MAPREDUCE-3991 into branch-0.23.
(harsh)
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1299053 13f79535-47bb-0310-9956-ffa450edef68
---
hadoop-mapreduce-project/CHANGES.txt | 2 ++
.../src/docs/src/documentation/content/xdocs/streaming.xml | 2 +-
2 files changed, 3 insertions(+), 1 deletion(-)
diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt
index 039a157a75b..6a7d597c264 100644
--- a/hadoop-mapreduce-project/CHANGES.txt
+++ b/hadoop-mapreduce-project/CHANGES.txt
@@ -28,6 +28,8 @@ Release 0.23.3 - UNRELEASED
MAPREDUCE-3935. Annotate Counters.Counter and Counters.Group as @Public.
(tomwhite)
+ MAPREDUCE-3991. Streaming FAQ has some wrong instructions about input files splitting. (harsh)
+
OPTIMIZATIONS
BUG FIXES
diff --git a/hadoop-mapreduce-project/src/docs/src/documentation/content/xdocs/streaming.xml b/hadoop-mapreduce-project/src/docs/src/documentation/content/xdocs/streaming.xml
index a1013e8dc9e..2ae6858b708 100644
--- a/hadoop-mapreduce-project/src/docs/src/documentation/content/xdocs/streaming.xml
+++ b/hadoop-mapreduce-project/src/docs/src/documentation/content/xdocs/streaming.xml
@@ -750,7 +750,7 @@ You can use Hadoop Streaming to do this.
As an example, consider the problem of zipping (compressing) a set of files across the hadoop cluster. You can achieve this using either of these methods:
- Hadoop Streaming and custom mapper script:
- - Generate a file containing the full HDFS path of the input files. Each map task would get one file name as input.
+ - Generate files listing the full HDFS paths of the files to be processed. Each list file is the input for an individual map task which processes the files listed.
- Create a mapper script which, given a filename, will get the file to local disk, gzip the file and put it back in the desired output directory
- The existing Hadoop Framework: