From 517eccb0c7956454174bad37408ba2b5c1ed85c7 Mon Sep 17 00:00:00 2001 From: Arpit Agarwal Date: Thu, 22 May 2014 20:41:44 +0000 Subject: [PATCH] HADOOP-10618: Merging r1596964 from trunk to branch-2. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1596965 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 3 + .../src/site/apt/SingleNodeSetup.apt.vm | 209 +----------------- 2 files changed, 6 insertions(+), 206 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 240664bf280..4d9f67315d6 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -54,6 +54,9 @@ Release 2.5.0 - UNRELEASED HADOOP-10614. CBZip2InputStream is not threadsafe (Xiangrui Meng via Sandy Ryza) + HADOOP-10618. Remove SingleNodeSetup.apt.vm. (Akira Ajisaka via + Arpit Agarwal) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-common-project/hadoop-common/src/site/apt/SingleNodeSetup.apt.vm b/hadoop-common-project/hadoop-common/src/site/apt/SingleNodeSetup.apt.vm index ebfb1466bdb..eb0c8015116 100644 --- a/hadoop-common-project/hadoop-common/src/site/apt/SingleNodeSetup.apt.vm +++ b/hadoop-common-project/hadoop-common/src/site/apt/SingleNodeSetup.apt.vm @@ -18,210 +18,7 @@ Single Node Setup -%{toc|section=1|fromDepth=0} + This page will be removed in the next major release. -* Purpose - - This document describes how to set up and configure a single-node - Hadoop installation so that you can quickly perform simple operations - using Hadoop MapReduce and the Hadoop Distributed File System (HDFS). - -* Prerequisites - -** Supported Platforms - - * GNU/Linux is supported as a development and production platform. - Hadoop has been demonstrated on GNU/Linux clusters with 2000 nodes. - - * Windows is also a supported platform. - -** Required Software - - Required software for Linux and Windows include: - - [[1]] Java^TM 1.6.x, preferably from Sun, must be installed. - - [[2]] ssh must be installed and sshd must be running to use the Hadoop - scripts that manage remote Hadoop daemons. - -** Installing Software - - If your cluster doesn't have the requisite software you will need to - install it. - - For example on Ubuntu Linux: - ----- - $ sudo apt-get install ssh - $ sudo apt-get install rsync ----- - -* Download - - To get a Hadoop distribution, download a recent stable release from one - of the Apache Download Mirrors. - -* Prepare to Start the Hadoop Cluster - - Unpack the downloaded Hadoop distribution. In the distribution, edit - the file <<>> to define at least <<>> to be the root - of your Java installation. - - Try the following command: - ----- - $ bin/hadoop ----- - - This will display the usage documentation for the hadoop script. - - Now you are ready to start your Hadoop cluster in one of the three - supported modes: - - * Local (Standalone) Mode - - * Pseudo-Distributed Mode - - * Fully-Distributed Mode - -* Standalone Operation - - By default, Hadoop is configured to run in a non-distributed mode, as a - single Java process. This is useful for debugging. - - The following example copies the unpacked conf directory to use as - input and then finds and displays every match of the given regular - expression. Output is written to the given output directory. - ----- - $ mkdir input - $ cp conf/*.xml input - $ bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+' - $ cat output/* ---- - -* Pseudo-Distributed Operation - - Hadoop can also be run on a single-node in a pseudo-distributed mode - where each Hadoop daemon runs in a separate Java process. - -** Configuration - - Use the following: - - conf/core-site.xml: - ----- - - - fs.defaultFS - hdfs://localhost:9000 - - ----- - - conf/hdfs-site.xml: - ----- - - - dfs.replication - 1 - - ----- - - conf/mapred-site.xml: - ----- - - - mapred.job.tracker - localhost:9001 - - ----- - -** Setup passphraseless ssh - - Now check that you can ssh to the localhost without a passphrase: - ----- - $ ssh localhost ----- - - If you cannot ssh to localhost without a passphrase, execute the - following commands: - ----- - $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa - $ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys ----- - -** Execution - - Format a new distributed-filesystem: - ----- - $ bin/hadoop namenode -format ----- - - Start the hadoop daemons: - ----- - $ bin/start-all.sh ----- - - The hadoop daemon log output is written to the <<<${HADOOP_LOG_DIR}>>> - directory (defaults to <<<${HADOOP_PREFIX}/logs>>>). - - Browse the web interface for the NameNode and the JobTracker; by - default they are available at: - - * NameNode - <<>> - - * JobTracker - <<>> - - Copy the input files into the distributed filesystem: - ----- - $ bin/hadoop fs -put conf input ----- - - Run some of the examples provided: - ----- - $ bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+' ----- - - Examine the output files: - - Copy the output files from the distributed filesystem to the local - filesytem and examine them: - ----- - $ bin/hadoop fs -get output output - $ cat output/* ----- - - or - - View the output files on the distributed filesystem: - ----- - $ bin/hadoop fs -cat output/* ----- - - When you're done, stop the daemons with: - ----- - $ bin/stop-all.sh ----- - -* Fully-Distributed Operation - - For information on setting up fully-distributed, non-trivial clusters - see {{{./ClusterSetup.html}Cluster Setup}}. - - Java and JNI are trademarks or registered trademarks of Sun - Microsystems, Inc. in the United States and other countries. + See {{{./SingleCluster.html}Single Cluster Setup}} to set up and configure a + single-node Hadoop installation.