From 2207a84a7d294acd8b20f92195baa6c0be734e80 Mon Sep 17 00:00:00 2001 From: Brahma Reddy Battula Date: Tue, 22 Nov 2016 19:45:45 +0530 Subject: [PATCH] HADOOP-13646. Remove outdated overview.html. Contributed By Brahma Reddy Battula. (cherry picked from commit afcf8d38e750f935c06629e641a1321b79c4cace) --- .../hadoop-common/src/main/java/overview.html | 274 ------------------ .../hadoop-hdfs/src/main/java/overview.html | 274 ------------------ pom.xml | 25 -- 3 files changed, 573 deletions(-) delete mode 100644 hadoop-common-project/hadoop-common/src/main/java/overview.html delete mode 100644 hadoop-hdfs-project/hadoop-hdfs/src/main/java/overview.html diff --git a/hadoop-common-project/hadoop-common/src/main/java/overview.html b/hadoop-common-project/hadoop-common/src/main/java/overview.html deleted file mode 100644 index 5868617709b..00000000000 --- a/hadoop-common-project/hadoop-common/src/main/java/overview.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - Hadoop - - - -Hadoop is a distributed computing platform. - -

Hadoop primarily consists of the Hadoop Distributed FileSystem -(HDFS) and an -implementation of the -Map-Reduce programming paradigm.

- - -

Hadoop is a software framework that lets one easily write and run applications -that process vast amounts of data. Here's what makes Hadoop especially useful:

- - -

Requirements

- -

Platforms

- - - -

Requisite Software

- -
    -
  1. - Java 1.6.x, preferably from - Sun. - Set JAVA_HOME to the root of your Java installation. -
  2. -
  3. - ssh must be installed and sshd must be running to use Hadoop's - scripts to manage remote Hadoop daemons. -
  4. -
  5. - rsync may be installed to use Hadoop's scripts to manage remote - Hadoop installations. -
  6. -
- -

Installing Required Software

- -

If your platform does not have the required software listed above, you -will have to install it.

- -

For example on Ubuntu Linux:

-

-$ sudo apt-get install ssh
-$ sudo apt-get install rsync
-

- -

Getting Started

- -

First, you need to get a copy of the Hadoop code.

- -

Edit the file conf/hadoop-env.sh to define at least -JAVA_HOME.

- -

Try the following command:

-bin/hadoop -

This will display the documentation for the Hadoop command script.

- -

Standalone operation

- -

By default, Hadoop is configured to run things in a non-distributed -mode, as a single Java process. This is useful for debugging, and can -be demonstrated as follows:

- -mkdir input
-cp conf/*.xml input
-bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'
-cat output/* -
-

This will display counts for each match of the -regular expression.

- -

Note that input is specified as a directory containing input -files and that output is also specified as a directory where parts are -written.

- -

Distributed operation

- -To configure Hadoop for distributed operation you must specify the -following: - -
    - -
  1. The NameNode (Distributed Filesystem master) host. This is -specified with the configuration property fs.default.name. -
  2. - -
  3. The org.apache.hadoop.mapred.JobTracker (MapReduce master) -host and port. This is specified with the configuration property -mapred.job.tracker. -
  4. - -
  5. A slaves file that lists the names of all the hosts in -the cluster. The default slaves file is conf/slaves. - -
- -

Pseudo-distributed configuration

- -You can in fact run everything on a single host. To run things this -way, put the following in: -
-
-conf/core-site.xml: -<configuration> - - <property> - <name>fs.default.name</name> - <value>hdfs://localhost/</value> - </property> - -</configuration> - -conf/hdfs-site.xml: -<configuration> - - <property> - <name>dfs.replication</name> - <value>1</value> - </property> - -</configuration> - -conf/mapred-site.xml: -<configuration> - - <property> - <name>mapred.job.tracker</name> - <value>localhost:9001</value> - </property> - -</configuration> - -

(We also set the HDFS replication level to 1 in order to -reduce warnings when running on a single node.)

- -

Now check that the command
ssh localhost
does not -require a password. If it does, execute the following commands:

- -

ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
-cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys -

- -

Bootstrapping

- -

A new distributed filesystem must be formatted with the following -command, run on the master node:

- -

bin/hadoop namenode -format

- -

The Hadoop daemons are started with the following command:

- -

bin/start-all.sh

- -

Daemon log output is written to the logs/ directory.

- -

Input files are copied into the distributed filesystem as follows:

- -

bin/hadoop fs -put input input

- -

Distributed execution

- -

Things are run as before, but output must be copied locally to -examine it:

- - -bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'
-bin/hadoop fs -get output output -cat output/* -
- -

When you're done, stop the daemons with:

- -

bin/stop-all.sh

- -

Fully-distributed operation

- -

Fully distributed operation is just like the pseudo-distributed operation -described above, except, specify:

- -
    - -
  1. The hostname or IP address of your master server in the value -for fs.default.name, - as hdfs://master.example.com/ in conf/core-site.xml.
  2. - -
  3. The host and port of the your master server in the value -of mapred.job.tracker -as master.example.com:port in conf/mapred-site.xml.
  4. - -
  5. Directories for dfs.name.dir and -dfs.data.dir -in conf/hdfs-site.xml. -These are local directories used to hold distributed filesystem -data on the master node and slave nodes respectively. Note -that dfs.data.dir may contain a space- or comma-separated -list of directory names, so that data may be stored on multiple local -devices.
  6. - -
  7. mapred.local.dir - in conf/mapred-site.xml, the local directory where temporary - MapReduce data is stored. It also may be a list of directories.
  8. - -
  9. mapred.map.tasks -and mapred.reduce.tasks -in conf/mapred-site.xml. -As a rule of thumb, use 10x the -number of slave processors for mapred.map.tasks, and 2x the -number of slave processors for mapred.reduce.tasks.
  10. - -
- -

Finally, list all slave hostnames or IP addresses in your -conf/slaves file, one per line. Then format your filesystem -and start your cluster on your master node, as above. - - - - diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/overview.html b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/overview.html deleted file mode 100644 index 759c093aa59..00000000000 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/overview.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - Hadoop - - - -Hadoop is a distributed computing platform. - -

Hadoop primarily consists of the Hadoop Distributed FileSystem -(HDFS) and an -implementation of the -Map-Reduce programming paradigm.

- - -

Hadoop is a software framework that lets one easily write and run applications -that process vast amounts of data. Here's what makes Hadoop especially useful:

- - -

Requirements

- -

Platforms

- - - -

Requisite Software

- -
    -
  1. - Java 1.6.x, preferably from - Sun. - Set JAVA_HOME to the root of your Java installation. -
  2. -
  3. - ssh must be installed and sshd must be running to use Hadoop's - scripts to manage remote Hadoop daemons. -
  4. -
  5. - rsync may be installed to use Hadoop's scripts to manage remote - Hadoop installations. -
  6. -
- -

Installing Required Software

- -

If your platform does not have the required software listed above, you -will have to install it.

- -

For example on Ubuntu Linux:

-

-$ sudo apt-get install ssh
-$ sudo apt-get install rsync
-

- -

Getting Started

- -

First, you need to get a copy of the Hadoop code.

- -

Edit the file conf/hadoop-env.sh to define at least -JAVA_HOME.

- -

Try the following command:

-bin/hadoop -

This will display the documentation for the Hadoop command script.

- -

Standalone operation

- -

By default, Hadoop is configured to run things in a non-distributed -mode, as a single Java process. This is useful for debugging, and can -be demonstrated as follows:

- -mkdir input
-cp conf/*.xml input
-bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'
-cat output/* -
-

This will display counts for each match of the -regular expression.

- -

Note that input is specified as a directory containing input -files and that output is also specified as a directory where parts are -written.

- -

Distributed operation

- -To configure Hadoop for distributed operation you must specify the -following: - -
    - -
  1. The NameNode (Distributed Filesystem master) host. This is -specified with the configuration property fs.default.name. -
  2. - -
  3. The org.apache.hadoop.mapred.JobTracker (MapReduce master) -host and port. This is specified with the configuration property -mapred.job.tracker. -
  4. - -
  5. A slaves file that lists the names of all the hosts in -the cluster. The default slaves file is conf/slaves. - -
- -

Pseudo-distributed configuration

- -You can in fact run everything on a single host. To run things this -way, put the following in: -
-
-conf/core-site.xml: -<configuration> - - <property> - <name>fs.default.name</name> - <value>hdfs://localhost/</value> - </property> - -</configuration> - -conf/hdfs-site.xml: -<configuration> - - <property> - <name>dfs.replication</name> - <value>1</value> - </property> - -</configuration> - -conf/mapred-site.xml: -<configuration> - - <property> - <name>mapred.job.tracker</name> - <value>localhost:9001</value> - </property> - -</configuration> - -

(We also set the HDFS replication level to 1 in order to -reduce warnings when running on a single node.)

- -

Now check that the command
ssh localhost
does not -require a password. If it does, execute the following commands:

- -

ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
-cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys -

- -

Bootstrapping

- -

A new distributed filesystem must be formatted with the following -command, run on the master node:

- -

bin/hadoop namenode -format

- -

The Hadoop daemons are started with the following command:

- -

bin/start-all.sh

- -

Daemon log output is written to the logs/ directory.

- -

Input files are copied into the distributed filesystem as follows:

- -

bin/hadoop fs -put input input

- -

Distributed execution

- -

Things are run as before, but output must be copied locally to -examine it:

- - -bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'
-bin/hadoop fs -get output output -cat output/* -
- -

When you're done, stop the daemons with:

- -

bin/stop-all.sh

- -

Fully-distributed operation

- -

Fully distributed operation is just like the pseudo-distributed operation -described above, except, specify:

- -
    - -
  1. The hostname or IP address of your master server in the value -for fs.default.name, - as hdfs://master.example.com/ in conf/core-site.xml.
  2. - -
  3. The host and port of the your master server in the value -of mapred.job.tracker -as master.example.com:port in conf/mapred-site.xml.
  4. - -
  5. Directories for dfs.name.dir and -dfs.data.dir -in conf/hdfs-site.xml. -These are local directories used to hold distributed filesystem -data on the master node and slave nodes respectively. Note -that dfs.data.dir may contain a space- or comma-separated -list of directory names, so that data may be stored on multiple local -devices.
  6. - -
  7. mapred.local.dir - in conf/mapred-site.xml, the local directory where temporary - MapReduce data is stored. It also may be a list of directories.
  8. - -
  9. mapred.map.tasks -and mapred.reduce.tasks -in conf/mapred-site.xml. -As a rule of thumb, use 10x the -number of slave processors for mapred.map.tasks, and 2x the -number of slave processors for mapred.reduce.tasks.
  10. - -
- -

Finally, list all slave hostnames or IP addresses in your -conf/slaves file, one per line. Then format your filesystem -and start your cluster on your master node, as above. - - - - diff --git a/pom.xml b/pom.xml index 7bcfe80241d..0f02ab03490 100644 --- a/pom.xml +++ b/pom.xml @@ -443,31 +443,6 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs - - dist - - - - - org.apache.maven.plugins - maven-javadoc-plugin - false - - - - default-cli - - aggregate - - - hadoop-common-project/hadoop-common/src/main/java/overview.html - - - - - - - sign