hadoop/hadoop-mapreduce-project/hadoop-yarn
Arun Murthy 88f815fc54 Merge -c 1213575 from trunk to branch-0.23 to fix MAPREDUCE-3537.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.23@1213576 13f79535-47bb-0310-9956-ffa450edef68
2011-12-13 06:35:10 +00:00
..
bin MAPREDUCE-3456. $HADOOP_PREFIX/bin/yarn should set defaults for $HADOOP_*_HOME (Eric Payne via mahadev) - Merging r1209825 from trunk. 2011-12-03 03:44:01 +00:00
conf MAPREDUCE-2746. Yarn servers can't communicate with each other with hadoop.security.authorization set to true (acmurthy via mahadev) - Merging r1188522 from trunk 2011-10-25 06:12:14 +00:00
dev-support MAPREDUCE-3297. Moved log related components into yarn-common so that HistoryServer and clients can use them without depending on the yarn-server-nodemanager module. Contributed by Siddharth Seth. 2011-11-03 08:07:16 +00:00
hadoop-yarn-api MAPREDUCE-3485. DISKS_FAILED -101 error code should be defined in same location as ABORTED_CONTAINER_EXIT_STATUS. (Ravi Gummadi via mahadev) - Merging r1210192 from trunk. 2011-12-04 20:12:41 +00:00
hadoop-yarn-applications MAPREDUCE-3121. NodeManager should handle disk-failures (Ravi Gummadi via mahadev) - Merging r1208131 from trunk. 2011-11-29 23:28:16 +00:00
hadoop-yarn-common MAPREDUCE-3485. DISKS_FAILED -101 error code should be defined in same location as ABORTED_CONTAINER_EXIT_STATUS. (Ravi Gummadi via mahadev) - Merging r1210192 from trunk. 2011-12-04 20:12:41 +00:00
hadoop-yarn-server Merge -c 1213575 from trunk to branch-0.23 to fix MAPREDUCE-3537. 2011-12-13 06:35:10 +00:00
hadoop-yarn-site svn merge -c 1212299 from trunk for HDFS-2594. 2011-12-09 09:34:32 +00:00
README Merge -r 1180115:1180116 from trunk to branch. FIXES: MAPREDUCE-3014 2011-10-28 21:34:14 +00:00
pom.xml Changing branch-0.23 to 0.23.1-SNAPSHOT. 2011-11-22 23:47:40 +00:00

README

YARN (YET ANOTHER RESOURCE NEGOTIATOR or YARN Application Resource Negotiator)
------------------------------------------------------------------------------

Requirements
-------------
Java: JDK 1.6
Maven: Maven 3

Setup
-----
Install protobuf 2.4.0a or higher (Download from http://code.google.com/p/protobuf/downloads/list)
 - install the protoc executable (configure, make, make install)
 - install the maven artifact (cd java; mvn install)
Installing protoc requires gcc 4.1.x or higher.
If the make step fails with (Valid until a fix is released for protobuf 2.4.0a)
    ./google/protobuf/descriptor.h:1152: error:
    `google::protobuf::internal::Mutex*google::protobuf::DescriptorPool::mutex_'
    is private
  Replace descriptor.cc with http://protobuf.googlecode.com/svn-history/r380/trunk/src/google/protobuf/descriptor.cc


Quick Maven Tips
----------------
clean workspace: mvn clean
compile and test: mvn install
skip tests: mvn install -DskipTests
skip test execution but compile: mvn install -Dmaven.test.skip.exec=true
clean and test: mvn clean install
run selected test after compile: mvn test -Dtest=TestClassName (combined: mvn clean install -Dtest=TestClassName)
create runnable binaries after install: mvn assembly:assembly -Pnative (combined: mvn clean install assembly:assembly -Pnative)

Eclipse Projects
----------------
http://maven.apache.org/guides/mini/guide-ide-eclipse.html

1. Generate .project and .classpath files in all maven modules
mvn eclipse:eclipse
CAUTION: If the project structure has changed from your previous workspace, clean up all .project and .classpath files recursively. Then run:
mvn eclipse:eclipse

2. Import the projects in eclipse.

3. Set the environment variable M2_REPO to point to your .m2/repository location.

NetBeans Projects
-----------------

NetBeans has builtin support of maven projects. Just "Open Project..."
and everything is setup automatically. Verified with NetBeans 6.9.1.


Custom Hadoop Dependencies
--------------------------

By default Hadoop dependencies are specified in the top-level pom.xml
properties section. One can override them via -Dhadoop-common.version=...
on the command line. ~/.m2/settings.xml can also be used to specify
these properties in different profiles, which is useful for IDEs.

Modules
-------
YARN consists of multiple modules. The modules are listed below as per the directory structure:

hadoop-yarn-api - Yarn's cross platform external interface

hadoop-yarn-common - Utilities which can be used by yarn clients and server

hadoop-yarn-server - Implementation of the hadoop-yarn-api
	hadoop-yarn-server-common - APIs shared between resourcemanager and nodemanager
	hadoop-yarn-server-nodemanager (TaskTracker replacement)
	hadoop-yarn-server-resourcemanager (JobTracker replacement)

Utilities for understanding the code
------------------------------------
Almost all of the yarn components as well as the mapreduce framework use
state-machines for all the data objects. To understand those central pieces of
the code, a visual representation of the state-machines helps much. You can first
convert the state-machines into graphviz(.gv) format by
running:
   mvn compile -Pvisualize
Then you can use the dot program for generating directed graphs and convert the above
.gv files to images. The graphviz package has the needed dot program and related
utilites.For e.g., to generate png files you can run:
   dot -Tpng NodeManager.gv > NodeManager.png