diff --git a/hadoop-assemblies/src/main/resources/assemblies/hadoop-dist.xml b/hadoop-assemblies/src/main/resources/assemblies/hadoop-dist.xml index 8db090062eb..4d93b11a043 100644 --- a/hadoop-assemblies/src/main/resources/assemblies/hadoop-dist.xml +++ b/hadoop-assemblies/src/main/resources/assemblies/hadoop-dist.xml @@ -91,13 +91,19 @@ ${project.artifactId}-${project.version}.jar ${project.artifactId}-${project.version}-tests.jar - ${project.artifactId}-${project.version}-sources.jar - ${project.artifactId}-${project.version}-test-sources.jar hadoop-tools-dist-*.jar + + ${project.build.directory} + /share/hadoop/${hadoop.component}/sources + + ${project.artifactId}-${project.version}-sources.jar + ${project.artifactId}-${project.version}-test-sources.jar + + ${basedir}/dev-support/jdiff /share/hadoop/${hadoop.component}/jdiff diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 43a8e483345..90d2f9e0d4e 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -149,6 +149,8 @@ Trunk (Unreleased) HADOOP-8924. Add maven plugin alternative to shell script to save package-info.java. (Chris Nauroth via suresh) + HADOOP-9277. Improve javadoc for FileContext. (Andrew Wang via suresh) + BUG FIXES HADOOP-8419. Fixed GzipCode NPE reset for IBM JDK. (Yu Li via eyang) @@ -339,7 +341,26 @@ Trunk (Unreleased) HADOOP-9190. packaging docs is broken. (Andy Isaacson via atm) -Release 2.0.3-alpha - Unreleased +Release 2.0.4-beta - UNRELEASED + + INCOMPATIBLE CHANGES + + NEW FEATURES + + HADOOP-9283. Add support for running the Hadoop client on AIX. (atm) + + IMPROVEMENTS + + HADOOP-9253. Capture ulimit info in the logs at service start time. + (Arpit Gupta via suresh) + + OPTIMIZATIONS + + BUG FIXES + + HADOOP-9294. GetGroupsTestBase fails on Windows. (Chris Nauroth via suresh) + +Release 2.0.3-alpha - 2013-02-06 INCOMPATIBLE CHANGES @@ -467,6 +488,9 @@ Release 2.0.3-alpha - Unreleased HADOOP-9231. Parametrize staging URL for the uniformity of distributionManagement. (Konstantin Boudnik via suresh) + HADOOP-9276. Allow BoundedByteArrayOutputStream to be resettable. + (Arun Murthy via hitesh) + OPTIMIZATIONS HADOOP-8866. SampleQuantiles#query is O(N^2) instead of O(N). (Andrew Wang @@ -592,6 +616,22 @@ Release 2.0.3-alpha - Unreleased HADOOP-9221. Convert remaining xdocs to APT. (Andy Isaacson via atm) HADOOP-8981. TestMetricsSystemImpl fails on Windows. (Xuan Gong via suresh) + + HADOOP-9124. SortedMapWritable violates contract of Map interface for + equals() and hashCode(). (Surenkumar Nihalani via tomwhite) + + HADOOP-9252. In StringUtils, humanReadableInt(..) has a race condition and + the synchronization of limitDecimalTo2(double) can be avoided. (szetszwo) + + HADOOP-9260. Hadoop version may be not correct when starting name node or + data node. (Chris Nauroth via jlowe) + + HADOOP-9278. Fix the file handle leak in HarMetaData.parseMetaData() in + HarFileSystem. (Chris Nauroth via szetszwo) + + HADOOP-9289. FsShell rm -f fails for non-matching globs. (Daryn Sharp via + suresh) + Release 2.0.2-alpha - 2012-09-07 @@ -1294,6 +1334,9 @@ Release 0.23.7 - UNRELEASED HADOOP-8849. FileUtil#fullyDelete should grant the target directories +rwx permissions (Ivan A. Veselovsky via bobby) + HADOOP-9067. provide test for LocalFileSystem.reportChecksumFailure + (Ivan A. Veselovsky via bobby) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh b/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh index 3e5f20ba20c..bf03d195e06 100755 --- a/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh +++ b/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh @@ -83,7 +83,8 @@ fi if [ "$command" == "datanode" ] && [ "$EUID" -eq 0 ] && [ -n "$HADOOP_SECURE_DN_USER" ]; then export HADOOP_PID_DIR=$HADOOP_SECURE_DN_PID_DIR export HADOOP_LOG_DIR=$HADOOP_SECURE_DN_LOG_DIR - export HADOOP_IDENT_STRING=$HADOOP_SECURE_DN_USER + export HADOOP_IDENT_STRING=$HADOOP_SECURE_DN_USER + starting_secure_dn="true" fi if [ "$HADOOP_IDENT_STRING" = "" ]; then @@ -154,7 +155,17 @@ case $startStop in ;; esac echo $! > $pid - sleep 1; head "$log" + sleep 1 + # capture the ulimit output + if [ "true" = "$starting_secure_dn" ]; then + echo "ulimit -a for secure datanode user $HADOOP_SECURE_DN_USER" >> $log + # capture the ulimit info for the appropriate user + su --shell=/bin/bash $HADOOP_SECURE_DN_USER -c 'ulimit -a' >> $log 2>&1 + else + echo "ulimit -a for user $USER" >> $log + ulimit -a >> $log 2>&1 + fi + head -30 "$log" sleep 3; if ! ps -p $! > /dev/null ; then exit 1 diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java index 978bb1ba0dd..d4ff03785c1 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java @@ -57,70 +57,60 @@ import org.apache.hadoop.security.token.Token; import org.apache.hadoop.util.ShutdownHookManager; /** - * The FileContext class provides an interface to the application writer for - * using the Hadoop file system. - * It provides a set of methods for the usual operation: create, open, - * list, etc + * The FileContext class provides an interface for users of the Hadoop + * file system. It exposes a number of file system operations, e.g. create, + * open, list. * - *

- * *** Path Names *** - *

+ *

Path Names

* - * The Hadoop file system supports a URI name space and URI names. - * It offers a forest of file systems that can be referenced using fully - * qualified URIs. - * Two common Hadoop file systems implementations are + * The Hadoop file system supports a URI namespace and URI names. This enables + * multiple types of file systems to be referenced using fully-qualified URIs. + * Two common Hadoop file system implementations are * * - * While URI names are very flexible, it requires knowing the name or address - * of the server. For convenience one often wants to access the default system - * in one's environment without knowing its name/address. This has an - * additional benefit that it allows one to change one's default fs - * (e.g. admin moves application from cluster1 to cluster2). + * The Hadoop file system also supports additional naming schemes besides URIs. + * Hadoop has the concept of a default file system, which implies a + * default URI scheme and authority. This enables slash-relative names + * relative to the default FS, which are more convenient for users and + * application writers. The default FS is typically set by the user's + * environment, though it can also be manually specified. *

* - * To facilitate this, Hadoop supports a notion of a default file system. - * The user can set his default file system, although this is - * typically set up for you in your environment via your default config. - * A default file system implies a default scheme and authority; slash-relative - * names (such as /for/bar) are resolved relative to that default FS. - * Similarly a user can also have working-directory-relative names (i.e. names - * not starting with a slash). While the working directory is generally in the - * same default FS, the wd can be in a different FS. + * Hadoop also supports working-directory-relative names, which are paths + * relative to the current working directory (similar to Unix). The working + * directory can be in a different file system than the default FS. *

- * Hence Hadoop path names can be one of: - *

+ * Thus, Hadoop path names can be specified as one of the following: + * * Relative paths with scheme (scheme:foo/bar) are illegal. * - *

- * ****The Role of the FileContext and configuration defaults**** - *

- * The FileContext provides file namespace context for resolving file names; - * it also contains the umask for permissions, In that sense it is like the - * per-process file-related state in Unix system. - * These two properties - *

- * in general, are obtained from the default configuration file - * in your environment, (@see {@link Configuration}). - * - * No other configuration parameters are obtained from the default config as - * far as the file context layer is concerned. All file system instances - * (i.e. deployments of file systems) have default properties; we call these - * server side (SS) defaults. Operation like create allow one to select many - * properties: either pass them in as explicit parameters or use - * the SS properties. - *

- * The file system related SS defaults are + *

Role of FileContext and Configuration Defaults

+ * + * The FileContext is the analogue of per-process file-related state in Unix. It + * contains two properties: + * + * + * In general, these properties are obtained from the default configuration file + * in the user's environment (see {@link Configuration}). + * + * Further file system properties are specified on the server-side. File system + * operations default to using these server-side defaults unless otherwise + * specified. + *

+ * The file system related server-side defaults are: *

* - *

- * *** Usage Model for the FileContext class *** - *

+ *

Example Usage

+ * * Example 1: use the default config read from the $HADOOP_CONFIG/core.xml. * Unspecified values come from core-defaults.xml in the release jar. * * Example 2: Get a FileContext with a specific URI as the default FS * + *
  • ... + * * Example 3: FileContext with local file system as the default * * Example 4: Use a specific config, ignoring $HADOOP_CONFIG * Generally you should not need use a config unless you are doing *