From 0f3461e4894ab64237e30ea6cb56d18831e37d2a Mon Sep 17 00:00:00 2001 From: Arpit Agarwal Date: Thu, 30 Jan 2014 03:26:46 +0000 Subject: [PATCH 01/23] HDFS-5492. Port HDFS-2069 (Incorrect default trash interval in the docs) to trunk. (Contributed by Akira Ajisaka) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562683 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 + .../src/site/apt/HdfsDesign.apt.vm | 82 +++++++++---------- 2 files changed, 44 insertions(+), 41 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index feaee23b9e6..d665a41a288 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -301,6 +301,9 @@ Release 2.4.0 - UNRELEASED BUG FIXES + HDFS-5492. Port HDFS-2069 (Incorrect default trash interval in the + docs) to trunk. (Akira Ajisaka via Arpit Agarwal) + Release 2.3.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HdfsDesign.apt.vm b/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HdfsDesign.apt.vm index 21cb6450ddf..9cd95fac0f0 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HdfsDesign.apt.vm +++ b/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HdfsDesign.apt.vm @@ -17,11 +17,11 @@ --- ${maven.build.timestamp} -%{toc|section=1|fromDepth=0} - HDFS Architecture -Introduction +%{toc|section=1|fromDepth=0} + +* Introduction The Hadoop Distributed File System (HDFS) is a distributed file system designed to run on commodity hardware. It has many similarities with @@ -35,9 +35,9 @@ Introduction is part of the Apache Hadoop Core project. The project URL is {{http://hadoop.apache.org/}}. -Assumptions and Goals +* Assumptions and Goals -Hardware Failure +** Hardware Failure Hardware failure is the norm rather than the exception. An HDFS instance may consist of hundreds or thousands of server machines, each @@ -47,7 +47,7 @@ Hardware Failure non-functional. Therefore, detection of faults and quick, automatic recovery from them is a core architectural goal of HDFS. -Streaming Data Access +** Streaming Data Access Applications that run on HDFS need streaming access to their data sets. They are not general purpose applications that typically run on general @@ -58,7 +58,7 @@ Streaming Data Access targeted for HDFS. POSIX semantics in a few key areas has been traded to increase data throughput rates. -Large Data Sets +** Large Data Sets Applications that run on HDFS have large data sets. A typical file in HDFS is gigabytes to terabytes in size. Thus, HDFS is tuned to support @@ -66,7 +66,7 @@ Large Data Sets to hundreds of nodes in a single cluster. It should support tens of millions of files in a single instance. -Simple Coherency Model +** Simple Coherency Model HDFS applications need a write-once-read-many access model for files. A file once created, written, and closed need not be changed. This @@ -75,7 +75,7 @@ Simple Coherency Model perfectly with this model. There is a plan to support appending-writes to files in the future. -“Moving Computation is Cheaper than Moving Data” +** “Moving Computation is Cheaper than Moving Data” A computation requested by an application is much more efficient if it is executed near the data it operates on. This is especially true when @@ -86,13 +86,13 @@ Simple Coherency Model running. HDFS provides interfaces for applications to move themselves closer to where the data is located. -Portability Across Heterogeneous Hardware and Software Platforms +** Portability Across Heterogeneous Hardware and Software Platforms HDFS has been designed to be easily portable from one platform to another. This facilitates widespread adoption of HDFS as a platform of choice for a large set of applications. -NameNode and DataNodes +* NameNode and DataNodes HDFS has a master/slave architecture. An HDFS cluster consists of a single NameNode, a master server that manages the file system namespace @@ -127,7 +127,7 @@ NameNode and DataNodes repository for all HDFS metadata. The system is designed in such a way that user data never flows through the NameNode. -The File System Namespace +* The File System Namespace HDFS supports a traditional hierarchical file organization. A user or an application can create directories and store files inside these @@ -145,7 +145,7 @@ The File System Namespace replication factor of that file. This information is stored by the NameNode. -Data Replication +* Data Replication HDFS is designed to reliably store very large files across machines in a large cluster. It stores each file as a sequence of blocks; all @@ -164,7 +164,7 @@ Data Replication [images/hdfsdatanodes.png] HDFS DataNodes -Replica Placement: The First Baby Steps +** Replica Placement: The First Baby Steps The placement of replicas is critical to HDFS reliability and performance. Optimizing replica placement distinguishes HDFS from most @@ -210,7 +210,7 @@ Replica Placement: The First Baby Steps The current, default replica placement policy described here is a work in progress. -Replica Selection +** Replica Selection To minimize global bandwidth consumption and read latency, HDFS tries to satisfy a read request from a replica that is closest to the reader. @@ -219,7 +219,7 @@ Replica Selection cluster spans multiple data centers, then a replica that is resident in the local data center is preferred over any remote replica. -Safemode +** Safemode On startup, the NameNode enters a special state called Safemode. Replication of data blocks does not occur when the NameNode is in the @@ -234,7 +234,7 @@ Safemode blocks (if any) that still have fewer than the specified number of replicas. The NameNode then replicates these blocks to other DataNodes. -The Persistence of File System Metadata +* The Persistence of File System Metadata The HDFS namespace is stored by the NameNode. The NameNode uses a transaction log called the EditLog to persistently record every change @@ -273,7 +273,7 @@ The Persistence of File System Metadata each of these local files and sends this report to the NameNode: this is the Blockreport. -The Communication Protocols +* The Communication Protocols All HDFS communication protocols are layered on top of the TCP/IP protocol. A client establishes a connection to a configurable TCP port @@ -284,13 +284,13 @@ The Communication Protocols RPCs. Instead, it only responds to RPC requests issued by DataNodes or clients. -Robustness +* Robustness The primary objective of HDFS is to store data reliably even in the presence of failures. The three common types of failures are NameNode failures, DataNode failures and network partitions. -Data Disk Failure, Heartbeats and Re-Replication +** Data Disk Failure, Heartbeats and Re-Replication Each DataNode sends a Heartbeat message to the NameNode periodically. A network partition can cause a subset of DataNodes to lose connectivity @@ -306,7 +306,7 @@ Data Disk Failure, Heartbeats and Re-Replication corrupted, a hard disk on a DataNode may fail, or the replication factor of a file may be increased. -Cluster Rebalancing +** Cluster Rebalancing The HDFS architecture is compatible with data rebalancing schemes. A scheme might automatically move data from one DataNode to another if @@ -316,7 +316,7 @@ Cluster Rebalancing cluster. These types of data rebalancing schemes are not yet implemented. -Data Integrity +** Data Integrity It is possible that a block of data fetched from a DataNode arrives corrupted. This corruption can occur because of faults in a storage @@ -330,7 +330,7 @@ Data Integrity to retrieve that block from another DataNode that has a replica of that block. -Metadata Disk Failure +** Metadata Disk Failure The FsImage and the EditLog are central data structures of HDFS. A corruption of these files can cause the HDFS instance to be @@ -350,16 +350,16 @@ Metadata Disk Failure Currently, automatic restart and failover of the NameNode software to another machine is not supported. -Snapshots +** Snapshots Snapshots support storing a copy of data at a particular instant of time. One usage of the snapshot feature may be to roll back a corrupted HDFS instance to a previously known good point in time. HDFS does not currently support snapshots but will in a future release. -Data Organization +* Data Organization -Data Blocks +** Data Blocks HDFS is designed to support very large files. Applications that are compatible with HDFS are those that deal with large data sets. These @@ -370,7 +370,7 @@ Data Blocks chunks, and if possible, each chunk will reside on a different DataNode. -Staging +** Staging A client request to create a file does not reach the NameNode immediately. In fact, initially the HDFS client caches the file data @@ -397,7 +397,7 @@ Staging side caching to improve performance. A POSIX requirement has been relaxed to achieve higher performance of data uploads. -Replication Pipelining +** Replication Pipelining When a client is writing data to an HDFS file, its data is first written to a local file as explained in the previous section. Suppose @@ -406,7 +406,7 @@ Replication Pipelining DataNodes from the NameNode. This list contains the DataNodes that will host a replica of that block. The client then flushes the data block to the first DataNode. The first DataNode starts receiving the data in - small portions (4 KB), writes each portion to its local repository and + small portions, writes each portion to its local repository and transfers that portion to the second DataNode in the list. The second DataNode, in turn starts receiving each portion of the data block, writes that portion to its repository and then flushes that portion to @@ -416,7 +416,7 @@ Replication Pipelining the next one in the pipeline. Thus, the data is pipelined from one DataNode to the next. -Accessibility +* Accessibility HDFS can be accessed from applications in many different ways. Natively, HDFS provides a @@ -426,7 +426,7 @@ Accessibility of an HDFS instance. Work is in progress to expose HDFS through the WebDAV protocol. -FS Shell +** FS Shell HDFS allows user data to be organized in the form of files and directories. It provides a commandline interface called FS shell that @@ -447,7 +447,7 @@ FS Shell FS shell is targeted for applications that need a scripting language to interact with the stored data. -DFSAdmin +** DFSAdmin The DFSAdmin command set is used for administering an HDFS cluster. These are commands that are used only by an HDFS administrator. Here @@ -463,16 +463,16 @@ DFSAdmin |Recommission or decommission DataNode(s) | <<>> *---------+---------+ -Browser Interface +** Browser Interface A typical HDFS install configures a web server to expose the HDFS namespace through a configurable TCP port. This allows a user to navigate the HDFS namespace and view the contents of its files using a web browser. -Space Reclamation +* Space Reclamation -File Deletes and Undeletes +** File Deletes and Undeletes When a file is deleted by a user or an application, it is not immediately removed from HDFS. Instead, HDFS first renames it to a file @@ -490,12 +490,12 @@ File Deletes and Undeletes file. The <<>> directory contains only the latest copy of the file that was deleted. The <<>> directory is just like any other directory with one special feature: HDFS applies specified policies to - automatically delete files from this directory. The current default - policy is to delete files from <<>> that are more than 6 hours old. - In the future, this policy will be configurable through a well defined - interface. + automatically delete files from this directory. Current default trash + interval is set to 0 (Deletes file without storing in trash). This value is + configurable parameter stored as <<>> stored in + core-site.xml. -Decrease Replication Factor +** Decrease Replication Factor When the replication factor of a file is reduced, the NameNode selects excess replicas that can be deleted. The next Heartbeat transfers this @@ -505,7 +505,7 @@ Decrease Replication Factor of the setReplication API call and the appearance of free space in the cluster. -References +* References Hadoop {{{http://hadoop.apache.org/docs/current/api/}JavaDoc API}}. From 285854fc2ec9e9ef3cbb3e27f5e944bd7926e7d6 Mon Sep 17 00:00:00 2001 From: Aaron Myers Date: Thu, 30 Jan 2014 15:49:47 +0000 Subject: [PATCH 02/23] HADOOP-10310. SaslRpcServer should be initialized even when no secret manager present. Contributed by Aaron T. Myers. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562863 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-common-project/hadoop-common/CHANGES.txt | 3 +++ .../src/main/java/org/apache/hadoop/ipc/Server.java | 2 +- 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 56a8c9cb28b..87d859710b8 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -665,6 +665,9 @@ Release 2.3.0 - UNRELEASED HADOOP-10288. Explicit reference to Log4JLogger breaks non-log4j users (todd) + HADOOP-10310. SaslRpcServer should be initialized even when no secret + manager present. (atm) + Release 2.2.0 - 2013-10-13 INCOMPATIBLE CHANGES diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java index 8c578fd107e..c7cada825d9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java @@ -2206,7 +2206,7 @@ protected Server(String bindAddress, int port, // Create the responder here responder = new Responder(); - if (secretManager != null) { + if (secretManager != null || UserGroupInformation.isSecurityEnabled()) { SaslRpcServer.init(conf); } From 21f797edeb2203347d55e68575419e23ad4b06a4 Mon Sep 17 00:00:00 2001 From: Vinod Kumar Vavilapalli Date: Thu, 30 Jan 2014 18:57:40 +0000 Subject: [PATCH 03/23] Updating YARN's CHANGES.txt to reflect the recent rebase of unreleased 2.3 to more latest branch-2. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562923 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 204 ++++++++++++++++---------------- 1 file changed, 102 insertions(+), 102 deletions(-) diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 8452f6e3bfc..bc5a40ae00b 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -116,6 +116,28 @@ Release 2.4.0 - UNRELEASED INCOMPATIBLE CHANGES + NEW FEATURES + + IMPROVEMENTS + + YARN-1630. Introduce timeout for async polling operations in YarnClientImpl + (Aditya Acharya via Sandy Ryza) + + OPTIMIZATIONS + + BUG FIXES + + YARN-1618. Fix invalid RMApp transition from NEW to FINAL_SAVING (kasha) + + YARN-1600. RM does not startup when security is enabled without spnego + configured (Haohui Mai via jlowe) + + YARN-1642. RMDTRenewer#getRMClient should use ClientRMProxy (kasha) + +Release 2.3.0 - UNRELEASED + + INCOMPATIBLE CHANGES + NEW FEATURES YARN-649. Added a new NM web-service to serve container logs in plain text @@ -169,6 +191,30 @@ Release 2.4.0 - UNRELEASED IMPROVEMENTS + YARN-305. Fair scheduler logs too many "Node offered to app" messages. + (Lohit Vijayarenu via Sandy Ryza) + + YARN-1258. Allow configuring the Fair Scheduler root queue (Sandy Ryza) + + YARN-1288. Make Fair Scheduler ACLs more user friendly (Sandy Ryza) + + YARN-1315. TestQueueACLs should also test FairScheduler (Sandy Ryza) + + YARN-1335. Move duplicate code from FSSchedulerApp and FiCaSchedulerApp + into SchedulerApplication (Sandy Ryza) + + YARN-1333. Support blacklisting in the Fair Scheduler (Tsuyoshi Ozawa via + Sandy Ryza) + + YARN-1109. Demote NodeManager "Sending out status for container" logs to + debug (haosdent via Sandy Ryza) + + YARN-1321. Changed NMTokenCache to support both singleton and an instance + usage. (Alejandro Abdelnur via vinodkv) + + YARN-1388. Fair Scheduler page always displays blank fair share (Liyin Liang + via Sandy Ryza) + YARN-7. Support CPU resource for DistributedShell. (Junping Du via llu) YARN-905. Add state filters to nodes CLI (Wei Yan via Sandy Ryza) @@ -334,13 +380,66 @@ Release 2.4.0 - UNRELEASED YARN-1573. ZK store should use a private password for root-node-acls. (kasha). - YARN-1630. Introduce timeout for async polling operations in YarnClientImpl - (Aditya Acharya via Sandy Ryza) - OPTIMIZATIONS BUG FIXES + YARN-1284. LCE: Race condition leaves dangling cgroups entries for killed + containers. (Alejandro Abdelnur via Sandy Ryza) + + YARN-1283. Fixed RM to give a fully-qualified proxy URL for an application + so that clients don't need to do scheme-mangling. (Omkar Vinit Joshi via + vinodkv) + + YARN-879. Fixed tests w.r.t o.a.h.y.server.resourcemanager.Application. + (Junping Du via devaraj) + + YARN-1265. Fair Scheduler chokes on unhealthy node reconnect (Sandy Ryza) + + YARN-1044. used/min/max resources do not display info in the scheduler page + (Sangjin Lee via Sandy Ryza) + + YARN-1259. In Fair Scheduler web UI, queue num pending and num active apps + switched. (Robert Kanter via Sandy Ryza) + + YARN-1295. In UnixLocalWrapperScriptBuilder, using bash -c can cause Text + file busy errors (Sandy Ryza) + + YARN-1185. Fixed FileSystemRMStateStore to not leave partial files that + prevent subsequent ResourceManager recovery. (Omkar Vinit Joshi via vinodkv) + + YARN-1331. yarn.cmd exits with NoClassDefFoundError trying to run rmadmin or + logs. (cnauroth) + + YARN-1330. Fair Scheduler: defaultQueueSchedulingPolicy does not take effect + (Sandy Ryza) + + YARN-1022. Unnecessary INFO logs in AMRMClientAsync (haosdent via bikas) + + YARN-1349. yarn.cmd does not support passthrough to any arbitrary class. + (cnauroth) + + YARN-1357. TestContainerLaunch.testContainerEnvVariables fails on Windows. + (Chuan Liu via cnauroth) + + YARN-1358. TestYarnCLI fails on Windows due to line endings. (Chuan Liu via + cnauroth) + + YARN-1343. NodeManagers additions/restarts are not reported as node updates + in AllocateResponse responses to AMs. (tucu) + + YARN-1381. Same relaxLocality appears twice in exception message of + AMRMClientImpl#checkLocalityRelaxationConflict() (Ted Yu via Sandy Ryza) + + YARN-1407. RM Web UI and REST APIs should uniformly use + YarnApplicationState (Sandy Ryza) + + YARN-1438. Ensure container diagnostics includes exception from container + launch. (stevel via acmurthy) + + YARN-1138. yarn.application.classpath is set to point to $HADOOP_CONF_DIR + etc., which does not work on Windows. (Chuan Liu via cnauroth) + YARN-461. Fair scheduler should not accept apps with empty string queue name. (ywskycn via tucu) @@ -466,107 +565,8 @@ Release 2.4.0 - UNRELEASED YARN-1575. Public localizer crashes with "Localized unkown resource" (jlowe) - YARN-1642. RMDTRenewer#getRMClient should use ClientRMProxy (kasha) - YARN-1629. IndexOutOfBoundsException in MaxRunningAppsEnforcer (Sandy Ryza) - YARN-1618. Fix invalid RMApp transition from NEW to FINAL_SAVING (kasha) - -Release 2.3.0 - UNRELEASED - - INCOMPATIBLE CHANGES - - NEW FEATURES - - IMPROVEMENTS - - YARN-305. Fair scheduler logs too many "Node offered to app" messages. - (Lohit Vijayarenu via Sandy Ryza) - - YARN-1258. Allow configuring the Fair Scheduler root queue (Sandy Ryza) - - YARN-1288. Make Fair Scheduler ACLs more user friendly (Sandy Ryza) - - YARN-1315. TestQueueACLs should also test FairScheduler (Sandy Ryza) - - YARN-1335. Move duplicate code from FSSchedulerApp and FiCaSchedulerApp - into SchedulerApplication (Sandy Ryza) - - YARN-1333. Support blacklisting in the Fair Scheduler (Tsuyoshi Ozawa via - Sandy Ryza) - - YARN-1109. Demote NodeManager "Sending out status for container" logs to - debug (haosdent via Sandy Ryza) - - YARN-1321. Changed NMTokenCache to support both singleton and an instance - usage. (Alejandro Abdelnur via vinodkv) - - YARN-1388. Fair Scheduler page always displays blank fair share (Liyin Liang - via Sandy Ryza) - - OPTIMIZATIONS - - BUG FIXES - - YARN-1284. LCE: Race condition leaves dangling cgroups entries for killed - containers. (Alejandro Abdelnur via Sandy Ryza) - - YARN-1283. Fixed RM to give a fully-qualified proxy URL for an application - so that clients don't need to do scheme-mangling. (Omkar Vinit Joshi via - vinodkv) - - YARN-879. Fixed tests w.r.t o.a.h.y.server.resourcemanager.Application. - (Junping Du via devaraj) - - YARN-1265. Fair Scheduler chokes on unhealthy node reconnect (Sandy Ryza) - - YARN-1044. used/min/max resources do not display info in the scheduler page - (Sangjin Lee via Sandy Ryza) - - YARN-1259. In Fair Scheduler web UI, queue num pending and num active apps - switched. (Robert Kanter via Sandy Ryza) - - YARN-1295. In UnixLocalWrapperScriptBuilder, using bash -c can cause Text - file busy errors (Sandy Ryza) - - YARN-1185. Fixed FileSystemRMStateStore to not leave partial files that - prevent subsequent ResourceManager recovery. (Omkar Vinit Joshi via vinodkv) - - YARN-1331. yarn.cmd exits with NoClassDefFoundError trying to run rmadmin or - logs. (cnauroth) - - YARN-1330. Fair Scheduler: defaultQueueSchedulingPolicy does not take effect - (Sandy Ryza) - - YARN-1022. Unnecessary INFO logs in AMRMClientAsync (haosdent via bikas) - - YARN-1349. yarn.cmd does not support passthrough to any arbitrary class. - (cnauroth) - - YARN-1357. TestContainerLaunch.testContainerEnvVariables fails on Windows. - (Chuan Liu via cnauroth) - - YARN-1358. TestYarnCLI fails on Windows due to line endings. (Chuan Liu via - cnauroth) - - YARN-1343. NodeManagers additions/restarts are not reported as node updates - in AllocateResponse responses to AMs. (tucu) - - YARN-1381. Same relaxLocality appears twice in exception message of - AMRMClientImpl#checkLocalityRelaxationConflict() (Ted Yu via Sandy Ryza) - - YARN-1407. RM Web UI and REST APIs should uniformly use - YarnApplicationState (Sandy Ryza) - - YARN-1438. Ensure container diagnostics includes exception from container - launch. (stevel via acmurthy) - - YARN-1138. yarn.application.classpath is set to point to $HADOOP_CONF_DIR - etc., which does not work on Windows. (Chuan Liu via cnauroth) - - YARN-1600. RM does not startup when security is enabled without spnego - configured (Haohui Mai via jlowe) - Release 2.2.0 - 2013-10-13 INCOMPATIBLE CHANGES From 95074f0867822f615a8c8a55783b413d3a2fdfd9 Mon Sep 17 00:00:00 2001 From: Vinod Kumar Vavilapalli Date: Thu, 30 Jan 2014 19:01:47 +0000 Subject: [PATCH 04/23] Updating trunk's MAPREDUCE CHANGES.txt to reflect the recent rebase of unreleased 2.3 to more latest branch-2. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562925 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-mapreduce-project/CHANGES.txt | 124 +++++++++++++-------------- 1 file changed, 62 insertions(+), 62 deletions(-) diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 53931dddb41..542a9a87ab9 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -152,6 +152,21 @@ Release 2.4.0 - UNRELEASED INCOMPATIBLE CHANGES + NEW FEATURES + + IMPROVEMENTS + + MAPREDUCE-5464. Add analogs of the SLOTS_MILLIS counters that jive with the + YARN resource model (Sandy Ryza) + + OPTIMIZATIONS + + BUG FIXES + +Release 2.3.0 - UNRELEASED + + INCOMPATIBLE CHANGES + NEW FEATURES MAPREDUCE-5265. History server admin service to refresh user and superuser @@ -169,6 +184,19 @@ Release 2.4.0 - UNRELEASED MAPREDUCE-5332. Support token-preserving restart of history server (jlowe) IMPROVEMENTS + + MAPREDUCE-5329. Allow MR applications to use additional AuxServices, + which are compatible with the default MapReduce shuffle. + (Avner BenHanoch via sseth) + + MAPREDUCE-5463. Deprecate SLOTS_MILLIS counters (Tzuyoshi Ozawa via Sandy + Ryza) + + MAPREDUCE-5457. Add a KeyOnlyTextOutputReader to enable streaming to write + out text files without separators (Sandy Ryza) + + MAPREDUCE-5596. Allow configuring the number of threads used to serve + shuffle connections (Sandy Ryza via jlowe) MAPREDUCE-434. LocalJobRunner limited to single reducer (Sandy Ryza and Aaron Kimball via Sandy Ryza) @@ -206,14 +234,14 @@ Release 2.4.0 - UNRELEASED MAPREDUCE-5725. Make explicit that TestNetworkedJob relies on the Capacity Scheduler (Sandy Ryza) - MAPREDUCE-5464. Add analogs of the SLOTS_MILLIS counters that jive with the - YARN resource model (Sandy Ryza) - MAPREDUCE-5732. Report proper queue when job has been automatically placed (Sandy Ryza) OPTIMIZATIONS + MAPREDUCE-4680. Job history cleaner should only check timestamps of files in + old enough directories (Robert Kanter via Sandy Ryza) + MAPREDUCE-5484. YarnChild unnecessarily loads job conf twice (Sandy Ryza) MAPREDUCE-5487. In task processes, JobConf is unnecessarily loaded again @@ -224,6 +252,37 @@ Release 2.4.0 - UNRELEASED BUG FIXES + MAPREDUCE-5569. FloatSplitter is not generating correct splits (Nathan + Roberts via jlowe) + + MAPREDUCE-5546. mapred.cmd on Windows set HADOOP_OPTS incorrectly (Chuan Liu + via cnauroth) + + MAPREDUCE-5518. Fixed typo "can't read paritions file". (Albert Chu + via devaraj) + + MAPREDUCE-5561. org.apache.hadoop.mapreduce.v2.app.job.impl.TestJobImpl + testcase failing on trunk (Karthik Kambatla via jlowe) + + MAPREDUCE-5598. TestUserDefinedCounters.testMapReduceJob is flakey + (Robert Kanter via jlowe) + + MAPREDUCE-5604. TestMRAMWithNonNormalizedCapabilities fails on Windows due to + exceeding max path length. (cnauroth) + + MAPREDUCE-5451. MR uses LD_LIBRARY_PATH which doesn't mean anything in + Windows. (Yingda Chen via cnauroth) + + MAPREDUCE-5409. MRAppMaster throws InvalidStateTransitonException: Invalid + event: TA_TOO_MANY_FETCH_FAILURE at KILLED for TaskAttemptImpl (Gera + Shegalov via jlowe) + + MAPREDUCE-5674. Missing start and finish time in mapred.JobStatus. + (Chuan Liu via cnauroth) + + MAPREDUCE-5650. Job fails when hprof mapreduce.task.profile.map/reduce.params + is specified (Gera Shegalov via Sandy Ryza) + MAPREDUCE-5316. job -list-attempt-ids command does not handle illegal task-state (Ashwin Shankar via jlowe) @@ -297,65 +356,6 @@ Release 2.4.0 - UNRELEASED MAPREDUCE-5723. MR AM container log can be truncated or empty. (Mohammad Kamrul Islam via kasha) -Release 2.3.0 - UNRELEASED - - INCOMPATIBLE CHANGES - - NEW FEATURES - - IMPROVEMENTS - - MAPREDUCE-5329. Allow MR applications to use additional AuxServices, - which are compatible with the default MapReduce shuffle. - (Avner BenHanoch via sseth) - - MAPREDUCE-5463. Deprecate SLOTS_MILLIS counters (Tzuyoshi Ozawa via Sandy - Ryza) - - MAPREDUCE-5457. Add a KeyOnlyTextOutputReader to enable streaming to write - out text files without separators (Sandy Ryza) - - MAPREDUCE-5596. Allow configuring the number of threads used to serve - shuffle connections (Sandy Ryza via jlowe) - - OPTIMIZATIONS - - MAPREDUCE-4680. Job history cleaner should only check timestamps of files in - old enough directories (Robert Kanter via Sandy Ryza) - - BUG FIXES - - MAPREDUCE-5569. FloatSplitter is not generating correct splits (Nathan - Roberts via jlowe) - - MAPREDUCE-5546. mapred.cmd on Windows set HADOOP_OPTS incorrectly (Chuan Liu - via cnauroth) - - MAPREDUCE-5518. Fixed typo "can't read paritions file". (Albert Chu - via devaraj) - - MAPREDUCE-5561. org.apache.hadoop.mapreduce.v2.app.job.impl.TestJobImpl - testcase failing on trunk (Karthik Kambatla via jlowe) - - MAPREDUCE-5598. TestUserDefinedCounters.testMapReduceJob is flakey - (Robert Kanter via jlowe) - - MAPREDUCE-5604. TestMRAMWithNonNormalizedCapabilities fails on Windows due to - exceeding max path length. (cnauroth) - - MAPREDUCE-5451. MR uses LD_LIBRARY_PATH which doesn't mean anything in - Windows. (Yingda Chen via cnauroth) - - MAPREDUCE-5409. MRAppMaster throws InvalidStateTransitonException: Invalid - event: TA_TOO_MANY_FETCH_FAILURE at KILLED for TaskAttemptImpl (Gera - Shegalov via jlowe) - - MAPREDUCE-5674. Missing start and finish time in mapred.JobStatus. - (Chuan Liu via cnauroth) - - MAPREDUCE-5650. Job fails when hprof mapreduce.task.profile.map/reduce.params - is specified (Gera Shegalov via Sandy Ryza) - Release 2.2.0 - 2013-10-13 INCOMPATIBLE CHANGES From 3d9ad8e3b60dd21db45466f4736abe6b1812b522 Mon Sep 17 00:00:00 2001 From: Jing Zhao Date: Thu, 30 Jan 2014 19:15:13 +0000 Subject: [PATCH 05/23] HDFS-5843. DFSClient.getFileChecksum() throws IOException if checksum is disabled. Contributed by Laurent Goujon. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562927 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 +++ .../hadoop/hdfs/server/datanode/DataXceiver.java | 5 +++-- .../org/apache/hadoop/hdfs/TestFSOutputSummer.java | 13 +++++++++++-- 3 files changed, 17 insertions(+), 4 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index d665a41a288..0f0054fde97 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -304,6 +304,9 @@ Release 2.4.0 - UNRELEASED HDFS-5492. Port HDFS-2069 (Incorrect default trash interval in the docs) to trunk. (Akira Ajisaka via Arpit Agarwal) + HDFS-5843. DFSClient.getFileChecksum() throws IOException if checksum is + disabled. (Laurent Goujon via jing9) + Release 2.3.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java index 13b411e9ef6..da38b91ac13 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java @@ -655,8 +655,9 @@ public void blockChecksum(final ExtendedBlock block, final BlockMetadataHeader header = BlockMetadataHeader.readHeader(checksumIn); final DataChecksum checksum = header.getChecksum(); final int bytesPerCRC = checksum.getBytesPerChecksum(); - final long crcPerBlock = (metadataIn.getLength() - - BlockMetadataHeader.getHeaderSize())/checksum.getChecksumSize(); + final long crcPerBlock = checksum.getChecksumSize() > 0 + ? (metadataIn.getLength() - BlockMetadataHeader.getHeaderSize())/checksum.getChecksumSize() + : 0; //compute block checksum final MD5Hash md5 = MD5Hash.digest(checksumIn); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFSOutputSummer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFSOutputSummer.java index 88a83715ff0..047f67f380a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFSOutputSummer.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFSOutputSummer.java @@ -71,7 +71,7 @@ private void writeFile2(Path name) throws Exception { cleanupFile(name); } - /* create a file, write data with vairable amount of data */ + /* create a file, write data with variable amount of data */ private void writeFile3(Path name) throws Exception { FSDataOutputStream stm = fileSys.create(name, true, fileSys.getConf().getInt(IO_FILE_BUFFER_SIZE_KEY, 4096), @@ -103,6 +103,8 @@ private void checkFile(Path name) throws Exception { stm.readFully(0, actual); checkAndEraseData(actual, 0, expected, "Read Sanity Test"); stm.close(); + // do a sanity check. Get the file checksum + fileSys.getFileChecksum(name); } private void cleanupFile(Path name) throws IOException { @@ -112,13 +114,20 @@ private void cleanupFile(Path name) throws IOException { } /** - * Test write opeation for output stream in DFS. + * Test write operation for output stream in DFS. */ @Test public void testFSOutputSummer() throws Exception { + doTestFSOutputSummer("CRC32"); + doTestFSOutputSummer("CRC32C"); + doTestFSOutputSummer("NULL"); + } + + private void doTestFSOutputSummer(String checksumType) throws Exception { Configuration conf = new HdfsConfiguration(); conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, BLOCK_SIZE); conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, BYTES_PER_CHECKSUM); + conf.set(DFSConfigKeys.DFS_CHECKSUM_TYPE_KEY, checksumType); MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf) .numDataNodes(NUM_OF_DATANODES) .build(); From f874b200617a4d2a9846487c201199ad4f1a6abd Mon Sep 17 00:00:00 2001 From: Arpit Agarwal Date: Thu, 30 Jan 2014 19:28:56 +0000 Subject: [PATCH 06/23] HADOOP-10139. Update and improve the Single Cluster Setup document. (Contributed by Akira Ajisaka) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562931 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 3 + .../src/site/apt/SingleCluster.apt.vm | 409 +++++++++++------- 2 files changed, 254 insertions(+), 158 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 87d859710b8..c487f677e46 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -301,6 +301,9 @@ Release 2.4.0 - UNRELEASED IMPROVEMENTS + HADOOP-10139. Update and improve the Single Cluster Setup document. + (Akira Ajisaka via Arpit Agarwal) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-common-project/hadoop-common/src/site/apt/SingleCluster.apt.vm b/hadoop-common-project/hadoop-common/src/site/apt/SingleCluster.apt.vm index cf8390e4610..ef7532a9e52 100644 --- a/hadoop-common-project/hadoop-common/src/site/apt/SingleCluster.apt.vm +++ b/hadoop-common-project/hadoop-common/src/site/apt/SingleCluster.apt.vm @@ -20,174 +20,267 @@ Hadoop MapReduce Next Generation - Setting up a Single Node Cluster. %{toc|section=1|fromDepth=0} -* Mapreduce Tarball +* Purpose - You should be able to obtain the MapReduce tarball from the release. - If not, you should be able to create a tarball from the source. + This document describes how to set up and configure a single-node Hadoop + installation so that you can quickly perform simple operations using Hadoop + MapReduce and the Hadoop Distributed File System (HDFS). + +* Prerequisites + +** Supported Platforms + + * GNU/Linux is supported as a development and production platform. + Hadoop has been demonstrated on GNU/Linux clusters with 2000 nodes. + + * Windows is also a supported platform but the followings steps + are for Linux only. To set up Hadoop on Windows, see + {{{http://wiki.apache.org/hadoop/Hadoop2OnWindows}wiki page}}. + +** Required Software + + Required software for Linux include: + + [[1]] Java\u2122 must be installed. Recommended Java versions are described + at {{{http://wiki.apache.org/hadoop/HadoopJavaVersions} + HadoopJavaVersions}}. + + [[2]] ssh must be installed and sshd must be running to use the Hadoop + scripts that manage remote Hadoop daemons. + +** Installing Software + + If your cluster doesn't have the requisite software you will need to install + it. + + For example on Ubuntu Linux: + +---- + $ sudo apt-get install ssh + $ sudo apt-get install rsync +---- + +* Download + + To get a Hadoop distribution, download a recent stable release from one of + the {{{http://www.apache.org/dyn/closer.cgi/hadoop/common/} + Apache Download Mirrors}}. + +* Prepare to Start the Hadoop Cluster + + Unpack the downloaded Hadoop distribution. In the distribution, edit + the file <<>> to define some parameters as + follows: + +---- + # set to the root of your Java installation + export JAVA_HOME=/usr/java/latest + + # Assuming your installation directory is /usr/local/hadoop + export HADOOP_PREFIX=/usr/local/hadoop +---- + + Try the following command: + +---- + $ bin/hadoop +---- + + This will display the usage documentation for the hadoop script. + + Now you are ready to start your Hadoop cluster in one of the three supported + modes: + + * {{{Standalone Operation}Local (Standalone) Mode}} + + * {{{Pseudo-Distributed Operation}Pseudo-Distributed Mode}} + + * {{{Fully-Distributed Operation}Fully-Distributed Mode}} + +* Standalone Operation + + By default, Hadoop is configured to run in a non-distributed mode, as a + single Java process. This is useful for debugging. + + The following example copies the unpacked conf directory to use as input + and then finds and displays every match of the given regular expression. + Output is written to the given output directory. + +---- + $ mkdir input + $ cp etc/hadoop/*.xml input + $ bin/hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-${project.version}.jar grep input output 'dfs[a-z.]+' + $ cat output/* +---- + +* Pseudo-Distributed Operation + + Hadoop can also be run on a single-node in a pseudo-distributed mode where + each Hadoop daemon runs in a separate Java process. + +** Configuration + + Use the following: + + etc/hadoop/core-site.xml: +---+ -$ mvn clean install -DskipTests -$ cd hadoop-mapreduce-project -$ mvn clean install assembly:assembly -Pnative -+---+ - <> You will need {{{http://code.google.com/p/protobuf}protoc 2.5.0}} - installed. - - To ignore the native builds in mapreduce you can omit the <<<-Pnative>>> argument - for maven. The tarball should be available in <<>> directory. - - -* Setting up the environment. - - Assuming you have installed hadoop-common/hadoop-hdfs and exported - <<$HADOOP_COMMON_HOME>>/<<$HADOOP_HDFS_HOME>>, untar hadoop mapreduce - tarball and set environment variable <<$HADOOP_MAPRED_HOME>> to the - untarred directory. Set <<$HADOOP_YARN_HOME>> the same as <<$HADOOP_MAPRED_HOME>>. - - <> The following instructions assume you have hdfs running. - -* Setting up Configuration. - - To start the ResourceManager and NodeManager, you will have to update the configs. - Assuming your $HADOOP_CONF_DIR is the configuration directory and has the installed - configs for HDFS and <<>>. There are 2 config files you will have to setup - <<>> and <<>>. - -** Setting up <<>> - - Add the following configs to your <<>>. - -+---+ - - mapreduce.cluster.temp.dir - - No description - true - - - - mapreduce.cluster.local.dir - - No description - true - + + + fs.defaultFS + hdfs://localhost:9000 + + +---+ -** Setting up <<>> - -Add the following configs to your <<>> - -+---+ - - yarn.resourcemanager.resource-tracker.address - host:port - host is the hostname of the resource manager and - port is the port on which the NodeManagers contact the Resource Manager. - - - - - yarn.resourcemanager.scheduler.address - host:port - host is the hostname of the resourcemanager and port is the port - on which the Applications in the cluster talk to the Resource Manager. - - - - - yarn.resourcemanager.scheduler.class - org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler - In case you do not want to use the default scheduler - - - - yarn.resourcemanager.address - host:port - the host is the hostname of the ResourceManager and the port is the port on - which the clients can talk to the Resource Manager. - - - - yarn.nodemanager.local-dirs - - the local directories used by the nodemanager - - - - yarn.nodemanager.address - 0.0.0.0:port - the nodemanagers bind to this port - - - - yarn.nodemanager.resource.memory-mb - 10240 - the amount of memory on the NodeManager in GB - - - - yarn.nodemanager.remote-app-log-dir - /app-logs - directory on hdfs where the application logs are moved to - - - - yarn.nodemanager.log-dirs - - the directories used by Nodemanagers as log directories - - - - yarn.nodemanager.aux-services - mapreduce_shuffle - shuffle service that needs to be set for Map Reduce to run - -+---+ - -* Setting up <<>> - - Make sure you populate the root queues in <<>>. - -+---+ - - yarn.scheduler.capacity.root.queues - unfunded,default - - - - yarn.scheduler.capacity.root.capacity - 100 - - - - yarn.scheduler.capacity.root.unfunded.capacity - 50 - - - - yarn.scheduler.capacity.root.default.capacity - 50 - -+---+ - -* Running daemons. - - Assuming that the environment variables <<$HADOOP_COMMON_HOME>>, <<$HADOOP_HDFS_HOME>>, <<$HADOO_MAPRED_HOME>>, - <<$HADOOP_YARN_HOME>>, <<$JAVA_HOME>> and <<$HADOOP_CONF_DIR>> have been set appropriately. - Set $<<$YARN_CONF_DIR>> the same as $<> - - Run ResourceManager and NodeManager as: + etc/hadoop/hdfs-site.xml: +---+ -$ cd $HADOOP_MAPRED_HOME -$ sbin/yarn-daemon.sh start resourcemanager -$ sbin/yarn-daemon.sh start nodemanager + + + dfs.replication + 1 + + +---+ - You should be up and running. You can run randomwriter as: +** Setup passphraseless ssh + + Now check that you can ssh to the localhost without a passphrase: + +---- + $ ssh localhost +---- + + If you cannot ssh to localhost without a passphrase, execute the + following commands: + +---- + $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa + $ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys +---- + +** Execution + + The following instructions are to run a MapReduce job locally. + If you want to execute a job on YARN, see {{YARN on Single Node}}. + + [[1]] Format the filesystem: + +---- + $ bin/hdfs namenode -format +---- + + [[2]] Start NameNode daemon and DataNode daemon: + +---- + $ sbin/start-dfs.sh +---- + + The hadoop daemon log output is written to the <<<${HADOOP_LOG_DIR}>>> + directory (defaults to <<<${HADOOP_HOME}/logs>>>). + + [[3]] Browse the web interface for the NameNode; by default it is + available at: + + * NameNode - <<>> + + [[4]] Make the HDFS directories required to execute MapReduce jobs: + +---- + $ bin/hdfs dfs -mkdir /user + $ bin/hdfs dfs -mkdir /user/ +---- + + [[5]] Copy the input files into the distributed filesystem: + +---- + $ bin/hdfs dfs -put etc/hadoop input +---- + + [[6]] Run some of the examples provided: + +---- + $ bin/hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-${project.version}.jar grep input output 'dfs[a-z.]+' +---- + + [[7]] Examine the output files: + + Copy the output files from the distributed filesystem to the local + filesystem and examine them: + +---- + $ bin/hdfs dfs -get output output + $ cat output/* +---- + + or + + View the output files on the distributed filesystem: + +---- + $ bin/hdfs dfs -cat output/* +---- + + [[8]] When you're done, stop the daemons with: + +---- + $ sbin/stop-dfs.sh +---- + +** YARN on Single Node + + You can run a MapReduce job on YARN in a pseudo-distributed mode by setting + a few parameters and running ResourceManager daemon and NodeManager daemon + in addition. + + The following instructions assume that 1. ~ 4. steps of + {{{Execution}the above instructions}} are already executed. + + [[1]] Configure parameters as follows: + + etc/hadoop/mapred-site.xml: +---+ -$ $HADOOP_COMMON_HOME/bin/hadoop jar hadoop-examples.jar randomwriter out + + + mapreduce.framework.name + yarn + + +---+ -Good luck. + etc/hadoop/yarn-site.xml: + ++---+ + + + yarn.nodemanager.aux-services + mapreduce_shuffle + + ++---+ + + [[2]] Start ResourceManager daemon and NodeManager daemon: + +---- + $ sbin/start-yarn.sh +---- + + [[3]] Browse the web interface for the ResourceManager; by default it is + available at: + + * ResourceManager - <<>> + + [[4]] Run a MapReduce job. + + [[5]] When you're done, stop the daemons with: + +---- + $ sbin/stop-yarn.sh +---- + +* Fully-Distributed Operation + + For information on setting up fully-distributed, non-trivial clusters + see {{{./ClusterSetup.html}Cluster Setup}}. From d0b4c7d6b18276193e32a7b0ebcd9b57e29be02c Mon Sep 17 00:00:00 2001 From: Vinod Kumar Vavilapalli Date: Thu, 30 Jan 2014 20:16:56 +0000 Subject: [PATCH 07/23] Updating branch-2's MAPREDUCE CHANGES.txt to reflect the recent rebase of unreleased 2.3 to more latest branch-2. Updating trunk's MAPREDUCE CHANGES.txt to fix more differences with branch-2's CHANGES.txt MAPREDUCE's CHANGES.txt is foobar'ed. There are still a bunch of misplaced CHANGES.txt entries - not fixing them now. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562945 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-mapreduce-project/CHANGES.txt | 23 +++++++++++++---------- 1 file changed, 13 insertions(+), 10 deletions(-) diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 542a9a87ab9..831b003954c 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -82,9 +82,6 @@ Trunk (Unreleased) BUG FIXES - MAPREDUCE-4272. SortedRanges.Range#compareTo is not spec compliant. - (Yu Gao via llu) - MAPREDUCE-3194. "mapred mradmin" command is broken in mrv2 (Jason Lowe via bobby) @@ -130,15 +127,9 @@ Trunk (Unreleased) MAPREDUCE-4574. Fix TotalOrderParitioner to work with non-WritableComparable key types. (harsh) - MAPREDUCE-4884. Streaming tests fail to start MiniMRCluster due to missing - queue configuration. (Chris Nauroth via suresh) - MAPREDUCE-5012. Typo in javadoc for IdentityMapper class. (Adam Monsen via suresh) - MAPREDUCE-4885. Streaming tests have multiple failures on Windows. (Chris - Nauroth via bikas) - MAPREDUCE-4987. TestMRJobs#testDistributedCache fails on Windows due to classpath problems and unexpected behavior of symlinks (Chris Nauroth via bikas) @@ -1002,9 +993,15 @@ Release 2.1.0-beta - 2013-08-22 HADOOP-9372. Fix bad timeout annotations on tests. (Arpit Agarwal via suresh) + MAPREDUCE-4885. Streaming tests have multiple failures on Windows. (Chris + Nauroth via bikas) + MAPREDUCE-5177. Use common utils FileUtil#setReadable/Writable/Executable & FileUtil#canRead/Write/Execute. (Ivan Mitic via suresh) + MAPREDUCE-5349. TestClusterMapReduceTestCase and TestJobName fail on Windows + in branch-2. (Chuan Liu via cnauroth) + MAPREDUCE-5355. MiniMRYarnCluster with localFs does not work on Windows. (Chuan Liu via cnauroth) @@ -1155,6 +1152,9 @@ Release 2.0.3-alpha - 2013-02-06 BUG FIXES + MAPREDUCE-4272. SortedRanges.Range#compareTo is not spec compliant. + (Yu Gao via llu) + MAPREDUCE-4607. Race condition in ReduceTask completion can result in Task being incorrectly failed. (Bikas Saha via tomwhite) @@ -1217,6 +1217,9 @@ Release 2.0.3-alpha - 2013-02-06 MAPREDUCE-4969. TestKeyValueTextInputFormat test fails with Open JDK 7. (Arpit Agarwal via suresh) + MAPREDUCE-4884. Streaming tests fail to start MiniMRCluster due to missing + queue configuration. (Chris Nauroth via suresh) + MAPREDUCE-4953. HadoopPipes misuses fprintf. (Andy Isaacson via atm) Release 2.0.2-alpha - 2012-09-07 @@ -1225,7 +1228,7 @@ Release 2.0.2-alpha - 2012-09-07 NEW FEATURES - MAPREDUCE-987. Exposing MiniDFS and MiniMR clusters as a single process + MAPREDUCE-987. Exposing MiniDFS and MiniMR clusters as a single process command-line. (ahmed via tucu) MAPREDUCE-4417. add support for encrypted shuffle (tucu) From 10f09017640075f578bec85219cd9054662a0524 Mon Sep 17 00:00:00 2001 From: Vinod Kumar Vavilapalli Date: Thu, 30 Jan 2014 20:30:12 +0000 Subject: [PATCH 08/23] Updating branch-2.3's YARN's CHANGES.txt to reflect the recent rebase of unreleased 2.3 to more latest branch-2. Fixing entry for MAPREDUCE-5732 in trunk and branch-2. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562949 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-mapreduce-project/CHANGES.txt | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 831b003954c..fae27e4929a 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -150,6 +150,9 @@ Release 2.4.0 - UNRELEASED MAPREDUCE-5464. Add analogs of the SLOTS_MILLIS counters that jive with the YARN resource model (Sandy Ryza) + MAPREDUCE-5732. Report proper queue when job has been automatically placed + (Sandy Ryza) + OPTIMIZATIONS BUG FIXES @@ -225,9 +228,6 @@ Release 2.3.0 - UNRELEASED MAPREDUCE-5725. Make explicit that TestNetworkedJob relies on the Capacity Scheduler (Sandy Ryza) - MAPREDUCE-5732. Report proper queue when job has been automatically placed - (Sandy Ryza) - OPTIMIZATIONS MAPREDUCE-4680. Job history cleaner should only check timestamps of files in From 6ada3c7d424519487e10e0158805e570ddd51889 Mon Sep 17 00:00:00 2001 From: Vinod Kumar Vavilapalli Date: Thu, 30 Jan 2014 20:37:23 +0000 Subject: [PATCH 09/23] Updating trunk's YARN CHANGES.txt after YARN-321 merge. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562950 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 36 ++++++++++++++++----------------- 1 file changed, 18 insertions(+), 18 deletions(-) diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index bc5a40ae00b..4cf6290d8f0 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -9,6 +9,21 @@ Trunk - Unreleased YARN-1496. Protocol additions to allow moving apps between queues (Sandy Ryza) + IMPROVEMENTS + + OPTIMIZATIONS + + BUG FIXES + + YARN-524 TestYarnVersionInfo failing if generated properties doesn't + include an SVN URL. (stevel) + +Release 2.4.0 - UNRELEASED + + INCOMPATIBLE CHANGES + + NEW FEATURES + YARN-930. Bootstrapping ApplicationHistoryService module. (vinodkv) YARN-947. Implementing the data objects to be used by the History reader @@ -70,13 +85,13 @@ Trunk - Unreleased YARN-987. Added ApplicationHistoryManager responsible for exposing reports to all clients. (Mayank Bansal via vinodkv) + YARN-1630. Introduce timeout for async polling operations in YarnClientImpl + (Aditya Acharya via Sandy Ryza) + OPTIMIZATIONS BUG FIXES - YARN-524 TestYarnVersionInfo failing if generated properties doesn't - include an SVN URL. (stevel) - YARN-935. Correcting pom.xml to build applicationhistoryserver module successfully. (Zhijie Shen via vinodkv) @@ -112,21 +127,6 @@ Trunk - Unreleased YARN-1613. Fixed the typo with the configuration name YARN_HISTORY_SERVICE_ENABLED. (Akira Ajisaka via vinodkv) -Release 2.4.0 - UNRELEASED - - INCOMPATIBLE CHANGES - - NEW FEATURES - - IMPROVEMENTS - - YARN-1630. Introduce timeout for async polling operations in YarnClientImpl - (Aditya Acharya via Sandy Ryza) - - OPTIMIZATIONS - - BUG FIXES - YARN-1618. Fix invalid RMApp transition from NEW to FINAL_SAVING (kasha) YARN-1600. RM does not startup when security is enabled without spnego From fd2ba4911bec60386a106c85aedd5cc99cd8cab3 Mon Sep 17 00:00:00 2001 From: Todd Lipcon Date: Thu, 30 Jan 2014 21:21:15 +0000 Subject: [PATCH 10/23] HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery. Contributed by Todd Lipcon. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1562970 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 ++ .../hdfs/server/namenode/FSNamesystem.java | 8 ++--- .../hdfs/server/namenode/LeaseManager.java | 36 ------------------- 3 files changed, 7 insertions(+), 40 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 0f0054fde97..4c6cf2e3634 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -299,6 +299,9 @@ Release 2.4.0 - UNRELEASED OPTIMIZATIONS + HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery + (todd) + BUG FIXES HDFS-5492. Port HDFS-2069 (Incorrect default trash interval in the diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index cd7bcf43769..ef2c28622f9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -3800,7 +3800,7 @@ private void commitOrCompleteLastBlock(final INodeFile fileINode, final long diff = fileINode.getPreferredBlockSize() - commitBlock.getNumBytes(); if (diff > 0) { try { - String path = leaseManager.findPath(fileINode); + String path = fileINode.getFullPathName(); dir.updateSpaceConsumed(path, 0, -diff*fileINode.getFileReplication()); } catch (IOException e) { LOG.warn("Unexpected exception while updating disk space.", e); @@ -4002,7 +4002,7 @@ void commitBlockSynchronization(ExtendedBlock lastblock, @VisibleForTesting String closeFileCommitBlocks(INodeFile pendingFile, BlockInfo storedBlock) throws IOException { - String src = leaseManager.findPath(pendingFile); + String src = pendingFile.getFullPathName(); // commit the last block and complete it if it has minimum replicas commitOrCompleteLastBlock(pendingFile, storedBlock); @@ -4024,7 +4024,7 @@ String closeFileCommitBlocks(INodeFile pendingFile, BlockInfo storedBlock) @VisibleForTesting String persistBlocks(INodeFile pendingFile, boolean logRetryCache) throws IOException { - String src = leaseManager.findPath(pendingFile); + String src = pendingFile.getFullPathName(); dir.persistBlocks(src, pendingFile, logRetryCache); return src; } @@ -5952,7 +5952,7 @@ private void updatePipelineInternal(String clientName, ExtendedBlock oldBlock, .getDatanodeStorageInfos(newNodes, newStorageIDs); blockinfo.setExpectedLocations(storages); - String src = leaseManager.findPath(pendingFile); + String src = pendingFile.getFullPathName(); dir.persistBlocks(src, pendingFile, logRetryCache); } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/LeaseManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/LeaseManager.java index 8b5fb81815f..b9f9d10396e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/LeaseManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/LeaseManager.java @@ -179,24 +179,6 @@ synchronized Lease reassignLease(Lease lease, String src, String newHolder) { return addLease(newHolder, src); } - /** - * Finds the pathname for the specified pendingFile - */ - public synchronized String findPath(INodeFile pendingFile) - throws IOException { - FileUnderConstructionFeature uc = pendingFile.getFileUnderConstructionFeature(); - Preconditions.checkArgument(uc != null); - Lease lease = getLease(uc.getClientName()); - if (lease != null) { - String src = lease.findPath(pendingFile); - if (src != null) { - return src; - } - } - throw new IOException("pendingFile (=" + pendingFile + ") not found." - + "(lease=" + lease + ")"); - } - /** * Renew the lease(s) held by the given client */ @@ -252,24 +234,6 @@ public boolean expiredSoftLimit() { return now() - lastUpdate > softLimit; } - /** - * @return the path associated with the pendingFile and null if not found. - */ - private String findPath(INodeFile pendingFile) { - try { - for (String src : paths) { - INode node = fsnamesystem.dir.getINode(src); - if (node == pendingFile - || (node.isFile() && node.asFile() == pendingFile)) { - return src; - } - } - } catch (UnresolvedLinkException e) { - throw new AssertionError("Lease files should reside on this FS"); - } - return null; - } - /** Does this lease contain any path? */ boolean hasPath() {return !paths.isEmpty();} From 1cac66ce68d5f8d5f73920133eb91fa43e9bdb1c Mon Sep 17 00:00:00 2001 From: Sanford Ryza Date: Thu, 30 Jan 2014 22:40:22 +0000 Subject: [PATCH 11/23] YARN-1617. Remove ancient comment and surround LOG.debug in AppSchedulingInfo.allocate (Sandy Ryza) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563004 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 3 ++ .../scheduler/AppSchedulingInfo.java | 34 ++++--------------- 2 files changed, 10 insertions(+), 27 deletions(-) diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 4cf6290d8f0..f90d5cb286d 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -88,6 +88,9 @@ Release 2.4.0 - UNRELEASED YARN-1630. Introduce timeout for async polling operations in YarnClientImpl (Aditya Acharya via Sandy Ryza) + YARN-1617. Remove ancient comment and surround LOG.debug in + AppSchedulingInfo.allocate (Sandy Ryza) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java index 7ac1c0a6714..3270dea4498 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java @@ -262,8 +262,13 @@ synchronized public void allocate(NodeType type, SchedulerNode node, pending = false; metrics.runAppAttempt(applicationId, user); } - LOG.debug("allocate: user: " + user + ", memory: " - + request.getCapability()); + if (LOG.isDebugEnabled()) { + LOG.debug("allocate: applicationId=" + applicationId + + " container=" + container.getId() + + " host=" + container.getNodeId().toString() + + " user=" + user + + " resource=" + request.getCapability()); + } metrics.allocateResources(user, 1, request.getCapability()); } @@ -277,9 +282,6 @@ synchronized public void allocate(NodeType type, SchedulerNode node, synchronized private void allocateNodeLocal( SchedulerNode node, Priority priority, ResourceRequest nodeLocalRequest, Container container) { - // Update consumption and track allocations - allocate(container); - // Update future requirements nodeLocalRequest.setNumContainers(nodeLocalRequest.getNumContainers() - 1); if (nodeLocalRequest.getNumContainers() == 0) { @@ -306,10 +308,6 @@ synchronized private void allocateNodeLocal( synchronized private void allocateRackLocal( SchedulerNode node, Priority priority, ResourceRequest rackLocalRequest, Container container) { - - // Update consumption and track allocations - allocate(container); - // Update future requirements rackLocalRequest.setNumContainers(rackLocalRequest.getNumContainers() - 1); if (rackLocalRequest.getNumContainers() == 0) { @@ -329,10 +327,6 @@ synchronized private void allocateRackLocal( synchronized private void allocateOffSwitch( SchedulerNode node, Priority priority, ResourceRequest offSwitchRequest, Container container) { - - // Update consumption and track allocations - allocate(container); - // Update future requirements decrementOutstanding(offSwitchRequest); } @@ -365,20 +359,6 @@ synchronized private void checkForDeactivation() { } } - synchronized private void allocate(Container container) { - // Update consumption and track allocations - //TODO: fixme sharad - /* try { - store.storeContainer(container); - } catch (IOException ie) { - // TODO fix this. we shouldnt ignore - }*/ - - LOG.debug("allocate: applicationId=" + applicationId + " container=" - + container.getId() + " host=" - + container.getNodeId().toString()); - } - synchronized public void stop(RMAppAttemptState rmAppAttemptFinalState) { // clear pending resources metrics for the application QueueMetrics metrics = queue.getMetrics(); From 067d52b98c1d17a73b142bb53acc8aaa9c041f38 Mon Sep 17 00:00:00 2001 From: Jing Zhao Date: Thu, 30 Jan 2014 23:53:23 +0000 Subject: [PATCH 12/23] HADOOP-10295. Allow distcp to automatically identify the checksum type of source files and use it for the target. Contributed by Jing Zhao and Laurent Goujon. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563019 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 3 + .../org/apache/hadoop/fs/FileChecksum.java | 13 +- .../hadoop/fs/MD5MD5CRC32FileChecksum.java | 15 +- .../hadoop/tools/DistCpOptionSwitch.java | 15 +- .../apache/hadoop/tools/DistCpOptions.java | 2 +- .../apache/hadoop/tools/OptionsParser.java | 6 +- .../hadoop/tools/mapred/CopyMapper.java | 6 +- .../mapred/RetriableFileCopyCommand.java | 109 +++++++++---- .../apache/hadoop/tools/util/DistCpUtils.java | 14 +- .../hadoop/tools/TestOptionsParser.java | 35 +++- .../hadoop/tools/mapred/TestCopyMapper.java | 150 +++++++++++++----- 11 files changed, 257 insertions(+), 111 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index c487f677e46..54011a9dfed 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -304,6 +304,9 @@ Release 2.4.0 - UNRELEASED HADOOP-10139. Update and improve the Single Cluster Setup document. (Akira Ajisaka via Arpit Agarwal) + HADOOP-10295. Allow distcp to automatically identify the checksum type of + source files and use it for the target. (jing9 and Laurent Goujon) + OPTIMIZATIONS BUG FIXES diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileChecksum.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileChecksum.java index 149a3e3a4a0..62f1a9b3f48 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileChecksum.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileChecksum.java @@ -21,21 +21,26 @@ import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; +import org.apache.hadoop.fs.Options.ChecksumOpt; import org.apache.hadoop.io.Writable; /** An abstract class representing file checksums for files. */ @InterfaceAudience.Public @InterfaceStability.Stable public abstract class FileChecksum implements Writable { - /** The checksum algorithm name */ + /** The checksum algorithm name */ public abstract String getAlgorithmName(); - /** The length of the checksum in bytes */ + /** The length of the checksum in bytes */ public abstract int getLength(); - /** The value of the checksum in bytes */ + /** The value of the checksum in bytes */ public abstract byte[] getBytes(); + public ChecksumOpt getChecksumOpt() { + return null; + } + /** Return true if both the algorithms and the values are the same. */ @Override public boolean equals(Object other) { @@ -50,7 +55,7 @@ public boolean equals(Object other) { return this.getAlgorithmName().equals(that.getAlgorithmName()) && Arrays.equals(this.getBytes(), that.getBytes()); } - + @Override public int hashCode() { return getAlgorithmName().hashCode() ^ Arrays.hashCode(getBytes()); diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.java index 0eb77ddd96e..59189956728 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.java @@ -56,7 +56,7 @@ public MD5MD5CRC32FileChecksum(int bytesPerCRC, long crcPerBlock, MD5Hash md5) { this.crcPerBlock = crcPerBlock; this.md5 = md5; } - + @Override public String getAlgorithmName() { return "MD5-of-" + crcPerBlock + "MD5-of-" + bytesPerCRC + @@ -73,10 +73,10 @@ public static DataChecksum.Type getCrcTypeFromAlgorithmName(String algorithm) throw new IOException("Unknown checksum type in " + algorithm); } - + @Override public int getLength() {return LENGTH;} - + @Override public byte[] getBytes() { return WritableUtils.toByteArray(this); @@ -88,6 +88,7 @@ public DataChecksum.Type getCrcType() { return DataChecksum.Type.CRC32; } + @Override public ChecksumOpt getChecksumOpt() { return new ChecksumOpt(getCrcType(), bytesPerCRC); } @@ -98,12 +99,12 @@ public void readFields(DataInput in) throws IOException { crcPerBlock = in.readLong(); md5 = MD5Hash.read(in); } - + @Override public void write(DataOutput out) throws IOException { out.writeInt(bytesPerCRC); out.writeLong(crcPerBlock); - md5.write(out); + md5.write(out); } /** Write that object to xml output. */ @@ -157,11 +158,11 @@ public static MD5MD5CRC32FileChecksum valueOf(Attributes attrs } } catch (Exception e) { throw new SAXException("Invalid attributes: bytesPerCRC=" + bytesPerCRC - + ", crcPerBlock=" + crcPerBlock + ", crcType=" + crcType + + ", crcPerBlock=" + crcPerBlock + ", crcType=" + crcType + ", md5=" + md5, e); } } - + @Override public String toString() { return getAlgorithmName() + ":" + md5; diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptionSwitch.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptionSwitch.java index 6cecf0618f8..c3630cc4fb4 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptionSwitch.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptionSwitch.java @@ -37,15 +37,16 @@ public enum DistCpOptionSwitch { /** * Preserves status of file/path in the target. * Default behavior with -p, is to preserve replication, - * block size, user, group and permission on the target file + * block size, user, group, permission and checksum type on the target file. + * Note that when preserving checksum type, block size is also preserved. * - * If any of the optional switches are present among rbugp, then - * only the corresponding file attribute is preserved + * If any of the optional switches are present among rbugpc, then + * only the corresponding file attribute is preserved. * */ PRESERVE_STATUS(DistCpConstants.CONF_LABEL_PRESERVE_STATUS, - new Option("p", true, "preserve status (rbugp)" + - "(replication, block-size, user, group, permission)")), + new Option("p", true, "preserve status (rbugpc)" + + "(replication, block-size, user, group, permission, checksum-type)")), /** * Update target location by copying only files that are missing @@ -53,7 +54,7 @@ public enum DistCpOptionSwitch { * across source and target. Typically used with DELETE_MISSING * Incompatible with ATOMIC_COMMIT */ - SYNC_FOLDERS(DistCpConstants.CONF_LABEL_SYNC_FOLDERS, + SYNC_FOLDERS(DistCpConstants.CONF_LABEL_SYNC_FOLDERS, new Option("update", false, "Update target, copying only missing" + "files or directories")), @@ -80,7 +81,7 @@ public enum DistCpOptionSwitch { * Max number of maps to use during copy. DistCp will split work * as equally as possible among these maps */ - MAX_MAPS(DistCpConstants.CONF_LABEL_MAX_MAPS, + MAX_MAPS(DistCpConstants.CONF_LABEL_MAX_MAPS, new Option("m", true, "Max number of concurrent maps to use for copy")), /** diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptions.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptions.java index f605767e1cc..b85dafa22a5 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptions.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/DistCpOptions.java @@ -61,7 +61,7 @@ public class DistCpOptions { private Path targetPath; public static enum FileAttribute{ - REPLICATION, BLOCKSIZE, USER, GROUP, PERMISSION; + REPLICATION, BLOCKSIZE, USER, GROUP, PERMISSION, CHECKSUMTYPE; public static FileAttribute getAttribute(char symbol) { for (FileAttribute attribute : values()) { diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/OptionsParser.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/OptionsParser.java index 79487e59b90..d36311d17b8 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/OptionsParser.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/OptionsParser.java @@ -34,7 +34,7 @@ public class OptionsParser { private static final Log LOG = LogFactory.getLog(OptionsParser.class); - private static final Options cliOptions = new Options(); + private static final Options cliOptions = new Options(); static { for (DistCpOptionSwitch option : DistCpOptionSwitch.values()) { @@ -50,7 +50,7 @@ private static class CustomParser extends GnuParser { protected String[] flatten(Options options, String[] arguments, boolean stopAtNonOption) { for (int index = 0; index < arguments.length; index++) { if (arguments[index].equals("-" + DistCpOptionSwitch.PRESERVE_STATUS.getSwitch())) { - arguments[index] = "-prbugp"; + arguments[index] = "-prbugpc"; } } return super.flatten(options, arguments, stopAtNonOption); @@ -125,7 +125,7 @@ public static DistCpOptions parse(String args[]) throws IllegalArgumentException option.setAtomicWorkPath(new Path(workPath)); } } else if (command.hasOption(DistCpOptionSwitch.WORK_PATH.getSwitch())) { - throw new IllegalArgumentException("-tmp work-path can only be specified along with -atomic"); + throw new IllegalArgumentException("-tmp work-path can only be specified along with -atomic"); } if (command.hasOption(DistCpOptionSwitch.LOG_PATH.getSwitch())) { diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/CopyMapper.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/CopyMapper.java index 7fe0b99a706..9386f6e9630 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/CopyMapper.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/CopyMapper.java @@ -111,7 +111,7 @@ public void setup(Context context) throws IOException, InterruptedException { */ private void initializeSSLConf(Context context) throws IOException { LOG.info("Initializing SSL configuration"); - + String workDir = conf.get(JobContext.JOB_LOCAL_DIR) + "/work"; Path[] cacheFiles = context.getLocalCacheFiles(); @@ -294,7 +294,7 @@ private void handleFailures(IOException exception, RetriableFileCopyCommand.CopyReadException) { incrementCounter(context, Counter.FAIL, 1); incrementCounter(context, Counter.BYTESFAILED, sourceFileStatus.getLen()); - context.write(null, new Text("FAIL: " + sourceFileStatus.getPath() + " - " + + context.write(null, new Text("FAIL: " + sourceFileStatus.getPath() + " - " + StringUtils.stringifyException(exception))); } else @@ -322,7 +322,7 @@ private boolean mustUpdate(FileSystem sourceFS, FileStatus source, Path target) targetFileStatus.getLen() != source.getLen() || (!skipCrc && !DistCpUtils.checksumsAreEqual(sourceFS, - source.getPath(), targetFS, target)) + source.getPath(), null, targetFS, target)) || (source.getBlockSize() != targetFileStatus.getBlockSize() && preserve.contains(FileAttribute.BLOCKSIZE)) ); diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/RetriableFileCopyCommand.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/RetriableFileCopyCommand.java index 580229cf8e8..1b735262581 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/RetriableFileCopyCommand.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/RetriableFileCopyCommand.java @@ -18,23 +18,33 @@ package org.apache.hadoop.tools.mapred; -import org.apache.hadoop.tools.util.RetriableCommand; -import org.apache.hadoop.tools.util.ThrottledInputStream; -import org.apache.hadoop.tools.util.DistCpUtils; -import org.apache.hadoop.tools.DistCpOptions.*; -import org.apache.hadoop.tools.DistCpConstants; -import org.apache.hadoop.fs.*; -import org.apache.hadoop.mapreduce.Mapper; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.io.IOUtils; +import java.io.BufferedInputStream; +import java.io.BufferedOutputStream; +import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; +import java.util.EnumSet; + import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CreateFlag; +import org.apache.hadoop.fs.FileChecksum; +import org.apache.hadoop.fs.FileStatus; +import org.apache.hadoop.fs.FileSystem; +import org.apache.hadoop.fs.Options.ChecksumOpt; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.permission.FsPermission; +import org.apache.hadoop.io.IOUtils; +import org.apache.hadoop.mapreduce.Mapper; +import org.apache.hadoop.tools.DistCpConstants; +import org.apache.hadoop.tools.DistCpOptions.FileAttribute; +import org.apache.hadoop.tools.util.DistCpUtils; +import org.apache.hadoop.tools.util.RetriableCommand; +import org.apache.hadoop.tools.util.ThrottledInputStream; import com.google.common.annotations.VisibleForTesting; -import java.io.*; -import java.util.EnumSet; - /** * This class extends RetriableCommand to implement the copy of files, * with retries on failure. @@ -44,7 +54,7 @@ public class RetriableFileCopyCommand extends RetriableCommand { private static Log LOG = LogFactory.getLog(RetriableFileCopyCommand.class); private static int BUFFER_SIZE = 8 * 1024; private boolean skipCrc = false; - + /** * Constructor, taking a description of the action. * @param description Verbose description of the copy operation. @@ -52,7 +62,7 @@ public class RetriableFileCopyCommand extends RetriableCommand { public RetriableFileCopyCommand(String description) { super(description); } - + /** * Create a RetriableFileCopyCommand. * @@ -99,15 +109,21 @@ private long doCopy(FileStatus sourceFileStatus, Path target, LOG.debug("Copying " + sourceFileStatus.getPath() + " to " + target); LOG.debug("Tmp-file path: " + tmpTargetPath); } - FileSystem sourceFS = sourceFileStatus.getPath().getFileSystem( - configuration); - long bytesRead = copyToTmpFile(tmpTargetPath, targetFS, sourceFileStatus, - context, fileAttributes); + final Path sourcePath = sourceFileStatus.getPath(); + final FileSystem sourceFS = sourcePath.getFileSystem(configuration); + final FileChecksum sourceChecksum = fileAttributes + .contains(FileAttribute.CHECKSUMTYPE) ? sourceFS + .getFileChecksum(sourcePath) : null; - compareFileLengths(sourceFileStatus, tmpTargetPath, configuration, bytesRead); + long bytesRead = copyToTmpFile(tmpTargetPath, targetFS, sourceFileStatus, + context, fileAttributes, sourceChecksum); + + compareFileLengths(sourceFileStatus, tmpTargetPath, configuration, + bytesRead); //At this point, src&dest lengths are same. if length==0, we skip checksum if ((bytesRead != 0) && (!skipCrc)) { - compareCheckSums(sourceFS, sourceFileStatus.getPath(), targetFS, tmpTargetPath); + compareCheckSums(sourceFS, sourceFileStatus.getPath(), sourceChecksum, + targetFS, tmpTargetPath); } promoteTmpToTarget(tmpTargetPath, target, targetFS); return bytesRead; @@ -118,14 +134,33 @@ private long doCopy(FileStatus sourceFileStatus, Path target, } } + /** + * @return the checksum spec of the source checksum if checksum type should be + * preserved + */ + private ChecksumOpt getChecksumOpt(EnumSet fileAttributes, + FileChecksum sourceChecksum) { + if (fileAttributes.contains(FileAttribute.CHECKSUMTYPE) + && sourceChecksum != null) { + return sourceChecksum.getChecksumOpt(); + } + return null; + } + private long copyToTmpFile(Path tmpTargetPath, FileSystem targetFS, - FileStatus sourceFileStatus, Mapper.Context context, - EnumSet fileAttributes) - throws IOException { - OutputStream outStream = new BufferedOutputStream(targetFS.create( - tmpTargetPath, true, BUFFER_SIZE, - getReplicationFactor(fileAttributes, sourceFileStatus, targetFS, tmpTargetPath), - getBlockSize(fileAttributes, sourceFileStatus, targetFS, tmpTargetPath), context)); + FileStatus sourceFileStatus, Mapper.Context context, + EnumSet fileAttributes, final FileChecksum sourceChecksum) + throws IOException { + FsPermission permission = FsPermission.getFileDefault().applyUMask( + FsPermission.getUMask(targetFS.getConf())); + OutputStream outStream = new BufferedOutputStream( + targetFS.create(tmpTargetPath, permission, + EnumSet.of(CreateFlag.CREATE, CreateFlag.OVERWRITE), BUFFER_SIZE, + getReplicationFactor(fileAttributes, sourceFileStatus, targetFS, + tmpTargetPath), + getBlockSize(fileAttributes, sourceFileStatus, targetFS, + tmpTargetPath), + context, getChecksumOpt(fileAttributes, sourceChecksum))); return copyBytes(sourceFileStatus, outStream, BUFFER_SIZE, context); } @@ -140,9 +175,10 @@ private void compareFileLengths(FileStatus sourceFileStatus, Path target, } private void compareCheckSums(FileSystem sourceFS, Path source, - FileSystem targetFS, Path target) - throws IOException { - if (!DistCpUtils.checksumsAreEqual(sourceFS, source, targetFS, target)) { + FileChecksum sourceChecksum, FileSystem targetFS, Path target) + throws IOException { + if (!DistCpUtils.checksumsAreEqual(sourceFS, source, sourceChecksum, + targetFS, target)) { StringBuilder errorMessage = new StringBuilder("Check-sum mismatch between ") .append(source).append(" and ").append(target).append("."); if (sourceFS.getFileStatus(source).getBlockSize() != targetFS.getFileStatus(target).getBlockSize()) { @@ -249,11 +285,18 @@ private static short getReplicationFactor( sourceFile.getReplication() : targetFS.getDefaultReplication(tmpTargetPath); } + /** + * @return the block size of the source file if we need to preserve either + * the block size or the checksum type. Otherwise the default block + * size of the target FS. + */ private static long getBlockSize( EnumSet fileAttributes, FileStatus sourceFile, FileSystem targetFS, Path tmpTargetPath) { - return fileAttributes.contains(FileAttribute.BLOCKSIZE)? - sourceFile.getBlockSize() : targetFS.getDefaultBlockSize(tmpTargetPath); + boolean preserve = fileAttributes.contains(FileAttribute.BLOCKSIZE) + || fileAttributes.contains(FileAttribute.CHECKSUMTYPE); + return preserve ? sourceFile.getBlockSize() : targetFS + .getDefaultBlockSize(tmpTargetPath); } /** @@ -261,7 +304,7 @@ private static long getBlockSize( * failures from other kinds of IOExceptions. * The failure to read from source is dealt with specially, in the CopyMapper. * Such failures may be skipped if the DistCpOptions indicate so. - * Write failures are intolerable, and amount to CopyMapper failure. + * Write failures are intolerable, and amount to CopyMapper failure. */ public static class CopyReadException extends IOException { public CopyReadException(Throwable rootCause) { diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java index 24c2c947e66..170e2ed23de 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java @@ -125,7 +125,7 @@ public static Class getStrategy(Configuration conf, * @param sourceRootPath - Source root path * @param childPath - Path for which relative path is required * @return - Relative portion of the child path (always prefixed with / - * unless it is empty + * unless it is empty */ public static String getRelativePath(Path sourceRootPath, Path childPath) { String childPathString = childPath.toUri().getPath(); @@ -277,9 +277,11 @@ public static String getStringDescriptionFor(long nBytes) { * If checksums's can't be retrieved, it doesn't fail the test * Only time the comparison would fail is when checksums are * available and they don't match - * + * * @param sourceFS FileSystem for the source path. * @param source The source path. + * @param sourceChecksum The checksum of the source file. If it is null we + * still need to retrieve it through sourceFS. * @param targetFS FileSystem for the target path. * @param target The target path. * @return If either checksum couldn't be retrieved, the function returns @@ -288,12 +290,12 @@ public static String getStringDescriptionFor(long nBytes) { * @throws IOException if there's an exception while retrieving checksums. */ public static boolean checksumsAreEqual(FileSystem sourceFS, Path source, - FileSystem targetFS, Path target) - throws IOException { - FileChecksum sourceChecksum = null; + FileChecksum sourceChecksum, FileSystem targetFS, Path target) + throws IOException { FileChecksum targetChecksum = null; try { - sourceChecksum = sourceFS.getFileChecksum(source); + sourceChecksum = sourceChecksum != null ? sourceChecksum : sourceFS + .getFileChecksum(source); targetChecksum = targetFS.getFileChecksum(target); } catch (IOException e) { LOG.error("Unable to retrieve checksum for " + source + " or " + target, e); diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestOptionsParser.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestOptionsParser.java index fb6e29212eb..90ae25b2d22 100644 --- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestOptionsParser.java +++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestOptionsParser.java @@ -110,7 +110,7 @@ public void testParsebandwidth() { "hdfs://localhost:8020/target/"}); Assert.assertEquals(options.getMapBandwidth(), 11); } - + @Test(expected=IllegalArgumentException.class) public void testParseNonPositiveBandwidth() { OptionsParser.parse(new String[] { @@ -119,7 +119,7 @@ public void testParseNonPositiveBandwidth() { "hdfs://localhost:8020/source/first", "hdfs://localhost:8020/target/"}); } - + @Test(expected=IllegalArgumentException.class) public void testParseZeroBandwidth() { OptionsParser.parse(new String[] { @@ -397,6 +397,7 @@ public void testPreserve() { Assert.assertFalse(options.shouldPreserve(FileAttribute.PERMISSION)); Assert.assertFalse(options.shouldPreserve(FileAttribute.USER)); Assert.assertFalse(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); options = OptionsParser.parse(new String[] { "-p", @@ -408,6 +409,7 @@ public void testPreserve() { Assert.assertTrue(options.shouldPreserve(FileAttribute.PERMISSION)); Assert.assertTrue(options.shouldPreserve(FileAttribute.USER)); Assert.assertTrue(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); options = OptionsParser.parse(new String[] { "-p", @@ -418,6 +420,7 @@ public void testPreserve() { Assert.assertTrue(options.shouldPreserve(FileAttribute.PERMISSION)); Assert.assertTrue(options.shouldPreserve(FileAttribute.USER)); Assert.assertTrue(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); options = OptionsParser.parse(new String[] { "-pbr", @@ -429,6 +432,7 @@ public void testPreserve() { Assert.assertFalse(options.shouldPreserve(FileAttribute.PERMISSION)); Assert.assertFalse(options.shouldPreserve(FileAttribute.USER)); Assert.assertFalse(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); options = OptionsParser.parse(new String[] { "-pbrgup", @@ -440,6 +444,31 @@ public void testPreserve() { Assert.assertTrue(options.shouldPreserve(FileAttribute.PERMISSION)); Assert.assertTrue(options.shouldPreserve(FileAttribute.USER)); Assert.assertTrue(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); + + options = OptionsParser.parse(new String[] { + "-pbrgupc", + "-f", + "hdfs://localhost:8020/source/first", + "hdfs://localhost:8020/target/"}); + Assert.assertTrue(options.shouldPreserve(FileAttribute.BLOCKSIZE)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.REPLICATION)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.PERMISSION)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.USER)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); + + options = OptionsParser.parse(new String[] { + "-pc", + "-f", + "hdfs://localhost:8020/source/first", + "hdfs://localhost:8020/target/"}); + Assert.assertFalse(options.shouldPreserve(FileAttribute.BLOCKSIZE)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.REPLICATION)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.PERMISSION)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.USER)); + Assert.assertFalse(options.shouldPreserve(FileAttribute.GROUP)); + Assert.assertTrue(options.shouldPreserve(FileAttribute.CHECKSUMTYPE)); options = OptionsParser.parse(new String[] { "-p", @@ -452,7 +481,7 @@ public void testPreserve() { attribIterator.next(); i++; } - Assert.assertEquals(i, 5); + Assert.assertEquals(i, 6); try { OptionsParser.parse(new String[] { diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java index f05743f41cb..9568e6131a8 100644 --- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java +++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java @@ -18,29 +18,6 @@ package org.apache.hadoop.tools.mapred; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileStatus; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.fs.permission.FsAction; -import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.hdfs.MiniDFSCluster; -import org.apache.hadoop.io.IOUtils; -import org.apache.hadoop.io.Text; -import org.apache.hadoop.mapreduce.*; -import org.apache.hadoop.security.AccessControlException; -import org.apache.hadoop.security.UserGroupInformation; -import org.apache.hadoop.tools.DistCpConstants; -import org.apache.hadoop.tools.DistCpOptionSwitch; -import org.apache.hadoop.tools.DistCpOptions; -import org.apache.hadoop.tools.StubContext; -import org.apache.hadoop.tools.util.DistCpUtils; -import org.junit.Assert; -import org.junit.BeforeClass; -import org.junit.Test; - import java.io.DataOutputStream; import java.io.IOException; import java.io.OutputStream; @@ -49,11 +26,38 @@ import java.util.EnumSet; import java.util.List; +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CreateFlag; +import org.apache.hadoop.fs.FileStatus; +import org.apache.hadoop.fs.FileSystem; +import org.apache.hadoop.fs.Options.ChecksumOpt; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.permission.FsAction; +import org.apache.hadoop.fs.permission.FsPermission; +import org.apache.hadoop.hdfs.MiniDFSCluster; +import org.apache.hadoop.io.IOUtils; +import org.apache.hadoop.io.Text; +import org.apache.hadoop.mapreduce.Mapper; +import org.apache.hadoop.security.AccessControlException; +import org.apache.hadoop.security.UserGroupInformation; +import org.apache.hadoop.tools.DistCpConstants; +import org.apache.hadoop.tools.DistCpOptionSwitch; +import org.apache.hadoop.tools.DistCpOptions; +import org.apache.hadoop.tools.StubContext; +import org.apache.hadoop.tools.util.DistCpUtils; +import org.apache.hadoop.util.DataChecksum; +import org.junit.Assert; +import org.junit.BeforeClass; +import org.junit.Test; + public class TestCopyMapper { private static final Log LOG = LogFactory.getLog(TestCopyMapper.class); private static List pathList = new ArrayList(); private static int nFiles = 0; private static final int DEFAULT_FILE_SIZE = 1024; + private static final long NON_DEFAULT_BLOCK_SIZE = 4096; private static MiniDFSCluster cluster; @@ -119,12 +123,27 @@ private static void createSourceDataWithDifferentBlockSize() throws Exception { mkdirs(SOURCE_PATH + "/2/3/4"); mkdirs(SOURCE_PATH + "/2/3"); mkdirs(SOURCE_PATH + "/5"); - touchFile(SOURCE_PATH + "/5/6", true); + touchFile(SOURCE_PATH + "/5/6", true, null); mkdirs(SOURCE_PATH + "/7"); mkdirs(SOURCE_PATH + "/7/8"); touchFile(SOURCE_PATH + "/7/8/9"); } + private static void createSourceDataWithDifferentChecksumType() + throws Exception { + mkdirs(SOURCE_PATH + "/1"); + mkdirs(SOURCE_PATH + "/2"); + mkdirs(SOURCE_PATH + "/2/3/4"); + mkdirs(SOURCE_PATH + "/2/3"); + mkdirs(SOURCE_PATH + "/5"); + touchFile(SOURCE_PATH + "/5/6", new ChecksumOpt(DataChecksum.Type.CRC32, + 512)); + mkdirs(SOURCE_PATH + "/7"); + mkdirs(SOURCE_PATH + "/7/8"); + touchFile(SOURCE_PATH + "/7/8/9", new ChecksumOpt(DataChecksum.Type.CRC32C, + 512)); + } + private static void mkdirs(String path) throws Exception { FileSystem fileSystem = cluster.getFileSystem(); final Path qualifiedPath = new Path(path).makeQualified(fileSystem.getUri(), @@ -134,21 +153,31 @@ private static void mkdirs(String path) throws Exception { } private static void touchFile(String path) throws Exception { - touchFile(path, false); + touchFile(path, false, null); } - private static void touchFile(String path, boolean createMultipleBlocks) throws Exception { - final long NON_DEFAULT_BLOCK_SIZE = 4096; + private static void touchFile(String path, ChecksumOpt checksumOpt) + throws Exception { + // create files with specific checksum opt and non-default block size + touchFile(path, true, checksumOpt); + } + + private static void touchFile(String path, boolean createMultipleBlocks, + ChecksumOpt checksumOpt) throws Exception { FileSystem fs; DataOutputStream outputStream = null; try { fs = cluster.getFileSystem(); final Path qualifiedPath = new Path(path).makeQualified(fs.getUri(), - fs.getWorkingDirectory()); - final long blockSize = createMultipleBlocks? NON_DEFAULT_BLOCK_SIZE : fs.getDefaultBlockSize(qualifiedPath) * 2; - outputStream = fs.create(qualifiedPath, true, 0, - (short)(fs.getDefaultReplication(qualifiedPath)*2), - blockSize); + fs.getWorkingDirectory()); + final long blockSize = createMultipleBlocks ? NON_DEFAULT_BLOCK_SIZE : fs + .getDefaultBlockSize(qualifiedPath) * 2; + FsPermission permission = FsPermission.getFileDefault().applyUMask( + FsPermission.getUMask(fs.getConf())); + outputStream = fs.create(qualifiedPath, permission, + EnumSet.of(CreateFlag.CREATE, CreateFlag.OVERWRITE), 0, + (short) (fs.getDefaultReplication(qualifiedPath) * 2), blockSize, + null, checksumOpt); byte[] bytes = new byte[DEFAULT_FILE_SIZE]; outputStream.write(bytes); long fileSize = DEFAULT_FILE_SIZE; @@ -171,17 +200,40 @@ private static void touchFile(String path, boolean createMultipleBlocks) throws } } + @Test + public void testCopyWithDifferentChecksumType() throws Exception { + testCopy(true); + } + @Test(timeout=40000) public void testRun() { + testCopy(false); + } + + private void testCopy(boolean preserveChecksum) { try { deleteState(); - createSourceData(); + if (preserveChecksum) { + createSourceDataWithDifferentChecksumType(); + } else { + createSourceData(); + } FileSystem fs = cluster.getFileSystem(); CopyMapper copyMapper = new CopyMapper(); StubContext stubContext = new StubContext(getConfiguration(), null, 0); Mapper.Context context = stubContext.getContext(); + + Configuration configuration = context.getConfiguration(); + EnumSet fileAttributes + = EnumSet.of(DistCpOptions.FileAttribute.REPLICATION); + if (preserveChecksum) { + fileAttributes.add(DistCpOptions.FileAttribute.CHECKSUMTYPE); + } + configuration.set(DistCpOptionSwitch.PRESERVE_STATUS.getConfigLabel(), + DistCpUtils.packAttributes(fileAttributes)); + copyMapper.setup(context); for (Path path: pathList) { @@ -195,19 +247,29 @@ public void testRun() { .replaceAll(SOURCE_PATH, TARGET_PATH)); Assert.assertTrue(fs.exists(targetPath)); Assert.assertTrue(fs.isFile(targetPath) == fs.isFile(path)); - Assert.assertEquals(fs.getFileStatus(path).getReplication(), - fs.getFileStatus(targetPath).getReplication()); - Assert.assertEquals(fs.getFileStatus(path).getBlockSize(), - fs.getFileStatus(targetPath).getBlockSize()); - Assert.assertTrue(!fs.isFile(targetPath) || - fs.getFileChecksum(targetPath).equals( - fs.getFileChecksum(path))); + FileStatus sourceStatus = fs.getFileStatus(path); + FileStatus targetStatus = fs.getFileStatus(targetPath); + Assert.assertEquals(sourceStatus.getReplication(), + targetStatus.getReplication()); + if (preserveChecksum) { + Assert.assertEquals(sourceStatus.getBlockSize(), + targetStatus.getBlockSize()); + } + Assert.assertTrue(!fs.isFile(targetPath) + || fs.getFileChecksum(targetPath).equals(fs.getFileChecksum(path))); } Assert.assertEquals(pathList.size(), stubContext.getReporter().getCounter(CopyMapper.Counter.COPY).getValue()); - Assert.assertEquals(nFiles * DEFAULT_FILE_SIZE, - stubContext.getReporter().getCounter(CopyMapper.Counter.BYTESCOPIED).getValue()); + if (!preserveChecksum) { + Assert.assertEquals(nFiles * DEFAULT_FILE_SIZE, stubContext + .getReporter().getCounter(CopyMapper.Counter.BYTESCOPIED) + .getValue()); + } else { + Assert.assertEquals(nFiles * NON_DEFAULT_BLOCK_SIZE * 2, stubContext + .getReporter().getCounter(CopyMapper.Counter.BYTESCOPIED) + .getValue()); + } testCopyingExistingFiles(fs, copyMapper, context); for (Text value : stubContext.getWriter().values()) { @@ -309,7 +371,7 @@ public void testPreserve() { UserGroupInformation tmpUser = UserGroupInformation.createRemoteUser("guest"); final CopyMapper copyMapper = new CopyMapper(); - + final Mapper.Context context = tmpUser. doAs(new PrivilegedAction.Context>() { @Override @@ -535,7 +597,7 @@ public StubContext run() { final Mapper.Context context = stubContext.getContext(); - + context.getConfiguration().set(DistCpConstants.CONF_LABEL_PRESERVE_STATUS, DistCpUtils.packAttributes(preserveStatus)); From 981679e6d7650141fc8737c7e13b16d8b795b408 Mon Sep 17 00:00:00 2001 From: Sanford Ryza Date: Fri, 31 Jan 2014 00:14:47 +0000 Subject: [PATCH 13/23] YARN-1498. Common scheduler changes for moving apps between queues (Sandy Ryza) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563021 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 3 + .../scheduler/AppSchedulingInfo.java | 37 +++- .../resourcemanager/scheduler/Queue.java | 2 + .../scheduler/QueueMetrics.java | 49 +++++- .../SchedulerApplicationAttempt.java | 25 ++- .../scheduler/fair/FSLeafQueue.java | 9 + .../scheduler/fair/FSParentQueue.java | 8 +- .../scheduler/fifo/FifoScheduler.java | 5 + .../scheduler/TestQueueMetrics.java | 12 +- .../TestSchedulerApplicationAttempt.java | 160 ++++++++++++++++++ 10 files changed, 284 insertions(+), 26 deletions(-) create mode 100644 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerApplicationAttempt.java diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index f90d5cb286d..a7f4c706b95 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -9,6 +9,9 @@ Trunk - Unreleased YARN-1496. Protocol additions to allow moving apps between queues (Sandy Ryza) + YARN-1498. Common scheduler changes for moving apps between queues (Sandy + Ryza) + IMPROVEMENTS OPTIMIZATIONS diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java index 3270dea4498..de71f71419d 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java @@ -64,7 +64,7 @@ public class AppSchedulingInfo { private Set blacklist = new HashSet(); //private final ApplicationStore store; - private final ActiveUsersManager activeUsersManager; + private ActiveUsersManager activeUsersManager; /* Allocated by scheduler */ boolean pending = true; // for app metrics @@ -171,11 +171,10 @@ synchronized public void updateResourceRequests( .getNumContainers() : 0; Resource lastRequestCapability = lastRequest != null ? lastRequest .getCapability() : Resources.none(); - metrics.incrPendingResources(user, request.getNumContainers() - - lastRequestContainers, Resources.subtractFrom( // save a clone - Resources.multiply(request.getCapability(), request - .getNumContainers()), Resources.multiply(lastRequestCapability, - lastRequestContainers))); + metrics.incrPendingResources(user, request.getNumContainers(), + request.getCapability()); + metrics.decrPendingResources(user, lastRequestContainers, + lastRequestCapability); } } } @@ -262,6 +261,7 @@ synchronized public void allocate(NodeType type, SchedulerNode node, pending = false; metrics.runAppAttempt(applicationId, user); } + if (LOG.isDebugEnabled()) { LOG.debug("allocate: applicationId=" + applicationId + " container=" + container.getId() @@ -269,7 +269,7 @@ synchronized public void allocate(NodeType type, SchedulerNode node, + " user=" + user + " resource=" + request.getCapability()); } - metrics.allocateResources(user, 1, request.getCapability()); + metrics.allocateResources(user, 1, request.getCapability(), true); } /** @@ -359,6 +359,26 @@ synchronized private void checkForDeactivation() { } } + synchronized public void move(Queue newQueue) { + QueueMetrics oldMetrics = queue.getMetrics(); + QueueMetrics newMetrics = newQueue.getMetrics(); + for (Map asks : requests.values()) { + ResourceRequest request = asks.get(ResourceRequest.ANY); + if (request != null) { + oldMetrics.decrPendingResources(user, request.getNumContainers(), + request.getCapability()); + newMetrics.incrPendingResources(user, request.getNumContainers(), + request.getCapability()); + } + } + oldMetrics.moveAppFrom(this); + newMetrics.moveAppTo(this); + activeUsersManager.deactivateApplication(user, applicationId); + activeUsersManager = newQueue.getActiveUsersManager(); + activeUsersManager.activateApplication(user, applicationId); + this.queue = newQueue; + } + synchronized public void stop(RMAppAttemptState rmAppAttemptFinalState) { // clear pending resources metrics for the application QueueMetrics metrics = queue.getMetrics(); @@ -366,8 +386,7 @@ synchronized public void stop(RMAppAttemptState rmAppAttemptFinalState) { ResourceRequest request = asks.get(ResourceRequest.ANY); if (request != null) { metrics.decrPendingResources(user, request.getNumContainers(), - Resources.multiply(request.getCapability(), request - .getNumContainers())); + request.getCapability()); } } metrics.finishAppAttempt(applicationId, pending, user); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/Queue.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/Queue.java index 0380d323824..c51f819af58 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/Queue.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/Queue.java @@ -58,4 +58,6 @@ public interface Queue { List getQueueUserAclInfo(UserGroupInformation user); boolean hasAccess(QueueACL acl, UserGroupInformation user); + + public ActiveUsersManager getActiveUsersManager(); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/QueueMetrics.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/QueueMetrics.java index 13066141dc5..a0519e2fa01 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/QueueMetrics.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/QueueMetrics.java @@ -280,6 +280,36 @@ public void finishApp(String user, RMAppState rmAppFinalState) { parent.finishApp(user, rmAppFinalState); } } + + public void moveAppFrom(AppSchedulingInfo app) { + if (app.isPending()) { + appsPending.decr(); + } else { + appsRunning.decr(); + } + QueueMetrics userMetrics = getUserMetrics(app.getUser()); + if (userMetrics != null) { + userMetrics.moveAppFrom(app); + } + if (parent != null) { + parent.moveAppFrom(app); + } + } + + public void moveAppTo(AppSchedulingInfo app) { + if (app.isPending()) { + appsPending.incr(); + } else { + appsRunning.incr(); + } + QueueMetrics userMetrics = getUserMetrics(app.getUser()); + if (userMetrics != null) { + userMetrics.moveAppTo(app); + } + if (parent != null) { + parent.moveAppTo(app); + } + } /** * Set available resources. To be called by scheduler periodically as @@ -324,8 +354,8 @@ public void incrPendingResources(String user, int containers, Resource res) { private void _incrPendingResources(int containers, Resource res) { pendingContainers.incr(containers); - pendingMB.incr(res.getMemory()); - pendingVCores.incr(res.getVirtualCores()); + pendingMB.incr(res.getMemory() * containers); + pendingVCores.incr(res.getVirtualCores() * containers); } public void decrPendingResources(String user, int containers, Resource res) { @@ -341,22 +371,25 @@ public void decrPendingResources(String user, int containers, Resource res) { private void _decrPendingResources(int containers, Resource res) { pendingContainers.decr(containers); - pendingMB.decr(res.getMemory()); - pendingVCores.decr(res.getVirtualCores()); + pendingMB.decr(res.getMemory() * containers); + pendingVCores.decr(res.getVirtualCores() * containers); } - public void allocateResources(String user, int containers, Resource res) { + public void allocateResources(String user, int containers, Resource res, + boolean decrPending) { allocatedContainers.incr(containers); aggregateContainersAllocated.incr(containers); allocatedMB.incr(res.getMemory() * containers); allocatedVCores.incr(res.getVirtualCores() * containers); - _decrPendingResources(containers, Resources.multiply(res, containers)); + if (decrPending) { + _decrPendingResources(containers, res); + } QueueMetrics userMetrics = getUserMetrics(user); if (userMetrics != null) { - userMetrics.allocateResources(user, containers, res); + userMetrics.allocateResources(user, containers, res, decrPending); } if (parent != null) { - parent.allocateResources(user, containers, res); + parent.allocateResources(user, containers, res, decrPending); } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApplicationAttempt.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApplicationAttempt.java index 38753cbbeda..7e7fdb765e1 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApplicationAttempt.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApplicationAttempt.java @@ -57,7 +57,7 @@ */ @Private @Unstable -public abstract class SchedulerApplicationAttempt { +public class SchedulerApplicationAttempt { private static final Log LOG = LogFactory .getLog(SchedulerApplicationAttempt.class); @@ -91,7 +91,7 @@ public abstract class SchedulerApplicationAttempt { protected Map lastScheduledContainer = new HashMap(); - protected final Queue queue; + protected Queue queue; protected boolean isStopped = false; protected final RMContext rmContext; @@ -431,4 +431,25 @@ public synchronized void transferStateFromPreviousAttempt( this.appSchedulingInfo .transferStateFromPreviousAppSchedulingInfo(appAttempt.appSchedulingInfo); } + + public void move(Queue newQueue) { + QueueMetrics oldMetrics = queue.getMetrics(); + QueueMetrics newMetrics = newQueue.getMetrics(); + String user = getUser(); + for (RMContainer liveContainer : liveContainers.values()) { + Resource resource = liveContainer.getContainer().getResource(); + oldMetrics.releaseResources(user, 1, resource); + newMetrics.allocateResources(user, 1, resource, false); + } + for (Map map : reservedContainers.values()) { + for (RMContainer reservedContainer : map.values()) { + Resource resource = reservedContainer.getReservedResource(); + oldMetrics.unreserveResource(user, resource); + newMetrics.reserveResource(user, resource); + } + } + + appSchedulingInfo.move(newQueue); + this.queue = newQueue; + } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSLeafQueue.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSLeafQueue.java index 17d585ab33f..e842a6a3557 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSLeafQueue.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSLeafQueue.java @@ -33,6 +33,7 @@ import org.apache.hadoop.yarn.api.records.QueueACL; import org.apache.hadoop.yarn.api.records.QueueUserACLInfo; import org.apache.hadoop.yarn.api.records.Resource; +import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerAppUtils; import org.apache.hadoop.yarn.util.resource.Resources; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplication; @@ -54,11 +55,14 @@ public class FSLeafQueue extends FSQueue { private long lastTimeAtMinShare; private long lastTimeAtHalfFairShare; + private final ActiveUsersManager activeUsersManager; + public FSLeafQueue(String name, FairScheduler scheduler, FSParentQueue parent) { super(name, scheduler, parent); this.lastTimeAtMinShare = scheduler.getClock().getTime(); this.lastTimeAtHalfFairShare = scheduler.getClock().getTime(); + activeUsersManager = new ActiveUsersManager(getMetrics()); } public void addApp(FSSchedulerApp app, boolean runnable) { @@ -245,4 +249,9 @@ public void setLastTimeAtHalfFairShare(long lastTimeAtHalfFairShare) { public int getNumRunnableApps() { return runnableAppScheds.size(); } + + @Override + public ActiveUsersManager getActiveUsersManager() { + return activeUsersManager; + } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSParentQueue.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSParentQueue.java index 070780a97d4..427cb864579 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSParentQueue.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSParentQueue.java @@ -33,7 +33,7 @@ import org.apache.hadoop.yarn.api.records.QueueUserACLInfo; import org.apache.hadoop.yarn.api.records.Resource; import org.apache.hadoop.yarn.util.resource.Resources; -import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt; +import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ActiveUsersManager; @Private @Unstable @@ -194,4 +194,10 @@ public void collectSchedulerApplications( childQueue.collectSchedulerApplications(apps); } } + + @Override + public ActiveUsersManager getActiveUsersManager() { + // Should never be called since all applications are submitted to LeafQueues + return null; + } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java index 696a64c3c02..026f22c49c3 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java @@ -184,6 +184,11 @@ public List getQueueUserAclInfo( public boolean hasAccess(QueueACL acl, UserGroupInformation user) { return getQueueAcls().get(acl).isUserAllowed(user); } + + @Override + public ActiveUsersManager getActiveUsersManager() { + return activeUsersManager; + } }; @Override diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestQueueMetrics.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestQueueMetrics.java index d0a8f7235fd..8ad71d231bc 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestQueueMetrics.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestQueueMetrics.java @@ -73,7 +73,7 @@ public void setUp() { checkApps(queueSource, 1, 1, 0, 0, 0, 0, true); metrics.setAvailableResourcesToQueue(Resources.createResource(100*GB, 100)); - metrics.incrPendingResources(user, 5, Resources.createResource(15*GB, 15)); + metrics.incrPendingResources(user, 5, Resources.createResource(3*GB, 3)); // Available resources is set externally, as it depends on dynamic // configurable cluster/queue resources checkResources(queueSource, 0, 0, 0, 0, 0, 100*GB, 100, 15*GB, 15, 5, 0, 0, 0); @@ -81,7 +81,7 @@ public void setUp() { metrics.runAppAttempt(app.getApplicationId(), user); checkApps(queueSource, 1, 0, 1, 0, 0, 0, true); - metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2)); + metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2), true); checkResources(queueSource, 6*GB, 6, 3, 3, 0, 100*GB, 100, 9*GB, 9, 2, 0, 0, 0); metrics.releaseResources(user, 1, Resources.createResource(2*GB, 2)); @@ -171,7 +171,7 @@ public void testQueueAppMetricsForMultipleFailures() { metrics.setAvailableResourcesToQueue(Resources.createResource(100*GB, 100)); metrics.setAvailableResourcesToUser(user, Resources.createResource(10*GB, 10)); - metrics.incrPendingResources(user, 5, Resources.createResource(15*GB, 15)); + metrics.incrPendingResources(user, 5, Resources.createResource(3*GB, 3)); // Available resources is set externally, as it depends on dynamic // configurable cluster/queue resources checkResources(queueSource, 0, 0, 0, 0, 0, 100*GB, 100, 15*GB, 15, 5, 0, 0, 0); @@ -181,7 +181,7 @@ public void testQueueAppMetricsForMultipleFailures() { checkApps(queueSource, 1, 0, 1, 0, 0, 0, true); checkApps(userSource, 1, 0, 1, 0, 0, 0, true); - metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2)); + metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2), true); checkResources(queueSource, 6*GB, 6, 3, 3, 0, 100*GB, 100, 9*GB, 9, 2, 0, 0, 0); checkResources(userSource, 6*GB, 6, 3, 3, 0, 10*GB, 10, 9*GB, 9, 2, 0, 0, 0); @@ -232,7 +232,7 @@ public void testQueueAppMetricsForMultipleFailures() { metrics.setAvailableResourcesToQueue(Resources.createResource(100*GB, 100)); parentMetrics.setAvailableResourcesToUser(user, Resources.createResource(10*GB, 10)); metrics.setAvailableResourcesToUser(user, Resources.createResource(10*GB, 10)); - metrics.incrPendingResources(user, 5, Resources.createResource(15*GB, 15)); + metrics.incrPendingResources(user, 5, Resources.createResource(3*GB, 3)); checkResources(queueSource, 0, 0, 0, 0, 0, 100*GB, 100, 15*GB, 15, 5, 0, 0, 0); checkResources(parentQueueSource, 0, 0, 0, 0, 0, 100*GB, 100, 15*GB, 15, 5, 0, 0, 0); checkResources(userSource, 0, 0, 0, 0, 0, 10*GB, 10, 15*GB, 15, 5, 0, 0, 0); @@ -242,7 +242,7 @@ public void testQueueAppMetricsForMultipleFailures() { checkApps(queueSource, 1, 0, 1, 0, 0, 0, true); checkApps(userSource, 1, 0, 1, 0, 0, 0, true); - metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2)); + metrics.allocateResources(user, 3, Resources.createResource(2*GB, 2), true); metrics.reserveResource(user, Resources.createResource(3*GB, 3)); // Available resources is set externally, as it depends on dynamic // configurable cluster/queue resources diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerApplicationAttempt.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerApplicationAttempt.java new file mode 100644 index 00000000000..93fd30027e9 --- /dev/null +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerApplicationAttempt.java @@ -0,0 +1,160 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.yarn.server.resourcemanager.scheduler; + +import static org.junit.Assert.assertEquals; +import static org.mockito.Mockito.*; + +import java.util.Arrays; +import java.util.HashMap; +import java.util.Map; + +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem; +import org.apache.hadoop.yarn.api.records.ApplicationAttemptId; +import org.apache.hadoop.yarn.api.records.ApplicationId; +import org.apache.hadoop.yarn.api.records.Container; +import org.apache.hadoop.yarn.api.records.ContainerId; +import org.apache.hadoop.yarn.api.records.NodeId; +import org.apache.hadoop.yarn.api.records.Priority; +import org.apache.hadoop.yarn.api.records.Resource; +import org.apache.hadoop.yarn.api.records.ResourceRequest; +import org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainer; +import org.junit.After; +import org.junit.Test; + +public class TestSchedulerApplicationAttempt { + + private static final NodeId nodeId = NodeId.newInstance("somehost", 5); + + private Configuration conf = new Configuration(); + + @After + public void tearDown() { + QueueMetrics.clearQueueMetrics(); + DefaultMetricsSystem.shutdown(); + } + + @Test + public void testMove() { + final String user = "user1"; + Queue parentQueue = createQueue("parent", null); + Queue oldQueue = createQueue("old", parentQueue); + Queue newQueue = createQueue("new", parentQueue); + QueueMetrics parentMetrics = parentQueue.getMetrics(); + QueueMetrics oldMetrics = oldQueue.getMetrics(); + QueueMetrics newMetrics = newQueue.getMetrics(); + + ApplicationAttemptId appAttId = createAppAttemptId(0, 0); + SchedulerApplicationAttempt app = new SchedulerApplicationAttempt(appAttId, + user, oldQueue, oldQueue.getActiveUsersManager(), null); + oldMetrics.submitApp(user); + + // Resource request + Resource requestedResource = Resource.newInstance(1536, 2); + Priority requestedPriority = Priority.newInstance(2); + ResourceRequest request = ResourceRequest.newInstance(requestedPriority, + ResourceRequest.ANY, requestedResource, 3); + app.updateResourceRequests(Arrays.asList(request)); + + // Allocated container + RMContainer container1 = createRMContainer(appAttId, 1, requestedResource); + app.liveContainers.put(container1.getContainerId(), container1); + SchedulerNode node = createNode(); + app.appSchedulingInfo.allocate(NodeType.OFF_SWITCH, node, requestedPriority, + request, container1.getContainer()); + + // Reserved container + Priority prio1 = Priority.newInstance(1); + Resource reservedResource = Resource.newInstance(2048, 3); + RMContainer container2 = createReservedRMContainer(appAttId, 1, reservedResource, + node.getNodeID(), prio1); + Map reservations = new HashMap(); + reservations.put(node.getNodeID(), container2); + app.reservedContainers.put(prio1, reservations); + oldMetrics.reserveResource(user, reservedResource); + + checkQueueMetrics(oldMetrics, 1, 1, 1536, 2, 2048, 3, 3072, 4); + checkQueueMetrics(newMetrics, 0, 0, 0, 0, 0, 0, 0, 0); + checkQueueMetrics(parentMetrics, 1, 1, 1536, 2, 2048, 3, 3072, 4); + + app.move(newQueue); + + checkQueueMetrics(oldMetrics, 0, 0, 0, 0, 0, 0, 0, 0); + checkQueueMetrics(newMetrics, 1, 1, 1536, 2, 2048, 3, 3072, 4); + checkQueueMetrics(parentMetrics, 1, 1, 1536, 2, 2048, 3, 3072, 4); + } + + private void checkQueueMetrics(QueueMetrics metrics, int activeApps, + int runningApps, int allocMb, int allocVcores, int reservedMb, + int reservedVcores, int pendingMb, int pendingVcores) { + assertEquals(activeApps, metrics.getActiveApps()); + assertEquals(runningApps, metrics.getAppsRunning()); + assertEquals(allocMb, metrics.getAllocatedMB()); + assertEquals(allocVcores, metrics.getAllocatedVirtualCores()); + assertEquals(reservedMb, metrics.getReservedMB()); + assertEquals(reservedVcores, metrics.getReservedVirtualCores()); + assertEquals(pendingMb, metrics.getPendingMB()); + assertEquals(pendingVcores, metrics.getPendingVirtualCores()); + } + + private SchedulerNode createNode() { + SchedulerNode node = mock(SchedulerNode.class); + when(node.getNodeName()).thenReturn("somehost"); + when(node.getRackName()).thenReturn("somerack"); + when(node.getNodeID()).thenReturn(nodeId); + return node; + } + + private RMContainer createReservedRMContainer(ApplicationAttemptId appAttId, + int id, Resource resource, NodeId nodeId, Priority reservedPriority) { + RMContainer container = createRMContainer(appAttId, id, resource); + when(container.getReservedResource()).thenReturn(resource); + when(container.getReservedPriority()).thenReturn(reservedPriority); + when(container.getReservedNode()).thenReturn(nodeId); + return container; + } + + private RMContainer createRMContainer(ApplicationAttemptId appAttId, int id, + Resource resource) { + ContainerId containerId = ContainerId.newInstance(appAttId, id); + RMContainer rmContainer = mock(RMContainer.class); + Container container = mock(Container.class); + when(container.getResource()).thenReturn(resource); + when(container.getNodeId()).thenReturn(nodeId); + when(rmContainer.getContainer()).thenReturn(container); + when(rmContainer.getContainerId()).thenReturn(containerId); + return rmContainer; + } + + private Queue createQueue(String name, Queue parent) { + QueueMetrics metrics = QueueMetrics.forQueue(name, parent, false, conf); + ActiveUsersManager activeUsersManager = new ActiveUsersManager(metrics); + Queue queue = mock(Queue.class); + when(queue.getMetrics()).thenReturn(metrics); + when(queue.getActiveUsersManager()).thenReturn(activeUsersManager); + return queue; + } + + private ApplicationAttemptId createAppAttemptId(int appId, int attemptId) { + ApplicationId appIdImpl = ApplicationId.newInstance(0, appId); + ApplicationAttemptId attId = + ApplicationAttemptId.newInstance(appIdImpl, attemptId); + return attId; + } +} From f9699ce09e546e38633d6c971f0b537fc40f881a Mon Sep 17 00:00:00 2001 From: Andrew Wang Date: Fri, 31 Jan 2014 02:10:28 +0000 Subject: [PATCH 14/23] HADOOP-10317. Rename branch-2.3 release version from 2.4.0-SNAPSHOT to 2.3.0-SNAPSHOT. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563035 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-common-project/hadoop-common/CHANGES.txt | 3 +++ 1 file changed, 3 insertions(+) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 54011a9dfed..792cea3d9bb 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -467,6 +467,9 @@ Release 2.3.0 - UNRELEASED "rpc.metrics.percentiles.intervals" to core-default.xml. (Akira Ajisaka via wang) + HADOOP-10317. Rename branch-2.3 release version from 2.4.0-SNAPSHOT + to 2.3.0-SNAPSHOT. (wang) + OPTIMIZATIONS HADOOP-10142. Avoid groups lookup for unprivileged users such as "dr.who" From edad2f62bbc1160e72d9fa6c2bde11c3498e14a4 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Fri, 31 Jan 2014 03:56:20 +0000 Subject: [PATCH 15/23] HADOOP-10313. Script and jenkins job to produce Hadoop release artifacts. (tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563043 13f79535-47bb-0310-9956-ffa450edef68 --- dev-support/create-release.sh | 124 ++++++++++++++++++ .../hadoop-common/CHANGES.txt | 3 + 2 files changed, 127 insertions(+) create mode 100755 dev-support/create-release.sh diff --git a/dev-support/create-release.sh b/dev-support/create-release.sh new file mode 100755 index 00000000000..3976a789ac4 --- /dev/null +++ b/dev-support/create-release.sh @@ -0,0 +1,124 @@ +#!/bin/bash +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + +# Function to probe the exit code of the script commands, +# and stop in the case of failure with an contextual error +# message. +run() { + echo "\$ ${@}" + "${@}" + exitCode=$? + if [[ $exitCode != 0 ]]; then + echo + echo "Failed! running ${@} in `pwd`" + echo + exit $exitCode + fi +} + +doMD5() { + MD5CMD="md5sum" + which $MD5CMD + if [[ $? != 0 ]]; then + MD5CMD="md5" + fi + run $MD5CMD ${1} > ${1}.md5 +} + +# If provided, the created release artifacts will be tagged with it +# (use RC#, i.e: RC0). Do not use a label to create the final release +# artifact. +RC_LABEL=$1 + +# Extract Hadoop version from POM +HADOOP_VERSION=`cat pom.xml | grep "" | head -1 | sed 's|^ *||' | sed 's|.*$||'` + +echo +echo "*****************************************************************" +echo +echo "Hadoop version to create release artifacts: ${HADOOP_VERSION}" +echo +echo "Release Candidate Label: ${RC_LABEL}" +echo +echo "*****************************************************************" +echo + +if [[ ! -z ${RC_LABEL} ]]; then + RC_LABEL="-${RC_LABEL}" +fi + +# Get Maven command +if [ -z "$MAVEN_HOME" ]; then + MVN=mvn +else + MVN=$MAVEN_HOME/bin/mvn +fi + +ARTIFACTS_DIR="target/artifacts" + +# Create staging dir for release artifacts + +run mkdir -p ${ARTIFACTS_DIR} + +# Create RAT report +run ${MVN} apache-rat:check + +# Create SRC and BIN tarballs for release, +# Using 'install’ goal instead of 'package' so artifacts are available +# in the Maven local cache for the site generation +run ${MVN} install -Pdist,docs,src,native -DskipTests -Dtar + +# Create site for release +run ${MVN} site site:stage -Pdist -Psrc +run mv target/staging/hadoop-project target/r${HADOOP_VERSION}/ +run cd target/ +run tar czf hadoop-site-${HADOOP_VERSION}.tar.gz r${HADOOP_VERSION}/* +run cd .. + +# Stage RAT report +find . -name rat.txt | xargs -I% cat % > ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-rat.txt + +# Stage CHANGES.txt files +run cp ./hadoop-common-project/hadoop-common/CHANGES.txt ${ARTIFACTS_DIR}/CHANGES-COMMON-${HADOOP_VERSION}${RC_LABEL}.txt +run cp ./hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt ${ARTIFACTS_DIR}/CHANGES-HDFS--${HADOOP_VERSION}${RC_LABEL}.txt +run cp ./hadoop-mapreduce-project/CHANGES.txt ${ARTIFACTS_DIR}/CHANGES-MAPREDUCE-${HADOOP_VERSION}${RC_LABEL}.txt +run cp ./hadoop-yarn-project/CHANGES.txt ${ARTIFACTS_DIR}/CHANGES-YARN-${HADOOP_VERSION}${RC_LABEL}.txt + +# Stage BIN tarball +run mv hadoop-dist/target/hadoop-${HADOOP_VERSION}.tar.gz ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}.tar.gz + +# State SRC tarball +run mv hadoop-dist/target/hadoop-${HADOOP_VERSION}-src.tar.gz ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-src.tar.gz + +# Stage SITE tarball +run mv target/hadoop-site-${HADOOP_VERSION}.tar.gz ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-site.tar.gz + +# MD5 SRC and BIN tarballs +doMD5 ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}.tar.gz +doMD5 ${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-src.tar.gz + +run cd ${ARTIFACTS_DIR} +ARTIFACTS_DIR=`pwd` +echo +echo "Congratulations, you have successfully built the release" +echo "artifacts for Apache Hadoop ${HADOOP_VERSION}${RC_LABEL}" +echo +echo "The artifacts for this run are available at ${ARTIFACTS_DIR}:" +run ls -1 ${ARTIFACTS_DIR} +echo +echo "Remember to sign them before staging them on the open" +echo diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 792cea3d9bb..3a79dd4100c 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -470,6 +470,9 @@ Release 2.3.0 - UNRELEASED HADOOP-10317. Rename branch-2.3 release version from 2.4.0-SNAPSHOT to 2.3.0-SNAPSHOT. (wang) + HADOOP-10313. Script and jenkins job to produce Hadoop release artifacts. + (tucu) + OPTIMIZATIONS HADOOP-10142. Avoid groups lookup for unprivileged users such as "dr.who" From ed097d24fe789844079a650eba05587e2973b158 Mon Sep 17 00:00:00 2001 From: Tsz-wo Sze Date: Fri, 31 Jan 2014 05:33:51 +0000 Subject: [PATCH 16/23] Add .classpath, .project, .settings and target to svn:ignore. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563063 13f79535-47bb-0310-9956-ffa450edef68 From 7039b776c64cd0b1c444d27ba2ae118b5a812ab2 Mon Sep 17 00:00:00 2001 From: Suresh Srinivas Date: Fri, 31 Jan 2014 05:35:30 +0000 Subject: [PATCH 17/23] HDFS-5856. DataNode.checkDiskError might throw NPE. Contributed by Josh Elser. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563064 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 ++ .../hadoop/hdfs/server/datanode/DataNode.java | 31 ++++++++++++++----- .../hdfs/server/datanode/TestDiskError.java | 16 ++++++++++ 3 files changed, 43 insertions(+), 7 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 4c6cf2e3634..1fb5c1c190c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -310,6 +310,9 @@ Release 2.4.0 - UNRELEASED HDFS-5843. DFSClient.getFileChecksum() throws IOException if checksum is disabled. (Laurent Goujon via jing9) + HDFS-5856. DataNode.checkDiskError might throw NPE. + (Josh Elser via suresh) + Release 2.3.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java index 0e0d79b28bd..6bf98b69465 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java @@ -36,6 +36,7 @@ import java.net.URI; import java.net.UnknownHostException; import java.nio.channels.ClosedByInterruptException; +import java.nio.channels.ClosedChannelException; import java.nio.channels.SocketChannel; import java.security.PrivilegedExceptionAction; import java.util.ArrayList; @@ -51,7 +52,6 @@ import javax.management.ObjectName; - import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; @@ -1324,12 +1324,7 @@ public void shutdown() { protected void checkDiskError(Exception e ) throws IOException { LOG.warn("checkDiskError: exception: ", e); - if (e instanceof SocketException || e instanceof SocketTimeoutException - || e instanceof ClosedByInterruptException - || e.getMessage().startsWith("An established connection was aborted") - || e.getMessage().startsWith("Broken pipe") - || e.getMessage().startsWith("Connection reset") - || e.getMessage().contains("java.nio.channels.SocketChannel")) { + if (isNetworkRelatedException(e)) { LOG.info("Not checking disk as checkDiskError was called on a network" + " related exception"); return; @@ -1342,6 +1337,28 @@ protected void checkDiskError(Exception e ) throws IOException { } } + /** + * Check if the provided exception looks like it's from a network error + * @param e the exception from a checkDiskError call + * @return true if this exception is network related, false otherwise + */ + protected boolean isNetworkRelatedException(Exception e) { + if (e instanceof SocketException + || e instanceof SocketTimeoutException + || e instanceof ClosedChannelException + || e instanceof ClosedByInterruptException) { + return true; + } + + String msg = e.getMessage(); + + return null != msg + && (msg.startsWith("An established connection was aborted") + || msg.startsWith("Broken pipe") + || msg.startsWith("Connection reset") + || msg.contains("java.nio.channels.SocketChannel")); + } + /** * Check if there is a disk failure and if so, handle the error */ diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDiskError.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDiskError.java index 6b70cbfc599..e36005bafc6 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDiskError.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDiskError.java @@ -18,12 +18,16 @@ package org.apache.hadoop.hdfs.server.datanode; import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import java.io.DataOutputStream; import java.io.File; import java.net.InetSocketAddress; import java.net.Socket; +import java.net.SocketException; +import java.net.SocketTimeoutException; +import java.nio.channels.ClosedChannelException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; @@ -196,4 +200,16 @@ public void testLocalDirs() throws Exception { } } } + + @Test + public void testNetworkErrorsIgnored() { + DataNode dn = cluster.getDataNodes().iterator().next(); + + assertTrue(dn.isNetworkRelatedException(new SocketException())); + assertTrue(dn.isNetworkRelatedException(new SocketTimeoutException())); + assertTrue(dn.isNetworkRelatedException(new ClosedChannelException())); + assertTrue(dn.isNetworkRelatedException(new Exception("Broken pipe foo bar"))); + assertFalse(dn.isNetworkRelatedException(new Exception())); + assertFalse(dn.isNetworkRelatedException(new Exception("random problem"))); + } } From 5d4bd6a15a606486c02fc58d23ea0ca74ba453c9 Mon Sep 17 00:00:00 2001 From: Jonathan Turner Eagles Date: Fri, 31 Jan 2014 16:00:39 +0000 Subject: [PATCH 18/23] YARN-1632. TestApplicationMasterServices should be under org.apache.hadoop.yarn.server.resourcemanager package (Chen He via jeagles) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563157 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-yarn-project/CHANGES.txt | 3 ++ .../TestApplicationMasterService.java | 38 +------------------ 2 files changed, 4 insertions(+), 37 deletions(-) rename hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/{applicationmasterservice => }/TestApplicationMasterService.java (78%) diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index a7f4c706b95..914966a28d4 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -140,6 +140,9 @@ Release 2.4.0 - UNRELEASED YARN-1642. RMDTRenewer#getRMClient should use ClientRMProxy (kasha) + YARN-1632. TestApplicationMasterServices should be under + org.apache.hadoop.yarn.server.resourcemanager package (Chen He via jeagles) + Release 2.3.0 - UNRELEASED INCOMPATIBLE CHANGES diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationMasterService.java similarity index 78% rename from hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java rename to hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationMasterService.java index 6c12141dc4c..4025ab76b52 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationmasterservice/TestApplicationMasterService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationMasterService.java @@ -16,28 +16,21 @@ * limitations under the License. */ -package org.apache.hadoop.yarn.server.resourcemanager.applicationmasterservice; +package org.apache.hadoop.yarn.server.resourcemanager; import junit.framework.Assert; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse; -import org.apache.hadoop.yarn.api.protocolrecords.RegisterApplicationMasterResponse; import org.apache.hadoop.yarn.api.records.Container; import org.apache.hadoop.yarn.api.records.ContainerId; import org.apache.hadoop.yarn.conf.YarnConfiguration; import org.apache.hadoop.yarn.exceptions.InvalidContainerReleaseException; -import org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException; import org.apache.hadoop.yarn.security.ContainerTokenIdentifier; -import org.apache.hadoop.yarn.server.resourcemanager.MockAM; -import org.apache.hadoop.yarn.server.resourcemanager.MockNM; -import org.apache.hadoop.yarn.server.resourcemanager.MockRM; -import org.apache.hadoop.yarn.server.resourcemanager.TestFifoScheduler; import org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMApp; import org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttempt; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ResourceScheduler; -import org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler; import org.apache.hadoop.yarn.server.utils.BuilderUtils; import org.junit.BeforeClass; @@ -152,33 +145,4 @@ public void testInvalidContainerReleaseRequest() throws Exception { } } } - - @Test (timeout = 60000) - public void testNotifyAMOfPlacedQueue() throws Exception { - // By default, FairScheduler assigns queue by user name - conf.setClass(YarnConfiguration.RM_SCHEDULER, FairScheduler.class, - ResourceScheduler.class); - MockRM rm = new MockRM(conf); - try { - rm.start(); - - // Register node1 - MockNM nm1 = rm.registerNode("127.0.0.1:1234", 6 * GB); - - // Submit an application - RMApp app1 = rm.submitApp(1024, "somename", "user1"); - - // kick the scheduling - nm1.nodeHeartbeat(true); - RMAppAttempt attempt1 = app1.getCurrentAppAttempt(); - MockAM am1 = rm.sendAMLaunched(attempt1.getAppAttemptId()); - - RegisterApplicationMasterResponse response = am1.registerAppAttempt(); - Assert.assertEquals("root.user1", response.getQueue()); - } finally { - if (rm != null) { - rm.stop(); - } - } - } } From 1e449178887b1f4f1ef62bca2720fe5e89b507f3 Mon Sep 17 00:00:00 2001 From: Jonathan Turner Eagles Date: Fri, 31 Jan 2014 16:14:31 +0000 Subject: [PATCH 19/23] YARN-1632. TestApplicationMasterServices should be under org.apache.hadoop.yarn.server.resourcemanager package (Chen He via jeagles) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563160 13f79535-47bb-0310-9956-ffa450edef68 From 22c7344e822fc0b9c41bbdf01971e281ce4e2a13 Mon Sep 17 00:00:00 2001 From: Chris Nauroth Date: Fri, 31 Jan 2014 20:05:53 +0000 Subject: [PATCH 20/23] =?UTF-8?q?HADOOP-10320.=20Javadoc=20in=20InterfaceS?= =?UTF-8?q?tability.java=20lacks=20final=20.=20Contributed=20by=20Ren?= =?UTF-8?q?=C3=A9=20Nyffenegger.?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563237 13f79535-47bb-0310-9956-ffa450edef68 --- .../org/apache/hadoop/classification/InterfaceStability.java | 1 + hadoop-common-project/hadoop-common/CHANGES.txt | 3 +++ 2 files changed, 4 insertions(+) diff --git a/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/InterfaceStability.java b/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/InterfaceStability.java index f78acc296cc..697327e77ec 100644 --- a/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/InterfaceStability.java +++ b/hadoop-common-project/hadoop-annotations/src/main/java/org/apache/hadoop/classification/InterfaceStability.java @@ -33,6 +33,7 @@ *
  • Classes that are {@link Private} are to be considered unstable unless * a different InterfaceStability annotation states otherwise.
  • *
  • Incompatible changes must not be made to classes marked as stable.
  • + * */ @InterfaceAudience.Public @InterfaceStability.Evolving diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 3a79dd4100c..0fc903acd13 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -311,6 +311,9 @@ Release 2.4.0 - UNRELEASED BUG FIXES + HADOOP-10320. Javadoc in InterfaceStability.java lacks final . + (René Nyffenegger via cnauroth) + Release 2.3.0 - UNRELEASED INCOMPATIBLE CHANGES From 5a2428d39fd382866c2c3377e012d6dc49572319 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Fri, 31 Jan 2014 20:11:29 +0000 Subject: [PATCH 21/23] HADOOP-10311. Cleanup vendor names from the code base. (tucu) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563239 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop-common/CHANGES.txt | 2 + .../src/main/data/2jobs2min-rumen-jh.json | 820 +++++++++--------- 2 files changed, 412 insertions(+), 410 deletions(-) diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 0fc903acd13..c858d7d4266 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -683,6 +683,8 @@ Release 2.3.0 - UNRELEASED HADOOP-10310. SaslRpcServer should be initialized even when no secret manager present. (atm) + HADOOP-10311. Cleanup vendor names from the code base. (tucu) + Release 2.2.0 - 2013-10-13 INCOMPATIBLE CHANGES diff --git a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json index 41d4c53224f..230e4fd8245 100644 --- a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json +++ b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json @@ -13,9 +13,9 @@ "finishTime" : 1371222078206, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061163, "result" : "SUCCESS", "finishTime" : 1371222078206, @@ -60,9 +60,9 @@ "finishTime" : 1371222076092, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061164, "result" : "SUCCESS", "finishTime" : 1371222076092, @@ -107,9 +107,9 @@ "finishTime" : 1371222076916, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061159, "result" : "SUCCESS", "finishTime" : 1371222076916, @@ -154,9 +154,9 @@ "finishTime" : 1371222078156, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061160, "result" : "SUCCESS", "finishTime" : 1371222078156, @@ -201,9 +201,9 @@ "finishTime" : 1371222077984, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061161, "result" : "SUCCESS", "finishTime" : 1371222077984, @@ -248,9 +248,9 @@ "finishTime" : 1371222078263, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222061161, "result" : "SUCCESS", "finishTime" : 1371222078263, @@ -295,9 +295,9 @@ "finishTime" : 1371222080381, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061165, "result" : "SUCCESS", "finishTime" : 1371222080381, @@ -342,9 +342,9 @@ "finishTime" : 1371222080174, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061168, "result" : "SUCCESS", "finishTime" : 1371222080174, @@ -389,9 +389,9 @@ "finishTime" : 1371222080738, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061150, "result" : "SUCCESS", "finishTime" : 1371222080738, @@ -436,9 +436,9 @@ "finishTime" : 1371222080294, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061167, "result" : "SUCCESS", "finishTime" : 1371222080294, @@ -483,9 +483,9 @@ "finishTime" : 1371222079719, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061175, "result" : "SUCCESS", "finishTime" : 1371222079719, @@ -530,9 +530,9 @@ "finishTime" : 1371222080741, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061174, "result" : "SUCCESS", "finishTime" : 1371222080741, @@ -577,9 +577,9 @@ "finishTime" : 1371222080631, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061173, "result" : "SUCCESS", "finishTime" : 1371222080631, @@ -624,9 +624,9 @@ "finishTime" : 1371222080047, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222061177, "result" : "SUCCESS", "finishTime" : 1371222080047, @@ -671,9 +671,9 @@ "finishTime" : 1371222080403, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061182, "result" : "SUCCESS", "finishTime" : 1371222080403, @@ -718,9 +718,9 @@ "finishTime" : 1371222080385, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061182, "result" : "SUCCESS", "finishTime" : 1371222080385, @@ -765,9 +765,9 @@ "finishTime" : 1371222080314, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061183, "result" : "SUCCESS", "finishTime" : 1371222080314, @@ -812,9 +812,9 @@ "finishTime" : 1371222079863, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061183, "result" : "SUCCESS", "finishTime" : 1371222079863, @@ -859,9 +859,9 @@ "finishTime" : 1371222080024, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061180, "result" : "SUCCESS", "finishTime" : 1371222080024, @@ -906,9 +906,9 @@ "finishTime" : 1371222081077, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061181, "result" : "SUCCESS", "finishTime" : 1371222081077, @@ -953,9 +953,9 @@ "finishTime" : 1371222080321, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061184, "result" : "SUCCESS", "finishTime" : 1371222080321, @@ -1000,9 +1000,9 @@ "finishTime" : 1371222081794, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222061178, "result" : "SUCCESS", "finishTime" : 1371222081794, @@ -1047,9 +1047,9 @@ "finishTime" : 1371222080729, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061185, "result" : "SUCCESS", "finishTime" : 1371222080729, @@ -1094,9 +1094,9 @@ "finishTime" : 1371222080052, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061186, "result" : "SUCCESS", "finishTime" : 1371222080052, @@ -1141,9 +1141,9 @@ "finishTime" : 1371222080302, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061188, "result" : "SUCCESS", "finishTime" : 1371222080302, @@ -1188,9 +1188,9 @@ "finishTime" : 1371222080632, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061188, "result" : "SUCCESS", "finishTime" : 1371222080632, @@ -1235,9 +1235,9 @@ "finishTime" : 1371222080672, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061191, "result" : "SUCCESS", "finishTime" : 1371222080672, @@ -1282,9 +1282,9 @@ "finishTime" : 1371222081541, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061193, "result" : "SUCCESS", "finishTime" : 1371222081541, @@ -1329,9 +1329,9 @@ "finishTime" : 1371222080630, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061190, "result" : "SUCCESS", "finishTime" : 1371222080630, @@ -1376,9 +1376,9 @@ "finishTime" : 1371222080196, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222061192, "result" : "SUCCESS", "finishTime" : 1371222080196, @@ -1423,9 +1423,9 @@ "finishTime" : 1371222095399, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222077057, "result" : "SUCCESS", "finishTime" : 1371222095399, @@ -1470,9 +1470,9 @@ "finishTime" : 1371222101752, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222078069, "result" : "SUCCESS", "finishTime" : 1371222101752, @@ -1517,9 +1517,9 @@ "finishTime" : 1371222102468, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222080073, "result" : "SUCCESS", "finishTime" : 1371222102468, @@ -1564,9 +1564,9 @@ "finishTime" : 1371222101998, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222080074, "result" : "SUCCESS", "finishTime" : 1371222101998, @@ -1611,9 +1611,9 @@ "finishTime" : 1371222102416, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222080074, "result" : "SUCCESS", "finishTime" : 1371222102416, @@ -1658,9 +1658,9 @@ "finishTime" : 1371222102415, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222081100, "result" : "SUCCESS", "finishTime" : 1371222102415, @@ -1705,9 +1705,9 @@ "finishTime" : 1371222125169, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222081101, "result" : "SUCCESS", "finishTime" : 1371222125169, @@ -1752,9 +1752,9 @@ "finishTime" : 1371222103766, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222081107, "result" : "SUCCESS", "finishTime" : 1371222103766, @@ -1799,9 +1799,9 @@ "finishTime" : 1371222102985, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222081108, "result" : "SUCCESS", "finishTime" : 1371222102985, @@ -1846,9 +1846,9 @@ "finishTime" : 1371222102412, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222081122, "result" : "SUCCESS", "finishTime" : 1371222102412, @@ -1893,9 +1893,9 @@ "finishTime" : 1371222103554, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222081118, "result" : "SUCCESS", "finishTime" : 1371222103554, @@ -1940,9 +1940,9 @@ "finishTime" : 1371222101748, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222081109, "result" : "SUCCESS", "finishTime" : 1371222101748, @@ -1987,9 +1987,9 @@ "finishTime" : 1371222103767, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222081119, "result" : "SUCCESS", "finishTime" : 1371222103767, @@ -2034,9 +2034,9 @@ "finishTime" : 1371222103881, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222081121, "result" : "SUCCESS", "finishTime" : 1371222103881, @@ -2081,9 +2081,9 @@ "finishTime" : 1371222102868, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222081121, "result" : "SUCCESS", "finishTime" : 1371222102868, @@ -2128,9 +2128,9 @@ "finishTime" : 1371222127471, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222081121, "result" : "SUCCESS", "finishTime" : 1371222127471, @@ -2175,9 +2175,9 @@ "finishTime" : 1371222124895, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222082113, "result" : "SUCCESS", "finishTime" : 1371222124895, @@ -2222,9 +2222,9 @@ "finishTime" : 1371222102985, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222082114, "result" : "SUCCESS", "finishTime" : 1371222102985, @@ -2269,9 +2269,9 @@ "finishTime" : 1371222102897, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222082113, "result" : "SUCCESS", "finishTime" : 1371222102897, @@ -2316,9 +2316,9 @@ "finishTime" : 1371222129137, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222082116, "result" : "SUCCESS", "finishTime" : 1371222129137, @@ -2363,9 +2363,9 @@ "finishTime" : 1371222103702, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222082115, "result" : "SUCCESS", "finishTime" : 1371222103702, @@ -2410,9 +2410,9 @@ "finishTime" : 1371222114762, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222082116, "result" : "SUCCESS", "finishTime" : 1371222114762, @@ -2457,9 +2457,9 @@ "finishTime" : 1371222116084, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222082115, "result" : "SUCCESS", "finishTime" : 1371222116084, @@ -2504,9 +2504,9 @@ "finishTime" : 1371222103694, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222083110, "result" : "SUCCESS", "finishTime" : 1371222103694, @@ -2551,9 +2551,9 @@ "finishTime" : 1371222127757, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222083125, "result" : "SUCCESS", "finishTime" : 1371222127757, @@ -2598,9 +2598,9 @@ "finishTime" : 1371222103714, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222083112, "result" : "SUCCESS", "finishTime" : 1371222103714, @@ -2645,9 +2645,9 @@ "finishTime" : 1371222103849, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222083165, "result" : "SUCCESS", "finishTime" : 1371222103849, @@ -2692,9 +2692,9 @@ "finishTime" : 1371222103712, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222083127, "result" : "SUCCESS", "finishTime" : 1371222103712, @@ -2739,9 +2739,9 @@ "finishTime" : 1371222122471, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222083128, "result" : "SUCCESS", "finishTime" : 1371222122471, @@ -2786,9 +2786,9 @@ "finishTime" : 1371222124673, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222083120, "result" : "SUCCESS", "finishTime" : 1371222124673, @@ -2833,9 +2833,9 @@ "finishTime" : 1371222116128, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222097139, "result" : "SUCCESS", "finishTime" : 1371222116128, @@ -2880,9 +2880,9 @@ "finishTime" : 1371222116509, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222103155, "result" : "SUCCESS", "finishTime" : 1371222116509, @@ -2927,9 +2927,9 @@ "finishTime" : 1371222120045, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222103156, "result" : "SUCCESS", "finishTime" : 1371222120045, @@ -2974,9 +2974,9 @@ "finishTime" : 1371222117644, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222103154, "result" : "SUCCESS", "finishTime" : 1371222117644, @@ -3021,9 +3021,9 @@ "finishTime" : 1371222117257, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222103156, "result" : "SUCCESS", "finishTime" : 1371222117257, @@ -3068,9 +3068,9 @@ "finishTime" : 1371222142884, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222103154, "result" : "SUCCESS", "finishTime" : 1371222142884, @@ -3115,9 +3115,9 @@ "finishTime" : 1371222118279, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222104206, "result" : "SUCCESS", "finishTime" : 1371222118279, @@ -3162,9 +3162,9 @@ "finishTime" : 1371222122382, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222104212, "result" : "SUCCESS", "finishTime" : 1371222122382, @@ -3209,9 +3209,9 @@ "finishTime" : 1371222139789, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222104213, "result" : "SUCCESS", "finishTime" : 1371222139789, @@ -3256,9 +3256,9 @@ "finishTime" : 1371222117823, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222104211, "result" : "SUCCESS", "finishTime" : 1371222117823, @@ -3303,9 +3303,9 @@ "finishTime" : 1371222120103, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222104209, "result" : "SUCCESS", "finishTime" : 1371222120103, @@ -3350,9 +3350,9 @@ "finishTime" : 1371222116746, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222104212, "result" : "SUCCESS", "finishTime" : 1371222116746, @@ -3397,9 +3397,9 @@ "finishTime" : 1371222119279, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222104210, "result" : "SUCCESS", "finishTime" : 1371222119279, @@ -3444,9 +3444,9 @@ "finishTime" : 1371222116354, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222105211, "result" : "SUCCESS", "finishTime" : 1371222116354, @@ -3491,9 +3491,9 @@ "finishTime" : 1371222117482, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222105212, "result" : "SUCCESS", "finishTime" : 1371222117482, @@ -3538,9 +3538,9 @@ "finishTime" : 1371222122401, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222105213, "result" : "SUCCESS", "finishTime" : 1371222122401, @@ -3585,9 +3585,9 @@ "finishTime" : 1371222122853, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222106215, "result" : "SUCCESS", "finishTime" : 1371222122853, @@ -3632,9 +3632,9 @@ "finishTime" : 1371222123218, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222106215, "result" : "SUCCESS", "finishTime" : 1371222123218, @@ -3679,9 +3679,9 @@ "finishTime" : 1371222127755, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222107218, "result" : "SUCCESS", "finishTime" : 1371222127755, @@ -3726,9 +3726,9 @@ "finishTime" : 1371222127589, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222107220, "result" : "SUCCESS", "finishTime" : 1371222127589, @@ -3773,9 +3773,9 @@ "finishTime" : 1371222123105, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222107219, "result" : "SUCCESS", "finishTime" : 1371222123105, @@ -3820,9 +3820,9 @@ "finishTime" : 1371222132159, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222116240, "result" : "SUCCESS", "finishTime" : 1371222132159, @@ -3867,9 +3867,9 @@ "finishTime" : 1371222132250, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222117244, "result" : "SUCCESS", "finishTime" : 1371222132250, @@ -3914,9 +3914,9 @@ "finishTime" : 1371222134319, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222117244, "result" : "SUCCESS", "finishTime" : 1371222134319, @@ -3961,9 +3961,9 @@ "finishTime" : 1371222133861, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222118252, "result" : "SUCCESS", "finishTime" : 1371222133861, @@ -4008,9 +4008,9 @@ "finishTime" : 1371222139704, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222118252, "result" : "SUCCESS", "finishTime" : 1371222139704, @@ -4055,9 +4055,9 @@ "finishTime" : 1371222136354, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222118251, "result" : "SUCCESS", "finishTime" : 1371222136354, @@ -4102,9 +4102,9 @@ "finishTime" : 1371222139789, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222118251, "result" : "SUCCESS", "finishTime" : 1371222139789, @@ -4149,9 +4149,9 @@ "finishTime" : 1371222136362, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222119253, "result" : "SUCCESS", "finishTime" : 1371222136362, @@ -4196,9 +4196,9 @@ "finishTime" : 1371222132666, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222119254, "result" : "SUCCESS", "finishTime" : 1371222132666, @@ -4243,9 +4243,9 @@ "finishTime" : 1371222141107, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222119254, "result" : "SUCCESS", "finishTime" : 1371222141107, @@ -4290,9 +4290,9 @@ "finishTime" : 1371222135965, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222120259, "result" : "SUCCESS", "finishTime" : 1371222135965, @@ -4337,9 +4337,9 @@ "finishTime" : 1371222141263, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222120260, "result" : "SUCCESS", "finishTime" : 1371222141263, @@ -4384,9 +4384,9 @@ "finishTime" : 1371222140945, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222121262, "result" : "SUCCESS", "finishTime" : 1371222140945, @@ -4431,9 +4431,9 @@ "finishTime" : 1371222136213, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222122266, "result" : "SUCCESS", "finishTime" : 1371222136213, @@ -4478,9 +4478,9 @@ "finishTime" : 1371222141112, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222123269, "result" : "SUCCESS", "finishTime" : 1371222141112, @@ -4549,7 +4549,7 @@ "s3.blocksize" : "67108864", "mapreduce.task.io.sort.factor" : "10", "yarn.nodemanager.disk-health-checker.interval-ms" : "120000", - "mapreduce.job.working.dir" : "hdfs://a2115.halxg.cloudera.com:8020/user/jenkins", + "mapreduce.job.working.dir" : "hdfs://a2115.smile.com:8020/user/jenkins", "yarn.admin.acl" : "*", "mapreduce.job.speculative.speculativecap" : "0.1", "dfs.namenode.num.checkpoints.retained" : "2", @@ -4558,7 +4558,7 @@ "io.map.index.interval" : "128", "s3.client-write-packet-size" : "65536", "mapreduce.task.files.preserve.failedtasks" : "false", - "dfs.namenode.http-address" : "a2115.halxg.cloudera.com:20101", + "dfs.namenode.http-address" : "a2115.smile.com:20101", "ha.zookeeper.session-timeout.ms" : "5000", "hadoop.hdfs.configuration.version" : "1", "s3.replication" : "3", @@ -4575,7 +4575,7 @@ "dfs.namenode.checkpoint.dir" : "file://${hadoop.tmp.dir}/dfs/namesecondary", "dfs.namenode.replication.work.multiplier.per.iteration" : "2", "fs.trash.interval" : "1", - "yarn.resourcemanager.admin.address" : "a2115.halxg.cloudera.com:8033", + "yarn.resourcemanager.admin.address" : "a2115.smile.com:8033", "ha.health-monitor.check-interval.ms" : "1000", "mapreduce.job.outputformat.class" : "org.apache.hadoop.examples.terasort.TeraOutputFormat", "hadoop.jetty.logs.serve.aliases" : "true", @@ -4727,7 +4727,7 @@ "dfs.https.port" : "20102", "dfs.namenode.replication.interval" : "3", "mapreduce.task.skip.start.attempts" : "2", - "dfs.namenode.https-address" : "a2115.halxg.cloudera.com:20102", + "dfs.namenode.https-address" : "a2115.smile.com:20102", "mapreduce.jobtracker.persist.jobstatus.dir" : "/jobtracker/jobsInfo", "ipc.client.kill.max" : "10", "dfs.ha.automatic-failover.enabled" : "false", @@ -4757,7 +4757,7 @@ "ftp.replication" : "3", "mapreduce.jobtracker.http.address" : "0.0.0.0:50030", "yarn.nodemanager.health-checker.script.timeout-ms" : "1200000", - "mapreduce.jobhistory.address" : "a2115.halxg.cloudera.com:10020", + "mapreduce.jobhistory.address" : "a2115.smile.com:10020", "mapreduce.jobtracker.taskcache.levels" : "2", "dfs.datanode.dns.nameserver" : "default", "mapreduce.application.classpath" : "$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*", @@ -4799,13 +4799,13 @@ "hadoop.fuse.timer.period" : "5", "dfs.namenode.num.extra.edits.retained" : "1000000", "hadoop.rpc.socket.factory.class.default" : "org.apache.hadoop.net.StandardSocketFactory", - "mapreduce.job.submithostname" : "a2115.halxg.cloudera.com", + "mapreduce.job.submithostname" : "a2115.smile.com", "dfs.namenode.handler.count" : "10", "fs.automatic.close" : "false", "mapreduce.job.submithostaddress" : "10.20.206.115", "mapreduce.tasktracker.healthchecker.script.timeout" : "600000", "dfs.datanode.directoryscan.interval" : "21600", - "yarn.resourcemanager.address" : "a2115.halxg.cloudera.com:8032", + "yarn.resourcemanager.address" : "a2115.smile.com:8032", "yarn.nodemanager.health-checker.interval-ms" : "600000", "dfs.client.file-block-storage-locations.num-threads" : "10", "yarn.resourcemanager.container-tokens.master-key-rolling-interval-secs" : "86400", @@ -4820,7 +4820,7 @@ "ftp.stream-buffer-size" : "4096", "dfs.namenode.avoid.write.stale.datanode" : "false", "hadoop.security.group.mapping.ldap.search.attr.member" : "member", - "mapreduce.output.fileoutputformat.outputdir" : "hdfs://a2115.halxg.cloudera.com:8020/user/jenkins/tera-gen-1", + "mapreduce.output.fileoutputformat.outputdir" : "hdfs://a2115.smile.com:8020/user/jenkins/tera-gen-1", "dfs.blockreport.initialDelay" : "0", "yarn.nm.liveness-monitor.expiry-interval-ms" : "600000", "hadoop.http.authentication.token.validity" : "36000", @@ -4836,7 +4836,7 @@ "mapreduce.task.profile" : "false", "mapreduce.tasktracker.http.address" : "0.0.0.0:50060", "mapreduce.tasktracker.instrumentation" : "org.apache.hadoop.mapred.TaskTrackerMetricsInst", - "mapreduce.jobhistory.webapp.address" : "a2115.halxg.cloudera.com:19888", + "mapreduce.jobhistory.webapp.address" : "a2115.smile.com:19888", "ha.failover-controller.graceful-fence.rpc-timeout.ms" : "5000", "yarn.ipc.rpc.class" : "org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC", "mapreduce.job.name" : "TeraGen", @@ -4866,14 +4866,14 @@ "hadoop.security.auth_to_local" : "DEFAULT", "dfs.secondary.namenode.kerberos.internal.spnego.principal" : "${dfs.web.authentication.kerberos.principal}", "ftp.client-write-packet-size" : "65536", - "fs.defaultFS" : "hdfs://a2115.halxg.cloudera.com:8020", + "fs.defaultFS" : "hdfs://a2115.smile.com:8020", "yarn.nodemanager.address" : "0.0.0.0:0", "yarn.scheduler.fair.assignmultiple" : "true", "yarn.resourcemanager.scheduler.client.thread-count" : "50", "mapreduce.task.merge.progress.records" : "10000", "file.client-write-packet-size" : "65536", "yarn.nodemanager.delete.thread-count" : "4", - "yarn.resourcemanager.scheduler.address" : "a2115.halxg.cloudera.com:8030", + "yarn.resourcemanager.scheduler.address" : "a2115.smile.com:8030", "fs.trash.checkpoint.interval" : "0", "hadoop.http.authentication.signature.secret.file" : "${user.home}/hadoop-http-auth-signature-secret", "s3native.stream-buffer-size" : "4096", @@ -4915,7 +4915,7 @@ "mapreduce.map.skip.maxrecords" : "0", "dfs.datanode.https.address" : "0.0.0.0:50475", "file.replication" : "1", - "yarn.resourcemanager.resource-tracker.address" : "a2115.halxg.cloudera.com:8031", + "yarn.resourcemanager.resource-tracker.address" : "a2115.smile.com:8031", "dfs.datanode.drop.cache.behind.reads" : "false", "hadoop.fuse.connection.timeout" : "300", "hadoop.work.around.non.threadsafe.getpwuid" : "false", @@ -4940,7 +4940,7 @@ "mapreduce.task.profile.maps" : "0-2", "dfs.namenode.decommission.interval" : "30", "dfs.image.compression.codec" : "org.apache.hadoop.io.compress.DefaultCodec", - "yarn.resourcemanager.webapp.address" : "a2115.halxg.cloudera.com:8088", + "yarn.resourcemanager.webapp.address" : "a2115.smile.com:8088", "mapreduce.jobtracker.system.dir" : "${hadoop.tmp.dir}/mapred/system", "hadoop.ssl.hostname.verifier" : "DEFAULT", "yarn.nodemanager.vmem-pmem-ratio" : "2.1", @@ -5127,9 +5127,9 @@ "finishTime" : 1371222188484, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166321, "result" : "SUCCESS", "finishTime" : 1371222188484, @@ -5174,9 +5174,9 @@ "finishTime" : 1371222185876, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166287, "result" : "SUCCESS", "finishTime" : 1371222185876, @@ -5221,9 +5221,9 @@ "finishTime" : 1371222187434, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166324, "result" : "SUCCESS", "finishTime" : 1371222187434, @@ -5268,9 +5268,9 @@ "finishTime" : 1371222186091, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166325, "result" : "SUCCESS", "finishTime" : 1371222186091, @@ -5315,9 +5315,9 @@ "finishTime" : 1371222189053, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166287, "result" : "SUCCESS", "finishTime" : 1371222189053, @@ -5362,9 +5362,9 @@ "finishTime" : 1371222184979, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166323, "result" : "SUCCESS", "finishTime" : 1371222184979, @@ -5409,9 +5409,9 @@ "finishTime" : 1371222185246, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166324, "result" : "SUCCESS", "finishTime" : 1371222185246, @@ -5456,9 +5456,9 @@ "finishTime" : 1371222185092, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222166325, "result" : "SUCCESS", "finishTime" : 1371222185092, @@ -5503,9 +5503,9 @@ "finishTime" : 1371222188434, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166282, "result" : "SUCCESS", "finishTime" : 1371222188434, @@ -5550,9 +5550,9 @@ "finishTime" : 1371222186078, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166289, "result" : "SUCCESS", "finishTime" : 1371222186078, @@ -5597,9 +5597,9 @@ "finishTime" : 1371222185273, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166293, "result" : "SUCCESS", "finishTime" : 1371222185273, @@ -5644,9 +5644,9 @@ "finishTime" : 1371222188434, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166290, "result" : "SUCCESS", "finishTime" : 1371222188434, @@ -5691,9 +5691,9 @@ "finishTime" : 1371222185278, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166295, "result" : "SUCCESS", "finishTime" : 1371222185278, @@ -5738,9 +5738,9 @@ "finishTime" : 1371222185273, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166291, "result" : "SUCCESS", "finishTime" : 1371222185273, @@ -5785,9 +5785,9 @@ "finishTime" : 1371222185953, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166297, "result" : "SUCCESS", "finishTime" : 1371222185953, @@ -5832,9 +5832,9 @@ "finishTime" : 1371222189058, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222166333, "result" : "SUCCESS", "finishTime" : 1371222189058, @@ -5879,9 +5879,9 @@ "finishTime" : 1371222185112, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166298, "result" : "SUCCESS", "finishTime" : 1371222185112, @@ -5926,9 +5926,9 @@ "finishTime" : 1371222185125, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166299, "result" : "SUCCESS", "finishTime" : 1371222185125, @@ -5973,9 +5973,9 @@ "finishTime" : 1371222185117, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166301, "result" : "SUCCESS", "finishTime" : 1371222185117, @@ -6020,9 +6020,9 @@ "finishTime" : 1371222186145, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166303, "result" : "SUCCESS", "finishTime" : 1371222186145, @@ -6067,9 +6067,9 @@ "finishTime" : 1371222188294, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166304, "result" : "SUCCESS", "finishTime" : 1371222188294, @@ -6114,9 +6114,9 @@ "finishTime" : 1371222185112, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166306, "result" : "SUCCESS", "finishTime" : 1371222185112, @@ -6161,9 +6161,9 @@ "finishTime" : 1371222186222, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166307, "result" : "SUCCESS", "finishTime" : 1371222186222, @@ -6208,9 +6208,9 @@ "finishTime" : 1371222186145, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222166345, "result" : "SUCCESS", "finishTime" : 1371222186145, @@ -6255,9 +6255,9 @@ "finishTime" : 1371222189046, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166312, "result" : "SUCCESS", "finishTime" : 1371222189046, @@ -6302,9 +6302,9 @@ "finishTime" : 1371222185135, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166313, "result" : "SUCCESS", "finishTime" : 1371222185135, @@ -6349,9 +6349,9 @@ "finishTime" : 1371222185121, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166319, "result" : "SUCCESS", "finishTime" : 1371222185121, @@ -6396,9 +6396,9 @@ "finishTime" : 1371222186069, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166318, "result" : "SUCCESS", "finishTime" : 1371222186069, @@ -6443,9 +6443,9 @@ "finishTime" : 1371222185115, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166327, "result" : "SUCCESS", "finishTime" : 1371222185115, @@ -6490,9 +6490,9 @@ "finishTime" : 1371222186085, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222166330, "result" : "SUCCESS", "finishTime" : 1371222186085, @@ -6537,9 +6537,9 @@ "finishTime" : 1371222209777, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222186223, "result" : "SUCCESS", "finishTime" : 1371222209777, @@ -6584,9 +6584,9 @@ "finishTime" : 1371222209814, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222186224, "result" : "SUCCESS", "finishTime" : 1371222209814, @@ -6631,9 +6631,9 @@ "finishTime" : 1371222209796, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222186230, "result" : "SUCCESS", "finishTime" : 1371222209796, @@ -6678,9 +6678,9 @@ "finishTime" : 1371222209114, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222186269, "result" : "SUCCESS", "finishTime" : 1371222209114, @@ -6725,9 +6725,9 @@ "finishTime" : 1371222205072, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222186230, "result" : "SUCCESS", "finishTime" : 1371222205072, @@ -6772,9 +6772,9 @@ "finishTime" : 1371222209265, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222186231, "result" : "SUCCESS", "finishTime" : 1371222209265, @@ -6819,9 +6819,9 @@ "finishTime" : 1371222205027, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222186270, "result" : "SUCCESS", "finishTime" : 1371222205027, @@ -6866,9 +6866,9 @@ "finishTime" : 1371222204802, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222186229, "result" : "SUCCESS", "finishTime" : 1371222204802, @@ -6913,9 +6913,9 @@ "finishTime" : 1371222205071, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222186231, "result" : "SUCCESS", "finishTime" : 1371222205071, @@ -6960,9 +6960,9 @@ "finishTime" : 1371222204844, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222186270, "result" : "SUCCESS", "finishTime" : 1371222204844, @@ -7007,9 +7007,9 @@ "finishTime" : 1371222208051, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222187226, "result" : "SUCCESS", "finishTime" : 1371222208051, @@ -7054,9 +7054,9 @@ "finishTime" : 1371222209764, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222187225, "result" : "SUCCESS", "finishTime" : 1371222209764, @@ -7101,9 +7101,9 @@ "finishTime" : 1371222210013, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222188236, "result" : "SUCCESS", "finishTime" : 1371222210013, @@ -7148,9 +7148,9 @@ "finishTime" : 1371222209829, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222188247, "result" : "SUCCESS", "finishTime" : 1371222209829, @@ -7195,9 +7195,9 @@ "finishTime" : 1371222205147, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222188243, "result" : "SUCCESS", "finishTime" : 1371222205147, @@ -7242,9 +7242,9 @@ "finishTime" : 1371222208888, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222188284, "result" : "SUCCESS", "finishTime" : 1371222208888, @@ -7289,9 +7289,9 @@ "finishTime" : 1371222209764, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222188246, "result" : "SUCCESS", "finishTime" : 1371222209764, @@ -7336,9 +7336,9 @@ "finishTime" : 1371222209770, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222188244, "result" : "SUCCESS", "finishTime" : 1371222209770, @@ -7383,9 +7383,9 @@ "finishTime" : 1371222209758, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222188248, "result" : "SUCCESS", "finishTime" : 1371222209758, @@ -7430,9 +7430,9 @@ "finishTime" : 1371222209877, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222188245, "result" : "SUCCESS", "finishTime" : 1371222209877, @@ -7477,9 +7477,9 @@ "finishTime" : 1371222209871, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222188246, "result" : "SUCCESS", "finishTime" : 1371222209871, @@ -7524,9 +7524,9 @@ "finishTime" : 1371222209798, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222189249, "result" : "SUCCESS", "finishTime" : 1371222209798, @@ -7571,9 +7571,9 @@ "finishTime" : 1371222209546, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222189252, "result" : "SUCCESS", "finishTime" : 1371222209546, @@ -7618,9 +7618,9 @@ "finishTime" : 1371222211732, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222189248, "result" : "SUCCESS", "finishTime" : 1371222211732, @@ -7665,9 +7665,9 @@ "finishTime" : 1371222209794, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222189250, "result" : "SUCCESS", "finishTime" : 1371222209794, @@ -7712,9 +7712,9 @@ "finishTime" : 1371222209829, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222190266, "result" : "SUCCESS", "finishTime" : 1371222209829, @@ -7759,9 +7759,9 @@ "finishTime" : 1371222209766, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222190265, "result" : "SUCCESS", "finishTime" : 1371222209766, @@ -7806,9 +7806,9 @@ "finishTime" : 1371222209780, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222190273, "result" : "SUCCESS", "finishTime" : 1371222209780, @@ -7853,9 +7853,9 @@ "finishTime" : 1371222209289, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222190266, "result" : "SUCCESS", "finishTime" : 1371222209289, @@ -7900,9 +7900,9 @@ "finishTime" : 1371222211562, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222192248, "result" : "SUCCESS", "finishTime" : 1371222211562, @@ -7947,9 +7947,9 @@ "finishTime" : 1371222225441, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222206294, "result" : "SUCCESS", "finishTime" : 1371222225441, @@ -7994,9 +7994,9 @@ "finishTime" : 1371222221672, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222206296, "result" : "SUCCESS", "finishTime" : 1371222221672, @@ -8041,9 +8041,9 @@ "finishTime" : 1371222221958, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222206298, "result" : "SUCCESS", "finishTime" : 1371222221958, @@ -8088,9 +8088,9 @@ "finishTime" : 1371222227163, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222207290, "result" : "SUCCESS", "finishTime" : 1371222227163, @@ -8135,9 +8135,9 @@ "finishTime" : 1371222227254, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222207289, "result" : "SUCCESS", "finishTime" : 1371222227254, @@ -8182,9 +8182,9 @@ "finishTime" : 1371222230890, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222207289, "result" : "SUCCESS", "finishTime" : 1371222230890, @@ -8229,9 +8229,9 @@ "finishTime" : 1371222241231, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222209293, "result" : "SUCCESS", "finishTime" : 1371222241231, @@ -8276,9 +8276,9 @@ "finishTime" : 1371222225138, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222209294, "result" : "SUCCESS", "finishTime" : 1371222225138, @@ -8323,9 +8323,9 @@ "finishTime" : 1371222230905, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222210298, "result" : "SUCCESS", "finishTime" : 1371222230905, @@ -8370,9 +8370,9 @@ "finishTime" : 1371222225328, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222210302, "result" : "SUCCESS", "finishTime" : 1371222225328, @@ -8417,9 +8417,9 @@ "finishTime" : 1371222230923, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222210301, "result" : "SUCCESS", "finishTime" : 1371222230923, @@ -8464,9 +8464,9 @@ "finishTime" : 1371222243147, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222210300, "result" : "SUCCESS", "finishTime" : 1371222243147, @@ -8511,9 +8511,9 @@ "finishTime" : 1371222230925, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222210302, "result" : "SUCCESS", "finishTime" : 1371222230925, @@ -8558,9 +8558,9 @@ "finishTime" : 1371222233643, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222211323, "result" : "SUCCESS", "finishTime" : 1371222233643, @@ -8605,9 +8605,9 @@ "finishTime" : 1371222231021, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222211318, "result" : "SUCCESS", "finishTime" : 1371222231021, @@ -8652,9 +8652,9 @@ "finishTime" : 1371222230806, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222211320, "result" : "SUCCESS", "finishTime" : 1371222230806, @@ -8699,9 +8699,9 @@ "finishTime" : 1371222231013, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222211322, "result" : "SUCCESS", "finishTime" : 1371222231013, @@ -8746,9 +8746,9 @@ "finishTime" : 1371222230986, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222211321, "result" : "SUCCESS", "finishTime" : 1371222230986, @@ -8793,9 +8793,9 @@ "finishTime" : 1371222230909, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222211322, "result" : "SUCCESS", "finishTime" : 1371222230909, @@ -8840,9 +8840,9 @@ "finishTime" : 1371222241634, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212318, "result" : "SUCCESS", "finishTime" : 1371222241634, @@ -8887,9 +8887,9 @@ "finishTime" : 1371222234375, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212320, "result" : "SUCCESS", "finishTime" : 1371222234375, @@ -8934,9 +8934,9 @@ "finishTime" : 1371222233436, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212365, "result" : "SUCCESS", "finishTime" : 1371222233436, @@ -8981,9 +8981,9 @@ "finishTime" : 1371222234234, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212323, "result" : "SUCCESS", "finishTime" : 1371222234234, @@ -9028,9 +9028,9 @@ "finishTime" : 1371222234374, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212325, "result" : "SUCCESS", "finishTime" : 1371222234374, @@ -9075,9 +9075,9 @@ "finishTime" : 1371222233257, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222212365, "result" : "SUCCESS", "finishTime" : 1371222233257, @@ -9122,9 +9122,9 @@ "finishTime" : 1371222240854, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222212327, "result" : "SUCCESS", "finishTime" : 1371222240854, @@ -9169,9 +9169,9 @@ "finishTime" : 1371222230895, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222212325, "result" : "SUCCESS", "finishTime" : 1371222230895, @@ -9216,9 +9216,9 @@ "finishTime" : 1371222243205, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222212324, "result" : "SUCCESS", "finishTime" : 1371222243205, @@ -9263,9 +9263,9 @@ "finishTime" : 1371222232772, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222212326, "result" : "SUCCESS", "finishTime" : 1371222232772, @@ -9310,9 +9310,9 @@ "finishTime" : 1371222232612, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2117.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2117.smile.com" ] }, - "hostName" : "/default-rack/a2117.halxg.cloudera.com", + "hostName" : "/default-rack/a2117.smile.com", "startTime" : 1371222213322, "result" : "SUCCESS", "finishTime" : 1371222232612, @@ -9357,9 +9357,9 @@ "finishTime" : 1371222235284, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222223387, "result" : "SUCCESS", "finishTime" : 1371222235284, @@ -9404,9 +9404,9 @@ "finishTime" : 1371222237538, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2115.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2115.smile.com" ] }, - "hostName" : "/default-rack/a2115.halxg.cloudera.com", + "hostName" : "/default-rack/a2115.smile.com", "startTime" : 1371222223387, "result" : "SUCCESS", "finishTime" : 1371222237538, @@ -9451,9 +9451,9 @@ "finishTime" : 1371222239274, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222225390, "result" : "SUCCESS", "finishTime" : 1371222239274, @@ -9498,9 +9498,9 @@ "finishTime" : 1371222249913, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2116.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2116.smile.com" ] }, - "hostName" : "/default-rack/a2116.halxg.cloudera.com", + "hostName" : "/default-rack/a2116.smile.com", "startTime" : 1371222226395, "result" : "SUCCESS", "finishTime" : 1371222249913, @@ -9545,9 +9545,9 @@ "finishTime" : 1371222241825, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222227397, "result" : "SUCCESS", "finishTime" : 1371222241825, @@ -9592,9 +9592,9 @@ "finishTime" : 1371222241726, "attempts" : [ { "location" : { - "layers" : [ "default-rack", "a2118.halxg.cloudera.com" ] + "layers" : [ "default-rack", "a2118.smile.com" ] }, - "hostName" : "/default-rack/a2118.halxg.cloudera.com", + "hostName" : "/default-rack/a2118.smile.com", "startTime" : 1371222228399, "result" : "SUCCESS", "finishTime" : 1371222241726, @@ -9663,7 +9663,7 @@ "s3.blocksize" : "67108864", "mapreduce.task.io.sort.factor" : "10", "yarn.nodemanager.disk-health-checker.interval-ms" : "120000", - "mapreduce.job.working.dir" : "hdfs://a2115.halxg.cloudera.com:8020/user/jenkins", + "mapreduce.job.working.dir" : "hdfs://a2115.smile.com:8020/user/jenkins", "yarn.admin.acl" : "*", "mapreduce.job.speculative.speculativecap" : "0.1", "dfs.namenode.num.checkpoints.retained" : "2", @@ -9672,7 +9672,7 @@ "io.map.index.interval" : "128", "s3.client-write-packet-size" : "65536", "mapreduce.task.files.preserve.failedtasks" : "false", - "dfs.namenode.http-address" : "a2115.halxg.cloudera.com:20101", + "dfs.namenode.http-address" : "a2115.smile.com:20101", "ha.zookeeper.session-timeout.ms" : "5000", "hadoop.hdfs.configuration.version" : "1", "s3.replication" : "3", @@ -9689,7 +9689,7 @@ "dfs.namenode.checkpoint.dir" : "file://${hadoop.tmp.dir}/dfs/namesecondary", "dfs.namenode.replication.work.multiplier.per.iteration" : "2", "fs.trash.interval" : "1", - "yarn.resourcemanager.admin.address" : "a2115.halxg.cloudera.com:8033", + "yarn.resourcemanager.admin.address" : "a2115.smile.com:8033", "ha.health-monitor.check-interval.ms" : "1000", "mapreduce.job.outputformat.class" : "org.apache.hadoop.examples.terasort.TeraOutputFormat", "hadoop.jetty.logs.serve.aliases" : "true", @@ -9841,7 +9841,7 @@ "dfs.https.port" : "20102", "dfs.namenode.replication.interval" : "3", "mapreduce.task.skip.start.attempts" : "2", - "dfs.namenode.https-address" : "a2115.halxg.cloudera.com:20102", + "dfs.namenode.https-address" : "a2115.smile.com:20102", "mapreduce.jobtracker.persist.jobstatus.dir" : "/jobtracker/jobsInfo", "ipc.client.kill.max" : "10", "dfs.ha.automatic-failover.enabled" : "false", @@ -9871,7 +9871,7 @@ "ftp.replication" : "3", "mapreduce.jobtracker.http.address" : "0.0.0.0:50030", "yarn.nodemanager.health-checker.script.timeout-ms" : "1200000", - "mapreduce.jobhistory.address" : "a2115.halxg.cloudera.com:10020", + "mapreduce.jobhistory.address" : "a2115.smile.com:10020", "mapreduce.jobtracker.taskcache.levels" : "2", "dfs.datanode.dns.nameserver" : "default", "mapreduce.application.classpath" : "$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*", @@ -9913,13 +9913,13 @@ "hadoop.fuse.timer.period" : "5", "dfs.namenode.num.extra.edits.retained" : "1000000", "hadoop.rpc.socket.factory.class.default" : "org.apache.hadoop.net.StandardSocketFactory", - "mapreduce.job.submithostname" : "a2115.halxg.cloudera.com", + "mapreduce.job.submithostname" : "a2115.smile.com", "dfs.namenode.handler.count" : "10", "fs.automatic.close" : "false", "mapreduce.job.submithostaddress" : "10.20.206.115", "mapreduce.tasktracker.healthchecker.script.timeout" : "600000", "dfs.datanode.directoryscan.interval" : "21600", - "yarn.resourcemanager.address" : "a2115.halxg.cloudera.com:8032", + "yarn.resourcemanager.address" : "a2115.smile.com:8032", "yarn.nodemanager.health-checker.interval-ms" : "600000", "dfs.client.file-block-storage-locations.num-threads" : "10", "yarn.resourcemanager.container-tokens.master-key-rolling-interval-secs" : "86400", @@ -9934,7 +9934,7 @@ "ftp.stream-buffer-size" : "4096", "dfs.namenode.avoid.write.stale.datanode" : "false", "hadoop.security.group.mapping.ldap.search.attr.member" : "member", - "mapreduce.output.fileoutputformat.outputdir" : "hdfs://a2115.halxg.cloudera.com:8020/user/jenkins/tera-gen-2", + "mapreduce.output.fileoutputformat.outputdir" : "hdfs://a2115.smile.com:8020/user/jenkins/tera-gen-2", "dfs.blockreport.initialDelay" : "0", "yarn.nm.liveness-monitor.expiry-interval-ms" : "600000", "hadoop.http.authentication.token.validity" : "36000", @@ -9950,7 +9950,7 @@ "mapreduce.task.profile" : "false", "mapreduce.tasktracker.http.address" : "0.0.0.0:50060", "mapreduce.tasktracker.instrumentation" : "org.apache.hadoop.mapred.TaskTrackerMetricsInst", - "mapreduce.jobhistory.webapp.address" : "a2115.halxg.cloudera.com:19888", + "mapreduce.jobhistory.webapp.address" : "a2115.smile.com:19888", "ha.failover-controller.graceful-fence.rpc-timeout.ms" : "5000", "yarn.ipc.rpc.class" : "org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC", "mapreduce.job.name" : "TeraGen", @@ -9980,14 +9980,14 @@ "hadoop.security.auth_to_local" : "DEFAULT", "dfs.secondary.namenode.kerberos.internal.spnego.principal" : "${dfs.web.authentication.kerberos.principal}", "ftp.client-write-packet-size" : "65536", - "fs.defaultFS" : "hdfs://a2115.halxg.cloudera.com:8020", + "fs.defaultFS" : "hdfs://a2115.smile.com:8020", "yarn.nodemanager.address" : "0.0.0.0:0", "yarn.scheduler.fair.assignmultiple" : "true", "yarn.resourcemanager.scheduler.client.thread-count" : "50", "mapreduce.task.merge.progress.records" : "10000", "file.client-write-packet-size" : "65536", "yarn.nodemanager.delete.thread-count" : "4", - "yarn.resourcemanager.scheduler.address" : "a2115.halxg.cloudera.com:8030", + "yarn.resourcemanager.scheduler.address" : "a2115.smile.com:8030", "fs.trash.checkpoint.interval" : "0", "hadoop.http.authentication.signature.secret.file" : "${user.home}/hadoop-http-auth-signature-secret", "s3native.stream-buffer-size" : "4096", @@ -10029,7 +10029,7 @@ "mapreduce.map.skip.maxrecords" : "0", "dfs.datanode.https.address" : "0.0.0.0:50475", "file.replication" : "1", - "yarn.resourcemanager.resource-tracker.address" : "a2115.halxg.cloudera.com:8031", + "yarn.resourcemanager.resource-tracker.address" : "a2115.smile.com:8031", "dfs.datanode.drop.cache.behind.reads" : "false", "hadoop.fuse.connection.timeout" : "300", "hadoop.work.around.non.threadsafe.getpwuid" : "false", @@ -10054,7 +10054,7 @@ "mapreduce.task.profile.maps" : "0-2", "dfs.namenode.decommission.interval" : "30", "dfs.image.compression.codec" : "org.apache.hadoop.io.compress.DefaultCodec", - "yarn.resourcemanager.webapp.address" : "a2115.halxg.cloudera.com:8088", + "yarn.resourcemanager.webapp.address" : "a2115.smile.com:8088", "mapreduce.jobtracker.system.dir" : "${hadoop.tmp.dir}/mapred/system", "hadoop.ssl.hostname.verifier" : "DEFAULT", "yarn.nodemanager.vmem-pmem-ratio" : "2.1", From 5beeb3016954a3ee0c1fb10a2083ffd540cd2c14 Mon Sep 17 00:00:00 2001 From: Arpit Agarwal Date: Fri, 31 Jan 2014 21:00:33 +0000 Subject: [PATCH 22/23] HDFS-5153. Datanode should send block reports for each storage in a separate message. (Arpit Agarwal) git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563254 13f79535-47bb-0310-9956-ffa450edef68 --- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 + .../org/apache/hadoop/hdfs/DFSConfigKeys.java | 2 + .../server/blockmanagement/BlockManager.java | 29 ++- .../blockmanagement/DatanodeDescriptor.java | 11 + .../hdfs/server/datanode/BPServiceActor.java | 169 +++++++------ .../hadoop/hdfs/server/datanode/DNConf.java | 5 + .../server/namenode/NameNodeRpcServer.java | 9 +- .../src/main/resources/hdfs-default.xml | 14 ++ ...ckReport.java => BlockReportTestBase.java} | 236 ++++++------------ ...stDnRespectsBlockReportSplitThreshold.java | 205 +++++++++++++++ .../TestNNHandlesBlockReportPerStorage.java | 42 ++++ .../TestNNHandlesCombinedBlockReport.java | 39 +++ 12 files changed, 516 insertions(+), 248 deletions(-) rename hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/{TestBlockReport.java => BlockReportTestBase.java} (85%) create mode 100644 hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDnRespectsBlockReportSplitThreshold.java create mode 100644 hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesBlockReportPerStorage.java create mode 100644 hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesCombinedBlockReport.java diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 1fb5c1c190c..7b21e11d852 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -297,6 +297,9 @@ Release 2.4.0 - UNRELEASED HDFS-5781. Use an array to record the mapping between FSEditLogOpCode and the corresponding byte value. (jing9) + HDFS-5153. Datanode should send block reports for each storage in a + separate message. (Arpit Agarwal) + OPTIMIZATIONS HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java index e82eb524bd0..38c40c21f38 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java @@ -399,6 +399,8 @@ public class DFSConfigKeys extends CommonConfigurationKeys { public static final long DFS_BLOCKREPORT_INTERVAL_MSEC_DEFAULT = 60 * 60 * 1000; public static final String DFS_BLOCKREPORT_INITIAL_DELAY_KEY = "dfs.blockreport.initialDelay"; public static final int DFS_BLOCKREPORT_INITIAL_DELAY_DEFAULT = 0; + public static final String DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY = "dfs.blockreport.split.threshold"; + public static final long DFS_BLOCKREPORT_SPLIT_THRESHOLD_DEFAULT = 1000 * 1000; public static final String DFS_CACHEREPORT_INTERVAL_MSEC_KEY = "dfs.cachereport.intervalMsec"; public static final long DFS_CACHEREPORT_INTERVAL_MSEC_DEFAULT = 10 * 1000; public static final String DFS_BLOCK_INVALIDATE_LIMIT_KEY = "dfs.block.invalidate.limit"; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java index 2e1eb6cc947..8ff314aeb96 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java @@ -1616,15 +1616,19 @@ public String toString() { /** * The given storage is reporting all its blocks. * Update the (storage-->block list) and (block-->storage list) maps. + * + * @return true if all known storages of the given DN have finished reporting. + * @throws IOException */ - public void processReport(final DatanodeID nodeID, + public boolean processReport(final DatanodeID nodeID, final DatanodeStorage storage, final String poolId, final BlockListAsLongs newReport) throws IOException { namesystem.writeLock(); final long startTime = Time.now(); //after acquiring write lock final long endTime; + DatanodeDescriptor node; try { - final DatanodeDescriptor node = datanodeManager.getDatanode(nodeID); + node = datanodeManager.getDatanode(nodeID); if (node == null || !node.isAlive) { throw new IOException( "ProcessReport from dead or unregistered node: " + nodeID); @@ -1632,13 +1636,21 @@ public void processReport(final DatanodeID nodeID, // To minimize startup time, we discard any second (or later) block reports // that we receive while still in startup phase. - final DatanodeStorageInfo storageInfo = node.updateStorage(storage); + DatanodeStorageInfo storageInfo = node.getStorageInfo(storage.getStorageID()); + + if (storageInfo == null) { + // We handle this for backwards compatibility. + storageInfo = node.updateStorage(storage); + LOG.warn("Unknown storageId " + storage.getStorageID() + + ", updating storageMap. This indicates a buggy " + + "DataNode that isn't heartbeating correctly."); + } if (namesystem.isInStartupSafeMode() && storageInfo.getBlockReportCount() > 0) { blockLog.info("BLOCK* processReport: " + "discarded non-initial block report from " + nodeID + " because namenode still in startup phase"); - return; + return !node.hasStaleStorages(); } if (storageInfo.numBlocks() == 0) { @@ -1655,7 +1667,7 @@ public void processReport(final DatanodeID nodeID, storageInfo.receivedBlockReport(); if (staleBefore && !storageInfo.areBlockContentsStale()) { LOG.info("BLOCK* processReport: Received first block report from " - + node + " after starting up or becoming active. Its block " + + storage + " after starting up or becoming active. Its block " + "contents are no longer considered stale"); rescanPostponedMisreplicatedBlocks(); } @@ -1670,9 +1682,10 @@ public void processReport(final DatanodeID nodeID, if (metrics != null) { metrics.addBlockReport((int) (endTime - startTime)); } - blockLog.info("BLOCK* processReport: from " - + nodeID + ", blocks: " + newReport.getNumberOfBlocks() + blockLog.info("BLOCK* processReport: from storage " + storage.getStorageID() + + " node " + nodeID + ", blocks: " + newReport.getNumberOfBlocks() + ", processing time: " + (endTime - startTime) + " msecs"); + return !node.hasStaleStorages(); } /** @@ -1827,7 +1840,7 @@ private void reportDiff(DatanodeDescriptor dn, DatanodeStorage storage, Collection toCorrupt, // add to corrupt replicas list Collection toUC) { // add to under-construction list - final DatanodeStorageInfo storageInfo = dn.updateStorage(storage); + final DatanodeStorageInfo storageInfo = dn.getStorageInfo(storage.getStorageID()); // place a delimiter in the list which separates blocks // that have been reported from those that have not diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java index fc4bf885092..c18f435f53e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java @@ -257,6 +257,17 @@ DatanodeStorageInfo[] getStorageInfos() { } } + boolean hasStaleStorages() { + synchronized (storageMap) { + for (DatanodeStorageInfo storage : storageMap.values()) { + if (storage.areBlockContentsStale()) { + return true; + } + } + return false; + } + } + /** * Remove block from the list of blocks belonging to the data-node. Remove * data-node from the block. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java index c91fca381fa..a1c2cb2f50c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java @@ -22,11 +22,9 @@ import java.io.IOException; import java.net.InetSocketAddress; import java.net.SocketTimeoutException; -import java.util.ArrayList; -import java.util.Collection; -import java.util.List; -import java.util.Map; +import java.util.*; +import com.google.common.base.Joiner; import org.apache.commons.logging.Log; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.ha.HAServiceProtocol.HAServiceState; @@ -437,75 +435,100 @@ void triggerDeletionReportForTests() { /** * Report the list blocks to the Namenode + * @return DatanodeCommands returned by the NN. May be null. * @throws IOException */ - DatanodeCommand blockReport() throws IOException { + List blockReport() throws IOException { // send block report if timer has expired. - DatanodeCommand cmd = null; - long startTime = now(); - if (startTime - lastBlockReport > dnConf.blockReportInterval) { - - // Flush any block information that precedes the block report. Otherwise - // we have a chance that we will miss the delHint information - // or we will report an RBW replica after the BlockReport already reports - // a FINALIZED one. - reportReceivedDeletedBlocks(); - - // Send one block report per known storage. - - // Create block report - long brCreateStartTime = now(); - long totalBlockCount = 0; - - Map perVolumeBlockLists = - dn.getFSDataset().getBlockReports(bpos.getBlockPoolId()); - - // Send block report - long brSendStartTime = now(); - StorageBlockReport[] reports = - new StorageBlockReport[perVolumeBlockLists.size()]; - - int i = 0; - for(Map.Entry kvPair : perVolumeBlockLists.entrySet()) { - DatanodeStorage dnStorage = kvPair.getKey(); - BlockListAsLongs blockList = kvPair.getValue(); - totalBlockCount += blockList.getNumberOfBlocks(); - - reports[i++] = - new StorageBlockReport( - dnStorage, blockList.getBlockListAsLongs()); - } - - cmd = bpNamenode.blockReport(bpRegistration, bpos.getBlockPoolId(), reports); - - // Log the block report processing stats from Datanode perspective - long brSendCost = now() - brSendStartTime; - long brCreateCost = brSendStartTime - brCreateStartTime; - dn.getMetrics().addBlockReport(brSendCost); - LOG.info("BlockReport of " + totalBlockCount - + " blocks took " + brCreateCost + " msec to generate and " - + brSendCost + " msecs for RPC and NN processing"); - - // If we have sent the first block report, then wait a random - // time before we start the periodic block reports. - if (resetBlockReportTime) { - lastBlockReport = startTime - DFSUtil.getRandom().nextInt((int)(dnConf.blockReportInterval)); - resetBlockReportTime = false; - } else { - /* say the last block report was at 8:20:14. The current report - * should have started around 9:20:14 (default 1 hour interval). - * If current time is : - * 1) normal like 9:20:18, next report should be at 10:20:14 - * 2) unexpected like 11:35:43, next report should be at 12:20:14 - */ - lastBlockReport += (now() - lastBlockReport) / - dnConf.blockReportInterval * dnConf.blockReportInterval; - } - LOG.info("sent block report, processed command:" + cmd); + final long startTime = now(); + if (startTime - lastBlockReport <= dnConf.blockReportInterval) { + return null; } - return cmd; + + ArrayList cmds = new ArrayList(); + + // Flush any block information that precedes the block report. Otherwise + // we have a chance that we will miss the delHint information + // or we will report an RBW replica after the BlockReport already reports + // a FINALIZED one. + reportReceivedDeletedBlocks(); + lastDeletedReport = startTime; + + long brCreateStartTime = now(); + Map perVolumeBlockLists = + dn.getFSDataset().getBlockReports(bpos.getBlockPoolId()); + + // Convert the reports to the format expected by the NN. + int i = 0; + int totalBlockCount = 0; + StorageBlockReport reports[] = + new StorageBlockReport[perVolumeBlockLists.size()]; + + for(Map.Entry kvPair : perVolumeBlockLists.entrySet()) { + BlockListAsLongs blockList = kvPair.getValue(); + reports[i++] = new StorageBlockReport( + kvPair.getKey(), blockList.getBlockListAsLongs()); + totalBlockCount += blockList.getNumberOfBlocks(); + } + + // Send the reports to the NN. + int numReportsSent; + long brSendStartTime = now(); + if (totalBlockCount < dnConf.blockReportSplitThreshold) { + // Below split threshold, send all reports in a single message. + numReportsSent = 1; + DatanodeCommand cmd = + bpNamenode.blockReport(bpRegistration, bpos.getBlockPoolId(), reports); + if (cmd != null) { + cmds.add(cmd); + } + } else { + // Send one block report per message. + numReportsSent = i; + for (StorageBlockReport report : reports) { + StorageBlockReport singleReport[] = { report }; + DatanodeCommand cmd = bpNamenode.blockReport( + bpRegistration, bpos.getBlockPoolId(), singleReport); + if (cmd != null) { + cmds.add(cmd); + } + } + } + + // Log the block report processing stats from Datanode perspective + long brSendCost = now() - brSendStartTime; + long brCreateCost = brSendStartTime - brCreateStartTime; + dn.getMetrics().addBlockReport(brSendCost); + LOG.info("Sent " + numReportsSent + " blockreports " + totalBlockCount + + " blocks total. Took " + brCreateCost + + " msec to generate and " + brSendCost + + " msecs for RPC and NN processing. " + + " Got back commands " + + (cmds.size() == 0 ? "none" : Joiner.on("; ").join(cmds))); + + scheduleNextBlockReport(startTime); + return cmds.size() == 0 ? null : cmds; } - + + private void scheduleNextBlockReport(long previousReportStartTime) { + // If we have sent the first set of block reports, then wait a random + // time before we start the periodic block reports. + if (resetBlockReportTime) { + lastBlockReport = previousReportStartTime - + DFSUtil.getRandom().nextInt((int)(dnConf.blockReportInterval)); + resetBlockReportTime = false; + } else { + /* say the last block report was at 8:20:14. The current report + * should have started around 9:20:14 (default 1 hour interval). + * If current time is : + * 1) normal like 9:20:18, next report should be at 10:20:14 + * 2) unexpected like 11:35:43, next report should be at 12:20:14 + */ + lastBlockReport += (now() - lastBlockReport) / + dnConf.blockReportInterval * dnConf.blockReportInterval; + } + } + DatanodeCommand cacheReport() throws IOException { // If caching is disabled, do not send a cache report if (dn.getFSDataset().getCacheCapacity() == 0) { @@ -513,7 +536,7 @@ DatanodeCommand cacheReport() throws IOException { } // send cache report if timer has expired. DatanodeCommand cmd = null; - long startTime = Time.monotonicNow(); + final long startTime = Time.monotonicNow(); if (startTime - lastCacheReport > dnConf.cacheReportInterval) { if (LOG.isDebugEnabled()) { LOG.debug("Sending cacheReport from service actor: " + this); @@ -613,7 +636,7 @@ private void offerService() throws Exception { // while (shouldRun()) { try { - long startTime = now(); + final long startTime = now(); // // Every so often, send heartbeat or block-report @@ -659,10 +682,10 @@ private void offerService() throws Exception { lastDeletedReport = startTime; } - DatanodeCommand cmd = blockReport(); - processCommand(new DatanodeCommand[]{ cmd }); + List cmds = blockReport(); + processCommand(cmds == null ? null : cmds.toArray(new DatanodeCommand[cmds.size()])); - cmd = cacheReport(); + DatanodeCommand cmd = cacheReport(); processCommand(new DatanodeCommand[]{ cmd }); // Now safe to start scanning the block pool. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DNConf.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DNConf.java index 5d7afc7ecfd..73f3661182f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DNConf.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DNConf.java @@ -23,6 +23,8 @@ import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_INITIAL_DELAY_KEY; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_DEFAULT; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY; +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY; +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_SPLIT_THRESHOLD_DEFAULT; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CACHEREPORT_INTERVAL_MSEC_DEFAULT; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CACHEREPORT_INTERVAL_MSEC_KEY; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY; @@ -70,6 +72,7 @@ public class DNConf { final long readaheadLength; final long heartBeatInterval; final long blockReportInterval; + final long blockReportSplitThreshold; final long deleteReportInterval; final long initialBlockReportDelay; final long cacheReportInterval; @@ -117,6 +120,8 @@ public DNConf(Configuration conf) { DFSConfigKeys.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT); this.blockReportInterval = conf.getLong(DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, DFS_BLOCKREPORT_INTERVAL_MSEC_DEFAULT); + this.blockReportSplitThreshold = conf.getLong(DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY, + DFS_BLOCKREPORT_SPLIT_THRESHOLD_DEFAULT); this.cacheReportInterval = conf.getLong(DFS_CACHEREPORT_INTERVAL_MSEC_KEY, DFS_CACHEREPORT_INTERVAL_MSEC_DEFAULT); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java index af7262605ac..fa91f865ee3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeRpcServer.java @@ -982,13 +982,18 @@ public DatanodeCommand blockReport(DatanodeRegistration nodeReg, + "from " + nodeReg + ", reports.length=" + reports.length); } final BlockManager bm = namesystem.getBlockManager(); + boolean hasStaleStorages = true; for(StorageBlockReport r : reports) { final BlockListAsLongs blocks = new BlockListAsLongs(r.getBlocks()); - bm.processReport(nodeReg, r.getStorage(), poolId, blocks); + hasStaleStorages = bm.processReport(nodeReg, r.getStorage(), poolId, blocks); } - if (nn.getFSImage().isUpgradeFinalized() && !nn.isStandbyState()) + if (nn.getFSImage().isUpgradeFinalized() && + !nn.isStandbyState() && + !hasStaleStorages) { return new FinalizeCommand(poolId); + } + return null; } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml b/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml index 24f0b03c0b1..b0019ccb61d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml @@ -482,6 +482,20 @@ Delay for first block report in seconds. + + dfs.blockreport.split.threshold + 1000000 + If the number of blocks on the DataNode is below this + threshold then it will send block reports for all Storage Directories + in a single message. + + If the number of blocks exceeds this threshold then the DataNode will + send block reports for each Storage Directory in separate messages. + + Set to zero to always split. + + + dfs.datanode.directoryscan.interval 21600 diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReport.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/BlockReportTestBase.java similarity index 85% rename from hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReport.java rename to hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/BlockReportTestBase.java index d777697c21a..bf74b321c64 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReport.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/BlockReportTestBase.java @@ -52,7 +52,6 @@ import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; import org.apache.hadoop.hdfs.server.namenode.NameNode; -import org.apache.hadoop.hdfs.server.protocol.DatanodeCommand; import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; import org.apache.hadoop.hdfs.server.protocol.DatanodeStorage; import org.apache.hadoop.hdfs.server.protocol.StorageBlockReport; @@ -69,20 +68,16 @@ import org.mockito.invocation.InvocationOnMock; /** - * This test simulates a variety of situations when blocks are being - * intentionally corrupted, unexpectedly modified, and so on before a block - * report is happening. + * This is the base class for simulating a variety of situations + * when blocks are being intentionally corrupted, unexpectedly modified, + * and so on before a block report is happening. * - * For each test case it runs two variations: - * #1 - For a given DN, the first variation sends block reports for all - * storages in a single call to the NN. - * #2 - For a given DN, the second variation sends block reports for each - * storage in a separate call. - * - * The behavior should be the same in either variation. + * By overriding {@link #sendBlockReports}, derived classes can test + * different variations of how block reports are split across storages + * and messages. */ -public class TestBlockReport { - public static final Log LOG = LogFactory.getLog(TestBlockReport.class); +public abstract class BlockReportTestBase { + public static final Log LOG = LogFactory.getLog(BlockReportTestBase.class); private static short REPL_FACTOR = 1; private static final int RAND_LIMIT = 2000; @@ -91,12 +86,11 @@ public class TestBlockReport { private static final int DN_N0 = 0; private static final int FILE_START = 0; - static final int BLOCK_SIZE = 1024; - static final int NUM_BLOCKS = 10; - static final int FILE_SIZE = NUM_BLOCKS * BLOCK_SIZE + 1; - static String bpid; + private static final int BLOCK_SIZE = 1024; + private static final int NUM_BLOCKS = 10; + private static final int FILE_SIZE = NUM_BLOCKS * BLOCK_SIZE + 1; - private MiniDFSCluster cluster; + protected MiniDFSCluster cluster; private DistributedFileSystem fs; private static Random rand = new Random(RAND_LIMIT); @@ -112,8 +106,7 @@ public class TestBlockReport { public void startUpCluster() throws IOException { REPL_FACTOR = 1; //Reset if case a test has modified the value cluster = new MiniDFSCluster.Builder(conf).numDataNodes(REPL_FACTOR).build(); - fs = (DistributedFileSystem) cluster.getFileSystem(); - bpid = cluster.getNamesystem().getBlockPoolId(); + fs = cluster.getFileSystem(); } @After @@ -123,6 +116,15 @@ public void shutDownCluster() throws IOException { cluster.shutdown(); } + protected static void resetConfiguration() { + conf = new Configuration(); + int customPerChecksumSize = 512; + int customBlockSize = customPerChecksumSize * 3; + conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, customPerChecksumSize); + conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, customBlockSize); + conf.setLong(DFSConfigKeys.DFS_DATANODE_DIRECTORYSCAN_INTERVAL_KEY, DN_RESCAN_INTERVAL); + } + // Generate a block report, optionally corrupting the generation // stamp and/or length of one block. private static StorageBlockReport[] getBlockReports( @@ -172,106 +174,11 @@ private static StorageBlockReport[] getBlockReports( * @param dnR * @param poolId * @param reports - * @param needtoSplit * @throws IOException */ - private void sendBlockReports(DatanodeRegistration dnR, String poolId, - StorageBlockReport[] reports, boolean needtoSplit) throws IOException { - if (!needtoSplit) { - LOG.info("Sending combined block reports for " + dnR); - cluster.getNameNodeRpc().blockReport(dnR, poolId, reports); - } else { - for (StorageBlockReport report : reports) { - LOG.info("Sending block report for storage " + report.getStorage().getStorageID()); - StorageBlockReport[] singletonReport = { report }; - cluster.getNameNodeRpc().blockReport(dnR, poolId, singletonReport); - } - } - } + protected abstract void sendBlockReports(DatanodeRegistration dnR, String poolId, + StorageBlockReport[] reports) throws IOException; - /** - * Test variations blockReport_01 through blockReport_09 with combined - * and split block reports. - */ - @Test - public void blockReportCombined_01() throws IOException { - blockReport_01(false); - } - - @Test - public void blockReportSplit_01() throws IOException { - blockReport_01(true); - } - - @Test - public void blockReportCombined_02() throws IOException { - blockReport_02(false); - } - - @Test - public void blockReportSplit_02() throws IOException { - blockReport_02(true); - } - - @Test - public void blockReportCombined_03() throws IOException { - blockReport_03(false); - } - - @Test - public void blockReportSplit_03() throws IOException { - blockReport_03(true); - } - - @Test - public void blockReportCombined_04() throws IOException { - blockReport_04(false); - } - - @Test - public void blockReportSplit_04() throws IOException { - blockReport_04(true); - } - - @Test - public void blockReportCombined_06() throws Exception { - blockReport_06(false); - } - - @Test - public void blockReportSplit_06() throws Exception { - blockReport_06(true); - } - - @Test - public void blockReportCombined_07() throws Exception { - blockReport_07(false); - } - - @Test - public void blockReportSplit_07() throws Exception { - blockReport_07(true); - } - - @Test - public void blockReportCombined_08() throws Exception { - blockReport_08(false); - } - - @Test - public void blockReportSplit_08() throws Exception { - blockReport_08(true); - } - - @Test - public void blockReportCombined_09() throws Exception { - blockReport_09(false); - } - - @Test - public void blockReportSplit_09() throws Exception { - blockReport_09(true); - } /** * Test write a file, verifies and closes it. Then the length of the blocks * are messed up and BlockReport is forced. @@ -279,7 +186,8 @@ public void blockReportSplit_09() throws Exception { * * @throws java.io.IOException on an error */ - private void blockReport_01(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_01() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); @@ -312,7 +220,7 @@ private void blockReport_01(boolean splitBlockReports) throws IOException { String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, false, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); List blocksAfterReport = DFSTestUtil.getAllBlocks(fs.open(filePath)); @@ -338,7 +246,8 @@ private void blockReport_01(boolean splitBlockReports) throws IOException { * * @throws IOException in case of errors */ - private void blockReport_02(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_02() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); LOG.info("Running test " + METHOD_NAME); @@ -393,7 +302,7 @@ private void blockReport_02(boolean splitBlockReports) throws IOException { String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn0.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn0, poolId, false, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); BlockManagerTestUtil.getComputedDatanodeWork(cluster.getNamesystem() .getBlockManager()); @@ -414,17 +323,18 @@ private void blockReport_02(boolean splitBlockReports) throws IOException { * * @throws IOException in case of an error */ - private void blockReport_03(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_03() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); ArrayList blocks = writeFile(METHOD_NAME, FILE_SIZE, filePath); - + // all blocks belong to the same file, hence same BP DataNode dn = cluster.getDataNodes().get(DN_N0); String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, true, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertThat("Wrong number of corrupt blocks", @@ -441,7 +351,8 @@ private void blockReport_03(boolean splitBlockReports) throws IOException { * * @throws IOException in case of an error */ - private void blockReport_04(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_04() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); DFSTestUtil.createFile(fs, filePath, @@ -459,7 +370,7 @@ private void blockReport_04(boolean splitBlockReports) throws IOException { DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, false, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertThat("Wrong number of corrupt blocks", @@ -476,7 +387,8 @@ private void blockReport_04(boolean splitBlockReports) throws IOException { * * @throws IOException in case of an error */ - private void blockReport_06(boolean splitBlockReports) throws Exception { + @Test(timeout=300000) + public void blockReport_06() throws Exception { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); final int DN_N1 = DN_N0 + 1; @@ -489,7 +401,7 @@ private void blockReport_06(boolean splitBlockReports) throws Exception { String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, false, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertEquals("Wrong number of PendingReplication Blocks", 0, cluster.getNamesystem().getUnderReplicatedBlocks()); @@ -508,7 +420,8 @@ private void blockReport_06(boolean splitBlockReports) throws Exception { * * @throws IOException in case of an error */ - private void blockReport_07(boolean splitBlockReports) throws Exception { + @Test(timeout=300000) + public void blockReport_07() throws Exception { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); final int DN_N1 = DN_N0 + 1; @@ -522,7 +435,7 @@ private void blockReport_07(boolean splitBlockReports) throws Exception { String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, true, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertThat("Wrong number of corrupt blocks", @@ -533,7 +446,7 @@ private void blockReport_07(boolean splitBlockReports) throws Exception { cluster.getNamesystem().getPendingReplicationBlocks(), is(0L)); reports = getBlockReports(dn, poolId, true, true); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertThat("Wrong number of corrupt blocks", @@ -559,7 +472,8 @@ private void blockReport_07(boolean splitBlockReports) throws Exception { * * @throws IOException in case of an error */ - private void blockReport_08(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_08() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); final int DN_N1 = DN_N0 + 1; @@ -578,13 +492,13 @@ private void blockReport_08(boolean splitBlockReports) throws IOException { bc.start(); waitForTempReplica(bl, DN_N1); - + // all blocks belong to the same file, hence same BP DataNode dn = cluster.getDataNodes().get(DN_N1); String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, false, false); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertEquals("Wrong number of PendingReplication blocks", blocks.size(), cluster.getNamesystem().getPendingReplicationBlocks()); @@ -600,7 +514,8 @@ private void blockReport_08(boolean splitBlockReports) throws IOException { // Similar to BlockReport_08 but corrupts GS and len of the TEMPORARY's // replica block. Expect the same behaviour: NN should simply ignore this // block - private void blockReport_09(boolean splitBlockReports) throws IOException { + @Test(timeout=300000) + public void blockReport_09() throws IOException { final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); final int DN_N1 = DN_N0 + 1; @@ -620,17 +535,17 @@ private void blockReport_09(boolean splitBlockReports) throws IOException { bc.start(); waitForTempReplica(bl, DN_N1); - + // all blocks belong to the same file, hence same BP DataNode dn = cluster.getDataNodes().get(DN_N1); String poolId = cluster.getNamesystem().getBlockPoolId(); DatanodeRegistration dnR = dn.getDNRegistrationForBP(poolId); StorageBlockReport[] reports = getBlockReports(dn, poolId, true, true); - sendBlockReports(dnR, poolId, reports, splitBlockReports); + sendBlockReports(dnR, poolId, reports); printStats(); assertEquals("Wrong number of PendingReplication blocks", 2, cluster.getNamesystem().getPendingReplicationBlocks()); - + try { bc.join(); } catch (InterruptedException e) {} @@ -638,7 +553,7 @@ private void blockReport_09(boolean splitBlockReports) throws IOException { resetConfiguration(); // return the initial state of the configuration } } - + /** * Test for the case where one of the DNs in the pipeline is in the * process of doing a block report exactly when the block is closed. @@ -648,7 +563,7 @@ private void blockReport_09(boolean splitBlockReports) throws IOException { * corrupt. * This is a regression test for HDFS-2791. */ - @Test + @Test(timeout=300000) public void testOneReplicaRbwReportArrivesAfterBlockCompleted() throws Exception { final CountDownLatch brFinished = new CountDownLatch(1); DelayAnswer delayer = new GenericTestUtils.DelayAnswer(LOG) { @@ -663,7 +578,7 @@ protected Object passThrough(InvocationOnMock invocation) } } }; - + final String METHOD_NAME = GenericTestUtils.getMethodName(); Path filePath = new Path("/" + METHOD_NAME + ".dat"); @@ -671,9 +586,9 @@ protected Object passThrough(InvocationOnMock invocation) // what happens when one of the DNs is slowed for some reason. REPL_FACTOR = 2; startDNandWait(null, false); - + NameNode nn = cluster.getNameNode(); - + FSDataOutputStream out = fs.create(filePath, REPL_FACTOR); try { AppendTestUtil.write(out, 0, 10); @@ -684,19 +599,19 @@ protected Object passThrough(InvocationOnMock invocation) DataNode dn = cluster.getDataNodes().get(0); DatanodeProtocolClientSideTranslatorPB spy = DataNodeTestUtils.spyOnBposToNN(dn, nn); - + Mockito.doAnswer(delayer) .when(spy).blockReport( Mockito.anyObject(), Mockito.anyString(), Mockito.anyObject()); - + // Force a block report to be generated. The block report will have // an RBW replica in it. Wait for the RPC to be sent, but block // it before it gets to the NN. dn.scheduleAllBlockReport(0); delayer.waitForCall(); - + } finally { IOUtils.closeStream(out); } @@ -705,22 +620,22 @@ protected Object passThrough(InvocationOnMock invocation) // state. delayer.proceed(); brFinished.await(); - + // Verify that no replicas are marked corrupt, and that the // file is still readable. BlockManagerTestUtil.updateState(nn.getNamesystem().getBlockManager()); assertEquals(0, nn.getNamesystem().getCorruptReplicaBlocks()); DFSTestUtil.readFile(fs, filePath); - + // Ensure that the file is readable even from the DN that we futzed with. cluster.stopDataNode(1); - DFSTestUtil.readFile(fs, filePath); + DFSTestUtil.readFile(fs, filePath); } private void waitForTempReplica(Block bl, int DN_N1) throws IOException { final boolean tooLongWait = false; final int TIMEOUT = 40000; - + if(LOG.isDebugEnabled()) { LOG.debug("Wait for datanode " + DN_N1 + " to appear"); } @@ -731,7 +646,7 @@ private void waitForTempReplica(Block bl, int DN_N1) throws IOException { LOG.debug("Total number of DNs " + cluster.getDataNodes().size()); } cluster.waitActive(); - + // Look about specified DN for the replica of the block from 1st DN final DataNode dn1 = cluster.getDataNodes().get(DN_N1); String bpid = cluster.getNamesystem().getBlockPoolId(); @@ -789,7 +704,7 @@ private ArrayList writeFile(final String METHOD_NAME, return blocks; } - private void startDNandWait(Path filePath, boolean waitReplicas) + private void startDNandWait(Path filePath, boolean waitReplicas) throws IOException, InterruptedException, TimeoutException { if (LOG.isDebugEnabled()) { LOG.debug("Before next DN start: " + cluster.getDataNodes().size()); @@ -802,7 +717,7 @@ private void startDNandWait(Path filePath, boolean waitReplicas) if (LOG.isDebugEnabled()) { int lastDn = datanodes.size() - 1; LOG.debug("New datanode " - + cluster.getDataNodes().get(lastDn).getDisplayName() + + cluster.getDataNodes().get(lastDn).getDisplayName() + " has been started"); } if (waitReplicas) { @@ -898,7 +813,7 @@ private static void initLoggers() { ((Log4JLogger) NameNode.stateChangeLog).getLogger().setLevel(Level.ALL); ((Log4JLogger) LogFactory.getLog(FSNamesystem.class)).getLogger().setLevel(Level.ALL); ((Log4JLogger) DataNode.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger) TestBlockReport.LOG).getLogger().setLevel(Level.ALL); + ((Log4JLogger) BlockReportTestBase.LOG).getLogger().setLevel(Level.ALL); } private Block findBlock(Path path, long size) throws IOException { @@ -918,11 +833,11 @@ private Block findBlock(Path path, long size) throws IOException { private class BlockChecker extends Thread { Path filePath; - + public BlockChecker(final Path filePath) { this.filePath = filePath; } - + @Override public void run() { try { @@ -933,13 +848,4 @@ public void run() { } } } - - private static void resetConfiguration() { - conf = new Configuration(); - int customPerChecksumSize = 512; - int customBlockSize = customPerChecksumSize * 3; - conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, customPerChecksumSize); - conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, customBlockSize); - conf.setLong(DFSConfigKeys.DFS_DATANODE_DIRECTORYSCAN_INTERVAL_KEY, DN_RESCAN_INTERVAL); - } } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDnRespectsBlockReportSplitThreshold.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDnRespectsBlockReportSplitThreshold.java new file mode 100644 index 00000000000..989c33d2f09 --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDnRespectsBlockReportSplitThreshold.java @@ -0,0 +1,205 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hdfs.server.datanode; + +import java.io.IOException; +import java.util.List; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hdfs.*; +import org.apache.hadoop.hdfs.protocol.BlockListAsLongs; +import org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB; +import org.apache.hadoop.hdfs.server.namenode.NameNode; +import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; +import org.apache.hadoop.hdfs.server.protocol.StorageBlockReport; +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY; +import org.apache.hadoop.test.GenericTestUtils; + +import org.junit.After; +import org.junit.Test; +import org.mockito.ArgumentCaptor; +import org.mockito.Mockito; + +import static org.hamcrest.core.Is.is; +import static org.junit.Assert.assertThat; +import static org.mockito.Matchers.*; +import static org.mockito.Mockito.times; + +/** + * Tests that the DataNode respects + * {@link DFSConfigKeys#DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY} + */ +public class TestDnRespectsBlockReportSplitThreshold { + public static final Log LOG = LogFactory.getLog(TestStorageReport.class); + + private static final int BLOCK_SIZE = 1024; + private static final short REPL_FACTOR = 1; + private static final long seed = 0xFEEDFACE; + private static final int BLOCKS_IN_FILE = 5; + + private static Configuration conf; + private MiniDFSCluster cluster; + private DistributedFileSystem fs; + static String bpid; + + public void startUpCluster(long splitThreshold) throws IOException { + conf = new HdfsConfiguration(); + conf.setLong(DFS_BLOCKREPORT_SPLIT_THRESHOLD_KEY, splitThreshold); + cluster = new MiniDFSCluster.Builder(conf) + .numDataNodes(REPL_FACTOR) + .build(); + fs = cluster.getFileSystem(); + bpid = cluster.getNamesystem().getBlockPoolId(); + } + + @After + public void shutDownCluster() throws IOException { + if (cluster != null) { + fs.close(); + cluster.shutdown(); + cluster = null; + } + } + + private void createFile(String filenamePrefix, int blockCount) + throws IOException { + Path path = new Path("/" + filenamePrefix + ".dat"); + DFSTestUtil.createFile(fs, path, BLOCK_SIZE, + blockCount * BLOCK_SIZE, BLOCK_SIZE, REPL_FACTOR, seed); + } + + private void verifyCapturedArguments( + ArgumentCaptor captor, + int expectedReportsPerCall, + int expectedTotalBlockCount) { + + List listOfReports = captor.getAllValues(); + int numBlocksReported = 0; + for (StorageBlockReport[] reports : listOfReports) { + assertThat(reports.length, is(expectedReportsPerCall)); + + for (StorageBlockReport report : reports) { + BlockListAsLongs blockList = new BlockListAsLongs(report.getBlocks()); + numBlocksReported += blockList.getNumberOfBlocks(); + } + } + + assert(numBlocksReported >= expectedTotalBlockCount); + } + + /** + * Test that if splitThreshold is zero, then we always get a separate + * call per storage. + */ + @Test(timeout=300000) + public void testAlwaysSplit() throws IOException, InterruptedException { + startUpCluster(0); + NameNode nn = cluster.getNameNode(); + DataNode dn = cluster.getDataNodes().get(0); + + // Create a file with a few blocks. + createFile(GenericTestUtils.getMethodName(), BLOCKS_IN_FILE); + + // Insert a spy object for the NN RPC. + DatanodeProtocolClientSideTranslatorPB nnSpy = + DataNodeTestUtils.spyOnBposToNN(dn, nn); + + // Trigger a block report so there is an interaction with the spy + // object. + DataNodeTestUtils.triggerBlockReport(dn); + + ArgumentCaptor captor = + ArgumentCaptor.forClass(StorageBlockReport[].class); + + Mockito.verify(nnSpy, times(MiniDFSCluster.DIRS_PER_DATANODE)).blockReport( + any(DatanodeRegistration.class), + anyString(), + captor.capture()); + + verifyCapturedArguments(captor, 1, BLOCKS_IN_FILE); + } + + /** + * Tests the behavior when the count of blocks is exactly one less than + * the threshold. + */ + @Test(timeout=300000) + public void testCornerCaseUnderThreshold() throws IOException, InterruptedException { + startUpCluster(BLOCKS_IN_FILE + 1); + NameNode nn = cluster.getNameNode(); + DataNode dn = cluster.getDataNodes().get(0); + + // Create a file with a few blocks. + createFile(GenericTestUtils.getMethodName(), BLOCKS_IN_FILE); + + // Insert a spy object for the NN RPC. + DatanodeProtocolClientSideTranslatorPB nnSpy = + DataNodeTestUtils.spyOnBposToNN(dn, nn); + + // Trigger a block report so there is an interaction with the spy + // object. + DataNodeTestUtils.triggerBlockReport(dn); + + ArgumentCaptor captor = + ArgumentCaptor.forClass(StorageBlockReport[].class); + + Mockito.verify(nnSpy, times(1)).blockReport( + any(DatanodeRegistration.class), + anyString(), + captor.capture()); + + verifyCapturedArguments(captor, MiniDFSCluster.DIRS_PER_DATANODE, BLOCKS_IN_FILE); + } + + /** + * Tests the behavior when the count of blocks is exactly equal to the + * threshold. + */ + @Test(timeout=300000) + public void testCornerCaseAtThreshold() throws IOException, InterruptedException { + startUpCluster(BLOCKS_IN_FILE); + NameNode nn = cluster.getNameNode(); + DataNode dn = cluster.getDataNodes().get(0); + + // Create a file with a few blocks. + createFile(GenericTestUtils.getMethodName(), BLOCKS_IN_FILE); + + // Insert a spy object for the NN RPC. + DatanodeProtocolClientSideTranslatorPB nnSpy = + DataNodeTestUtils.spyOnBposToNN(dn, nn); + + // Trigger a block report so there is an interaction with the spy + // object. + DataNodeTestUtils.triggerBlockReport(dn); + + ArgumentCaptor captor = + ArgumentCaptor.forClass(StorageBlockReport[].class); + + Mockito.verify(nnSpy, times(MiniDFSCluster.DIRS_PER_DATANODE)).blockReport( + any(DatanodeRegistration.class), + anyString(), + captor.capture()); + + verifyCapturedArguments(captor, 1, BLOCKS_IN_FILE); + } + +} diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesBlockReportPerStorage.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesBlockReportPerStorage.java new file mode 100644 index 00000000000..1b03786fa39 --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesBlockReportPerStorage.java @@ -0,0 +1,42 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hdfs.server.datanode; + +import java.io.IOException; + +import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; +import org.apache.hadoop.hdfs.server.protocol.StorageBlockReport; + + +/** + * Runs all tests in BlockReportTestBase, sending one block per storage. + * This is the default DataNode behavior post HDFS-2832. + */ +public class TestNNHandlesBlockReportPerStorage extends BlockReportTestBase { + + @Override + protected void sendBlockReports(DatanodeRegistration dnR, String poolId, + StorageBlockReport[] reports) throws IOException { + for (StorageBlockReport report : reports) { + LOG.info("Sending block report for storage " + report.getStorage().getStorageID()); + StorageBlockReport[] singletonReport = { report }; + cluster.getNameNodeRpc().blockReport(dnR, poolId, singletonReport); + } + } +} diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesCombinedBlockReport.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesCombinedBlockReport.java new file mode 100644 index 00000000000..036b550c668 --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestNNHandlesCombinedBlockReport.java @@ -0,0 +1,39 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hdfs.server.datanode; + +import java.io.IOException; + +import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; +import org.apache.hadoop.hdfs.server.protocol.StorageBlockReport; + +/** + * Runs all tests in BlockReportTestBase, sending one block report + * per DataNode. This tests that the NN can handle the legacy DN + * behavior where it presents itself as a single logical storage. + */ +public class TestNNHandlesCombinedBlockReport extends BlockReportTestBase { + + @Override + protected void sendBlockReports(DatanodeRegistration dnR, String poolId, + StorageBlockReport[] reports) throws IOException { + LOG.info("Sending combined block reports for " + dnR); + cluster.getNameNodeRpc().blockReport(dnR, poolId, reports); + } +} From 1f7dd7811a1152d0a798524e831586eebbe110b6 Mon Sep 17 00:00:00 2001 From: Jing Zhao Date: Fri, 31 Jan 2014 23:00:34 +0000 Subject: [PATCH 23/23] HDFS-5804. HDFS NFS Gateway fails to mount and proxy when using Kerberos. Contributed by Abin Shahab. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1563323 13f79535-47bb-0310-9956-ffa450edef68 --- .../hadoop/hdfs/nfs/nfs3/DFSClientCache.java | 28 +++++++++++++++++-- .../hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java | 12 ++++---- .../apache/hadoop/hdfs/nfs/TestReaddir.java | 14 +++++++++- .../hdfs/nfs/nfs3/TestDFSClientCache.java | 25 +++++++++++++++++ .../hadoop/hdfs/nfs/nfs3/TestWrites.java | 9 ++++++ hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 ++ 6 files changed, 83 insertions(+), 8 deletions(-) diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/DFSClientCache.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/DFSClientCache.java index 174496557e4..b6822c41142 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/DFSClientCache.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/DFSClientCache.java @@ -26,6 +26,7 @@ import java.util.concurrent.ExecutionException; import java.util.concurrent.TimeUnit; +import com.google.common.base.Preconditions; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -163,8 +164,9 @@ private CacheLoader clientLoader() { return new CacheLoader() { @Override public DFSClient load(String userName) throws Exception { - UserGroupInformation ugi = UserGroupInformation - .createRemoteUser(userName); + UserGroupInformation ugi = getUserGroupInformation( + userName, + UserGroupInformation.getCurrentUser()); // Guava requires CacheLoader never returns null. return ugi.doAs(new PrivilegedExceptionAction() { @@ -177,6 +179,28 @@ public DFSClient run() throws IOException { }; } + /** + * This method uses the currentUser, and real user to create a proxy + * @param effectiveUser The user who is being proxied by the real user + * @param realUser The actual user who does the command + * @return Proxy UserGroupInformation + * @throws IOException If proxying fails + */ + UserGroupInformation getUserGroupInformation( + String effectiveUser, + UserGroupInformation realUser) + throws IOException { + Preconditions.checkNotNull(effectiveUser); + Preconditions.checkNotNull(realUser); + UserGroupInformation ugi = + UserGroupInformation.createProxyUser(effectiveUser, realUser); + if (LOG.isDebugEnabled()){ + LOG.debug(String.format("Created ugi:" + + " %s for username: %s", ugi, effectiveUser)); + } + return ugi; + } + private RemovalListener clientRemovalListener() { return new RemovalListener() { @Override diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java index 17670a9afe8..c2fc70990ca 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java @@ -479,9 +479,9 @@ public ACCESS3Response access(XDR xdr, SecurityHandler securityHandler, } try { - // Use superUserClient to get file attr since we don't know whether the - // NFS client user has access permission to the file - attrs = writeManager.getFileAttr(superUserClient, handle, iug); + // HDFS-5804 removed supserUserClient access + attrs = writeManager.getFileAttr(dfsClient, handle, iug); + if (attrs == null) { LOG.error("Can't get path for fileId:" + handle.getFileId()); return new ACCESS3Response(Nfs3Status.NFS3ERR_STALE); @@ -603,8 +603,10 @@ public READ3Response read(XDR xdr, SecurityHandler securityHandler, // Only do access check. try { // Don't read from cache. Client may not have read permission. - attrs = Nfs3Utils.getFileAttr(superUserClient, - Nfs3Utils.getFileIdPath(handle), iug); + attrs = Nfs3Utils.getFileAttr( + dfsClient, + Nfs3Utils.getFileIdPath(handle), + iug); } catch (IOException e) { if (LOG.isDebugEnabled()) { LOG.debug("Get error accessing file, fileId:" + handle.getFileId()); diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/TestReaddir.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/TestReaddir.java index e1493b60ea2..1ca43bc8abb 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/TestReaddir.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/TestReaddir.java @@ -22,6 +22,7 @@ import java.io.IOException; import java.net.InetAddress; +import java.util.Arrays; import java.util.List; import org.apache.hadoop.conf.Configuration; @@ -40,6 +41,9 @@ import org.apache.hadoop.nfs.nfs3.response.READDIRPLUS3Response.EntryPlus3; import org.apache.hadoop.oncrpc.XDR; import org.apache.hadoop.oncrpc.security.SecurityHandler; +import org.apache.hadoop.security.UserGroupInformation; +import org.apache.hadoop.security.authorize.ProxyUsers; +import org.apache.hadoop.util.StringUtils; import org.junit.AfterClass; import org.junit.Before; import org.junit.BeforeClass; @@ -58,9 +62,17 @@ public class TestReaddir { static RpcProgramNfs3 nfsd; static String testdir = "/tmp"; static SecurityHandler securityHandler; - + @BeforeClass public static void setup() throws Exception { + String currentUser = System.getProperty("user.name"); + config.set( + ProxyUsers.getProxySuperuserGroupConfKey(currentUser), + "*"); + config.set( + ProxyUsers.getProxySuperuserIpConfKey(currentUser), + "*"); + ProxyUsers.refreshSuperUserGroupsConfiguration(config); cluster = new MiniDFSCluster.Builder(config).numDataNodes(1).build(); cluster.waitActive(); hdfs = cluster.getFileSystem(); diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestDFSClientCache.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestDFSClientCache.java index 360bb145fee..face4fe6532 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestDFSClientCache.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestDFSClientCache.java @@ -20,12 +20,15 @@ import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; +import static org.junit.Assert.assertThat; +import static org.hamcrest.core.Is.is; import java.io.IOException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hdfs.DFSClient; +import org.apache.hadoop.security.UserGroupInformation; import org.junit.Test; public class TestDFSClientCache { @@ -49,6 +52,28 @@ public void testEviction() throws IOException { assertEquals(MAX_CACHE_SIZE - 1, cache.clientCache.size()); } + @Test + public void testGetUserGroupInformation() throws IOException { + String userName = "user1"; + String currentUser = "currentUser"; + + UserGroupInformation currentUserUgi = UserGroupInformation + .createUserForTesting(currentUser, new String[0]); + currentUserUgi.setAuthenticationMethod( + UserGroupInformation.AuthenticationMethod.KERBEROS); + Configuration conf = new Configuration(); + conf.set(FileSystem.FS_DEFAULT_NAME_KEY, "hdfs://localhost"); + DFSClientCache cache = new DFSClientCache(conf); + UserGroupInformation ugiResult + = cache.getUserGroupInformation(userName, currentUserUgi); + + assertThat(ugiResult.getUserName(), is(userName)); + assertThat(ugiResult.getRealUser(), is(currentUserUgi)); + assertThat( + ugiResult.getAuthenticationMethod(), + is(UserGroupInformation.AuthenticationMethod.PROXY)); + } + private static boolean isDfsClientClose(DFSClient c) { try { c.exists(""); diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestWrites.java b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestWrites.java index 2ef614a1edf..4771204a252 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestWrites.java +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestWrites.java @@ -50,6 +50,7 @@ import org.apache.hadoop.nfs.nfs3.response.READ3Response; import org.apache.hadoop.oncrpc.XDR; import org.apache.hadoop.oncrpc.security.SecurityHandler; +import org.apache.hadoop.security.authorize.ProxyUsers; import org.jboss.netty.channel.Channel; import org.junit.Assert; import org.junit.Test; @@ -285,6 +286,14 @@ public void testWriteStableHow() throws IOException, InterruptedException { SecurityHandler securityHandler = Mockito.mock(SecurityHandler.class); Mockito.when(securityHandler.getUser()).thenReturn( System.getProperty("user.name")); + String currentUser = System.getProperty("user.name"); + config.set( + ProxyUsers.getProxySuperuserGroupConfKey(currentUser), + "*"); + config.set( + ProxyUsers.getProxySuperuserIpConfKey(currentUser), + "*"); + ProxyUsers.refreshSuperUserGroupsConfiguration(config); try { cluster = new MiniDFSCluster.Builder(config).numDataNodes(1).build(); diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 7b21e11d852..6ae103f7c42 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -300,6 +300,9 @@ Release 2.4.0 - UNRELEASED HDFS-5153. Datanode should send block reports for each storage in a separate message. (Arpit Agarwal) + HDFS-5804. HDFS NFS Gateway fails to mount and proxy when using Kerberos. + (Abin Shahab via jing9) + OPTIMIZATIONS HDFS-5790. LeaseManager.findPath is very slow when many leases need recovery