From d7c48409e339d9964ec7681b91a076fc5bfa2c87 Mon Sep 17 00:00:00 2001 From: Alejandro Abdelnur Date: Thu, 10 Jan 2013 17:27:58 +0000 Subject: [PATCH] MAPREDUCE-4907. Ammendment, forgot to svn add testcase in original commit git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1431510 13f79535-47bb-0310-9956-ffa450edef68 --- .../TestClientDistributedCacheManager.java | 114 ++++++++++++++++++ 1 file changed, 114 insertions(+) create mode 100644 hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java new file mode 100644 index 00000000000..4824ba39e64 --- /dev/null +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java @@ -0,0 +1,114 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.mapreduce.filecache; + +import java.io.File; +import java.io.IOException; +import java.net.URI; +import java.util.HashMap; +import java.util.Map; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.FileStatus; +import org.apache.hadoop.fs.FileSystem; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.io.SequenceFile; +import org.apache.hadoop.io.Text; +import org.apache.hadoop.io.SequenceFile.CompressionType; +import org.apache.hadoop.mapreduce.Job; +import org.apache.hadoop.mapreduce.MRJobConfig; +import org.junit.After; +import org.junit.Assert; +import org.junit.Before; +import org.junit.Test; + +public class TestClientDistributedCacheManager { + private static final Log LOG = LogFactory.getLog( + TestClientDistributedCacheManager.class); + + private static final String TEST_ROOT_DIR = + new File(System.getProperty("test.build.data", "/tmp")).toURI() + .toString().replace(' ', '+'); + + private FileSystem fs; + private Path firstCacheFile; + private Path secondCacheFile; + private Configuration conf; + + @Before + public void setup() throws IOException { + conf = new Configuration(); + fs = FileSystem.get(conf); + firstCacheFile = new Path(TEST_ROOT_DIR, "firstcachefile"); + secondCacheFile = new Path(TEST_ROOT_DIR, "secondcachefile"); + createTempFile(firstCacheFile, conf); + createTempFile(secondCacheFile, conf); + } + + @After + public void tearDown() throws IOException { + if (!fs.delete(firstCacheFile, false)) { + LOG.warn("Failed to delete firstcachefile"); + } + if (!fs.delete(secondCacheFile, false)) { + LOG.warn("Failed to delete secondcachefile"); + } + } + + @Test + public void testDetermineTimestamps() throws IOException { + Job job = Job.getInstance(conf); + job.addCacheFile(firstCacheFile.toUri()); + job.addCacheFile(secondCacheFile.toUri()); + Configuration jobConf = job.getConfiguration(); + + Map statCache = new HashMap(); + ClientDistributedCacheManager.determineTimestamps(jobConf, statCache); + + FileStatus firstStatus = statCache.get(firstCacheFile.toUri()); + FileStatus secondStatus = statCache.get(secondCacheFile.toUri()); + + Assert.assertNotNull(firstStatus); + Assert.assertNotNull(secondStatus); + Assert.assertEquals(2, statCache.size()); + String expected = firstStatus.getModificationTime() + "," + + secondStatus.getModificationTime(); + Assert.assertEquals(expected, jobConf.get(MRJobConfig.CACHE_FILE_TIMESTAMPS)); + } + + @SuppressWarnings("deprecation") + void createTempFile(Path p, Configuration conf) throws IOException { + SequenceFile.Writer writer = null; + try { + writer = SequenceFile.createWriter(fs, conf, p, + Text.class, Text.class, + CompressionType.NONE); + writer.append(new Text("text"), new Text("moretext")); + } catch(Exception e) { + throw new IOException(e.getLocalizedMessage()); + } finally { + if (writer != null) { + writer.close(); + } + writer = null; + } + LOG.info("created: " + p); + } +}