HDFS-3121. Add HDFS tests for HADOOP-8014 change. Contributed by John George. Missed adding the file in the earlier commit.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1310631 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
de3fb5c23a
commit
02f5e43686
|
@ -0,0 +1,176 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.fs.viewfs;
|
||||
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
|
||||
import javax.security.auth.login.LoginException;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.FileSystemTestHelper;
|
||||
import org.apache.hadoop.fs.FsConstants;
|
||||
import org.apache.hadoop.fs.ContentSummary;
|
||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||
import org.apache.hadoop.io.DataInputBuffer;
|
||||
import org.apache.hadoop.io.DataOutputBuffer;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.fs.FsServerDefaults;
|
||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_FILE_BUFFER_SIZE_DEFAULT;
|
||||
import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_FILE_BUFFER_SIZE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_SIZE_DEFAULT;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_SIZE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_DEFAULT;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_SYNCONCLOSE_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_HEARTBEAT_RECHECK_INTERVAL_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_REPLICATION_MIN_KEY;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_DEFAULT;
|
||||
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_REPLICATION_KEY;
|
||||
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import static org.junit.Assert.*;
|
||||
|
||||
/**
|
||||
* Tests for viewfs implementation of default fs level values.
|
||||
* This tests for both passing in a path (based on mount point)
|
||||
* to obtain the default value of the fs that the path is mounted on
|
||||
* or just passing in no arguments.
|
||||
*/
|
||||
public class TestViewFsDefaultValue {
|
||||
|
||||
static final String testFileDir = "/tmp/test/";
|
||||
static final String testFileName = testFileDir + "testFileStatusSerialziation";
|
||||
private static MiniDFSCluster cluster;
|
||||
private static Configuration CONF = new Configuration();
|
||||
private static FileSystem fHdfs;
|
||||
private static FileSystem vfs;
|
||||
private static Path testFilePath;
|
||||
private static Path testFileDirPath;
|
||||
|
||||
@BeforeClass
|
||||
public static void clusterSetupAtBegining() throws IOException,
|
||||
LoginException, URISyntaxException {
|
||||
|
||||
CONF.setLong(DFS_BLOCK_SIZE_KEY, DFS_BLOCK_SIZE_DEFAULT);
|
||||
CONF.setInt(DFS_BYTES_PER_CHECKSUM_KEY, DFS_BYTES_PER_CHECKSUM_DEFAULT);
|
||||
CONF.setInt(DFS_CLIENT_WRITE_PACKET_SIZE_KEY,
|
||||
DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT);
|
||||
CONF.setInt(DFS_REPLICATION_KEY, DFS_REPLICATION_DEFAULT + 1);
|
||||
CONF.setInt(IO_FILE_BUFFER_SIZE_KEY, IO_FILE_BUFFER_SIZE_DEFAULT);
|
||||
|
||||
cluster = new MiniDFSCluster.Builder(CONF).numDataNodes(DFS_REPLICATION_DEFAULT + 1).build();
|
||||
cluster.waitClusterUp();
|
||||
fHdfs = cluster.getFileSystem();
|
||||
FileSystemTestHelper.createFile(fHdfs, testFileName);
|
||||
Configuration conf = ViewFileSystemTestSetup.createConfig();
|
||||
ConfigUtil.addLink(conf, "/tmp", new URI(fHdfs.getUri().toString() +
|
||||
"/tmp"));
|
||||
vfs = FileSystem.get(FsConstants.VIEWFS_URI, conf);
|
||||
testFileDirPath = new Path (testFileDir);
|
||||
testFilePath = new Path (testFileName);
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Test that default blocksize values can be retrieved on the client side.
|
||||
*/
|
||||
@Test
|
||||
public void testGetDefaultBlockSize()
|
||||
throws IOException, URISyntaxException {
|
||||
// createFile does not use defaultBlockSize to create the file,
|
||||
// but we are only looking at the defaultBlockSize, so this
|
||||
// test should still pass
|
||||
try {
|
||||
vfs.getDefaultBlockSize();
|
||||
fail("getServerDefaults on viewFs did not throw excetion!");
|
||||
} catch (NotInMountpointException e) {
|
||||
assertEquals(vfs.getDefaultBlockSize(testFilePath),
|
||||
DFS_BLOCK_SIZE_DEFAULT);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that default replication values can be retrieved on the client side.
|
||||
*/
|
||||
@Test
|
||||
public void testGetDefaultReplication()
|
||||
throws IOException, URISyntaxException {
|
||||
try {
|
||||
vfs.getDefaultReplication();
|
||||
fail("getDefaultReplication on viewFs did not throw excetion!");
|
||||
} catch (NotInMountpointException e) {
|
||||
assertEquals(vfs.getDefaultReplication(testFilePath),
|
||||
DFS_REPLICATION_DEFAULT+1);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Test that server default values can be retrieved on the client side.
|
||||
*/
|
||||
@Test
|
||||
public void testServerDefaults() throws IOException {
|
||||
try {
|
||||
FsServerDefaults serverDefaults = vfs.getServerDefaults();
|
||||
fail("getServerDefaults on viewFs did not throw excetion!");
|
||||
} catch (NotInMountpointException e) {
|
||||
FsServerDefaults serverDefaults = vfs.getServerDefaults(testFilePath);
|
||||
assertEquals(DFS_BLOCK_SIZE_DEFAULT, serverDefaults.getBlockSize());
|
||||
assertEquals(DFS_BYTES_PER_CHECKSUM_DEFAULT,
|
||||
serverDefaults.getBytesPerChecksum());
|
||||
assertEquals(DFS_CLIENT_WRITE_PACKET_SIZE_DEFAULT,
|
||||
serverDefaults.getWritePacketSize());
|
||||
assertEquals(IO_FILE_BUFFER_SIZE_DEFAULT,
|
||||
serverDefaults.getFileBufferSize());
|
||||
assertEquals(DFS_REPLICATION_DEFAULT + 1,
|
||||
serverDefaults.getReplication());
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that getContentSummary can be retrieved on the client side.
|
||||
*/
|
||||
@Test
|
||||
public void testGetContentSummary() throws IOException {
|
||||
FileSystem hFs = cluster.getFileSystem(0);
|
||||
final DistributedFileSystem dfs = (DistributedFileSystem)hFs;
|
||||
dfs.setQuota(testFileDirPath, 100, 500);
|
||||
ContentSummary cs = vfs.getContentSummary(testFileDirPath);
|
||||
assertEquals(100, cs.getQuota());
|
||||
assertEquals(500, cs.getSpaceQuota());
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void cleanup() throws IOException {
|
||||
fHdfs.delete(new Path(testFileName), true);
|
||||
}
|
||||
|
||||
}
|
Loading…
Reference in New Issue