HDFS-2219. Change fsck to support fully qualified paths so that a particular namenode in a federated cluster with multiple namenodes can be specified in the path parameter.
This commit is contained in:
parent
446545c496
commit
7fe0f25ad2
|
@ -502,6 +502,10 @@ Release 2.7.0 - UNRELEASED
|
||||||
|
|
||||||
HDFS-7600. Refine hdfs admin classes to reuse common code. (jing9)
|
HDFS-7600. Refine hdfs admin classes to reuse common code. (jing9)
|
||||||
|
|
||||||
|
HDFS-2219. Change fsck to support fully qualified paths so that a
|
||||||
|
particular namenode in a federated cluster with multiple namenodes
|
||||||
|
can be specified in the path parameter. (szetszwo)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
HDFS-7454. Reduce memory footprint for AclEntries in NameNode.
|
HDFS-7454. Reduce memory footprint for AclEntries in NameNode.
|
||||||
|
|
|
@ -32,6 +32,7 @@ import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.conf.Configured;
|
import org.apache.hadoop.conf.Configured;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hdfs.DFSUtil;
|
import org.apache.hadoop.hdfs.DFSUtil;
|
||||||
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
import org.apache.hadoop.hdfs.DistributedFileSystem;
|
||||||
import org.apache.hadoop.hdfs.HAUtil;
|
import org.apache.hadoop.hdfs.HAUtil;
|
||||||
|
@ -229,14 +230,14 @@ public class DFSck extends Configured implements Tool {
|
||||||
* @return Returns http address or null if failure.
|
* @return Returns http address or null if failure.
|
||||||
* @throws IOException if we can't determine the active NN address
|
* @throws IOException if we can't determine the active NN address
|
||||||
*/
|
*/
|
||||||
private URI getCurrentNamenodeAddress() throws IOException {
|
private URI getCurrentNamenodeAddress(Path target) throws IOException {
|
||||||
//String nnAddress = null;
|
//String nnAddress = null;
|
||||||
Configuration conf = getConf();
|
Configuration conf = getConf();
|
||||||
|
|
||||||
//get the filesystem object to verify it is an HDFS system
|
//get the filesystem object to verify it is an HDFS system
|
||||||
FileSystem fs;
|
final FileSystem fs;
|
||||||
try {
|
try {
|
||||||
fs = FileSystem.get(conf);
|
fs = target.getFileSystem(conf);
|
||||||
} catch (IOException ioe) {
|
} catch (IOException ioe) {
|
||||||
System.err.println("FileSystem is inaccessible due to:\n"
|
System.err.println("FileSystem is inaccessible due to:\n"
|
||||||
+ StringUtils.stringifyException(ioe));
|
+ StringUtils.stringifyException(ioe));
|
||||||
|
@ -254,16 +255,6 @@ public class DFSck extends Configured implements Tool {
|
||||||
private int doWork(final String[] args) throws IOException {
|
private int doWork(final String[] args) throws IOException {
|
||||||
final StringBuilder url = new StringBuilder();
|
final StringBuilder url = new StringBuilder();
|
||||||
|
|
||||||
URI namenodeAddress = getCurrentNamenodeAddress();
|
|
||||||
if (namenodeAddress == null) {
|
|
||||||
//Error message already output in {@link #getCurrentNamenodeAddress()}
|
|
||||||
System.err.println("DFSck exiting.");
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
url.append(namenodeAddress.toString());
|
|
||||||
System.err.println("Connecting to namenode via " + url.toString());
|
|
||||||
|
|
||||||
url.append("/fsck?ugi=").append(ugi.getShortUserName());
|
url.append("/fsck?ugi=").append(ugi.getShortUserName());
|
||||||
String dir = null;
|
String dir = null;
|
||||||
boolean doListCorruptFileBlocks = false;
|
boolean doListCorruptFileBlocks = false;
|
||||||
|
@ -309,7 +300,20 @@ public class DFSck extends Configured implements Tool {
|
||||||
if (null == dir) {
|
if (null == dir) {
|
||||||
dir = "/";
|
dir = "/";
|
||||||
}
|
}
|
||||||
url.append("&path=").append(URLEncoder.encode(dir, "UTF-8"));
|
|
||||||
|
final Path dirpath = new Path(dir);
|
||||||
|
final URI namenodeAddress = getCurrentNamenodeAddress(dirpath);
|
||||||
|
if (namenodeAddress == null) {
|
||||||
|
//Error message already output in {@link #getCurrentNamenodeAddress()}
|
||||||
|
System.err.println("DFSck exiting.");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
url.insert(0, namenodeAddress.toString());
|
||||||
|
url.append("&path=").append(URLEncoder.encode(
|
||||||
|
Path.getPathWithoutSchemeAndAuthority(dirpath).toString(), "UTF-8"));
|
||||||
|
System.err.println("Connecting to namenode via " + url.toString());
|
||||||
|
|
||||||
if (doListCorruptFileBlocks) {
|
if (doListCorruptFileBlocks) {
|
||||||
return listCorruptFileBlocks(dir, url.toString());
|
return listCorruptFileBlocks(dir, url.toString());
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,141 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.namenode;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.Random;
|
||||||
|
import java.util.concurrent.TimeoutException;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.commons.logging.impl.Log4JLogger;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.hdfs.HdfsConfiguration;
|
||||||
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
|
import org.apache.hadoop.hdfs.MiniDFSNNTopology;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.ClientProtocol;
|
||||||
|
import org.apache.hadoop.hdfs.server.balancer.TestBalancer;
|
||||||
|
import org.apache.log4j.Level;
|
||||||
|
import org.junit.Assert;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test fsck with multiple NameNodes
|
||||||
|
*/
|
||||||
|
public class TestFsckWithMultipleNameNodes {
|
||||||
|
static final Log LOG = LogFactory.getLog(TestFsckWithMultipleNameNodes.class);
|
||||||
|
{
|
||||||
|
((Log4JLogger)NameNode.stateChangeLog).getLogger().setLevel(Level.OFF);
|
||||||
|
((Log4JLogger)LeaseManager.LOG).getLogger().setLevel(Level.OFF);
|
||||||
|
((Log4JLogger)LogFactory.getLog(FSNamesystem.class)).getLogger().setLevel(Level.OFF);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
private static final String FILE_NAME = "/tmp.txt";
|
||||||
|
private static final Path FILE_PATH = new Path(FILE_NAME);
|
||||||
|
|
||||||
|
private static final Random RANDOM = new Random();
|
||||||
|
|
||||||
|
static {
|
||||||
|
TestBalancer.initTestSetup();
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Common objects used in various methods. */
|
||||||
|
private static class Suite {
|
||||||
|
final MiniDFSCluster cluster;
|
||||||
|
final ClientProtocol[] clients;
|
||||||
|
final short replication;
|
||||||
|
|
||||||
|
Suite(MiniDFSCluster cluster, final int nNameNodes, final int nDataNodes)
|
||||||
|
throws IOException {
|
||||||
|
this.cluster = cluster;
|
||||||
|
clients = new ClientProtocol[nNameNodes];
|
||||||
|
for(int i = 0; i < nNameNodes; i++) {
|
||||||
|
clients[i] = cluster.getNameNode(i).getRpcServer();
|
||||||
|
}
|
||||||
|
replication = (short)Math.max(1, nDataNodes - 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** create a file with a length of <code>fileLen</code> */
|
||||||
|
private void createFile(int index, long len
|
||||||
|
) throws IOException, InterruptedException, TimeoutException {
|
||||||
|
final FileSystem fs = cluster.getFileSystem(index);
|
||||||
|
DFSTestUtil.createFile(fs, FILE_PATH, len, replication, RANDOM.nextLong());
|
||||||
|
DFSTestUtil.waitReplication(fs, FILE_PATH, replication);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
private static Configuration createConf() {
|
||||||
|
final Configuration conf = new HdfsConfiguration();
|
||||||
|
conf.setLong(DFSConfigKeys.DFS_NAMENODE_ACCESSTIME_PRECISION_KEY, 1L);
|
||||||
|
conf.setLong(DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, 10000L);
|
||||||
|
return conf;
|
||||||
|
}
|
||||||
|
|
||||||
|
private void runTest(final int nNameNodes, final int nDataNodes,
|
||||||
|
Configuration conf) throws Exception {
|
||||||
|
LOG.info("nNameNodes=" + nNameNodes + ", nDataNodes=" + nDataNodes);
|
||||||
|
|
||||||
|
LOG.info("RUN_TEST -1");
|
||||||
|
final MiniDFSCluster cluster = new MiniDFSCluster
|
||||||
|
.Builder(conf)
|
||||||
|
.nnTopology(MiniDFSNNTopology.simpleFederatedTopology(nNameNodes))
|
||||||
|
.numDataNodes(nDataNodes)
|
||||||
|
.build();
|
||||||
|
LOG.info("RUN_TEST 0");
|
||||||
|
DFSTestUtil.setFederatedConfiguration(cluster, conf);
|
||||||
|
|
||||||
|
try {
|
||||||
|
cluster.waitActive();
|
||||||
|
LOG.info("RUN_TEST 1");
|
||||||
|
final Suite s = new Suite(cluster, nNameNodes, nDataNodes);
|
||||||
|
for(int i = 0; i < nNameNodes; i++) {
|
||||||
|
s.createFile(i, 1024);
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.info("RUN_TEST 2");
|
||||||
|
final String[] urls = new String[nNameNodes];
|
||||||
|
for(int i = 0; i < urls.length; i++) {
|
||||||
|
urls[i] = cluster.getFileSystem(i).getUri() + FILE_NAME;
|
||||||
|
LOG.info("urls[" + i + "]=" + urls[i]);
|
||||||
|
final String result = TestFsck.runFsck(conf, 0, false, urls[i]);
|
||||||
|
LOG.info("result=" + result);
|
||||||
|
Assert.assertTrue(result.contains("Status: HEALTHY"));
|
||||||
|
}
|
||||||
|
} finally {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
LOG.info("RUN_TEST 6");
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Test a cluster with even distribution,
|
||||||
|
* then a new empty node is added to the cluster
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testFsck() throws Exception {
|
||||||
|
final Configuration conf = createConf();
|
||||||
|
runTest(3, 1, conf);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in New Issue