HBASE-21175 Partially initialized SnapshotHFileCleaner leads to NPE during TestHFileArchiving

Signed-off-by: tedyu <yuzhihong@gmail.com>
This commit is contained in:
Artem Ervits 2018-10-26 13:39:05 -04:00 committed by tedyu
parent e5ba79816a
commit 7cdb525192
2 changed files with 50 additions and 17 deletions

View File

@ -32,6 +32,8 @@ import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.Stoppable; import org.apache.hadoop.hbase.Stoppable;
import org.apache.hadoop.hbase.io.HFileLink; import org.apache.hadoop.hbase.io.HFileLink;
import org.apache.hadoop.hbase.master.HMaster;
import org.apache.hadoop.hbase.master.MasterServices;
import org.apache.hadoop.hbase.regionserver.StoreFileInfo; import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
import org.apache.hadoop.hbase.util.StealJobQueue; import org.apache.hadoop.hbase.util.StealJobQueue;
import org.apache.yetus.audience.InterfaceAudience; import org.apache.yetus.audience.InterfaceAudience;
@ -44,6 +46,7 @@ import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesti
*/ */
@InterfaceAudience.Private @InterfaceAudience.Private
public class HFileCleaner extends CleanerChore<BaseHFileCleanerDelegate> { public class HFileCleaner extends CleanerChore<BaseHFileCleanerDelegate> {
private MasterServices master;
public static final String MASTER_HFILE_CLEANER_PLUGINS = "hbase.master.hfilecleaner.plugins"; public static final String MASTER_HFILE_CLEANER_PLUGINS = "hbase.master.hfilecleaner.plugins";
@ -496,4 +499,10 @@ public class HFileCleaner extends CleanerChore<BaseHFileCleanerDelegate> {
t.interrupt(); t.interrupt();
} }
} }
public void init(Map<String, Object> params) {
if (params != null && params.containsKey(HMaster.MASTER)) {
this.master = (MasterServices) params.get(HMaster.MASTER);
}
}
} }

View File

@ -19,26 +19,33 @@ package org.apache.hadoop.hbase.backup;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import java.io.IOException; import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collections; import java.util.Collections;
import java.util.HashMap;
import java.util.List; import java.util.List;
import java.util.Map;
import java.util.Objects;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter; import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.ChoreService;
import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.HBaseClassTestRule;
import org.apache.hadoop.hbase.HBaseTestingUtility;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.Stoppable;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Admin;
import org.apache.hadoop.hbase.master.HMaster;
import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; import org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
import org.apache.hadoop.hbase.regionserver.CompactingMemStore;
import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.HRegionServer;
import org.apache.hadoop.hbase.regionserver.Region;
import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.MediumTests;
import org.apache.hadoop.hbase.testclassification.MiscTests; import org.apache.hadoop.hbase.testclassification.MiscTests;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
@ -177,10 +184,11 @@ public class TestHFileArchiving {
/** /**
* Test that the region directory is removed when we archive a region without store files, but * Test that the region directory is removed when we archive a region without store files, but
* still has hidden files. * still has hidden files.
* @throws Exception * @throws IOException throws an IOException if there's problem creating a table
* or if there's an issue with accessing FileSystem.
*/ */
@Test @Test
public void testDeleteRegionWithNoStoreFiles() throws Exception { public void testDeleteRegionWithNoStoreFiles() throws IOException {
final TableName tableName = TableName.valueOf(name.getMethodName()); final TableName tableName = TableName.valueOf(name.getMethodName());
UTIL.createTable(tableName, TEST_FAM); UTIL.createTable(tableName, TEST_FAM);
@ -209,7 +217,7 @@ public class TestHFileArchiving {
PathFilter nonHidden = new PathFilter() { PathFilter nonHidden = new PathFilter() {
@Override @Override
public boolean accept(Path file) { public boolean accept(Path file) {
return dirFilter.accept(file) && !file.getName().toString().startsWith("."); return dirFilter.accept(file) && !file.getName().startsWith(".");
} }
}; };
FileStatus[] storeDirs = FSUtils.listStatus(fs, regionDir, nonHidden); FileStatus[] storeDirs = FSUtils.listStatus(fs, regionDir, nonHidden);
@ -271,12 +279,14 @@ public class TestHFileArchiving {
assertArchiveFiles(fs, storeFiles, 30000); assertArchiveFiles(fs, storeFiles, 30000);
} }
private void assertArchiveFiles(FileSystem fs, List<String> storeFiles, long timeout) throws IOException { private void assertArchiveFiles(FileSystem fs, List<String> storeFiles, long timeout)
throws IOException {
long end = System.currentTimeMillis() + timeout; long end = System.currentTimeMillis() + timeout;
Path archiveDir = HFileArchiveUtil.getArchivePath(UTIL.getConfiguration()); Path archiveDir = HFileArchiveUtil.getArchivePath(UTIL.getConfiguration());
List<String> archivedFiles = new ArrayList<>(); List<String> archivedFiles = new ArrayList<>();
// We have to ensure that the DeleteTableHandler is finished. HBaseAdmin.deleteXXX() can return before all files // We have to ensure that the DeleteTableHandler is finished. HBaseAdmin.deleteXXX()
// can return before all files
// are archived. We should fix HBASE-5487 and fix synchronous operations from admin. // are archived. We should fix HBASE-5487 and fix synchronous operations from admin.
while (System.currentTimeMillis() < end) { while (System.currentTimeMillis() < end) {
archivedFiles = getAllFileNames(fs, archiveDir); archivedFiles = getAllFileNames(fs, archiveDir);
@ -304,10 +314,11 @@ public class TestHFileArchiving {
/** /**
* Test that the store files are archived when a column family is removed. * Test that the store files are archived when a column family is removed.
* @throws Exception * @throws java.io.IOException if there's a problem creating a table.
* @throws java.lang.InterruptedException problem getting a RegionServer.
*/ */
@Test @Test
public void testArchiveOnTableFamilyDelete() throws Exception { public void testArchiveOnTableFamilyDelete() throws IOException, InterruptedException {
final TableName tableName = TableName.valueOf(name.getMethodName()); final TableName tableName = TableName.valueOf(name.getMethodName());
UTIL.createTable(tableName, new byte[][] {TEST_FAM, Bytes.toBytes("fam2")}); UTIL.createTable(tableName, new byte[][] {TEST_FAM, Bytes.toBytes("fam2")});
@ -374,11 +385,10 @@ public class TestHFileArchiving {
Stoppable stoppable = new StoppableImplementation(); Stoppable stoppable = new StoppableImplementation();
// The cleaner should be looping without long pauses to reproduce the race condition. // The cleaner should be looping without long pauses to reproduce the race condition.
HFileCleaner cleaner = new HFileCleaner(1, stoppable, conf, fs, archiveDir); HFileCleaner cleaner = getHFileCleaner(stoppable, conf, fs, archiveDir);
assertFalse("cleaner should not be null", cleaner == null); assertNotNull("cleaner should not be null", cleaner);
try { try {
choreService.scheduleChore(cleaner); choreService.scheduleChore(cleaner);
// Keep creating/archiving new files while the cleaner is running in the other thread // Keep creating/archiving new files while the cleaner is running in the other thread
long startTime = System.currentTimeMillis(); long startTime = System.currentTimeMillis();
for (long fid = 0; (System.currentTimeMillis() - startTime) < TEST_TIME; ++fid) { for (long fid = 0; (System.currentTimeMillis() - startTime) < TEST_TIME; ++fid) {
@ -418,6 +428,16 @@ public class TestHFileArchiving {
} }
} }
// Avoid passing a null master to CleanerChore, see HBASE-21175
private HFileCleaner getHFileCleaner(Stoppable stoppable, Configuration conf,
FileSystem fs, Path archiveDir) throws IOException {
Map<String, Object> params = new HashMap<>();
params.put(HMaster.MASTER, UTIL.getMiniHBaseCluster().getMaster());
HFileCleaner cleaner = new HFileCleaner(1, stoppable, conf, fs, archiveDir);
cleaner.init(params);
return Objects.requireNonNull(cleaner);
}
private void clearArchiveDirectory() throws IOException { private void clearArchiveDirectory() throws IOException {
UTIL.getTestFileSystem().delete( UTIL.getTestFileSystem().delete(
new Path(UTIL.getDefaultRootDirPath(), HConstants.HFILE_ARCHIVE_DIRECTORY), true); new Path(UTIL.getDefaultRootDirPath(), HConstants.HFILE_ARCHIVE_DIRECTORY), true);
@ -428,7 +448,7 @@ public class TestHFileArchiving {
* @param fs the file system to inspect * @param fs the file system to inspect
* @param archiveDir the directory in which to look * @param archiveDir the directory in which to look
* @return a list of all files in the directory and sub-directories * @return a list of all files in the directory and sub-directories
* @throws IOException * @throws java.io.IOException throws IOException in case FS is unavailable
*/ */
private List<String> getAllFileNames(final FileSystem fs, Path archiveDir) throws IOException { private List<String> getAllFileNames(final FileSystem fs, Path archiveDir) throws IOException {
FileStatus[] files = FSUtils.listStatus(fs, archiveDir, new PathFilter() { FileStatus[] files = FSUtils.listStatus(fs, archiveDir, new PathFilter() {
@ -446,12 +466,16 @@ public class TestHFileArchiving {
/** Recursively lookup all the file names under the file[] array **/ /** Recursively lookup all the file names under the file[] array **/
private List<String> recurseOnFiles(FileSystem fs, FileStatus[] files, List<String> fileNames) private List<String> recurseOnFiles(FileSystem fs, FileStatus[] files, List<String> fileNames)
throws IOException { throws IOException {
if (files == null || files.length == 0) return fileNames; if (files == null || files.length == 0) {
return fileNames;
}
for (FileStatus file : files) { for (FileStatus file : files) {
if (file.isDirectory()) { if (file.isDirectory()) {
recurseOnFiles(fs, FSUtils.listStatus(fs, file.getPath(), null), fileNames); recurseOnFiles(fs, FSUtils.listStatus(fs, file.getPath(), null), fileNames);
} else fileNames.add(file.getPath().getName()); } else {
fileNames.add(file.getPath().getName());
}
} }
return fileNames; return fileNames;
} }