HBASE-27125 The batch size of cleaning expired mob files should have an upper bound (#4541)

Signed-off-by: Duo Zhang <zhangduo@apache.org>
This commit is contained in:
Xiaolin Ha 2022-06-17 23:35:26 +08:00 committed by haxiaolin
parent e21f164050
commit 15a88efb17
4 changed files with 58 additions and 32 deletions

View File

@ -49,6 +49,9 @@ public final class MobConstants {
public static final String MOB_CACHE_EVICT_REMAIN_RATIO = "hbase.mob.cache.evict.remain.ratio"; public static final String MOB_CACHE_EVICT_REMAIN_RATIO = "hbase.mob.cache.evict.remain.ratio";
public static final Tag MOB_REF_TAG = public static final Tag MOB_REF_TAG =
new ArrayBackedTag(TagType.MOB_REFERENCE_TAG_TYPE, HConstants.EMPTY_BYTE_ARRAY); new ArrayBackedTag(TagType.MOB_REFERENCE_TAG_TYPE, HConstants.EMPTY_BYTE_ARRAY);
public static final String MOB_CLEANER_BATCH_SIZE_UPPER_BOUND =
"hbase.master.mob.cleaner.batch.size.upper.bound";
public static final int DEFAULT_MOB_CLEANER_BATCH_SIZE_UPPER_BOUND = 10000;
public static final float DEFAULT_EVICT_REMAIN_RATIO = 0.5f; public static final float DEFAULT_EVICT_REMAIN_RATIO = 0.5f;
public static final long DEFAULT_MOB_CACHE_EVICT_PERIOD = 3600L; public static final long DEFAULT_MOB_CACHE_EVICT_PERIOD = 3600L;

View File

@ -17,6 +17,9 @@
*/ */
package org.apache.hadoop.hbase.mob; package org.apache.hadoop.hbase.mob;
import static org.apache.hadoop.hbase.mob.MobConstants.DEFAULT_MOB_CLEANER_BATCH_SIZE_UPPER_BOUND;
import static org.apache.hadoop.hbase.mob.MobConstants.MOB_CLEANER_BATCH_SIZE_UPPER_BOUND;
import java.io.FileNotFoundException; import java.io.FileNotFoundException;
import java.io.IOException; import java.io.IOException;
import java.text.ParseException; import java.text.ParseException;
@ -331,20 +334,30 @@ public final class MobUtils {
} }
filesToClean filesToClean
.add(new HStoreFile(fs, file.getPath(), conf, cacheConfig, BloomType.NONE, true)); .add(new HStoreFile(fs, file.getPath(), conf, cacheConfig, BloomType.NONE, true));
if (
filesToClean.size() >= conf.getInt(MOB_CLEANER_BATCH_SIZE_UPPER_BOUND,
DEFAULT_MOB_CLEANER_BATCH_SIZE_UPPER_BOUND)
) {
if (
removeMobFiles(conf, fs, tableName, mobTableDir, columnDescriptor.getName(),
filesToClean)
) {
deletedFileCount += filesToClean.size();
}
filesToClean.clear();
}
} }
} catch (Exception e) { } catch (Exception e) {
LOG.error("Cannot parse the fileName " + fileName, e); LOG.error("Cannot parse the fileName " + fileName, e);
} }
} }
if (!filesToClean.isEmpty()) { if (
try { !filesToClean.isEmpty() && removeMobFiles(conf, fs, tableName, mobTableDir,
removeMobFiles(conf, fs, tableName, mobTableDir, columnDescriptor.getName(), filesToClean); columnDescriptor.getName(), filesToClean)
deletedFileCount = filesToClean.size(); ) {
} catch (IOException e) { deletedFileCount += filesToClean.size();
LOG.error("Failed to delete the mob files " + filesToClean, e);
}
} }
LOG.info(deletedFileCount + " expired mob files are deleted"); LOG.info("Table {} {} expired mob files in total are deleted", tableName, deletedFileCount);
} }
/** /**
@ -487,10 +500,17 @@ public final class MobUtils {
* @param family The name of the column family. * @param family The name of the column family.
* @param storeFiles The files to be deleted. * @param storeFiles The files to be deleted.
*/ */
public static void removeMobFiles(Configuration conf, FileSystem fs, TableName tableName, public static boolean removeMobFiles(Configuration conf, FileSystem fs, TableName tableName,
Path tableDir, byte[] family, Collection<HStoreFile> storeFiles) throws IOException { Path tableDir, byte[] family, Collection<HStoreFile> storeFiles) {
HFileArchiver.archiveStoreFiles(conf, fs, getMobRegionInfo(tableName), tableDir, family, try {
storeFiles); HFileArchiver.archiveStoreFiles(conf, fs, getMobRegionInfo(tableName), tableDir, family,
storeFiles);
LOG.info("Table {} {} expired mob files are deleted", tableName, storeFiles.size());
return true;
} catch (IOException e) {
LOG.error("Failed to delete the mob files, table {}", tableName, e);
}
return false;
} }
/** /**

View File

@ -365,14 +365,14 @@ public class PartitionedMobCompactor extends MobCompactor {
+ "table='{}' and column='{}'", tableName, column.getNameAsString()); + "table='{}' and column='{}'", tableName, column.getNameAsString());
for (CompactionDelPartition delPartition : request.getDelPartitions()) { for (CompactionDelPartition delPartition : request.getDelPartitions()) {
LOG.info(Objects.toString(delPartition.listDelFiles())); LOG.info(Objects.toString(delPartition.listDelFiles()));
try { if (
MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(), !MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(),
delPartition.getStoreFiles()); delPartition.getStoreFiles())
} catch (IOException e) { ) {
LOG.error( LOG.error(
"Failed to archive the del files {} for partition {} table='{}' and " + "column='{}'", "Failed to archive the del files {} for partition {} table='{}' and " + "column='{}'",
delPartition.getStoreFiles(), delPartition.getId(), tableName, column.getNameAsString(), delPartition.getStoreFiles(), delPartition.getId(), tableName,
e); column.getNameAsString());
} }
} }
} }
@ -695,14 +695,10 @@ public class PartitionedMobCompactor extends MobCompactor {
} }
// archive the old mob files, do not archive the del files. // archive the old mob files, do not archive the del files.
try { closeStoreFileReaders(mobFilesToCompact);
closeStoreFileReaders(mobFilesToCompact); closeReaders = false;
closeReaders = false; MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(),
MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(), mobFilesToCompact);
mobFilesToCompact);
} catch (IOException e) {
LOG.error("Failed to archive the files " + mobFilesToCompact, e);
}
} finally { } finally {
if (closeReaders) { if (closeReaders) {
closeStoreFileReaders(mobFilesToCompact); closeStoreFileReaders(mobFilesToCompact);
@ -811,11 +807,7 @@ public class PartitionedMobCompactor extends MobCompactor {
// commit the new del file // commit the new del file
Path path = MobUtils.commitFile(conf, fs, filePath, mobFamilyDir, compactionCacheConfig); Path path = MobUtils.commitFile(conf, fs, filePath, mobFamilyDir, compactionCacheConfig);
// archive the old del files // archive the old del files
try { MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(), delFiles);
MobUtils.removeMobFiles(conf, fs, tableName, mobTableDir, column.getName(), delFiles);
} catch (IOException e) {
LOG.error("Failed to archive the old del files " + delFiles, e);
}
return path; return path;
} }

View File

@ -17,6 +17,7 @@
*/ */
package org.apache.hadoop.hbase.mob; package org.apache.hadoop.hbase.mob;
import static org.apache.hadoop.hbase.mob.MobConstants.MOB_CLEANER_BATCH_SIZE_UPPER_BOUND;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileStatus;
@ -53,6 +54,7 @@ public class TestExpiredMobFileCleaner {
private final static String family = "family"; private final static String family = "family";
private final static byte[] row1 = Bytes.toBytes("row1"); private final static byte[] row1 = Bytes.toBytes("row1");
private final static byte[] row2 = Bytes.toBytes("row2"); private final static byte[] row2 = Bytes.toBytes("row2");
private final static byte[] row3 = Bytes.toBytes("row3");
private final static byte[] qf = Bytes.toBytes("qf"); private final static byte[] qf = Bytes.toBytes("qf");
private static BufferedMutator table; private static BufferedMutator table;
@ -61,6 +63,7 @@ public class TestExpiredMobFileCleaner {
@BeforeClass @BeforeClass
public static void setUpBeforeClass() throws Exception { public static void setUpBeforeClass() throws Exception {
TEST_UTIL.getConfiguration().setInt("hfile.format.version", 3); TEST_UTIL.getConfiguration().setInt("hfile.format.version", 3);
TEST_UTIL.getConfiguration().setInt(MOB_CLEANER_BATCH_SIZE_UPPER_BOUND, 2);
} }
@AfterClass @AfterClass
@ -146,6 +149,14 @@ public class TestExpiredMobFileCleaner {
String f2 = secondFiles[1].getPath().getName(); String f2 = secondFiles[1].getPath().getName();
String secondFile = f1.equals(firstFile) ? f2 : f1; String secondFile = f1.equals(firstFile) ? f2 : f1;
ts = EnvironmentEdgeManager.currentTime() - 4 * secondsOfDay() * 1000; // 4 days before
putKVAndFlush(table, row3, dummyData, ts);
ts = EnvironmentEdgeManager.currentTime() - 4 * secondsOfDay() * 1000; // 4 days before
putKVAndFlush(table, row3, dummyData, ts);
FileStatus[] thirdFiles = TEST_UTIL.getTestFileSystem().listStatus(mobDirPath);
// now there are 4 mob files
assertEquals("Before cleanup without delay 3", 4, thirdFiles.length);
modifyColumnExpiryDays(2); // ttl = 2, make the first row expired modifyColumnExpiryDays(2); // ttl = 2, make the first row expired
// run the cleaner // run the cleaner
@ -156,7 +167,7 @@ public class TestExpiredMobFileCleaner {
FileStatus[] filesAfterClean = TEST_UTIL.getTestFileSystem().listStatus(mobDirPath); FileStatus[] filesAfterClean = TEST_UTIL.getTestFileSystem().listStatus(mobDirPath);
String lastFile = filesAfterClean[0].getPath().getName(); String lastFile = filesAfterClean[0].getPath().getName();
// the first mob fie is removed // there are 4 mob files in total, but only 3 need to be cleaned
assertEquals("After cleanup without delay 1", 1, filesAfterClean.length); assertEquals("After cleanup without delay 1", 1, filesAfterClean.length);
assertEquals("After cleanup without delay 2", secondFile, lastFile); assertEquals("After cleanup without delay 2", secondFile, lastFile);
} }