HBASE-9829 make the compaction logging less confusing

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1550866 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
sershe 2013-12-14 01:49:46 +00:00
parent 4ee8b16d67
commit 763e060db9
6 changed files with 73 additions and 82 deletions

View File

@ -56,8 +56,8 @@ public class CompactSplitThread implements CompactionRequestor {
private final HRegionServer server; private final HRegionServer server;
private final Configuration conf; private final Configuration conf;
private final ThreadPoolExecutor largeCompactions; private final ThreadPoolExecutor longCompactions;
private final ThreadPoolExecutor smallCompactions; private final ThreadPoolExecutor shortCompactions;
private final ThreadPoolExecutor splits; private final ThreadPoolExecutor splits;
private final ThreadPoolExecutor mergePool; private final ThreadPoolExecutor mergePool;
@ -88,28 +88,28 @@ public class CompactSplitThread implements CompactionRequestor {
final String n = Thread.currentThread().getName(); final String n = Thread.currentThread().getName();
this.largeCompactions = new ThreadPoolExecutor(largeThreads, largeThreads, this.longCompactions = new ThreadPoolExecutor(largeThreads, largeThreads,
60, TimeUnit.SECONDS, new PriorityBlockingQueue<Runnable>(), 60, TimeUnit.SECONDS, new PriorityBlockingQueue<Runnable>(),
new ThreadFactory() { new ThreadFactory() {
@Override @Override
public Thread newThread(Runnable r) { public Thread newThread(Runnable r) {
Thread t = new Thread(r); Thread t = new Thread(r);
t.setName(n + "-largeCompactions-" + System.currentTimeMillis()); t.setName(n + "-longCompactions-" + System.currentTimeMillis());
return t; return t;
} }
}); });
this.largeCompactions.setRejectedExecutionHandler(new Rejection()); this.longCompactions.setRejectedExecutionHandler(new Rejection());
this.smallCompactions = new ThreadPoolExecutor(smallThreads, smallThreads, this.shortCompactions = new ThreadPoolExecutor(smallThreads, smallThreads,
60, TimeUnit.SECONDS, new PriorityBlockingQueue<Runnable>(), 60, TimeUnit.SECONDS, new PriorityBlockingQueue<Runnable>(),
new ThreadFactory() { new ThreadFactory() {
@Override @Override
public Thread newThread(Runnable r) { public Thread newThread(Runnable r) {
Thread t = new Thread(r); Thread t = new Thread(r);
t.setName(n + "-smallCompactions-" + System.currentTimeMillis()); t.setName(n + "-shortCompactions-" + System.currentTimeMillis());
return t; return t;
} }
}); });
this.smallCompactions this.shortCompactions
.setRejectedExecutionHandler(new Rejection()); .setRejectedExecutionHandler(new Rejection());
this.splits = (ThreadPoolExecutor) this.splits = (ThreadPoolExecutor)
Executors.newFixedThreadPool(splitThreads, Executors.newFixedThreadPool(splitThreads,
@ -136,8 +136,8 @@ public class CompactSplitThread implements CompactionRequestor {
@Override @Override
public String toString() { public String toString() {
return "compaction_queue=(" return "compaction_queue=("
+ largeCompactions.getQueue().size() + ":" + longCompactions.getQueue().size() + ":"
+ smallCompactions.getQueue().size() + ")" + shortCompactions.getQueue().size() + ")"
+ ", split_queue=" + splits.getQueue().size() + ", split_queue=" + splits.getQueue().size()
+ ", merge_queue=" + mergePool.getQueue().size(); + ", merge_queue=" + mergePool.getQueue().size();
} }
@ -146,17 +146,17 @@ public class CompactSplitThread implements CompactionRequestor {
StringBuffer queueLists = new StringBuffer(); StringBuffer queueLists = new StringBuffer();
queueLists.append("Compaction/Split Queue dump:\n"); queueLists.append("Compaction/Split Queue dump:\n");
queueLists.append(" LargeCompation Queue:\n"); queueLists.append(" LargeCompation Queue:\n");
BlockingQueue<Runnable> lq = largeCompactions.getQueue(); BlockingQueue<Runnable> lq = longCompactions.getQueue();
Iterator it = lq.iterator(); Iterator it = lq.iterator();
while(it.hasNext()){ while(it.hasNext()){
queueLists.append(" "+it.next().toString()); queueLists.append(" "+it.next().toString());
queueLists.append("\n"); queueLists.append("\n");
} }
if( smallCompactions != null ){ if( shortCompactions != null ){
queueLists.append("\n"); queueLists.append("\n");
queueLists.append(" SmallCompation Queue:\n"); queueLists.append(" SmallCompation Queue:\n");
lq = smallCompactions.getQueue(); lq = shortCompactions.getQueue();
it = lq.iterator(); it = lq.iterator();
while(it.hasNext()){ while(it.hasNext()){
queueLists.append(" "+it.next().toString()); queueLists.append(" "+it.next().toString());
@ -312,10 +312,10 @@ public class CompactSplitThread implements CompactionRequestor {
// pool; we will do selection there, and move to large pool if necessary. // pool; we will do selection there, and move to large pool if necessary.
long size = selectNow ? compaction.getRequest().getSize() : 0; long size = selectNow ? compaction.getRequest().getSize() : 0;
ThreadPoolExecutor pool = (!selectNow && s.throttleCompaction(size)) ThreadPoolExecutor pool = (!selectNow && s.throttleCompaction(size))
? largeCompactions : smallCompactions; ? longCompactions : shortCompactions;
pool.execute(new CompactionRunner(s, r, compaction, pool)); pool.execute(new CompactionRunner(s, r, compaction, pool));
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
String type = (pool == smallCompactions) ? "Small " : "Large "; String type = (pool == shortCompactions) ? "Small " : "Large ";
LOG.debug(type + "Compaction requested: " + (selectNow ? compaction.toString() : "system") LOG.debug(type + "Compaction requested: " + (selectNow ? compaction.toString() : "system")
+ (why != null && !why.isEmpty() ? "; Because: " + why : "") + "; " + this); + (why != null && !why.isEmpty() ? "; Because: " + why : "") + "; " + this);
} }
@ -345,8 +345,8 @@ public class CompactSplitThread implements CompactionRequestor {
void interruptIfNecessary() { void interruptIfNecessary() {
splits.shutdown(); splits.shutdown();
mergePool.shutdown(); mergePool.shutdown();
largeCompactions.shutdown(); longCompactions.shutdown();
smallCompactions.shutdown(); shortCompactions.shutdown();
} }
private void waitFor(ThreadPoolExecutor t, String name) { private void waitFor(ThreadPoolExecutor t, String name) {
@ -367,8 +367,8 @@ public class CompactSplitThread implements CompactionRequestor {
void join() { void join() {
waitFor(splits, "Split Thread"); waitFor(splits, "Split Thread");
waitFor(mergePool, "Merge Thread"); waitFor(mergePool, "Merge Thread");
waitFor(largeCompactions, "Large Compaction Thread"); waitFor(longCompactions, "Large Compaction Thread");
waitFor(smallCompactions, "Small Compaction Thread"); waitFor(shortCompactions, "Small Compaction Thread");
} }
/** /**
@ -378,16 +378,16 @@ public class CompactSplitThread implements CompactionRequestor {
* @return The current size of the regions queue. * @return The current size of the regions queue.
*/ */
public int getCompactionQueueSize() { public int getCompactionQueueSize() {
return largeCompactions.getQueue().size() + smallCompactions.getQueue().size(); return longCompactions.getQueue().size() + shortCompactions.getQueue().size();
} }
public int getLargeCompactionQueueSize() { public int getLargeCompactionQueueSize() {
return largeCompactions.getQueue().size(); return longCompactions.getQueue().size();
} }
public int getSmallCompactionQueueSize() { public int getSmallCompactionQueueSize() {
return smallCompactions.getQueue().size(); return shortCompactions.getQueue().size();
} }
@ -455,7 +455,7 @@ public class CompactSplitThread implements CompactionRequestor {
// We might end up waiting for a while, so cancel the selection. // We might end up waiting for a while, so cancel the selection.
assert this.compaction.hasSelection(); assert this.compaction.hasSelection();
ThreadPoolExecutor pool = store.throttleCompaction( ThreadPoolExecutor pool = store.throttleCompaction(
compaction.getRequest().getSize()) ? largeCompactions : smallCompactions; compaction.getRequest().getSize()) ? longCompactions : shortCompactions;
if (this.parent != pool) { if (this.parent != pool) {
this.store.cancelRequestedCompaction(this.compaction); this.store.cancelRequestedCompaction(this.compaction);
this.compaction = null; this.compaction = null;

View File

@ -4527,19 +4527,6 @@ public class HRegion implements HeapSize { // , Writable{
return dstRegion; return dstRegion;
} }
/**
* @return True if needs a major compaction.
* @throws IOException
*/
boolean isMajorCompaction() throws IOException {
for (Store store : this.stores.values()) {
if (store.isMajorCompaction()) {
return true;
}
}
return false;
}
// //
// HBASE-880 // HBASE-880
// //
@ -5802,7 +5789,7 @@ public class HRegion implements HeapSize { // , Writable{
(isMajor ? majorInProgress : minorInProgress).incrementAndGet(); (isMajor ? majorInProgress : minorInProgress).incrementAndGet();
} }
public void reportCompactionRequestEnd(boolean isMajor, int numFiles, long filesSizeCompacted){ public void reportCompactionRequestEnd(boolean isMajor, int numFiles, long filesSizeCompacted) {
int newValue = (isMajor ? majorInProgress : minorInProgress).decrementAndGet(); int newValue = (isMajor ? majorInProgress : minorInProgress).decrementAndGet();
// metrics // metrics

View File

@ -1163,10 +1163,9 @@ public class HStore implements Store {
CompactionRequest cr, List<StoreFile> sfs, long compactionStartTime) { CompactionRequest cr, List<StoreFile> sfs, long compactionStartTime) {
long now = EnvironmentEdgeManager.currentTimeMillis(); long now = EnvironmentEdgeManager.currentTimeMillis();
StringBuilder message = new StringBuilder( StringBuilder message = new StringBuilder(
"Completed" + (cr.isMajor() ? " major " : " ") + "compaction of " "Completed" + (cr.isMajor() ? " major" : "") + " compaction of "
+ cr.getFiles().size() + " file(s) in " + this + " of " + cr.getFiles().size() + (cr.isAllFiles() ? " (all)" : "") + " file(s) in "
+ this.getRegionInfo().getRegionNameAsString() + this + " of " + this.getRegionInfo().getRegionNameAsString() + " into ");
+ " into ");
if (sfs.isEmpty()) { if (sfs.isEmpty()) {
message.append("none, "); message.append("none, ");
} else { } else {
@ -1342,6 +1341,7 @@ public class HStore implements Store {
} }
CompactionContext compaction = storeEngine.createCompaction(); CompactionContext compaction = storeEngine.createCompaction();
CompactionRequest request = null;
this.lock.readLock().lock(); this.lock.readLock().lock();
try { try {
synchronized (filesCompacting) { synchronized (filesCompacting) {
@ -1388,9 +1388,9 @@ public class HStore implements Store {
compaction.forceSelect( compaction.forceSelect(
baseRequest.combineWith(compaction.getRequest())); baseRequest.combineWith(compaction.getRequest()));
} }
// Finally, we have the resulting files list. Check if we have any files at all. // Finally, we have the resulting files list. Check if we have any files at all.
final Collection<StoreFile> selectedFiles = compaction.getRequest().getFiles(); request = compaction.getRequest();
final Collection<StoreFile> selectedFiles = request.getFiles();
if (selectedFiles.isEmpty()) { if (selectedFiles.isEmpty()) {
return null; return null;
} }
@ -1404,24 +1404,21 @@ public class HStore implements Store {
Collections.sort(filesCompacting, StoreFile.Comparators.SEQ_ID); Collections.sort(filesCompacting, StoreFile.Comparators.SEQ_ID);
// If we're enqueuing a major, clear the force flag. // If we're enqueuing a major, clear the force flag.
boolean isMajor = selectedFiles.size() == this.getStorefilesCount(); this.forceMajor = this.forceMajor && !request.isMajor();
this.forceMajor = this.forceMajor && !isMajor;
// Set common request properties. // Set common request properties.
// Set priority, either override value supplied by caller or from store. // Set priority, either override value supplied by caller or from store.
compaction.getRequest().setPriority( request.setPriority((priority != Store.NO_PRIORITY) ? priority : getCompactPriority());
(priority != Store.NO_PRIORITY) ? priority : getCompactPriority()); request.setDescription(getRegionInfo().getRegionNameAsString(), getColumnFamilyName());
compaction.getRequest().setIsMajor(isMajor);
compaction.getRequest().setDescription(
getRegionInfo().getRegionNameAsString(), getColumnFamilyName());
} }
} finally { } finally {
this.lock.readLock().unlock(); this.lock.readLock().unlock();
} }
LOG.debug(getRegionInfo().getEncodedName() + " - " + getColumnFamilyName() + ": Initiating " LOG.debug(getRegionInfo().getEncodedName() + " - " + getColumnFamilyName()
+ (compaction.getRequest().isMajor() ? "major" : "minor") + " compaction"); + ": Initiating " + (request.isMajor() ? "major" : "minor") + " compaction"
this.region.reportCompactionRequestStart(compaction.getRequest().isMajor()); + (request.isAllFiles() ? " (all files)" : ""));
this.region.reportCompactionRequestStart(request.isMajor());
return compaction; return compaction;
} }
@ -1732,9 +1729,6 @@ public class HStore implements Store {
this.forceMajor = true; this.forceMajor = true;
} }
boolean getForceMajorCompaction() {
return this.forceMajor;
}
////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////
// File administration // File administration

View File

@ -45,7 +45,8 @@ public class CompactionRequest implements Comparable<CompactionRequest> {
static final Log LOG = LogFactory.getLog(CompactionRequest.class); static final Log LOG = LogFactory.getLog(CompactionRequest.class);
// was this compaction promoted to an off-peak // was this compaction promoted to an off-peak
private boolean isOffPeak = false; private boolean isOffPeak = false;
private boolean isMajor = false; private enum DisplayCompactionType { MINOR, ALL_FILES, MAJOR }
private DisplayCompactionType isMajor = DisplayCompactionType.MINOR;
private int priority = Store.NO_PRIORITY; private int priority = Store.NO_PRIORITY;
private Collection<StoreFile> filesToCompact; private Collection<StoreFile> filesToCompact;
@ -156,8 +157,13 @@ public class CompactionRequest implements Comparable<CompactionRequest> {
return totalSize; return totalSize;
} }
public boolean isAllFiles() {
return this.isMajor == DisplayCompactionType.MAJOR
|| this.isMajor == DisplayCompactionType.ALL_FILES;
}
public boolean isMajor() { public boolean isMajor() {
return this.isMajor; return this.isMajor == DisplayCompactionType.MAJOR;
} }
/** Gets the priority for the request */ /** Gets the priority for the request */
@ -187,8 +193,10 @@ public class CompactionRequest implements Comparable<CompactionRequest> {
* @param isMajor <tt>true</tt> if the system determines that this compaction should be a major * @param isMajor <tt>true</tt> if the system determines that this compaction should be a major
* compaction * compaction
*/ */
public void setIsMajor(boolean isMajor) { public void setIsMajor(boolean isMajor, boolean isAllFiles) {
this.isMajor = isMajor; assert isAllFiles || !isMajor;
this.isMajor = !isAllFiles ? DisplayCompactionType.MINOR
: (isMajor ? DisplayCompactionType.MAJOR : DisplayCompactionType.ALL_FILES);
} }
@Override @Override

View File

@ -45,7 +45,7 @@ public class DefaultCompactor extends Compactor {
* Do a minor/major compaction on an explicit set of storefiles from a Store. * Do a minor/major compaction on an explicit set of storefiles from a Store.
*/ */
public List<Path> compact(final CompactionRequest request) throws IOException { public List<Path> compact(final CompactionRequest request) throws IOException {
FileDetails fd = getFileDetails(request.getFiles(), request.isMajor()); FileDetails fd = getFileDetails(request.getFiles(), request.isAllFiles());
this.progress = new CompactionProgress(fd.maxKeyCount); this.progress = new CompactionProgress(fd.maxKeyCount);
// Find the smallest read point across all the Scanners. // Find the smallest read point across all the Scanners.
@ -57,9 +57,9 @@ public class DefaultCompactor extends Compactor {
try { try {
InternalScanner scanner = null; InternalScanner scanner = null;
try { try {
/* Include deletes, unless we are doing a major compaction */ /* Include deletes, unless we are doing a compaction of all files */
ScanType scanType = ScanType scanType =
request.isMajor() ? ScanType.COMPACT_DROP_DELETES : ScanType.COMPACT_RETAIN_DELETES; request.isAllFiles() ? ScanType.COMPACT_DROP_DELETES : ScanType.COMPACT_RETAIN_DELETES;
scanner = preCreateCoprocScanner(request, scanType, fd.earliestPutTs, scanners); scanner = preCreateCoprocScanner(request, scanType, fd.earliestPutTs, scanners);
if (scanner == null) { if (scanner == null) {
scanner = createScanner(store, scanners, scanType, smallestReadPoint, fd.earliestPutTs); scanner = createScanner(store, scanners, scanType, smallestReadPoint, fd.earliestPutTs);
@ -89,7 +89,7 @@ public class DefaultCompactor extends Compactor {
} }
} finally { } finally {
if (writer != null) { if (writer != null) {
writer.appendMetadata(fd.maxSeqId, request.isMajor()); writer.appendMetadata(fd.maxSeqId, request.isAllFiles());
writer.close(); writer.close();
newFiles.add(writer.getPath()); newFiles.add(writer.getPath());
} }
@ -110,7 +110,7 @@ public class DefaultCompactor extends Compactor {
public List<Path> compactForTesting(final Collection<StoreFile> filesToCompact, boolean isMajor) public List<Path> compactForTesting(final Collection<StoreFile> filesToCompact, boolean isMajor)
throws IOException { throws IOException {
CompactionRequest cr = new CompactionRequest(filesToCompact); CompactionRequest cr = new CompactionRequest(filesToCompact);
cr.setIsMajor(isMajor); cr.setIsMajor(isMajor, isMajor);
return this.compact(cr); return this.compact(cr);
} }
} }

View File

@ -73,7 +73,7 @@ public class RatioBasedCompactionPolicy extends CompactionPolicy {
} }
/** /**
* @param candidateFiles candidate files, ordered from oldest to newest * @param candidateFiles candidate files, ordered from oldest to newest. All files in store.
* @return subset copy of candidate list that meets compaction criteria * @return subset copy of candidate list that meets compaction criteria
* @throws java.io.IOException * @throws java.io.IOException
*/ */
@ -93,9 +93,11 @@ public class RatioBasedCompactionPolicy extends CompactionPolicy {
filesCompacting.size() + " compacting, " + candidateSelection.size() + filesCompacting.size() + " compacting, " + candidateSelection.size() +
" eligible, " + storeConfigInfo.getBlockingFileCount() + " blocking"); " eligible, " + storeConfigInfo.getBlockingFileCount() + " blocking");
long cfTtl = this.storeConfigInfo.getStoreFileTtl(); // If we can't have all files, we cannot do major anyway
if (!forceMajor) { boolean isAllFiles = candidateFiles.size() == candidateSelection.size();
if (!(forceMajor && isAllFiles)) {
// If there are expired files, only select them so that compaction deletes them // If there are expired files, only select them so that compaction deletes them
long cfTtl = this.storeConfigInfo.getStoreFileTtl();
if (comConf.shouldDeleteExpired() && (cfTtl != Long.MAX_VALUE)) { if (comConf.shouldDeleteExpired() && (cfTtl != Long.MAX_VALUE)) {
ArrayList<StoreFile> expiredSelection = selectExpiredStoreFiles( ArrayList<StoreFile> expiredSelection = selectExpiredStoreFiles(
candidateSelection, EnvironmentEdgeManager.currentTimeMillis() - cfTtl); candidateSelection, EnvironmentEdgeManager.currentTimeMillis() - cfTtl);
@ -104,28 +106,28 @@ public class RatioBasedCompactionPolicy extends CompactionPolicy {
} }
} }
candidateSelection = skipLargeFiles(candidateSelection); candidateSelection = skipLargeFiles(candidateSelection);
isAllFiles = candidateFiles.size() == candidateSelection.size();
} }
// Force a major compaction if this is a user-requested major compaction, // Try a major compaction if this is a user-requested major compaction,
// or if we do not have too many files to compact and this was requested // or if we do not have too many files to compact and this was requested as a major compaction
// as a major compaction. boolean isTryingMajor = (forceMajor && isAllFiles && isUserCompaction)
|| (((forceMajor && isAllFiles) || isMajorCompaction(candidateSelection))
&& (candidateSelection.size() < comConf.getMaxFilesToCompact()));
// Or, if there are any references among the candidates. // Or, if there are any references among the candidates.
boolean majorCompaction = ( boolean isAfterSplit = StoreUtils.hasReferences(candidateSelection);
(forceMajor && isUserCompaction) if (!isTryingMajor && !isAfterSplit) {
|| ((forceMajor || isMajorCompaction(candidateSelection)) // We're are not compacting all files, let's see what files are applicable
&& (candidateSelection.size() < comConf.getMaxFilesToCompact()))
|| StoreUtils.hasReferences(candidateSelection)
);
if (!majorCompaction) {
// we're doing a minor compaction, let's see what files are applicable
candidateSelection = filterBulk(candidateSelection); candidateSelection = filterBulk(candidateSelection);
candidateSelection = applyCompactionPolicy(candidateSelection, mayUseOffPeak, mayBeStuck); candidateSelection = applyCompactionPolicy(candidateSelection, mayUseOffPeak, mayBeStuck);
candidateSelection = checkMinFilesCriteria(candidateSelection); candidateSelection = checkMinFilesCriteria(candidateSelection);
} }
candidateSelection = removeExcessFiles(candidateSelection, isUserCompaction, majorCompaction); candidateSelection = removeExcessFiles(candidateSelection, isUserCompaction, isTryingMajor);
// Now we have the final file list, so we can determine if we can do major/all files.
isAllFiles = (candidateFiles.size() == candidateSelection.size());
CompactionRequest result = new CompactionRequest(candidateSelection); CompactionRequest result = new CompactionRequest(candidateSelection);
result.setOffPeak(!candidateSelection.isEmpty() && !majorCompaction && mayUseOffPeak); result.setOffPeak(!candidateSelection.isEmpty() && !isAllFiles && mayUseOffPeak);
result.setIsMajor(isTryingMajor && isAllFiles, isAllFiles);
return result; return result;
} }