HBASE-16788 Guard HFile archiving under a separate lock
This commit is contained in:
parent
fcef2c02c9
commit
7493e79f15
|
@ -46,6 +46,7 @@ import java.util.concurrent.ExecutorCompletionService;
|
||||||
import java.util.concurrent.Future;
|
import java.util.concurrent.Future;
|
||||||
import java.util.concurrent.ThreadPoolExecutor;
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
import java.util.concurrent.atomic.AtomicBoolean;
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
import java.util.concurrent.locks.ReentrantLock;
|
||||||
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
|
@ -147,6 +148,19 @@ public class HStore implements Store {
|
||||||
* - completing a compaction
|
* - completing a compaction
|
||||||
*/
|
*/
|
||||||
final ReentrantReadWriteLock lock = new ReentrantReadWriteLock();
|
final ReentrantReadWriteLock lock = new ReentrantReadWriteLock();
|
||||||
|
/**
|
||||||
|
* Lock specific to archiving compacted store files. This avoids races around
|
||||||
|
* the combination of retrieving the list of compacted files and moving them to
|
||||||
|
* the archive directory. Since this is usually a background process (other than
|
||||||
|
* on close), we don't want to handle this with the store write lock, which would
|
||||||
|
* block readers and degrade performance.
|
||||||
|
*
|
||||||
|
* Locked by:
|
||||||
|
* - CompactedHFilesDispatchHandler via closeAndArchiveCompactedFiles()
|
||||||
|
* - close()
|
||||||
|
*/
|
||||||
|
final ReentrantLock archiveLock = new ReentrantLock();
|
||||||
|
|
||||||
private final boolean verifyBulkLoads;
|
private final boolean verifyBulkLoads;
|
||||||
|
|
||||||
private ScanInfo scanInfo;
|
private ScanInfo scanInfo;
|
||||||
|
@ -794,6 +808,7 @@ public class HStore implements Store {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ImmutableCollection<StoreFile> close() throws IOException {
|
public ImmutableCollection<StoreFile> close() throws IOException {
|
||||||
|
this.archiveLock.lock();
|
||||||
this.lock.writeLock().lock();
|
this.lock.writeLock().lock();
|
||||||
try {
|
try {
|
||||||
// Clear so metrics doesn't find them.
|
// Clear so metrics doesn't find them.
|
||||||
|
@ -849,6 +864,7 @@ public class HStore implements Store {
|
||||||
return result;
|
return result;
|
||||||
} finally {
|
} finally {
|
||||||
this.lock.writeLock().unlock();
|
this.lock.writeLock().unlock();
|
||||||
|
this.archiveLock.unlock();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2357,7 +2373,10 @@ public class HStore implements Store {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void closeAndArchiveCompactedFiles() throws IOException {
|
public synchronized void closeAndArchiveCompactedFiles() throws IOException {
|
||||||
|
// ensure other threads do not attempt to archive the same files on close()
|
||||||
|
archiveLock.lock();
|
||||||
|
try {
|
||||||
lock.readLock().lock();
|
lock.readLock().lock();
|
||||||
Collection<StoreFile> copyCompactedfiles = null;
|
Collection<StoreFile> copyCompactedfiles = null;
|
||||||
try {
|
try {
|
||||||
|
@ -2378,6 +2397,9 @@ public class HStore implements Store {
|
||||||
if (copyCompactedfiles != null && !copyCompactedfiles.isEmpty()) {
|
if (copyCompactedfiles != null && !copyCompactedfiles.isEmpty()) {
|
||||||
removeCompactedfiles(copyCompactedfiles);
|
removeCompactedfiles(copyCompactedfiles);
|
||||||
}
|
}
|
||||||
|
} finally {
|
||||||
|
archiveLock.unlock();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -0,0 +1,198 @@
|
||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.Stoppable;
|
||||||
|
import org.apache.hadoop.hbase.TableName;
|
||||||
|
import org.apache.hadoop.hbase.client.Put;
|
||||||
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
|
import org.apache.hadoop.hbase.testclassification.RegionServerTests;
|
||||||
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
|
import org.apache.hadoop.hbase.wal.WALFactory;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.experimental.categories.Category;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.io.InterruptedIOException;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Collection;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
import java.util.concurrent.atomic.AtomicReference;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tests a race condition between archiving of compacted files in CompactedHFilesDischarger chore
|
||||||
|
* and HRegion.close();
|
||||||
|
*/
|
||||||
|
@Category({RegionServerTests.class, MediumTests.class})
|
||||||
|
public class TestCompactionArchiveConcurrentClose {
|
||||||
|
public HBaseTestingUtility testUtil;
|
||||||
|
|
||||||
|
private Path testDir;
|
||||||
|
private AtomicBoolean archived = new AtomicBoolean();
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() throws Exception {
|
||||||
|
testUtil = HBaseTestingUtility.createLocalHTU();
|
||||||
|
testDir = testUtil.getDataTestDir("TestStoreFileRefresherChore");
|
||||||
|
FSUtils.setRootDir(testUtil.getConfiguration(), testDir);
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void tearDown() throws Exception {
|
||||||
|
testUtil.cleanupTestDir();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testStoreCloseAndDischargeRunningInParallel() throws Exception {
|
||||||
|
byte[] fam = Bytes.toBytes("f");
|
||||||
|
byte[] col = Bytes.toBytes("c");
|
||||||
|
byte[] val = Bytes.toBytes("val");
|
||||||
|
|
||||||
|
TableName tableName = TableName.valueOf(getClass().getSimpleName());
|
||||||
|
HTableDescriptor htd = new HTableDescriptor(tableName);
|
||||||
|
htd.addFamily(new HColumnDescriptor(fam));
|
||||||
|
HRegionInfo info = new HRegionInfo(tableName, null, null, false);
|
||||||
|
Region region = initHRegion(htd, info);
|
||||||
|
RegionServerServices rss = mock(RegionServerServices.class);
|
||||||
|
List<Region> regions = new ArrayList<Region>();
|
||||||
|
regions.add(region);
|
||||||
|
when(rss.getOnlineRegions()).thenReturn(regions);
|
||||||
|
|
||||||
|
// Create the cleaner object
|
||||||
|
CompactedHFilesDischarger cleaner =
|
||||||
|
new CompactedHFilesDischarger(1000, (Stoppable) null, rss, false);
|
||||||
|
// Add some data to the region and do some flushes
|
||||||
|
int batchSize = 10;
|
||||||
|
int fileCount = 10;
|
||||||
|
for (int f = 0; f < fileCount; f++) {
|
||||||
|
int start = f * batchSize;
|
||||||
|
for (int i = start; i < start + batchSize; i++) {
|
||||||
|
Put p = new Put(Bytes.toBytes("row" + i));
|
||||||
|
p.addColumn(fam, col, val);
|
||||||
|
region.put(p);
|
||||||
|
}
|
||||||
|
// flush them
|
||||||
|
region.flush(true);
|
||||||
|
}
|
||||||
|
|
||||||
|
Store store = region.getStore(fam);
|
||||||
|
assertEquals(fileCount, store.getStorefilesCount());
|
||||||
|
|
||||||
|
Collection<StoreFile> storefiles = store.getStorefiles();
|
||||||
|
// None of the files should be in compacted state.
|
||||||
|
for (StoreFile file : storefiles) {
|
||||||
|
assertFalse(file.isCompactedAway());
|
||||||
|
}
|
||||||
|
// Do compaction
|
||||||
|
region.compact(true);
|
||||||
|
|
||||||
|
// now run the cleaner with a concurrent close
|
||||||
|
Thread cleanerThread = new Thread() {
|
||||||
|
public void run() {
|
||||||
|
cleaner.chore();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
cleanerThread.start();
|
||||||
|
// wait for cleaner to pause
|
||||||
|
synchronized (archived) {
|
||||||
|
if (!archived.get()) {
|
||||||
|
archived.wait();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
final AtomicReference<Exception> closeException = new AtomicReference<>();
|
||||||
|
Thread closeThread = new Thread() {
|
||||||
|
public void run() {
|
||||||
|
// wait for the chore to complete and call close
|
||||||
|
try {
|
||||||
|
((HRegion) region).close();
|
||||||
|
} catch (IOException e) {
|
||||||
|
closeException.set(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
closeThread.start();
|
||||||
|
// no error should occur after the execution of the test
|
||||||
|
closeThread.join();
|
||||||
|
cleanerThread.join();
|
||||||
|
|
||||||
|
if (closeException.get() != null) {
|
||||||
|
throw closeException.get();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private Region initHRegion(HTableDescriptor htd, HRegionInfo info)
|
||||||
|
throws IOException {
|
||||||
|
Configuration conf = testUtil.getConfiguration();
|
||||||
|
Path tableDir = FSUtils.getTableDir(testDir, htd.getTableName());
|
||||||
|
|
||||||
|
HRegionFileSystem fs = new WaitingHRegionFileSystem(conf, tableDir.getFileSystem(conf),
|
||||||
|
tableDir, info);
|
||||||
|
final Configuration walConf = new Configuration(conf);
|
||||||
|
FSUtils.setRootDir(walConf, tableDir);
|
||||||
|
final WALFactory wals = new WALFactory(walConf, null, "log_" + info.getEncodedName());
|
||||||
|
HRegion region =
|
||||||
|
new HRegion(fs, wals.getWAL(info.getEncodedNameAsBytes(), info.getTable().getNamespace()),
|
||||||
|
conf, htd, null);
|
||||||
|
|
||||||
|
region.initialize();
|
||||||
|
|
||||||
|
return region;
|
||||||
|
}
|
||||||
|
|
||||||
|
private class WaitingHRegionFileSystem extends HRegionFileSystem {
|
||||||
|
|
||||||
|
public WaitingHRegionFileSystem(final Configuration conf, final FileSystem fs,
|
||||||
|
final Path tableDir, final HRegionInfo regionInfo) {
|
||||||
|
super(conf, fs, tableDir, regionInfo);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void removeStoreFiles(String familyName, Collection<StoreFile> storeFiles)
|
||||||
|
throws IOException {
|
||||||
|
super.removeStoreFiles(familyName, storeFiles);
|
||||||
|
archived.set(true);
|
||||||
|
synchronized (archived) {
|
||||||
|
archived.notifyAll();
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
// unfortunately we can't use a stronger barrier here as the fix synchronizing
|
||||||
|
// the race condition will then block
|
||||||
|
Thread.sleep(100);
|
||||||
|
} catch (InterruptedException ie) {
|
||||||
|
throw new InterruptedIOException("Interrupted waiting for latch");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue