HBASE-25458 HRegion methods cleanup (#2838)
Signed-off-by: meiyi <myimeiyi@gmail.com>
This commit is contained in:
parent
2b25fca6d8
commit
1719565aa7
@ -906,17 +906,19 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
void setHTableSpecificConf() {
|
||||
if (this.htableDescriptor == null) return;
|
||||
private void setHTableSpecificConf() {
|
||||
if (this.htableDescriptor == null) {
|
||||
return;
|
||||
}
|
||||
long flushSize = this.htableDescriptor.getMemStoreFlushSize();
|
||||
|
||||
if (flushSize <= 0) {
|
||||
flushSize = conf.getLong(HConstants.HREGION_MEMSTORE_FLUSH_SIZE,
|
||||
TableDescriptorBuilder.DEFAULT_MEMSTORE_FLUSH_SIZE);
|
||||
TableDescriptorBuilder.DEFAULT_MEMSTORE_FLUSH_SIZE);
|
||||
}
|
||||
this.memstoreFlushSize = flushSize;
|
||||
long mult = conf.getLong(HConstants.HREGION_MEMSTORE_BLOCK_MULTIPLIER,
|
||||
HConstants.DEFAULT_HREGION_MEMSTORE_BLOCK_MULTIPLIER);
|
||||
HConstants.DEFAULT_HREGION_MEMSTORE_BLOCK_MULTIPLIER);
|
||||
this.blockingMemStoreSize = this.memstoreFlushSize * mult;
|
||||
}
|
||||
|
||||
@ -1324,7 +1326,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* Increase the size of mem store in this region and the size of global mem
|
||||
* store
|
||||
*/
|
||||
void incMemStoreSize(MemStoreSize mss) {
|
||||
private void incMemStoreSize(MemStoreSize mss) {
|
||||
incMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize(),
|
||||
mss.getCellsCount());
|
||||
}
|
||||
@ -1344,7 +1346,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
mss.getCellsCount());
|
||||
}
|
||||
|
||||
void decrMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta,
|
||||
private void decrMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta,
|
||||
int cellsCountDelta) {
|
||||
if (this.rsAccounting != null) {
|
||||
rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta);
|
||||
@ -1964,7 +1966,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
protected ThreadPoolExecutor getStoreOpenAndCloseThreadPool(
|
||||
private ThreadPoolExecutor getStoreOpenAndCloseThreadPool(
|
||||
final String threadNamePrefix) {
|
||||
int numStores = Math.max(1, this.htableDescriptor.getColumnFamilyCount());
|
||||
int maxThreads = Math.min(numStores,
|
||||
@ -1973,7 +1975,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
return getOpenAndCloseThreadPool(maxThreads, threadNamePrefix);
|
||||
}
|
||||
|
||||
protected ThreadPoolExecutor getStoreFileOpenAndCloseThreadPool(
|
||||
ThreadPoolExecutor getStoreFileOpenAndCloseThreadPool(
|
||||
final String threadNamePrefix) {
|
||||
int numStores = Math.max(1, this.htableDescriptor.getColumnFamilyCount());
|
||||
int maxThreads = Math.max(1,
|
||||
@ -1983,7 +1985,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
return getOpenAndCloseThreadPool(maxThreads, threadNamePrefix);
|
||||
}
|
||||
|
||||
static ThreadPoolExecutor getOpenAndCloseThreadPool(int maxThreads,
|
||||
private static ThreadPoolExecutor getOpenAndCloseThreadPool(int maxThreads,
|
||||
final String threadNamePrefix) {
|
||||
return Threads.getBoundedCachedThreadPool(maxThreads, 30L, TimeUnit.SECONDS,
|
||||
new ThreadFactory() {
|
||||
@ -2437,11 +2439,11 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
boolean isCompactionNeeded();
|
||||
}
|
||||
|
||||
public FlushResultImpl flushcache(boolean flushAllStores, boolean writeFlushRequestWalMarker,
|
||||
FlushResultImpl flushcache(boolean flushAllStores, boolean writeFlushRequestWalMarker,
|
||||
FlushLifeCycleTracker tracker) throws IOException {
|
||||
List families = null;
|
||||
List<byte[]> families = null;
|
||||
if (flushAllStores) {
|
||||
families = new ArrayList();
|
||||
families = new ArrayList<>();
|
||||
families.addAll(this.getTableDescriptor().getColumnFamilyNames());
|
||||
}
|
||||
return this.flushcache(families, writeFlushRequestWalMarker, tracker);
|
||||
@ -2922,7 +2924,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
|
||||
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NN_NAKED_NOTIFY",
|
||||
justification="Intentional; notify is about completed flush")
|
||||
protected FlushResultImpl internalFlushCacheAndCommit(WAL wal, MonitoredTask status,
|
||||
FlushResultImpl internalFlushCacheAndCommit(WAL wal, MonitoredTask status,
|
||||
PrepareFlushResult prepareResult, Collection<HStore> storesToFlush) throws IOException {
|
||||
// prepare flush context is carried via PrepareFlushResult
|
||||
TreeMap<byte[], StoreFlushContext> storeFlushCtxs = prepareResult.storeFlushCtxs;
|
||||
@ -3119,12 +3121,6 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
protected RegionScanner instantiateRegionScanner(Scan scan,
|
||||
List<KeyValueScanner> additionalScanners) throws IOException {
|
||||
return instantiateRegionScanner(scan, additionalScanners, HConstants.NO_NONCE,
|
||||
HConstants.NO_NONCE);
|
||||
}
|
||||
|
||||
protected RegionScannerImpl instantiateRegionScanner(Scan scan,
|
||||
List<KeyValueScanner> additionalScanners, long nonceGroup, long nonce) throws IOException {
|
||||
if (scan.isReversed()) {
|
||||
@ -3139,9 +3135,8 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
/**
|
||||
* Prepare a delete for a row mutation processor
|
||||
* @param delete The passed delete is modified by this method. WARNING!
|
||||
* @throws IOException
|
||||
*/
|
||||
public void prepareDelete(Delete delete) throws IOException {
|
||||
private void prepareDelete(Delete delete) throws IOException {
|
||||
// Check to see if this is a deleteRow insert
|
||||
if(delete.getFamilyCellMap().isEmpty()){
|
||||
for(byte [] family : this.htableDescriptor.getColumnFamilyNames()){
|
||||
@ -3165,39 +3160,18 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
startRegionOperation(Operation.DELETE);
|
||||
try {
|
||||
// All edits for the given row (across all column families) must happen atomically.
|
||||
doBatchMutate(delete);
|
||||
mutate(delete);
|
||||
} finally {
|
||||
closeRegionOperation(Operation.DELETE);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Row needed by below method.
|
||||
*/
|
||||
private static final byte [] FOR_UNIT_TESTS_ONLY = Bytes.toBytes("ForUnitTestsOnly");
|
||||
|
||||
/**
|
||||
* This is used only by unit tests. Not required to be a public API.
|
||||
* @param familyMap map of family to edits for the given family.
|
||||
* @throws IOException
|
||||
*/
|
||||
void delete(NavigableMap<byte[], List<Cell>> familyMap,
|
||||
Durability durability) throws IOException {
|
||||
Delete delete = new Delete(FOR_UNIT_TESTS_ONLY);
|
||||
delete.setFamilyCellMap(familyMap);
|
||||
delete.setDurability(durability);
|
||||
doBatchMutate(delete);
|
||||
}
|
||||
|
||||
/**
|
||||
* Set up correct timestamps in the KVs in Delete object.
|
||||
* <p>Caller should have the row and region locks.
|
||||
* @param mutation
|
||||
* @param familyMap
|
||||
* @param byteNow
|
||||
* @throws IOException
|
||||
* <p/>
|
||||
* Caller should have the row and region locks.
|
||||
*/
|
||||
public void prepareDeleteTimestamps(Mutation mutation, Map<byte[], List<Cell>> familyMap,
|
||||
private void prepareDeleteTimestamps(Mutation mutation, Map<byte[], List<Cell>> familyMap,
|
||||
byte[] byteNow) throws IOException {
|
||||
for (Map.Entry<byte[], List<Cell>> e : familyMap.entrySet()) {
|
||||
|
||||
@ -3241,7 +3215,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
void updateDeleteLatestVersionTimestamp(Cell cell, Get get, int count, byte[] byteNow)
|
||||
private void updateDeleteLatestVersionTimestamp(Cell cell, Get get, int count, byte[] byteNow)
|
||||
throws IOException {
|
||||
List<Cell> result = get(get, false);
|
||||
|
||||
@ -3269,7 +3243,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
startRegionOperation(Operation.PUT);
|
||||
try {
|
||||
// All edits for the given row (across all column families) must happen atomically.
|
||||
doBatchMutate(put);
|
||||
mutate(put);
|
||||
} finally {
|
||||
closeRegionOperation(Operation.PUT);
|
||||
}
|
||||
@ -3316,7 +3290,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* Visitor interface for batch operations
|
||||
*/
|
||||
@FunctionalInterface
|
||||
public interface Visitor {
|
||||
interface Visitor {
|
||||
/**
|
||||
* @param index operation index
|
||||
* @return If true continue visiting remaining entries, break otherwise
|
||||
@ -3722,14 +3696,17 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
|
||||
|
||||
/**
|
||||
* Batch of mutation operations. Base class is shared with {@link ReplayBatchOperation} as most
|
||||
* of the logic is same.
|
||||
* Batch of mutation operations. Base class is shared with {@link ReplayBatchOperation} as most of
|
||||
* the logic is same.
|
||||
*/
|
||||
static class MutationBatchOperation extends BatchOperation<Mutation> {
|
||||
private static class MutationBatchOperation extends BatchOperation<Mutation> {
|
||||
|
||||
private long nonceGroup;
|
||||
|
||||
private long nonce;
|
||||
|
||||
public MutationBatchOperation(final HRegion region, Mutation[] operations, boolean atomic,
|
||||
long nonceGroup, long nonce) {
|
||||
long nonceGroup, long nonce) {
|
||||
super(region, operations);
|
||||
this.atomic = atomic;
|
||||
this.nonceGroup = nonceGroup;
|
||||
@ -4364,10 +4341,12 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* Batch of mutations for replay. Base class is shared with {@link MutationBatchOperation} as most
|
||||
* of the logic is same.
|
||||
*/
|
||||
static class ReplayBatchOperation extends BatchOperation<MutationReplay> {
|
||||
private static final class ReplayBatchOperation extends BatchOperation<MutationReplay> {
|
||||
|
||||
private long origLogSeqNum = 0;
|
||||
|
||||
public ReplayBatchOperation(final HRegion region, MutationReplay[] operations,
|
||||
long origLogSeqNum) {
|
||||
long origLogSeqNum) {
|
||||
super(region, operations);
|
||||
this.origLogSeqNum = origLogSeqNum;
|
||||
}
|
||||
@ -4475,12 +4454,12 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
public OperationStatus[] batchMutate(Mutation[] mutations, boolean atomic, long nonceGroup,
|
||||
long nonce) throws IOException {
|
||||
private OperationStatus[] batchMutate(Mutation[] mutations, boolean atomic, long nonceGroup,
|
||||
long nonce) throws IOException {
|
||||
// As it stands, this is used for 3 things
|
||||
// * batchMutate with single mutation - put/delete/increment/append, separate or from
|
||||
// checkAndMutate.
|
||||
// * coprocessor calls (see ex. BulkDeleteEndpoint).
|
||||
// * batchMutate with single mutation - put/delete/increment/append, separate or from
|
||||
// checkAndMutate.
|
||||
// * coprocessor calls (see ex. BulkDeleteEndpoint).
|
||||
// So nonces are not really ever used by HBase. They could be by coprocs, and checkAnd...
|
||||
return batchMutate(new MutationBatchOperation(this, mutations, atomic, nonceGroup, nonce));
|
||||
}
|
||||
@ -4488,8 +4467,12 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
@Override
|
||||
public OperationStatus[] batchMutate(Mutation[] mutations) throws IOException {
|
||||
// If the mutations has any Increment/Append operations, we need to do batchMutate atomically
|
||||
boolean atomic = Arrays.stream(mutations)
|
||||
.anyMatch(m -> m instanceof Increment || m instanceof Append);
|
||||
boolean atomic =
|
||||
Arrays.stream(mutations).anyMatch(m -> m instanceof Increment || m instanceof Append);
|
||||
return batchMutate(mutations, atomic);
|
||||
}
|
||||
|
||||
OperationStatus[] batchMutate(Mutation[] mutations, boolean atomic) throws IOException {
|
||||
return batchMutate(mutations, atomic, HConstants.NO_NONCE, HConstants.NO_NONCE);
|
||||
}
|
||||
|
||||
@ -4519,24 +4502,23 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
|
||||
/**
|
||||
* Perform a batch of mutations.
|
||||
*
|
||||
* <p/>
|
||||
* Operations in a batch are stored with highest durability specified of for all operations in a
|
||||
* batch, except for {@link Durability#SKIP_WAL}.
|
||||
*
|
||||
* <p>This function is called from {@link #batchReplay(WALSplitUtil.MutationReplay[], long)} with
|
||||
* <p/>
|
||||
* This function is called from {@link #batchReplay(WALSplitUtil.MutationReplay[], long)} with
|
||||
* {@link ReplayBatchOperation} instance and {@link #batchMutate(Mutation[])} with
|
||||
* {@link MutationBatchOperation} instance as an argument. As the processing of replay batch
|
||||
* and mutation batch is very similar, lot of code is shared by providing generic methods in
|
||||
* base class {@link BatchOperation}. The logic for this method and
|
||||
* {@link #doMiniBatchMutate(BatchOperation)} is implemented using methods in base class which
|
||||
* are overridden by derived classes to implement special behavior.
|
||||
*
|
||||
* {@link MutationBatchOperation} instance as an argument. As the processing of replay batch and
|
||||
* mutation batch is very similar, lot of code is shared by providing generic methods in base
|
||||
* class {@link BatchOperation}. The logic for this method and
|
||||
* {@link #doMiniBatchMutate(BatchOperation)} is implemented using methods in base class which are
|
||||
* overridden by derived classes to implement special behavior.
|
||||
* @param batchOp contains the list of mutations
|
||||
* @return an array of OperationStatus which internally contains the
|
||||
* OperationStatusCode and the exceptionMessage if any.
|
||||
* @return an array of OperationStatus which internally contains the OperationStatusCode and the
|
||||
* exceptionMessage if any.
|
||||
* @throws IOException if an IO problem is encountered
|
||||
*/
|
||||
OperationStatus[] batchMutate(BatchOperation<?> batchOp) throws IOException {
|
||||
private OperationStatus[] batchMutate(BatchOperation<?> batchOp) throws IOException {
|
||||
boolean initialized = false;
|
||||
batchOp.startRegionOperation();
|
||||
try {
|
||||
@ -4690,7 +4672,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* Returns effective durability from the passed durability and
|
||||
* the table descriptor.
|
||||
*/
|
||||
protected Durability getEffectiveDurability(Durability d) {
|
||||
private Durability getEffectiveDurability(Durability d) {
|
||||
return d == Durability.USE_DEFAULT ? this.regionDurability : d;
|
||||
}
|
||||
|
||||
@ -4879,7 +4861,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
// All edits for the given row (across all column families) must happen atomically.
|
||||
Result r;
|
||||
if (mutation != null) {
|
||||
r = doBatchMutate(mutation, true).getResult();
|
||||
r = mutate(mutation, true).getResult();
|
||||
} else {
|
||||
r = mutateRow(rowMutations);
|
||||
}
|
||||
@ -4939,27 +4921,26 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
return matches;
|
||||
}
|
||||
|
||||
private OperationStatus doBatchMutate(Mutation mutation) throws IOException {
|
||||
return doBatchMutate(mutation, false);
|
||||
private OperationStatus mutate(Mutation mutation) throws IOException {
|
||||
return mutate(mutation, false);
|
||||
}
|
||||
|
||||
private OperationStatus doBatchMutate(Mutation mutation, boolean atomic) throws IOException {
|
||||
return doBatchMutate(mutation, atomic, HConstants.NO_NONCE, HConstants.NO_NONCE);
|
||||
private OperationStatus mutate(Mutation mutation, boolean atomic) throws IOException {
|
||||
return mutate(mutation, atomic, HConstants.NO_NONCE, HConstants.NO_NONCE);
|
||||
}
|
||||
|
||||
private OperationStatus doBatchMutate(Mutation mutation, boolean atomic, long nonceGroup,
|
||||
long nonce) throws IOException {
|
||||
OperationStatus[] batchMutate = this.batchMutate(new Mutation[]{mutation}, atomic,
|
||||
nonceGroup, nonce);
|
||||
if (batchMutate[0].getOperationStatusCode().equals(OperationStatusCode.SANITY_CHECK_FAILURE)) {
|
||||
throw new FailedSanityCheckException(batchMutate[0].getExceptionMsg());
|
||||
} else if (batchMutate[0].getOperationStatusCode().equals(OperationStatusCode.BAD_FAMILY)) {
|
||||
throw new NoSuchColumnFamilyException(batchMutate[0].getExceptionMsg());
|
||||
} else if (batchMutate[0].getOperationStatusCode().equals(
|
||||
OperationStatusCode.STORE_TOO_BUSY)) {
|
||||
throw new RegionTooBusyException(batchMutate[0].getExceptionMsg());
|
||||
private OperationStatus mutate(Mutation mutation, boolean atomic, long nonceGroup, long nonce)
|
||||
throws IOException {
|
||||
OperationStatus[] status =
|
||||
this.batchMutate(new Mutation[] { mutation }, atomic, nonceGroup, nonce);
|
||||
if (status[0].getOperationStatusCode().equals(OperationStatusCode.SANITY_CHECK_FAILURE)) {
|
||||
throw new FailedSanityCheckException(status[0].getExceptionMsg());
|
||||
} else if (status[0].getOperationStatusCode().equals(OperationStatusCode.BAD_FAMILY)) {
|
||||
throw new NoSuchColumnFamilyException(status[0].getExceptionMsg());
|
||||
} else if (status[0].getOperationStatusCode().equals(OperationStatusCode.STORE_TOO_BUSY)) {
|
||||
throw new RegionTooBusyException(status[0].getExceptionMsg());
|
||||
}
|
||||
return batchMutate[0];
|
||||
return status[0];
|
||||
}
|
||||
|
||||
/**
|
||||
@ -5018,7 +4999,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
/**
|
||||
* Possibly rewrite incoming cell tags.
|
||||
*/
|
||||
void rewriteCellTags(Map<byte[], List<Cell>> familyMap, final Mutation m) {
|
||||
private void rewriteCellTags(Map<byte[], List<Cell>> familyMap, final Mutation m) {
|
||||
// Check if we have any work to do and early out otherwise
|
||||
// Update these checks as more logic is added here
|
||||
if (m.getTTL() == Long.MAX_VALUE) {
|
||||
@ -5040,15 +5021,17 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
/**
|
||||
* Check if resources to support an update.
|
||||
*
|
||||
* We throw RegionTooBusyException if above memstore limit
|
||||
* and expect client to retry using some kind of backoff
|
||||
*/
|
||||
void checkResources() throws RegionTooBusyException {
|
||||
* <p/>
|
||||
* We throw RegionTooBusyException if above memstore limit and expect client to retry using some
|
||||
* kind of backoff
|
||||
*/
|
||||
private void checkResources() throws RegionTooBusyException {
|
||||
// If catalog region, do not impose resource constraints or block updates.
|
||||
if (this.getRegionInfo().isMetaRegion()) return;
|
||||
if (this.getRegionInfo().isMetaRegion()) {
|
||||
return;
|
||||
}
|
||||
|
||||
MemStoreSize mss = this.memStoreSizing.getMemStoreSize();
|
||||
if (mss.getHeapSize() + mss.getOffHeapSize() > this.blockingMemStoreSize) {
|
||||
@ -5073,13 +5056,13 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
/**
|
||||
* @throws IOException Throws exception if region is in read-only mode.
|
||||
*/
|
||||
protected void checkReadOnly() throws IOException {
|
||||
private void checkReadOnly() throws IOException {
|
||||
if (isReadOnly()) {
|
||||
throw new DoNotRetryIOException("region is read only");
|
||||
}
|
||||
}
|
||||
|
||||
protected void checkReadsEnabled() throws IOException {
|
||||
private void checkReadsEnabled() throws IOException {
|
||||
if (!this.writestate.readsEnabled) {
|
||||
throw new IOException(getRegionInfo().getEncodedName()
|
||||
+ ": The region's reads are disabled. Cannot serve the request");
|
||||
@ -5093,23 +5076,6 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
this.writestate.setReadsEnabled(readsEnabled);
|
||||
}
|
||||
|
||||
/**
|
||||
* Add updates first to the wal and then add values to memstore.
|
||||
* Warning: Assumption is caller has lock on passed in row.
|
||||
* @param edits Cell updates by column
|
||||
* @throws IOException
|
||||
*/
|
||||
void put(final byte [] row, byte [] family, List<Cell> edits)
|
||||
throws IOException {
|
||||
NavigableMap<byte[], List<Cell>> familyMap;
|
||||
familyMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
|
||||
familyMap.put(family, edits);
|
||||
Put p = new Put(row);
|
||||
p.setFamilyCellMap(familyMap);
|
||||
doBatchMutate(p);
|
||||
}
|
||||
|
||||
/**
|
||||
* @param delta If we are doing delta changes -- e.g. increment/append -- then this flag will be
|
||||
* set; when set we will run operations that make sense in the increment/append scenario
|
||||
@ -6004,7 +5970,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* Currently, this method is used to drop memstore to prevent memory leak
|
||||
* when replaying recovered.edits while opening region.
|
||||
*/
|
||||
public MemStoreSize dropMemStoreContents() throws IOException {
|
||||
private MemStoreSize dropMemStoreContents() throws IOException {
|
||||
MemStoreSizing totalFreedSize = new NonThreadSafeMemStoreSizing();
|
||||
this.updatesLock.writeLock().lock();
|
||||
try {
|
||||
@ -8054,11 +8020,11 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
|
||||
/**
|
||||
* Open HRegion.
|
||||
* <p/>
|
||||
* Calls initialize and sets sequenceId.
|
||||
* @return Returns <code>this</code>
|
||||
*/
|
||||
protected HRegion openHRegion(final CancelableProgressable reporter)
|
||||
throws IOException {
|
||||
private HRegion openHRegion(final CancelableProgressable reporter) throws IOException {
|
||||
try {
|
||||
// Refuse to open the region if we are missing local compression support
|
||||
TableDescriptorChecker.checkCompression(htableDescriptor);
|
||||
@ -8200,7 +8166,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
return get(get, withCoprocessor, HConstants.NO_NONCE, HConstants.NO_NONCE);
|
||||
}
|
||||
|
||||
public List<Cell> get(Get get, boolean withCoprocessor, long nonceGroup, long nonce)
|
||||
private List<Cell> get(Get get, boolean withCoprocessor, long nonceGroup, long nonce)
|
||||
throws IOException {
|
||||
List<Cell> results = new ArrayList<>();
|
||||
long before = EnvironmentEdgeManager.currentTime();
|
||||
@ -8564,7 +8530,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
startRegionOperation(Operation.APPEND);
|
||||
try {
|
||||
// All edits for the given row (across all column families) must happen atomically.
|
||||
return doBatchMutate(append, true, nonceGroup, nonce).getResult();
|
||||
return mutate(append, true, nonceGroup, nonce).getResult();
|
||||
} finally {
|
||||
closeRegionOperation(Operation.APPEND);
|
||||
}
|
||||
@ -8581,7 +8547,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
startRegionOperation(Operation.INCREMENT);
|
||||
try {
|
||||
// All edits for the given row (across all column families) must happen atomically.
|
||||
return doBatchMutate(increment, true, nonceGroup, nonce).getResult();
|
||||
return mutate(increment, true, nonceGroup, nonce).getResult();
|
||||
} finally {
|
||||
closeRegionOperation(Operation.INCREMENT);
|
||||
}
|
||||
@ -9143,15 +9109,11 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
flushesQueued.increment();
|
||||
}
|
||||
|
||||
public long getReadPoint() {
|
||||
return getReadPoint(IsolationLevel.READ_COMMITTED);
|
||||
}
|
||||
|
||||
/**
|
||||
* If a handler thread is eligible for interrupt, make it ineligible. Should be paired
|
||||
* with {{@link #enableInterrupts()}.
|
||||
*/
|
||||
protected void disableInterrupts() {
|
||||
void disableInterrupts() {
|
||||
regionLockHolders.computeIfPresent(Thread.currentThread(), (t,b) -> false);
|
||||
}
|
||||
|
||||
@ -9159,7 +9121,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* If a handler thread was made ineligible for interrupt via {{@link #disableInterrupts()},
|
||||
* make it eligible again. No-op if interrupts are already enabled.
|
||||
*/
|
||||
protected void enableInterrupts() {
|
||||
void enableInterrupts() {
|
||||
regionLockHolders.computeIfPresent(Thread.currentThread(), (t,b) -> true);
|
||||
}
|
||||
|
||||
@ -9331,7 +9293,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||
* features
|
||||
* @param conf region configurations
|
||||
*/
|
||||
static void decorateRegionConfiguration(Configuration conf) {
|
||||
private static void decorateRegionConfiguration(Configuration conf) {
|
||||
if (ReplicationUtils.isReplicationForBulkLoadDataEnabled(conf)) {
|
||||
String plugins = conf.get(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY,"");
|
||||
String replicationCoprocessorClass = ReplicationObserver.class.getCanonicalName();
|
||||
|
@ -999,8 +999,7 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
||||
Arrays.sort(mArray, (v1, v2) -> Row.COMPARATOR.compare(v1, v2));
|
||||
}
|
||||
|
||||
OperationStatus[] codes = region.batchMutate(mArray, atomic, HConstants.NO_NONCE,
|
||||
HConstants.NO_NONCE);
|
||||
OperationStatus[] codes = region.batchMutate(mArray, atomic);
|
||||
|
||||
// When atomic is true, it indicates that the mutateRow API or the batch API with
|
||||
// RowMutations is called. In this case, we need to merge the results of the
|
||||
|
@ -709,8 +709,7 @@ public class TestCompactingMemStore extends TestDefaultMemStore {
|
||||
mss = memstore.getFlushableSize();
|
||||
MemStoreSnapshot snapshot = memstore.snapshot(); // push keys to snapshot
|
||||
// simulate flusher
|
||||
region.decrMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize(),
|
||||
mss.getCellsCount());
|
||||
region.decrMemStoreSize(mss);
|
||||
ImmutableSegment s = memstore.getSnapshot();
|
||||
assertEquals(7, s.getCellsCount());
|
||||
assertEquals(0, regionServicesForStores.getMemStoreSize());
|
||||
@ -787,8 +786,7 @@ public class TestCompactingMemStore extends TestDefaultMemStore {
|
||||
mss = memstore.getFlushableSize();
|
||||
MemStoreSnapshot snapshot = memstore.snapshot(); // push keys to snapshot
|
||||
// simulate flusher
|
||||
region.decrMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize(),
|
||||
mss.getCellsCount());
|
||||
region.decrMemStoreSize(mss);
|
||||
ImmutableSegment s = memstore.getSnapshot();
|
||||
assertEquals(4, s.getCellsCount());
|
||||
assertEquals(0, regionServicesForStores.getMemStoreSize());
|
||||
|
@ -282,8 +282,7 @@ public class TestCompactingToCellFlatMapMemStore extends TestCompactingMemStore
|
||||
mss = memstore.getFlushableSize();
|
||||
MemStoreSnapshot snapshot = memstore.snapshot(); // push keys to snapshot
|
||||
// simulate flusher
|
||||
region.decrMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize(),
|
||||
mss.getCellsCount());
|
||||
region.decrMemStoreSize(mss);
|
||||
ImmutableSegment s = memstore.getSnapshot();
|
||||
assertEquals(4, s.getCellsCount());
|
||||
assertEquals(0, regionServicesForStores.getMemStoreSize());
|
||||
|
@ -26,7 +26,9 @@ import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.List;
|
||||
import java.util.NavigableMap;
|
||||
import java.util.Objects;
|
||||
import java.util.TreeMap;
|
||||
import java.util.concurrent.atomic.AtomicLong;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
@ -968,28 +970,23 @@ public class TestDefaultMemStore {
|
||||
}
|
||||
|
||||
/**
|
||||
* Inserts a new region's meta information into the passed
|
||||
* <code>meta</code> region. Used by the HMaster bootstrap code adding
|
||||
* new table to hbase:meta table.
|
||||
*
|
||||
* Inserts a new region's meta information into the passed <code>meta</code> region.
|
||||
* @param meta hbase:meta HRegion to be updated
|
||||
* @param r HRegion to add to <code>meta</code>
|
||||
*
|
||||
* @throws IOException
|
||||
*/
|
||||
public static void addRegionToMETA(final HRegion meta, final HRegion r) throws IOException {
|
||||
meta.checkResources();
|
||||
private static void addRegionToMETA(final HRegion meta, final HRegion r) throws IOException {
|
||||
// The row key is the region name
|
||||
byte[] row = r.getRegionInfo().getRegionName();
|
||||
final long now = EnvironmentEdgeManager.currentTime();
|
||||
final List<Cell> cells = new ArrayList<>(2);
|
||||
cells.add(new KeyValue(row, HConstants.CATALOG_FAMILY,
|
||||
HConstants.REGIONINFO_QUALIFIER, now, RegionInfo.toByteArray(r.getRegionInfo())));
|
||||
cells.add(new KeyValue(row, HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER, now,
|
||||
RegionInfo.toByteArray(r.getRegionInfo())));
|
||||
// Set into the root table the version of the meta table.
|
||||
cells.add(new KeyValue(row, HConstants.CATALOG_FAMILY,
|
||||
HConstants.META_VERSION_QUALIFIER, now,
|
||||
cells.add(new KeyValue(row, HConstants.CATALOG_FAMILY, HConstants.META_VERSION_QUALIFIER, now,
|
||||
Bytes.toBytes(HConstants.META_VERSION)));
|
||||
meta.put(row, HConstants.CATALOG_FAMILY, cells);
|
||||
NavigableMap<byte[], List<Cell>> familyMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
familyMap.put(HConstants.CATALOG_FAMILY, cells);
|
||||
meta.put(new Put(row, HConstants.LATEST_TIMESTAMP, familyMap));
|
||||
}
|
||||
|
||||
private class EnvironmentEdgeForMemstoreTest implements EnvironmentEdge {
|
||||
|
@ -26,6 +26,7 @@ import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertNull;
|
||||
import static org.junit.Assert.assertThrows;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
import static org.mockito.ArgumentMatchers.any;
|
||||
@ -43,7 +44,6 @@ import static org.mockito.Mockito.when;
|
||||
import java.io.IOException;
|
||||
import java.io.InterruptedIOException;
|
||||
import java.math.BigDecimal;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.security.PrivilegedExceptionAction;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
@ -140,7 +140,6 @@ import org.apache.hadoop.hbase.io.hfile.HFile;
|
||||
import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
|
||||
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||
import org.apache.hadoop.hbase.regionserver.HRegion.MutationBatchOperation;
|
||||
import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
|
||||
import org.apache.hadoop.hbase.regionserver.Region.Operation;
|
||||
import org.apache.hadoop.hbase.regionserver.Region.RowLock;
|
||||
@ -1684,9 +1683,7 @@ public class TestHRegion {
|
||||
long syncs = prepareRegionForBachPut(puts, source, false);
|
||||
|
||||
// 1. Straight forward case, should succeed
|
||||
MutationBatchOperation batchOp = new MutationBatchOperation(region, puts, true,
|
||||
HConstants.NO_NONCE, HConstants.NO_NONCE);
|
||||
OperationStatus[] codes = this.region.batchMutate(batchOp);
|
||||
OperationStatus[] codes = this.region.batchMutate(puts, true);
|
||||
assertEquals(10, codes.length);
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertEquals(OperationStatusCode.SUCCESS, codes[i].getOperationStatusCode());
|
||||
@ -1700,15 +1697,11 @@ public class TestHRegion {
|
||||
MultithreadedTestUtil.TestContext ctx = new MultithreadedTestUtil.TestContext(CONF);
|
||||
final AtomicReference<IOException> retFromThread = new AtomicReference<>();
|
||||
final CountDownLatch finishedPuts = new CountDownLatch(1);
|
||||
final MutationBatchOperation finalBatchOp = new MutationBatchOperation(region, puts, true,
|
||||
HConstants
|
||||
.NO_NONCE,
|
||||
HConstants.NO_NONCE);
|
||||
TestThread putter = new TestThread(ctx) {
|
||||
@Override
|
||||
public void doWork() throws IOException {
|
||||
try {
|
||||
region.batchMutate(finalBatchOp);
|
||||
region.batchMutate(puts, true);
|
||||
} catch (IOException ioe) {
|
||||
LOG.error("test failed!", ioe);
|
||||
retFromThread.set(ioe);
|
||||
@ -1735,10 +1728,8 @@ public class TestHRegion {
|
||||
// 3. Exception thrown in validation
|
||||
LOG.info("Next a batch put with one invalid family");
|
||||
puts[5].addColumn(Bytes.toBytes("BAD_CF"), qual, value);
|
||||
batchOp = new MutationBatchOperation(region, puts, true, HConstants.NO_NONCE,
|
||||
HConstants.NO_NONCE);
|
||||
thrown.expect(NoSuchColumnFamilyException.class);
|
||||
this.region.batchMutate(batchOp);
|
||||
this.region.batchMutate(puts, true);
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -3177,23 +3168,19 @@ public class TestHRegion {
|
||||
List<Cell> kvs = new ArrayList<>();
|
||||
kvs.add(new KeyValue(row1, fam4, null, null));
|
||||
|
||||
byte[] forUnitTestsOnly = Bytes.toBytes("ForUnitTestsOnly");
|
||||
|
||||
// testing existing family
|
||||
byte[] family = fam2;
|
||||
NavigableMap<byte[], List<Cell>> deleteMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
deleteMap.put(family, kvs);
|
||||
region.delete(deleteMap, Durability.SYNC_WAL);
|
||||
deleteMap.put(fam2, kvs);
|
||||
region.delete(new Delete(forUnitTestsOnly, HConstants.LATEST_TIMESTAMP, deleteMap));
|
||||
|
||||
// testing non existing family
|
||||
boolean ok = false;
|
||||
family = fam4;
|
||||
try {
|
||||
deleteMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
deleteMap.put(family, kvs);
|
||||
region.delete(deleteMap, Durability.SYNC_WAL);
|
||||
} catch (Exception e) {
|
||||
ok = true;
|
||||
}
|
||||
assertTrue("Family " + new String(family, StandardCharsets.UTF_8) + " does exist", ok);
|
||||
NavigableMap<byte[], List<Cell>> deleteMap2 = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
deleteMap2.put(fam4, kvs);
|
||||
assertThrows("Family " + Bytes.toString(fam4) + " does exist",
|
||||
NoSuchColumnFamilyException.class,
|
||||
() -> region.delete(new Delete(forUnitTestsOnly, HConstants.LATEST_TIMESTAMP, deleteMap2)));
|
||||
}
|
||||
|
||||
@Test
|
||||
@ -3542,6 +3529,8 @@ public class TestHRegion {
|
||||
byte[] col2 = Bytes.toBytes("col2");
|
||||
byte[] col3 = Bytes.toBytes("col3");
|
||||
|
||||
byte[] forUnitTestsOnly = Bytes.toBytes("ForUnitTestsOnly");
|
||||
|
||||
// Setting up region
|
||||
this.region = initHRegion(tableName, method, CONF, fam1);
|
||||
// Building checkerList
|
||||
@ -3552,12 +3541,12 @@ public class TestHRegion {
|
||||
|
||||
NavigableMap<byte[], List<Cell>> deleteMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
|
||||
deleteMap.put(fam1, kvs);
|
||||
region.delete(deleteMap, Durability.SYNC_WAL);
|
||||
region.delete(new Delete(forUnitTestsOnly, HConstants.LATEST_TIMESTAMP, deleteMap));
|
||||
|
||||
// extract the key values out the memstore:
|
||||
// This is kinda hacky, but better than nothing...
|
||||
long now = System.currentTimeMillis();
|
||||
AbstractMemStore memstore = (AbstractMemStore)region.getStore(fam1).memstore;
|
||||
AbstractMemStore memstore = (AbstractMemStore) region.getStore(fam1).memstore;
|
||||
Cell firstCell = memstore.getActive().first();
|
||||
assertTrue(firstCell.getTimestamp() <= now);
|
||||
now = firstCell.getTimestamp();
|
||||
|
Loading…
x
Reference in New Issue
Block a user