HBASE-26142 NullPointerException when set 'hbase.hregion.memstore.mslab.indexchunksize.percent' to zero (#3531)

Signed-off-by: Duo Zhang <zhangduo@apache.org>
This commit is contained in:
chenglei 2021-08-05 20:48:18 +08:00 committed by GitHub
parent da950b9be2
commit d4aed4d59e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
6 changed files with 255 additions and 51 deletions

View File

@ -19,6 +19,8 @@ package org.apache.hadoop.hbase.regionserver;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicInteger;
import org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.yetus.audience.InterfaceAudience; import org.apache.yetus.audience.InterfaceAudience;
@ -49,6 +51,8 @@ public abstract class Chunk {
// The unique id associated with the chunk. // The unique id associated with the chunk.
private final int id; private final int id;
private final ChunkType chunkType;
// indicates if the chunk is formed by ChunkCreator#MemstorePool // indicates if the chunk is formed by ChunkCreator#MemstorePool
private final boolean fromPool; private final boolean fromPool;
@ -58,8 +62,8 @@ public abstract class Chunk {
* @param size in bytes * @param size in bytes
* @param id the chunk id * @param id the chunk id
*/ */
public Chunk(int size, int id) { public Chunk(int size, int id, ChunkType chunkType) {
this(size, id, false); this(size, id, chunkType, false);
} }
/** /**
@ -69,9 +73,10 @@ public abstract class Chunk {
* @param id the chunk id * @param id the chunk id
* @param fromPool if the chunk is formed by pool * @param fromPool if the chunk is formed by pool
*/ */
public Chunk(int size, int id, boolean fromPool) { public Chunk(int size, int id, ChunkType chunkType, boolean fromPool) {
this.size = size; this.size = size;
this.id = id; this.id = id;
this.chunkType = chunkType;
this.fromPool = fromPool; this.fromPool = fromPool;
} }
@ -79,16 +84,24 @@ public abstract class Chunk {
return this.id; return this.id;
} }
ChunkType getChunkType() {
return this.chunkType;
}
boolean isFromPool() { boolean isFromPool() {
return this.fromPool; return this.fromPool;
} }
boolean isJumbo() { boolean isJumbo() {
return size > ChunkCreator.getInstance().getChunkSize(); return chunkType == ChunkCreator.ChunkType.JUMBO_CHUNK;
} }
boolean isIndexChunk() { boolean isIndexChunk() {
return size == ChunkCreator.getInstance().getChunkSize(ChunkCreator.ChunkType.INDEX_CHUNK); return chunkType == ChunkCreator.ChunkType.INDEX_CHUNK;
}
boolean isDataChunk() {
return chunkType == ChunkCreator.ChunkType.DATA_CHUNK;
} }
/** /**

View File

@ -77,6 +77,7 @@ public class ChunkCreator {
static boolean chunkPoolDisabled = false; static boolean chunkPoolDisabled = false;
private MemStoreChunkPool dataChunksPool; private MemStoreChunkPool dataChunksPool;
private final int chunkSize; private final int chunkSize;
private int indexChunkSize;
private MemStoreChunkPool indexChunksPool; private MemStoreChunkPool indexChunksPool;
ChunkCreator(int chunkSize, boolean offheap, long globalMemStoreSize, float poolSizePercentage, ChunkCreator(int chunkSize, boolean offheap, long globalMemStoreSize, float poolSizePercentage,
@ -94,13 +95,14 @@ public class ChunkCreator {
HeapMemoryManager heapMemoryManager) { HeapMemoryManager heapMemoryManager) {
this.dataChunksPool = initializePool("data", globalMemStoreSize, this.dataChunksPool = initializePool("data", globalMemStoreSize,
(1 - indexChunkSizePercentage) * poolSizePercentage, (1 - indexChunkSizePercentage) * poolSizePercentage,
initialCountPercentage, chunkSize, heapMemoryManager); initialCountPercentage, chunkSize, ChunkType.DATA_CHUNK, heapMemoryManager);
// The index chunks pool is needed only when the index type is CCM. // The index chunks pool is needed only when the index type is CCM.
// Since the pools are not created at all when the index type isn't CCM, // Since the pools are not created at all when the index type isn't CCM,
// we don't need to check it here. // we don't need to check it here.
this.indexChunkSize = (int) (indexChunkSizePercentage * chunkSize);
this.indexChunksPool = initializePool("index", globalMemStoreSize, this.indexChunksPool = initializePool("index", globalMemStoreSize,
indexChunkSizePercentage * poolSizePercentage, indexChunkSizePercentage * poolSizePercentage,
initialCountPercentage, (int) (indexChunkSizePercentage * chunkSize), initialCountPercentage, this.indexChunkSize, ChunkType.INDEX_CHUNK,
heapMemoryManager); heapMemoryManager);
} }
@ -163,14 +165,20 @@ public class ChunkCreator {
Chunk getChunk(CompactingMemStore.IndexType chunkIndexType, ChunkType chunkType) { Chunk getChunk(CompactingMemStore.IndexType chunkIndexType, ChunkType chunkType) {
switch (chunkType) { switch (chunkType) {
case INDEX_CHUNK: case INDEX_CHUNK:
if (indexChunksPool != null) { if (indexChunksPool == null) {
return getChunk(chunkIndexType, indexChunksPool.getChunkSize()); if (indexChunkSize <= 0) {
throw new IllegalArgumentException(
"chunkType is INDEX_CHUNK but indexChunkSize is:[" + this.indexChunkSize + "]");
}
return getChunk(chunkIndexType, chunkType, indexChunkSize);
} else {
return getChunk(chunkIndexType, chunkType, indexChunksPool.getChunkSize());
} }
case DATA_CHUNK: case DATA_CHUNK:
if (dataChunksPool == null) { if (dataChunksPool == null) {
return getChunk(chunkIndexType, chunkSize); return getChunk(chunkIndexType, chunkType, chunkSize);
} else { } else {
return getChunk(chunkIndexType, dataChunksPool.getChunkSize()); return getChunk(chunkIndexType, chunkType, dataChunksPool.getChunkSize());
} }
default: default:
throw new IllegalArgumentException( throw new IllegalArgumentException(
@ -184,14 +192,14 @@ public class ChunkCreator {
* @param chunkIndexType whether the requested chunk is going to be used with CellChunkMap index * @param chunkIndexType whether the requested chunk is going to be used with CellChunkMap index
* @param size the size of the chunk to be allocated, in bytes * @param size the size of the chunk to be allocated, in bytes
*/ */
Chunk getChunk(CompactingMemStore.IndexType chunkIndexType, int size) { Chunk getChunk(CompactingMemStore.IndexType chunkIndexType, ChunkType chunkType, int size) {
Chunk chunk = null; Chunk chunk = null;
MemStoreChunkPool pool = null; MemStoreChunkPool pool = null;
// if the size is suitable for one of the pools // if it is one of the pools
if (dataChunksPool != null && size == dataChunksPool.getChunkSize()) { if (dataChunksPool != null && chunkType == ChunkType.DATA_CHUNK) {
pool = dataChunksPool; pool = dataChunksPool;
} else if (indexChunksPool != null && size == indexChunksPool.getChunkSize()) { } else if (indexChunksPool != null && chunkType == ChunkType.INDEX_CHUNK) {
pool = indexChunksPool; pool = indexChunksPool;
} }
@ -211,7 +219,7 @@ public class ChunkCreator {
if (chunk == null) { if (chunk == null) {
// the second parameter explains whether CellChunkMap index is requested, // the second parameter explains whether CellChunkMap index is requested,
// in that case, put allocated on demand chunk mapping into chunkIdMap // in that case, put allocated on demand chunk mapping into chunkIdMap
chunk = createChunk(false, chunkIndexType, size); chunk = createChunk(false, chunkIndexType, chunkType, size);
} }
// now we need to actually do the expensive memory allocation step in case of a new chunk, // now we need to actually do the expensive memory allocation step in case of a new chunk,
@ -228,14 +236,15 @@ public class ChunkCreator {
*/ */
Chunk getJumboChunk(int jumboSize) { Chunk getJumboChunk(int jumboSize) {
int allocSize = jumboSize + SIZEOF_CHUNK_HEADER; int allocSize = jumboSize + SIZEOF_CHUNK_HEADER;
if (allocSize <= dataChunksPool.getChunkSize()) {
if (allocSize <= this.getChunkSize(ChunkType.DATA_CHUNK)) {
LOG.warn("Jumbo chunk size " + jumboSize + " must be more than regular chunk size " LOG.warn("Jumbo chunk size " + jumboSize + " must be more than regular chunk size "
+ dataChunksPool.getChunkSize() + ". Converting to regular chunk."); + this.getChunkSize(ChunkType.DATA_CHUNK) + ". Converting to regular chunk.");
return getChunk(CompactingMemStore.IndexType.CHUNK_MAP); return getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
} }
// the new chunk is going to hold the jumbo cell data and needs to be referenced by // the new chunk is going to hold the jumbo cell data and needs to be referenced by
// a strong map. Therefore the CCM index type // a strong map. Therefore the CCM index type
return getChunk(CompactingMemStore.IndexType.CHUNK_MAP, allocSize); return getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.JUMBO_CHUNK, allocSize);
} }
/** /**
@ -245,15 +254,16 @@ public class ChunkCreator {
* @param size the size of the chunk to be allocated, in bytes * @param size the size of the chunk to be allocated, in bytes
* @return the chunk * @return the chunk
*/ */
private Chunk createChunk(boolean pool, CompactingMemStore.IndexType chunkIndexType, int size) { private Chunk createChunk(boolean pool, CompactingMemStore.IndexType chunkIndexType,
ChunkType chunkType, int size) {
Chunk chunk = null; Chunk chunk = null;
int id = chunkID.getAndIncrement(); int id = chunkID.getAndIncrement();
assert id > 0; assert id > 0;
// do not create offheap chunk on demand // do not create offheap chunk on demand
if (pool && this.offheap) { if (pool && this.offheap) {
chunk = new OffheapChunk(size, id, pool); chunk = new OffheapChunk(size, id, chunkType, pool);
} else { } else {
chunk = new OnheapChunk(size, id, pool); chunk = new OnheapChunk(size, id, chunkType, pool);
} }
if (pool || (chunkIndexType == CompactingMemStore.IndexType.CHUNK_MAP)) { if (pool || (chunkIndexType == CompactingMemStore.IndexType.CHUNK_MAP)) {
// put the pool chunk into the chunkIdMap so it is not GC-ed // put the pool chunk into the chunkIdMap so it is not GC-ed
@ -264,12 +274,13 @@ public class ChunkCreator {
// Chunks from pool are created covered with strong references anyway // Chunks from pool are created covered with strong references anyway
// TODO: change to CHUNK_MAP if it is generally defined // TODO: change to CHUNK_MAP if it is generally defined
private Chunk createChunkForPool(CompactingMemStore.IndexType chunkIndexType, int chunkSize) { private Chunk createChunkForPool(CompactingMemStore.IndexType chunkIndexType, ChunkType chunkType,
int chunkSize) {
if (chunkSize != dataChunksPool.getChunkSize() && if (chunkSize != dataChunksPool.getChunkSize() &&
chunkSize != indexChunksPool.getChunkSize()) { chunkSize != indexChunksPool.getChunkSize()) {
return null; return null;
} }
return createChunk(true, chunkIndexType, chunkSize); return createChunk(true, chunkIndexType, chunkType, chunkSize);
} }
// Used to translate the ChunkID into a chunk ref // Used to translate the ChunkID into a chunk ref
@ -309,6 +320,7 @@ public class ChunkCreator {
*/ */
private class MemStoreChunkPool implements HeapMemoryTuneObserver { private class MemStoreChunkPool implements HeapMemoryTuneObserver {
private final int chunkSize; private final int chunkSize;
private final ChunkType chunkType;
private int maxCount; private int maxCount;
// A queue of reclaimed chunks // A queue of reclaimed chunks
@ -323,15 +335,18 @@ public class ChunkCreator {
private final LongAdder reusedChunkCount = new LongAdder(); private final LongAdder reusedChunkCount = new LongAdder();
private final String label; private final String label;
MemStoreChunkPool(String label, int chunkSize, int maxCount, int initialCount, MemStoreChunkPool(String label, int chunkSize, ChunkType chunkType, int maxCount,
int initialCount,
float poolSizePercentage) { float poolSizePercentage) {
this.label = label; this.label = label;
this.chunkSize = chunkSize; this.chunkSize = chunkSize;
this.chunkType = chunkType;
this.maxCount = maxCount; this.maxCount = maxCount;
this.poolSizePercentage = poolSizePercentage; this.poolSizePercentage = poolSizePercentage;
this.reclaimedChunks = new LinkedBlockingQueue<>(); this.reclaimedChunks = new LinkedBlockingQueue<>();
for (int i = 0; i < initialCount; i++) { for (int i = 0; i < initialCount; i++) {
Chunk chunk = createChunk(true, CompactingMemStore.IndexType.ARRAY_MAP, chunkSize); Chunk chunk =
createChunk(true, CompactingMemStore.IndexType.ARRAY_MAP, chunkType, chunkSize);
chunk.init(); chunk.init();
reclaimedChunks.add(chunk); reclaimedChunks.add(chunk);
} }
@ -367,7 +382,7 @@ public class ChunkCreator {
long created = this.chunkCount.get(); long created = this.chunkCount.get();
if (created < this.maxCount) { if (created < this.maxCount) {
if (this.chunkCount.compareAndSet(created, created + 1)) { if (this.chunkCount.compareAndSet(created, created + 1)) {
chunk = createChunkForPool(chunkIndexType, chunkSize); chunk = createChunkForPool(chunkIndexType, chunkType, chunkSize);
break; break;
} }
} else { } else {
@ -465,7 +480,7 @@ public class ChunkCreator {
} }
private MemStoreChunkPool initializePool(String label, long globalMemStoreSize, private MemStoreChunkPool initializePool(String label, long globalMemStoreSize,
float poolSizePercentage, float initialCountPercentage, int chunkSize, float poolSizePercentage, float initialCountPercentage, int chunkSize, ChunkType chunkType,
HeapMemoryManager heapMemoryManager) { HeapMemoryManager heapMemoryManager) {
if (poolSizePercentage <= 0) { if (poolSizePercentage <= 0) {
LOG.info("{} poolSizePercentage is less than 0. So not using pool", label); LOG.info("{} poolSizePercentage is less than 0. So not using pool", label);
@ -486,8 +501,8 @@ public class ChunkCreator {
int initialCount = (int) (initialCountPercentage * maxCount); int initialCount = (int) (initialCountPercentage * maxCount);
LOG.info("Allocating {} MemStoreChunkPool with chunk size {}, max count {}, initial count {}", LOG.info("Allocating {} MemStoreChunkPool with chunk size {}, max count {}, initial count {}",
label, StringUtils.byteDesc(chunkSize), maxCount, initialCount); label, StringUtils.byteDesc(chunkSize), maxCount, initialCount);
MemStoreChunkPool memStoreChunkPool = new MemStoreChunkPool(label, chunkSize, maxCount, MemStoreChunkPool memStoreChunkPool = new MemStoreChunkPool(label, chunkSize, chunkType,
initialCount, poolSizePercentage); maxCount, initialCount, poolSizePercentage);
if (heapMemoryManager != null && memStoreChunkPool != null) { if (heapMemoryManager != null && memStoreChunkPool != null) {
// Register with Heap Memory manager // Register with Heap Memory manager
heapMemoryManager.registerTuneObserver(memStoreChunkPool); heapMemoryManager.registerTuneObserver(memStoreChunkPool);
@ -578,6 +593,8 @@ public class ChunkCreator {
case INDEX_CHUNK: case INDEX_CHUNK:
if (indexChunksPool != null) { if (indexChunksPool != null) {
return indexChunksPool.getChunkSize(); return indexChunksPool.getChunkSize();
} else {
return indexChunkSize;
} }
case DATA_CHUNK: case DATA_CHUNK:
if (dataChunksPool != null) { if (dataChunksPool != null) {
@ -606,7 +623,7 @@ public class ChunkCreator {
if (chunk != null) { if (chunk != null) {
if (chunk.isFromPool() && chunk.isIndexChunk()) { if (chunk.isFromPool() && chunk.isIndexChunk()) {
indexChunksPool.putbackChunks(chunk); indexChunksPool.putbackChunks(chunk);
} else if (chunk.isFromPool() && chunk.size == dataChunksPool.getChunkSize()) { } else if (chunk.isFromPool() && chunk.isDataChunk()) {
dataChunksPool.putbackChunks(chunk); dataChunksPool.putbackChunks(chunk);
} else { } else {
// chunks which are not from one of the pools // chunks which are not from one of the pools
@ -621,5 +638,13 @@ public class ChunkCreator {
return; return;
} }
MemStoreChunkPool getIndexChunksPool() {
return this.indexChunksPool;
}
MemStoreChunkPool getDataChunksPool() {
return this.dataChunksPool;
}
} }

View File

@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.regionserver;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType;
import org.apache.yetus.audience.InterfaceAudience; import org.apache.yetus.audience.InterfaceAudience;
/** /**
@ -27,13 +28,13 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Private @InterfaceAudience.Private
public class OffheapChunk extends Chunk { public class OffheapChunk extends Chunk {
OffheapChunk(int size, int id) { OffheapChunk(int size, int id, ChunkType chunkType) {
// better if this is always created fromPool. This should not be called // better if this is always created fromPool. This should not be called
super(size, id); super(size, id, chunkType);
} }
OffheapChunk(int size, int id, boolean fromPool) { OffheapChunk(int size, int id, ChunkType chunkType, boolean fromPool) {
super(size, id, fromPool); super(size, id, chunkType, fromPool);
assert fromPool == true; assert fromPool == true;
} }

View File

@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.regionserver;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType;
import org.apache.yetus.audience.InterfaceAudience; import org.apache.yetus.audience.InterfaceAudience;
/** /**
@ -27,12 +28,12 @@ import org.apache.yetus.audience.InterfaceAudience;
@InterfaceAudience.Private @InterfaceAudience.Private
public class OnheapChunk extends Chunk { public class OnheapChunk extends Chunk {
OnheapChunk(int size, int id) { OnheapChunk(int size, int id, ChunkType chunkType) {
super(size, id); super(size, id, chunkType);
} }
OnheapChunk(int size, int id, boolean fromPool) { OnheapChunk(int size, int id, ChunkType chunkType, boolean fromPool) {
super(size, id, fromPool); super(size, id, chunkType, fromPool);
} }
@Override @Override

View File

@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.HBaseClassTestRule;
import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.KeyValueUtil; import org.apache.hadoop.hbase.KeyValueUtil;
import org.apache.hadoop.hbase.io.util.MemorySizeUtil; import org.apache.hadoop.hbase.io.util.MemorySizeUtil;
import org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType;
import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests;
import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.testclassification.SmallTests;
import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.ByteBufferUtils;
@ -330,7 +331,9 @@ public class TestCellFlatSet {
// allocate new chunks and use the data JUMBO chunk to hold the full data of the cells // allocate new chunks and use the data JUMBO chunk to hold the full data of the cells
// and the normal index chunk to hold the cell-representations // and the normal index chunk to hold the cell-representations
Chunk dataJumboChunk = Chunk dataJumboChunk =
chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, smallChunkSize); chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.JUMBO_CHUNK,
smallChunkSize);
assertTrue(dataJumboChunk.isJumbo());
Chunk idxChunk = chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP); Chunk idxChunk = chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
// the array of index chunks to be used as a basis for CellChunkMap // the array of index chunks to be used as a basis for CellChunkMap
Chunk[] chunkArray = new Chunk[8]; // according to test currently written 8 is way enough Chunk[] chunkArray = new Chunk[8]; // according to test currently written 8 is way enough
@ -364,7 +367,10 @@ public class TestCellFlatSet {
// Jumbo chunks are working only with one cell per chunk, thus always allocate a new jumbo // Jumbo chunks are working only with one cell per chunk, thus always allocate a new jumbo
// data chunk for next cell // data chunk for next cell
dataJumboChunk = chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP,smallChunkSize); dataJumboChunk =
chunkCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.JUMBO_CHUNK,
smallChunkSize);
assertTrue(dataJumboChunk.isJumbo());
dataBuffer = dataJumboChunk.getData(); dataBuffer = dataJumboChunk.getData();
dataOffset = ChunkCreator.SIZEOF_CHUNK_HEADER; dataOffset = ChunkCreator.SIZEOF_CHUNK_HEADER;
} }

View File

@ -20,18 +20,23 @@ package org.apache.hadoop.hbase.regionserver;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
import java.io.IOException; import java.io.IOException;
import java.lang.management.ManagementFactory; import java.lang.management.ManagementFactory;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.util.Collections;
import java.util.List; import java.util.List;
import java.util.Random; import java.util.Random;
import java.util.concurrent.atomic.AtomicReference;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.ByteBufferKeyValue; import org.apache.hadoop.hbase.ByteBufferKeyValue;
import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.HBaseClassTestRule;
import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.exceptions.UnexpectedStateException; import org.apache.hadoop.hbase.exceptions.UnexpectedStateException;
import org.apache.hadoop.hbase.io.util.MemorySizeUtil; import org.apache.hadoop.hbase.io.util.MemorySizeUtil;
import org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType;
import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests;
import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.testclassification.SmallTests;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
@ -237,22 +242,30 @@ public class TestMemStoreChunkPool {
ChunkCreator newCreator = new ChunkCreator(chunkSize, false, 400, 1, 0.5f, null, 0); ChunkCreator newCreator = new ChunkCreator(chunkSize, false, 400, 1, 0.5f, null, 0);
assertEquals(initialCount, newCreator.getPoolSize()); assertEquals(initialCount, newCreator.getPoolSize());
assertEquals(maxCount, newCreator.getMaxCount()); assertEquals(maxCount, newCreator.getMaxCount());
ChunkCreator.instance = newCreator;// Replace the global ref with the new one we created. // Replace the global ref with the new one we created.
// Used it for the testing. Later in finally we put // Used it for the testing. Later in finally we put
// back the original // back the original
ChunkCreator.instance = newCreator;
final KeyValue kv = new KeyValue(Bytes.toBytes("r"), Bytes.toBytes("f"), Bytes.toBytes("q"), final KeyValue kv = new KeyValue(Bytes.toBytes("r"), Bytes.toBytes("f"), Bytes.toBytes("q"),
new byte[valSize]); new byte[valSize]);
final AtomicReference<Throwable> exceptionRef = new AtomicReference<Throwable>();
try { try {
Runnable r = new Runnable() { Runnable r = new Runnable() {
@Override @Override
public void run() { public void run() {
MemStoreLAB memStoreLAB = new MemStoreLABImpl(conf); try {
for (int i = 0; i < maxCount; i++) { MemStoreLAB memStoreLAB = new MemStoreLABImpl(conf);
memStoreLAB.copyCellInto(kv);// Try allocate size = chunkSize. Means every for (int i = 0; i < maxCount; i++) {
// allocate call will result in a new chunk // Try allocate size = chunkSize. Means every
// allocate call will result in a new chunk
memStoreLAB.copyCellInto(kv);
}
// Close MemStoreLAB so that all chunks will be tried to be put back to pool
memStoreLAB.close();
} catch (Throwable execption) {
exceptionRef.set(execption);
} }
// Close MemStoreLAB so that all chunks will be tried to be put back to pool
memStoreLAB.close();
} }
}; };
Thread t1 = new Thread(r); Thread t1 = new Thread(r);
@ -264,9 +277,154 @@ public class TestMemStoreChunkPool {
t1.join(); t1.join();
t2.join(); t2.join();
t3.join(); t3.join();
assertTrue(newCreator.getPoolSize() <= maxCount); assertTrue(exceptionRef.get() == null);
assertTrue(newCreator.getPoolSize() <= maxCount && newCreator.getPoolSize() > 0);
} finally { } finally {
ChunkCreator.instance = oldCreator; ChunkCreator.instance = oldCreator;
} }
} }
// This test is for HBASE-26142, which throws NPE when indexChunksPool is null.
@Test
public void testNoIndexChunksPoolOrNoDataChunksPool() throws Exception {
final int maxCount = 10;
final int initialCount = 5;
final int newChunkSize = 40;
final int valSize = 7;
ChunkCreator oldCreator = ChunkCreator.getInstance();
try {
// Test dataChunksPool is not null and indexChunksPool is null
ChunkCreator newCreator = new ChunkCreator(newChunkSize, false, 400, 1, 0.5f, null, 0);
assertEquals(initialCount, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
assertEquals(maxCount, newCreator.getMaxCount());
assertEquals(0, newCreator.getMaxCount(ChunkType.INDEX_CHUNK));
assertTrue(newCreator.getDataChunksPool() != null);
assertTrue(newCreator.getIndexChunksPool() == null);
ChunkCreator.instance = newCreator;
final KeyValue kv = new KeyValue(Bytes.toBytes("r"), Bytes.toBytes("f"), Bytes.toBytes("q"),
new byte[valSize]);
MemStoreLAB memStoreLAB = new MemStoreLABImpl(conf);
memStoreLAB.copyCellInto(kv);
memStoreLAB.close();
assertEquals(initialCount, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
Chunk dataChunk = newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
assertTrue(dataChunk.isDataChunk());
assertTrue(dataChunk.isFromPool());
assertEquals(initialCount - 1, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
newCreator.putbackChunks(Collections.singleton(dataChunk.getId()));
assertEquals(initialCount, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
// We set ChunkCreator.indexChunkSize to 0, but we want to get a IndexChunk
try {
newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.INDEX_CHUNK);
fail();
} catch (IllegalArgumentException e) {
}
Chunk jumboChunk = newCreator.getJumboChunk(newChunkSize + 10);
assertTrue(jumboChunk.isJumbo());
assertTrue(!jumboChunk.isFromPool());
assertEquals(initialCount, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
// Test both dataChunksPool and indexChunksPool are null
newCreator = new ChunkCreator(newChunkSize, false, 400, 0, 0.5f, null, 0);
assertEquals(0, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
assertEquals(0, newCreator.getMaxCount());
assertEquals(0, newCreator.getMaxCount(ChunkType.INDEX_CHUNK));
assertTrue(newCreator.getDataChunksPool() == null);
assertTrue(newCreator.getIndexChunksPool() == null);
ChunkCreator.instance = newCreator;
memStoreLAB = new MemStoreLABImpl(conf);
memStoreLAB.copyCellInto(kv);
memStoreLAB.close();
assertEquals(0, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
dataChunk = newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
assertTrue(dataChunk.isDataChunk());
assertTrue(!dataChunk.isFromPool());
assertEquals(0, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
try {
// We set ChunkCreator.indexChunkSize to 0, but we want to get a IndexChunk
newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.INDEX_CHUNK);
fail();
} catch (IllegalArgumentException e) {
}
jumboChunk = newCreator.getJumboChunk(newChunkSize + 10);
assertTrue(jumboChunk.isJumbo());
assertTrue(!jumboChunk.isFromPool());
assertEquals(0, newCreator.getPoolSize());
assertEquals(0, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
// Test dataChunksPool is null and indexChunksPool is not null
newCreator = new ChunkCreator(newChunkSize, false, 400, 1, 0.5f, null, 1);
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
assertEquals(0, newCreator.getMaxCount());
assertEquals(maxCount, newCreator.getMaxCount(ChunkType.INDEX_CHUNK));
assertTrue(newCreator.getDataChunksPool() == null);
assertTrue(newCreator.getIndexChunksPool() != null);
assertEquals(newCreator.getChunkSize(ChunkType.DATA_CHUNK),
newCreator.getChunkSize(ChunkType.INDEX_CHUNK));
ChunkCreator.instance = newCreator;
memStoreLAB = new MemStoreLABImpl(conf);
memStoreLAB.copyCellInto(kv);
memStoreLAB.close();
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
dataChunk = newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
assertTrue(dataChunk.isDataChunk());
assertTrue(!dataChunk.isFromPool());
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
Chunk indexChunk =
newCreator.getChunk(CompactingMemStore.IndexType.CHUNK_MAP, ChunkType.INDEX_CHUNK);
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount - 1, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
assertTrue(indexChunk.isIndexChunk());
assertTrue(indexChunk.isFromPool());
newCreator.putbackChunks(Collections.singleton(indexChunk.getId()));
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
jumboChunk = newCreator.getJumboChunk(newChunkSize + 10);
assertTrue(jumboChunk.isJumbo());
assertTrue(!jumboChunk.isFromPool());
assertEquals(0, newCreator.getPoolSize());
assertEquals(initialCount, newCreator.getPoolSize(ChunkType.INDEX_CHUNK));
} finally {
ChunkCreator.instance = oldCreator;
}
// Test both dataChunksPool and indexChunksPool are not null
assertTrue(ChunkCreator.getInstance().getDataChunksPool() != null);
assertTrue(ChunkCreator.getInstance().getIndexChunksPool() != null);
Chunk dataChunk = ChunkCreator.getInstance().getChunk(CompactingMemStore.IndexType.CHUNK_MAP);
assertTrue(dataChunk.isDataChunk());
assertTrue(dataChunk.isFromPool());
Chunk indexChunk = ChunkCreator.getInstance().getChunk(CompactingMemStore.IndexType.CHUNK_MAP,
ChunkType.INDEX_CHUNK);
assertTrue(indexChunk.isIndexChunk());
assertTrue(indexChunk.isFromPool());
Chunk jumboChunk =
ChunkCreator.getInstance().getJumboChunk(ChunkCreator.getInstance().getChunkSize() + 10);
assertTrue(jumboChunk.isJumbo());
assertTrue(!jumboChunk.isFromPool());
}
} }