HBASE-11550 Custom value for BUCKET_CACHE_BUCKETS_KEY should be sorted (Gustavo Anatoly)
Conflicts: hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketAllocator.java
This commit is contained in:
parent
65375f8258
commit
cd59a023c4
|
@ -484,7 +484,7 @@ public class CacheConfig {
|
|||
if (configuredBucketSizes != null) {
|
||||
bucketSizes = new int[configuredBucketSizes.length];
|
||||
for (int i = 0; i < configuredBucketSizes.length; i++) {
|
||||
bucketSizes[i] = Integer.parseInt(configuredBucketSizes[i]);
|
||||
bucketSizes[i] = Integer.parseInt(configuredBucketSizes[i].trim());
|
||||
}
|
||||
}
|
||||
BucketCache bucketCache = null;
|
||||
|
|
|
@ -21,13 +21,11 @@
|
|||
package org.apache.hadoop.hbase.io.hfile.bucket;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.concurrent.atomic.AtomicLong;
|
||||
|
||||
import com.google.common.base.Objects;
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
|
@ -36,6 +34,10 @@ import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
|||
import org.apache.hadoop.hbase.io.hfile.bucket.BucketCache.BucketEntry;
|
||||
import org.codehaus.jackson.annotate.JsonIgnoreProperties;
|
||||
|
||||
import com.google.common.base.Objects;
|
||||
import com.google.common.base.Preconditions;
|
||||
import com.google.common.primitives.Ints;
|
||||
|
||||
/**
|
||||
* This class is used to allocate a block with specified size and free the block
|
||||
* when evicting. It manages an array of buckets, each bucket is associated with
|
||||
|
@ -299,11 +301,8 @@ public final class BucketAllocator {
|
|||
BucketAllocator(long availableSpace, int[] bucketSizes)
|
||||
throws BucketAllocatorException {
|
||||
this.bucketSizes = bucketSizes == null ? DEFAULT_BUCKET_SIZES : bucketSizes;
|
||||
int largestBucket = this.bucketSizes[0];
|
||||
for (int i : this.bucketSizes) {
|
||||
largestBucket = Math.max(largestBucket, i);
|
||||
}
|
||||
this.bigItemSize = largestBucket;
|
||||
Arrays.sort(this.bucketSizes);
|
||||
this.bigItemSize = Ints.max(this.bucketSizes);
|
||||
this.bucketCapacity = FEWEST_ITEMS_IN_BUCKET * bigItemSize;
|
||||
buckets = new Bucket[(int) (availableSpace / bucketCapacity)];
|
||||
if (buckets.length < this.bucketSizes.length)
|
||||
|
|
|
@ -52,6 +52,7 @@ import org.apache.hadoop.classification.InterfaceAudience;
|
|||
import org.apache.hadoop.hbase.io.HeapSize;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockCache;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockCacheKey;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockCacheUtil;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockPriority;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockType;
|
||||
import org.apache.hadoop.hbase.io.hfile.CacheStats;
|
||||
|
@ -59,7 +60,6 @@ import org.apache.hadoop.hbase.io.hfile.Cacheable;
|
|||
import org.apache.hadoop.hbase.io.hfile.CacheableDeserializer;
|
||||
import org.apache.hadoop.hbase.io.hfile.CacheableDeserializerIdManager;
|
||||
import org.apache.hadoop.hbase.io.hfile.CachedBlock;
|
||||
import org.apache.hadoop.hbase.io.hfile.BlockCacheUtil;
|
||||
import org.apache.hadoop.hbase.io.hfile.CombinedBlockCache;
|
||||
import org.apache.hadoop.hbase.io.hfile.HFileBlock;
|
||||
import org.apache.hadoop.hbase.util.ConcurrentIndex;
|
||||
|
@ -166,7 +166,6 @@ public class BucketCache implements BlockCache, HeapSize {
|
|||
private long cacheCapacity;
|
||||
/** Approximate block size */
|
||||
private final long blockSize;
|
||||
private final int[] bucketSizes;
|
||||
|
||||
/** Duration of IO errors tolerated before we disable cache, 1 min as default */
|
||||
private final int ioErrorsTolerationDuration;
|
||||
|
@ -228,7 +227,6 @@ public class BucketCache implements BlockCache, HeapSize {
|
|||
this.cacheCapacity = capacity;
|
||||
this.persistencePath = persistencePath;
|
||||
this.blockSize = blockSize;
|
||||
this.bucketSizes = bucketSizes;
|
||||
this.ioErrorsTolerationDuration = ioErrorsTolerationDuration;
|
||||
|
||||
bucketAllocator = new BucketAllocator(capacity, bucketSizes);
|
||||
|
@ -244,7 +242,7 @@ public class BucketCache implements BlockCache, HeapSize {
|
|||
|
||||
if (ioEngine.isPersistent() && persistencePath != null) {
|
||||
try {
|
||||
retrieveFromFile();
|
||||
retrieveFromFile(bucketSizes);
|
||||
} catch (IOException ioex) {
|
||||
LOG.error("Can't restore from file because of", ioex);
|
||||
} catch (ClassNotFoundException cnfe) {
|
||||
|
@ -866,7 +864,7 @@ public class BucketCache implements BlockCache, HeapSize {
|
|||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private void retrieveFromFile() throws IOException, BucketAllocatorException,
|
||||
private void retrieveFromFile(int[] bucketSizes) throws IOException, BucketAllocatorException,
|
||||
ClassNotFoundException {
|
||||
File persistenceFile = new File(persistencePath);
|
||||
if (!persistenceFile.exists()) {
|
||||
|
|
Loading…
Reference in New Issue