mirror of https://github.com/apache/druid.git
fix populateCache flag not working
This commit is contained in:
parent
0feddc3831
commit
5b4c04510f
|
@ -176,32 +176,37 @@ public class CachingClusteredClient<T> implements QueryRunner<T>
|
|||
queryCacheKey = null;
|
||||
}
|
||||
|
||||
// Pull cached segments from cache and remove from set of segments to query
|
||||
if (useCache && queryCacheKey != null) {
|
||||
if (queryCacheKey != null) {
|
||||
Map<Pair<ServerSelector, SegmentDescriptor>, Cache.NamedKey> cacheKeys = Maps.newHashMap();
|
||||
for (Pair<ServerSelector, SegmentDescriptor> e : segments) {
|
||||
cacheKeys.put(e, computeSegmentCacheKey(e.lhs.getSegment().getIdentifier(), e.rhs, queryCacheKey));
|
||||
for (Pair<ServerSelector, SegmentDescriptor> segment : segments) {
|
||||
final Cache.NamedKey segmentCacheKey = computeSegmentCacheKey(
|
||||
segment.lhs.getSegment().getIdentifier(),
|
||||
segment.rhs,
|
||||
queryCacheKey
|
||||
);
|
||||
cacheKeys.put(segment, segmentCacheKey);
|
||||
}
|
||||
|
||||
Map<Cache.NamedKey, byte[]> cachedValues = cache.getBulk(cacheKeys.values());
|
||||
// Pull cached segments from cache and remove from set of segments to query
|
||||
final Map<Cache.NamedKey, byte[]> cachedValues;
|
||||
if (useCache) {
|
||||
cachedValues = cache.getBulk(cacheKeys.values());
|
||||
} else {
|
||||
cachedValues = ImmutableMap.of();
|
||||
}
|
||||
|
||||
for (Map.Entry<Pair<ServerSelector, SegmentDescriptor>, Cache.NamedKey> entry : cacheKeys.entrySet()) {
|
||||
Pair<ServerSelector, SegmentDescriptor> segment = entry.getKey();
|
||||
Cache.NamedKey segmentCacheKey = entry.getValue();
|
||||
|
||||
final ServerSelector selector = segment.lhs;
|
||||
final SegmentDescriptor descriptor = segment.rhs;
|
||||
final Interval segmentQueryInterval = descriptor.getInterval();
|
||||
final Interval segmentQueryInterval = segment.rhs.getInterval();
|
||||
|
||||
final byte[] cachedValue = cachedValues.get(segmentCacheKey);
|
||||
|
||||
if (cachedValue != null) {
|
||||
cachedResults.add(Pair.of(segmentQueryInterval.getStart(), cachedValue));
|
||||
|
||||
// remove cached segment from set of segments to query
|
||||
segments.remove(segment);
|
||||
} else {
|
||||
final String segmentIdentifier = selector.getSegment().getIdentifier();
|
||||
cachedResults.add(Pair.of(segmentQueryInterval.getStart(), cachedValue));
|
||||
} else if (populateCache) {
|
||||
final String segmentIdentifier = segment.lhs.getSegment().getIdentifier();
|
||||
cachePopulatorMap.put(
|
||||
String.format("%s_%s", segmentIdentifier, segmentQueryInterval),
|
||||
new CachePopulator(cache, objectMapper, segmentCacheKey)
|
||||
|
@ -331,9 +336,12 @@ public class CachingClusteredClient<T> implements QueryRunner<T>
|
|||
String segmentIdentifier = value.getSegmentId();
|
||||
final Iterable<T> segmentResults = value.getResults();
|
||||
|
||||
cachePopulatorMap.get(
|
||||
CachePopulator cachePopulator = cachePopulatorMap.get(
|
||||
String.format("%s_%s", segmentIdentifier, value.getInterval())
|
||||
).populate(Iterables.transform(segmentResults, prepareForCache));
|
||||
);
|
||||
if(cachePopulator != null) {
|
||||
cachePopulator.populate(Iterables.transform(segmentResults, prepareForCache));
|
||||
}
|
||||
|
||||
return Sequences.simple(
|
||||
Iterables.transform(
|
||||
|
|
Loading…
Reference in New Issue