mirror of https://github.com/apache/druid.git
Fix auto compaction to compact only same or abutting intervals (#6808)
* Fix auto compaction to compact only same or abutting intervals * fix test
This commit is contained in:
parent
5b8a221713
commit
a07e66c540
|
@ -269,6 +269,15 @@ public class NewestSegmentFirstIterator implements CompactionSegmentIterator
|
||||||
final List<PartitionChunk<DataSegment>> chunks = Lists.newArrayList(timeChunkHolder.getObject().iterator());
|
final List<PartitionChunk<DataSegment>> chunks = Lists.newArrayList(timeChunkHolder.getObject().iterator());
|
||||||
final long timeChunkSizeBytes = chunks.stream().mapToLong(chunk -> chunk.getObject().getSize()).sum();
|
final long timeChunkSizeBytes = chunks.stream().mapToLong(chunk -> chunk.getObject().getSize()).sum();
|
||||||
|
|
||||||
|
final boolean isSameOrAbuttingInterval;
|
||||||
|
final Interval lastInterval = segmentsToCompact.getIntervalOfLastSegment();
|
||||||
|
if (lastInterval == null) {
|
||||||
|
isSameOrAbuttingInterval = true;
|
||||||
|
} else {
|
||||||
|
final Interval currentInterval = chunks.get(0).getObject().getInterval();
|
||||||
|
isSameOrAbuttingInterval = currentInterval.isEqual(lastInterval) || currentInterval.abuts(lastInterval);
|
||||||
|
}
|
||||||
|
|
||||||
// The segments in a holder should be added all together or not.
|
// The segments in a holder should be added all together or not.
|
||||||
final boolean isCompactibleSize = SegmentCompactorUtil.isCompactibleSize(
|
final boolean isCompactibleSize = SegmentCompactorUtil.isCompactibleSize(
|
||||||
inputSegmentSize,
|
inputSegmentSize,
|
||||||
|
@ -280,7 +289,10 @@ public class NewestSegmentFirstIterator implements CompactionSegmentIterator
|
||||||
segmentsToCompact.getNumSegments(),
|
segmentsToCompact.getNumSegments(),
|
||||||
chunks.size()
|
chunks.size()
|
||||||
);
|
);
|
||||||
if (isCompactibleSize && isCompactibleNum && (!keepSegmentGranularity || segmentsToCompact.isEmpty())) {
|
if (isCompactibleSize
|
||||||
|
&& isCompactibleNum
|
||||||
|
&& isSameOrAbuttingInterval
|
||||||
|
&& (!keepSegmentGranularity || segmentsToCompact.isEmpty())) {
|
||||||
chunks.forEach(chunk -> segmentsToCompact.add(chunk.getObject()));
|
chunks.forEach(chunk -> segmentsToCompact.add(chunk.getObject()));
|
||||||
} else {
|
} else {
|
||||||
if (segmentsToCompact.getNumSegments() > 1) {
|
if (segmentsToCompact.getNumSegments() > 1) {
|
||||||
|
@ -514,6 +526,16 @@ public class NewestSegmentFirstIterator implements CompactionSegmentIterator
|
||||||
return segments.isEmpty();
|
return segments.isEmpty();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Nullable
|
||||||
|
private Interval getIntervalOfLastSegment()
|
||||||
|
{
|
||||||
|
if (segments.isEmpty()) {
|
||||||
|
return null;
|
||||||
|
} else {
|
||||||
|
return segments.get(segments.size() - 1).getInterval();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private int getNumSegments()
|
private int getNumSegments()
|
||||||
{
|
{
|
||||||
return segments.size();
|
return segments.size();
|
||||||
|
|
|
@ -226,23 +226,23 @@ public class DruidCoordinatorSegmentCompactorTest
|
||||||
expectedVersionSupplier
|
expectedVersionSupplier
|
||||||
);
|
);
|
||||||
|
|
||||||
// compact for 2017-01-07T12:00:00.000Z/2017-01-08T12:00:00.000Z
|
// compact for 2017-01-08T00:00:00.000Z/2017-01-08T12:00:00.000Z
|
||||||
expectedRemainingSegments -= 40;
|
expectedRemainingSegments -= 20;
|
||||||
assertCompactSegments(
|
assertCompactSegments(
|
||||||
compactor,
|
compactor,
|
||||||
keepSegmentGranularity,
|
keepSegmentGranularity,
|
||||||
Intervals.of("2017-01-%02dT12:00:00/2017-01-%02dT12:00:00", 4, 8),
|
Intervals.of("2017-01-%02dT00:00:00/2017-01-%02dT12:00:00", 8, 8),
|
||||||
expectedRemainingSegments,
|
expectedRemainingSegments,
|
||||||
expectedCompactTaskCount,
|
expectedCompactTaskCount,
|
||||||
expectedVersionSupplier
|
expectedVersionSupplier
|
||||||
);
|
);
|
||||||
|
|
||||||
for (int endDay = 4; endDay > 1; endDay -= 1) {
|
for (int endDay = 5; endDay > 1; endDay -= 1) {
|
||||||
expectedRemainingSegments -= 40;
|
expectedRemainingSegments -= 40;
|
||||||
assertCompactSegments(
|
assertCompactSegments(
|
||||||
compactor,
|
compactor,
|
||||||
keepSegmentGranularity,
|
keepSegmentGranularity,
|
||||||
Intervals.of("2017-01-%02dT12:00:00/2017-01-%02dT12:00:00", endDay - 1, endDay),
|
Intervals.of("2017-01-%02dT00:00:00/2017-01-%02dT00:00:00", endDay - 1, endDay),
|
||||||
expectedRemainingSegments,
|
expectedRemainingSegments,
|
||||||
expectedCompactTaskCount,
|
expectedCompactTaskCount,
|
||||||
expectedVersionSupplier
|
expectedVersionSupplier
|
||||||
|
@ -362,7 +362,7 @@ public class DruidCoordinatorSegmentCompactorTest
|
||||||
|
|
||||||
// One of dataSource is compacted
|
// One of dataSource is compacted
|
||||||
if (expectedRemainingSegments > 0) {
|
if (expectedRemainingSegments > 0) {
|
||||||
// If expectedRemainingSegments is positive, we check how many dataSources have the segments waiting
|
// If expectedRemainingSegments is positive, we check how many dataSources have the segments waiting for
|
||||||
// compaction.
|
// compaction.
|
||||||
long numDataSourceOfExpectedRemainingSegments = stats
|
long numDataSourceOfExpectedRemainingSegments = stats
|
||||||
.getDataSources(DruidCoordinatorSegmentCompactor.SEGMENT_SIZE_WAIT_COMPACT)
|
.getDataSources(DruidCoordinatorSegmentCompactor.SEGMENT_SIZE_WAIT_COMPACT)
|
||||||
|
|
|
@ -110,61 +110,18 @@ public class NewestSegmentFirstPolicyTest
|
||||||
assertCompactSegmentIntervals(
|
assertCompactSegmentIntervals(
|
||||||
iterator,
|
iterator,
|
||||||
segmentPeriod,
|
segmentPeriod,
|
||||||
Intervals.of("2017-11-16T21:00:00/2017-11-16T22:00:00"),
|
Intervals.of("2017-11-16T20:00:00/2017-11-16T21:00:00"),
|
||||||
Intervals.of("2017-11-17T02:00:00/2017-11-17T03:00:00"),
|
Intervals.of("2017-11-17T02:00:00/2017-11-17T03:00:00"),
|
||||||
false
|
false
|
||||||
);
|
);
|
||||||
|
|
||||||
final List<DataSegment> segments = iterator.next();
|
assertCompactSegmentIntervals(
|
||||||
Assert.assertNotNull(segments);
|
iterator,
|
||||||
|
segmentPeriod,
|
||||||
if (keepSegmentGranularity) {
|
Intervals.of("2017-11-14T00:00:00/2017-11-14T01:00:00"),
|
||||||
// If keepSegmentGranularity = true, the iterator returns the segments of only the next time chunk.
|
Intervals.of("2017-11-16T06:00:00/2017-11-16T07:00:00"),
|
||||||
Assert.assertEquals(4, segments.size());
|
true
|
||||||
|
);
|
||||||
final List<Interval> expectedIntervals = new ArrayList<>(segments.size());
|
|
||||||
for (int i = 0; i < 4; i++) {
|
|
||||||
expectedIntervals.add(Intervals.of("2017-11-16T20:00:00/2017-11-16T21:00:00"));
|
|
||||||
}
|
|
||||||
|
|
||||||
Assert.assertEquals(
|
|
||||||
expectedIntervals,
|
|
||||||
segments.stream().map(DataSegment::getInterval).collect(Collectors.toList())
|
|
||||||
);
|
|
||||||
|
|
||||||
assertCompactSegmentIntervals(
|
|
||||||
iterator,
|
|
||||||
segmentPeriod,
|
|
||||||
Intervals.of("2017-11-16T06:00:00/2017-11-16T07:00:00"),
|
|
||||||
Intervals.of("2017-11-16T06:00:00/2017-11-16T07:00:00"),
|
|
||||||
false
|
|
||||||
);
|
|
||||||
} else {
|
|
||||||
// If keepSegmentGranularity = false, the returned segments can span over multiple time chunks.
|
|
||||||
Assert.assertEquals(8, segments.size());
|
|
||||||
|
|
||||||
final List<Interval> expectedIntervals = new ArrayList<>(segments.size());
|
|
||||||
for (int i = 0; i < 4; i++) {
|
|
||||||
expectedIntervals.add(Intervals.of("2017-11-16T06:00:00/2017-11-16T07:00:00"));
|
|
||||||
}
|
|
||||||
for (int i = 0; i < 4; i++) {
|
|
||||||
expectedIntervals.add(Intervals.of("2017-11-16T20:00:00/2017-11-16T21:00:00"));
|
|
||||||
}
|
|
||||||
expectedIntervals.sort(Comparators.intervalsByStartThenEnd());
|
|
||||||
|
|
||||||
Assert.assertEquals(
|
|
||||||
expectedIntervals,
|
|
||||||
segments.stream().map(DataSegment::getInterval).collect(Collectors.toList())
|
|
||||||
);
|
|
||||||
|
|
||||||
assertCompactSegmentIntervals(
|
|
||||||
iterator,
|
|
||||||
segmentPeriod,
|
|
||||||
Intervals.of("2017-11-14T00:00:00/2017-11-14T01:00:00"),
|
|
||||||
Intervals.of("2017-11-16T05:00:00/2017-11-16T06:00:00"),
|
|
||||||
true
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
@ -588,6 +545,52 @@ public class NewestSegmentFirstPolicyTest
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testHoleInSearchInterval()
|
||||||
|
{
|
||||||
|
final Period segmentPeriod = new Period("PT1H");
|
||||||
|
final CompactionSegmentIterator iterator = policy.reset(
|
||||||
|
ImmutableMap.of(DATA_SOURCE, createCompactionConfig(10000, 100, new Period("PT1H"))),
|
||||||
|
ImmutableMap.of(
|
||||||
|
DATA_SOURCE,
|
||||||
|
createTimeline(
|
||||||
|
new SegmentGenerateSpec(Intervals.of("2017-11-16T00:00:00/2017-11-17T00:00:00"), segmentPeriod)
|
||||||
|
)
|
||||||
|
),
|
||||||
|
ImmutableMap.of(
|
||||||
|
DATA_SOURCE,
|
||||||
|
ImmutableList.of(
|
||||||
|
Intervals.of("2017-11-16T04:00:00/2017-11-16T10:00:00"),
|
||||||
|
Intervals.of("2017-11-16T14:00:00/2017-11-16T20:00:00")
|
||||||
|
)
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
assertCompactSegmentIntervals(
|
||||||
|
iterator,
|
||||||
|
segmentPeriod,
|
||||||
|
Intervals.of("2017-11-16T20:00:00/2017-11-16T21:00:00"),
|
||||||
|
Intervals.of("2017-11-16T22:00:00/2017-11-16T23:00:00"),
|
||||||
|
false
|
||||||
|
);
|
||||||
|
|
||||||
|
assertCompactSegmentIntervals(
|
||||||
|
iterator,
|
||||||
|
segmentPeriod,
|
||||||
|
Intervals.of("2017-11-16T10:00:00/2017-11-16T11:00:00"),
|
||||||
|
Intervals.of("2017-11-16T13:00:00/2017-11-16T14:00:00"),
|
||||||
|
false
|
||||||
|
);
|
||||||
|
|
||||||
|
assertCompactSegmentIntervals(
|
||||||
|
iterator,
|
||||||
|
segmentPeriod,
|
||||||
|
Intervals.of("2017-11-16T00:00:00/2017-11-16T01:00:00"),
|
||||||
|
Intervals.of("2017-11-16T03:00:00/2017-11-16T04:00:00"),
|
||||||
|
true
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
private static void assertCompactSegmentIntervals(
|
private static void assertCompactSegmentIntervals(
|
||||||
CompactionSegmentIterator iterator,
|
CompactionSegmentIterator iterator,
|
||||||
Period segmentPeriod,
|
Period segmentPeriod,
|
||||||
|
@ -600,6 +603,14 @@ public class NewestSegmentFirstPolicyTest
|
||||||
while (iterator.hasNext()) {
|
while (iterator.hasNext()) {
|
||||||
final List<DataSegment> segments = iterator.next();
|
final List<DataSegment> segments = iterator.next();
|
||||||
|
|
||||||
|
final Interval firstInterval = segments.get(0).getInterval();
|
||||||
|
Assert.assertTrue(
|
||||||
|
"Intervals should be same or abutting",
|
||||||
|
segments.stream().allMatch(
|
||||||
|
segment -> segment.getInterval().isEqual(firstInterval) || segment.getInterval().abuts(firstInterval)
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
final List<Interval> expectedIntervals = new ArrayList<>(segments.size());
|
final List<Interval> expectedIntervals = new ArrayList<>(segments.size());
|
||||||
for (int i = 0; i < segments.size(); i++) {
|
for (int i = 0; i < segments.size(); i++) {
|
||||||
if (i > 0 && i % DEFAULT_NUM_SEGMENTS_PER_SHARD == 0) {
|
if (i > 0 && i % DEFAULT_NUM_SEGMENTS_PER_SHARD == 0) {
|
||||||
|
|
Loading…
Reference in New Issue