mirror of https://github.com/apache/druid.git
Replace the deprecated usage of NoneShardSpec (#3166)
This commit is contained in:
parent
4cc39b2ee7
commit
45f553fc28
|
@ -43,7 +43,7 @@ public class DataSegmentPusherUtilTest
|
|||
loadSpec,
|
||||
Arrays.asList("dim1", "dim2"),
|
||||
Arrays.asList("met1", "met2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
null,
|
||||
1
|
||||
);
|
||||
|
|
|
@ -100,7 +100,7 @@ public class DataSegmentTest
|
|||
loadSpec,
|
||||
Arrays.asList("dim1", "dim2"),
|
||||
Arrays.asList("met1", "met2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
TEST_VERSION,
|
||||
1
|
||||
);
|
||||
|
@ -152,7 +152,7 @@ public class DataSegmentTest
|
|||
.dataSource("foo")
|
||||
.interval(new Interval("2012-01-01/2012-01-02"))
|
||||
.version(new DateTime("2012-01-01T11:22:33.444Z").toString())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.build();
|
||||
|
||||
Assert.assertEquals(
|
||||
|
|
|
@ -11,8 +11,8 @@ public class NoneShardSpecTest
|
|||
@Test
|
||||
public void testEqualsAndHashCode()
|
||||
{
|
||||
final ShardSpec one = new NoneShardSpec();
|
||||
final ShardSpec two = new NoneShardSpec();
|
||||
final ShardSpec one = NoneShardSpec.instance();
|
||||
final ShardSpec two = NoneShardSpec.instance();
|
||||
Assert.assertEquals(one, two);
|
||||
Assert.assertEquals(one.hashCode(), two.hashCode());
|
||||
}
|
||||
|
|
|
@ -383,4 +383,4 @@ public class SelectBenchmark
|
|||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -49,7 +49,7 @@ public class AzureDataSegmentKillerTest extends EasyMockSupport
|
|||
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
);
|
||||
|
|
|
@ -56,7 +56,7 @@ public class AzureDataSegmentPullerTest extends EasyMockSupport
|
|||
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
);
|
||||
|
|
|
@ -60,7 +60,7 @@ public class AzureDataSegmentPusherTest extends EasyMockSupport
|
|||
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
);
|
||||
|
@ -99,7 +99,7 @@ public class AzureDataSegmentPusherTest extends EasyMockSupport
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
size
|
||||
);
|
||||
|
|
|
@ -79,7 +79,7 @@ public class CloudFilesDataSegmentPusherTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
size
|
||||
);
|
||||
|
|
|
@ -122,7 +122,7 @@ public class HdfsDataSegmentKillerTest
|
|||
),
|
||||
ImmutableList.of("product"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
12334
|
||||
);
|
||||
|
|
|
@ -66,7 +66,7 @@ public class HdfsDataSegmentPusherTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
size
|
||||
);
|
||||
|
|
|
@ -53,7 +53,7 @@ public class S3DataSegmentMoverTest
|
|||
),
|
||||
ImmutableList.of("dim1", "dim1"),
|
||||
ImmutableList.of("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
);
|
||||
|
@ -128,7 +128,7 @@ public class S3DataSegmentMoverTest
|
|||
),
|
||||
ImmutableList.of("dim1", "dim1"),
|
||||
ImmutableList.of("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
), ImmutableMap.<String, Object>of("bucket", "DOES NOT EXIST", "baseKey", "baseKey"));
|
||||
|
@ -151,7 +151,7 @@ public class S3DataSegmentMoverTest
|
|||
),
|
||||
ImmutableList.of("dim1", "dim1"),
|
||||
ImmutableList.of("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1
|
||||
), ImmutableMap.<String, Object>of("bucket", "DOES NOT EXIST", "baseKey", "baseKey2"));
|
||||
|
|
|
@ -73,7 +73,7 @@ public class S3DataSegmentPusherTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
size
|
||||
);
|
||||
|
|
|
@ -171,7 +171,7 @@ public class DetermineHashedPartitionsJob implements Jobby
|
|||
|
||||
List<HadoopyShardSpec> actualSpecs = Lists.newArrayListWithExpectedSize(numberOfShards);
|
||||
if (numberOfShards == 1) {
|
||||
actualSpecs.add(new HadoopyShardSpec(new NoneShardSpec(), shardCount++));
|
||||
actualSpecs.add(new HadoopyShardSpec(NoneShardSpec.instance(), shardCount++));
|
||||
} else {
|
||||
for (int i = 0; i < numberOfShards; ++i) {
|
||||
actualSpecs.add(
|
||||
|
|
|
@ -645,7 +645,7 @@ public class DeterminePartitionsJob implements Jobby
|
|||
final ShardSpec shardSpec;
|
||||
|
||||
if (currentDimPartitions.partitions.isEmpty()) {
|
||||
shardSpec = new NoneShardSpec();
|
||||
shardSpec = NoneShardSpec.instance();
|
||||
} else {
|
||||
if (currentDimPartition.rows < config.getTargetPartitionSize() * SHARD_COMBINE_THRESHOLD) {
|
||||
// Combine with previous shard
|
||||
|
|
|
@ -80,7 +80,7 @@ public class HadoopDruidDetermineConfigurationJob implements Jobby
|
|||
shardSpecs.put(bucket, specs);
|
||||
log.info("DateTime[%s], spec[%s]", bucket, specs);
|
||||
} else {
|
||||
final HadoopyShardSpec spec = new HadoopyShardSpec(new NoneShardSpec(), shardCount++);
|
||||
final HadoopyShardSpec spec = new HadoopyShardSpec(NoneShardSpec.instance(), shardCount++);
|
||||
shardSpecs.put(bucket, Lists.newArrayList(spec));
|
||||
log.info("DateTime[%s], spec[%s]", bucket, spec);
|
||||
}
|
||||
|
|
|
@ -73,7 +73,7 @@ public class HadoopIngestionSpecUpdateDatasourcePathSpecSegmentsTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
2
|
||||
);
|
||||
|
|
|
@ -73,7 +73,7 @@ public class DatasourceInputFormatTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
2
|
||||
)
|
||||
|
@ -89,7 +89,7 @@ public class DatasourceInputFormatTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
11
|
||||
)
|
||||
|
@ -105,7 +105,7 @@ public class DatasourceInputFormatTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
4
|
||||
)
|
||||
|
|
|
@ -53,7 +53,7 @@ public class DatasourceInputSplitTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
12334
|
||||
),
|
||||
|
|
|
@ -44,7 +44,7 @@ public class WindowedDataSegmentTest
|
|||
),
|
||||
ImmutableList.of("host"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
2
|
||||
);
|
||||
|
|
|
@ -96,7 +96,7 @@ public class DatasourcePathSpecTest
|
|||
),
|
||||
ImmutableList.of("product"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
12334
|
||||
)
|
||||
|
@ -112,7 +112,7 @@ public class DatasourcePathSpecTest
|
|||
),
|
||||
ImmutableList.of("product"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
12335
|
||||
)
|
||||
|
|
|
@ -214,7 +214,7 @@ public class IndexTask extends AbstractFixedIntervalTask
|
|||
shardSpecs.add(new HashBasedNumberedShardSpec(i, numShards, null, jsonMapper));
|
||||
}
|
||||
} else {
|
||||
shardSpecs = ImmutableList.<ShardSpec>of(new NoneShardSpec());
|
||||
shardSpecs = ImmutableList.<ShardSpec>of(NoneShardSpec.instance());
|
||||
}
|
||||
}
|
||||
for (final ShardSpec shardSpec : shardSpecs) {
|
||||
|
@ -301,7 +301,7 @@ public class IndexTask extends AbstractFixedIntervalTask
|
|||
final List<ShardSpec> shardSpecs = Lists.newArrayList();
|
||||
|
||||
if (numberOfShards == 1) {
|
||||
shardSpecs.add(new NoneShardSpec());
|
||||
shardSpecs.add(NoneShardSpec.instance());
|
||||
} else {
|
||||
for (int i = 0; i < numberOfShards; ++i) {
|
||||
shardSpecs.add(new HashBasedNumberedShardSpec(i, numberOfShards, null, jsonMapper));
|
||||
|
|
|
@ -320,7 +320,7 @@ public abstract class MergeTaskBase extends AbstractFixedIntervalTask
|
|||
.interval(mergedInterval)
|
||||
.version(version)
|
||||
.binaryVersion(IndexIO.CURRENT_VERSION_ID)
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.dimensions(Lists.newArrayList(mergedDimensions))
|
||||
.metrics(Lists.newArrayList(mergedMetrics))
|
||||
.build();
|
||||
|
|
|
@ -60,7 +60,7 @@ public class ConvertSegmentTaskTest
|
|||
ImmutableMap.<String, Object>of(),
|
||||
ImmutableList.<String>of(),
|
||||
ImmutableList.<String>of(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
102937
|
||||
);
|
||||
|
|
|
@ -47,7 +47,7 @@ public class HadoopConverterTaskSerDeTest
|
|||
private static final Map<String, Object> LOAD_SPEC = ImmutableMap.<String, Object>of("someKey", "someVal");
|
||||
private static final List<String> DIMENSIONS = ImmutableList.of("dim1", "dim2");
|
||||
private static final List<String> METRICS = ImmutableList.of("metric1", "metric2");
|
||||
private static final ShardSpec SHARD_SPEC = new NoneShardSpec();
|
||||
private static final ShardSpec SHARD_SPEC = NoneShardSpec.instance();
|
||||
private static final int BINARY_VERSION = 34718;
|
||||
private static final long SEGMENT_SIZE = 7483901348790L;
|
||||
private static final IndexSpec INDEX_SPEC = new IndexSpec(new ConciseBitmapSerdeFactory(), "lz4", "lzf");
|
||||
|
|
|
@ -331,7 +331,7 @@ public class TaskSerdeTest
|
|||
null,
|
||||
null,
|
||||
1,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
indexSpec,
|
||||
null,
|
||||
0,
|
||||
|
@ -487,7 +487,7 @@ public class TaskSerdeTest
|
|||
ImmutableMap.<String, Object>of(),
|
||||
ImmutableList.of("dim1", "dim2"),
|
||||
ImmutableList.of("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
12345L
|
||||
),
|
||||
|
@ -511,7 +511,7 @@ public class TaskSerdeTest
|
|||
ImmutableMap.<String, Object>of(),
|
||||
ImmutableList.of("dim1", "dim2"),
|
||||
ImmutableList.of("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
12345L
|
||||
);
|
||||
|
|
|
@ -752,7 +752,7 @@ public class TaskLifecycleTest
|
|||
.version("2011-04-6T16:52:46.119-05:00")
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build();
|
||||
|
|
|
@ -169,4 +169,4 @@ public class LongGenericColumnPartSerde implements ColumnPartSerde
|
|||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -163,7 +163,7 @@ public class MultiSegmentSelectQueryTest
|
|||
interval.getStart(),
|
||||
interval.getEnd(),
|
||||
version,
|
||||
new NoneShardSpec()
|
||||
NoneShardSpec.instance()
|
||||
);
|
||||
}
|
||||
|
||||
|
|
|
@ -74,4 +74,4 @@ public class CompressedVSizeIndexedV3SupplierTest extends CompressedVSizeIndexed
|
|||
buffer, ByteOrder.nativeOrder()
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -445,7 +445,7 @@ public class SchemalessIndex
|
|||
Ordering.natural().nullsFirst()
|
||||
);
|
||||
|
||||
ShardSpec noneShardSpec = new NoneShardSpec();
|
||||
ShardSpec noneShardSpec = NoneShardSpec.instance();
|
||||
|
||||
for (int i = 0; i < intervals.size(); i++) {
|
||||
timeline.add(intervals.get(i), i, noneShardSpec.createChunk(filesToMap.get(i)));
|
||||
|
|
|
@ -165,4 +165,4 @@ public class CompressedIntsIndexedWriterTest
|
|||
vals = new int[0];
|
||||
checkSerializedSizeAndData(2);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -217,4 +217,4 @@ public class CompressedVSizeIndexedV3WriterTest
|
|||
vals = new ArrayList<>();
|
||||
checkSerializedSizeAndData(1, 2);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -161,4 +161,4 @@ public class CompressedVSizeIntsIndexedWriterTest
|
|||
vals = new int[0];
|
||||
checkSerializedSizeAndData(2);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -108,4 +108,4 @@ public class VSizeIndexedIntsWriterTest
|
|||
vals = new int[0];
|
||||
checkSerializedSizeAndData();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -96,4 +96,4 @@ public class OnheapIncrementalIndexTest
|
|||
|
||||
Assert.assertEquals(0, checkFailedCount.get());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -45,7 +45,7 @@ public class RealtimeTuningConfig implements TuningConfig, AppenderatorConfig
|
|||
private static final VersioningPolicy defaultVersioningPolicy = new IntervalStartVersioningPolicy();
|
||||
private static final RejectionPolicyFactory defaultRejectionPolicyFactory = new ServerTimeRejectionPolicyFactory();
|
||||
private static final int defaultMaxPendingPersists = 0;
|
||||
private static final ShardSpec defaultShardSpec = new NoneShardSpec();
|
||||
private static final ShardSpec defaultShardSpec = NoneShardSpec.instance();
|
||||
private static final IndexSpec defaultIndexSpec = new IndexSpec();
|
||||
private static final Boolean defaultBuildV9Directly = Boolean.FALSE;
|
||||
private static final Boolean defaultReportParseExceptions = Boolean.FALSE;
|
||||
|
|
|
@ -358,7 +358,7 @@ public class BrokerServerViewTest extends CuratorTestBase
|
|||
.version(version)
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build();
|
||||
|
|
|
@ -147,7 +147,7 @@ public class CachingClusteredClientFunctionalityTest {
|
|||
.dataSource("test")
|
||||
.interval(interval)
|
||||
.version(version)
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.build(),
|
||||
new TierSelectorStrategy() {
|
||||
@Override
|
||||
|
|
|
@ -2581,7 +2581,7 @@ public class CachingClusteredClientTest
|
|||
null,
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
null,
|
||||
-1
|
||||
);
|
||||
|
|
|
@ -353,7 +353,7 @@ public class CoordinatorServerViewTest extends CuratorTestBase
|
|||
.version(version)
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build();
|
||||
|
|
|
@ -61,7 +61,7 @@ public class DataSegmentTest
|
|||
loadSpec,
|
||||
Arrays.asList("dim1", "dim2"),
|
||||
Arrays.asList("met1", "met2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
1
|
||||
);
|
||||
|
@ -108,7 +108,7 @@ public class DataSegmentTest
|
|||
.dataSource("foo")
|
||||
.interval(new Interval("2012-01-01/2012-01-02"))
|
||||
.version(new DateTime("2012-01-01T11:22:33.444Z").toString())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.build();
|
||||
|
||||
Assert.assertEquals(
|
||||
|
|
|
@ -110,7 +110,7 @@ public class DirectDruidClientTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
0L
|
||||
),
|
||||
|
@ -220,7 +220,7 @@ public class DirectDruidClientTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
0L
|
||||
),
|
||||
|
@ -287,7 +287,7 @@ public class DirectDruidClientTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
0L
|
||||
);
|
||||
|
|
|
@ -50,7 +50,7 @@ public class ImmutableSegmentLoadInfoTest
|
|||
null,
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0, 0
|
||||
), Sets.newHashSet(new DruidServerMetadata("a", "host", 10, "type", "tier", 1))
|
||||
);
|
||||
|
|
|
@ -51,7 +51,7 @@ public class ServerSelectorTest
|
|||
.version("v1")
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build(),
|
||||
|
@ -82,7 +82,7 @@ public class ServerSelectorTest
|
|||
))
|
||||
.metrics(
|
||||
ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build()
|
||||
|
|
|
@ -123,7 +123,7 @@ public class TierSelectorStrategyTest
|
|||
com.google.common.collect.Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
0L
|
||||
),
|
||||
|
|
|
@ -78,7 +78,7 @@ public class IndexerSQLMetadataStorageCoordinatorTest
|
|||
ImmutableMap.<String, Object>of(),
|
||||
ImmutableList.of("dim1"),
|
||||
ImmutableList.of("m1"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
100
|
||||
);
|
||||
|
|
|
@ -57,7 +57,7 @@ public class MetadataSegmentManagerTest
|
|||
),
|
||||
ImmutableList.of("dim1", "dim2", "dim3"),
|
||||
ImmutableList.of("count", "value"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1234L
|
||||
);
|
||||
|
@ -73,7 +73,7 @@ public class MetadataSegmentManagerTest
|
|||
),
|
||||
ImmutableList.of("dim1", "dim2", "dim3"),
|
||||
ImmutableList.of("count", "value"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
1234L
|
||||
);
|
||||
|
|
|
@ -69,7 +69,7 @@ public class RealtimeTuningConfigTest
|
|||
Assert.assertEquals(0, config.getHandoffConditionTimeout());
|
||||
Assert.assertEquals(new IndexSpec(), config.getIndexSpec());
|
||||
Assert.assertEquals(new Period("PT10M"), config.getIntermediatePersistPeriod());
|
||||
Assert.assertEquals(new NoneShardSpec(), config.getShardSpec());
|
||||
Assert.assertEquals(NoneShardSpec.instance(), config.getShardSpec());
|
||||
Assert.assertEquals(0, config.getMaxPendingPersists());
|
||||
Assert.assertEquals(75000, config.getMaxRowsInMemory());
|
||||
Assert.assertEquals(0, config.getMergeThreadPriority());
|
||||
|
@ -111,7 +111,7 @@ public class RealtimeTuningConfigTest
|
|||
Assert.assertEquals(100, config.getHandoffConditionTimeout());
|
||||
Assert.assertEquals(new IndexSpec(), config.getIndexSpec());
|
||||
Assert.assertEquals(new Period("PT1H"), config.getIntermediatePersistPeriod());
|
||||
Assert.assertEquals(new NoneShardSpec(), config.getShardSpec());
|
||||
Assert.assertEquals(NoneShardSpec.instance(), config.getShardSpec());
|
||||
Assert.assertEquals(100, config.getMaxPendingPersists());
|
||||
Assert.assertEquals(100, config.getMaxRowsInMemory());
|
||||
Assert.assertEquals(100, config.getMergeThreadPriority());
|
||||
|
|
|
@ -111,7 +111,7 @@ public class LocalDataSegmentKillerTest
|
|||
),
|
||||
ImmutableList.of("product"),
|
||||
ImmutableList.of("visited_sum", "unique_hosts"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
9,
|
||||
12334
|
||||
);
|
||||
|
|
|
@ -54,7 +54,7 @@ public class LocalDataSegmentPusherTest
|
|||
null,
|
||||
null,
|
||||
null,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
null,
|
||||
-1
|
||||
);
|
||||
|
|
|
@ -146,7 +146,7 @@ public class SegmentLoaderLocalCacheManagerTest
|
|||
.version("2015-05-27T03:38:35.683Z")
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(0)
|
||||
.build();
|
||||
|
|
|
@ -354,4 +354,4 @@ public class CoordinatorBasedSegmentHandoffNotifierTest
|
|||
0, 0
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -52,7 +52,7 @@ public class SegmentChangeRequestDropTest
|
|||
loadSpec,
|
||||
Arrays.asList("dim1", "dim2"),
|
||||
Arrays.asList("met1", "met2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
1
|
||||
);
|
||||
|
|
|
@ -52,7 +52,7 @@ public class SegmentChangeRequestLoadTest
|
|||
loadSpec,
|
||||
Arrays.asList("dim1", "dim2"),
|
||||
Arrays.asList("met1", "met2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
1
|
||||
);
|
||||
|
|
|
@ -467,7 +467,7 @@ public class ServerManagerTest
|
|||
ImmutableMap.<String, Object>of("version", version, "interval", interval),
|
||||
Arrays.asList("dim1", "dim2", "dim3"),
|
||||
Arrays.asList("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
123L
|
||||
)
|
||||
|
@ -489,7 +489,7 @@ public class ServerManagerTest
|
|||
ImmutableMap.<String, Object>of("version", version, "interval", interval),
|
||||
Arrays.asList("dim1", "dim2", "dim3"),
|
||||
Arrays.asList("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
123L
|
||||
)
|
||||
|
|
|
@ -418,7 +418,7 @@ public class ZkCoordinatorTest extends CuratorTestBase
|
|||
ImmutableMap.<String, Object>of("version", version, "interval", interval, "cacheDir", infoDir),
|
||||
Arrays.asList("dim1", "dim2", "dim3"),
|
||||
Arrays.asList("metric1", "metric2"),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
123L
|
||||
);
|
||||
|
|
|
@ -102,7 +102,7 @@ public class DruidCoordinatorBalancerProfiler
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
4L
|
||||
)
|
||||
|
|
|
@ -78,7 +78,7 @@ public class DruidCoordinatorBalancerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
11L
|
||||
);
|
||||
|
@ -89,7 +89,7 @@ public class DruidCoordinatorBalancerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
7L
|
||||
);
|
||||
|
@ -100,7 +100,7 @@ public class DruidCoordinatorBalancerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
4L
|
||||
);
|
||||
|
@ -111,7 +111,7 @@ public class DruidCoordinatorBalancerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
8L
|
||||
);
|
||||
|
|
|
@ -79,7 +79,7 @@ public class DruidCoordinatorRuleRunnerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
IndexIO.CURRENT_VERSION_ID,
|
||||
1
|
||||
)
|
||||
|
@ -992,7 +992,7 @@ public class DruidCoordinatorRuleRunnerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
1,
|
||||
0
|
||||
);
|
||||
|
@ -1140,7 +1140,7 @@ public class DruidCoordinatorRuleRunnerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
1,
|
||||
0
|
||||
);
|
||||
|
|
|
@ -351,7 +351,7 @@ public class LoadQueuePeonTest extends CuratorTestBase
|
|||
.version("2015-05-27T03:38:35.683Z")
|
||||
.dimensions(ImmutableList.<String>of())
|
||||
.metrics(ImmutableList.<String>of())
|
||||
.shardSpec(new NoneShardSpec())
|
||||
.shardSpec(NoneShardSpec.instance())
|
||||
.binaryVersion(9)
|
||||
.size(1200)
|
||||
.build();
|
||||
|
|
|
@ -83,7 +83,7 @@ public class ReservoirSegmentSamplerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
11L
|
||||
);
|
||||
|
@ -94,7 +94,7 @@ public class ReservoirSegmentSamplerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
7L
|
||||
);
|
||||
|
@ -105,7 +105,7 @@ public class ReservoirSegmentSamplerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
4L
|
||||
);
|
||||
|
@ -116,7 +116,7 @@ public class ReservoirSegmentSamplerTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
8L
|
||||
);
|
||||
|
|
|
@ -94,7 +94,7 @@ public class LoadRuleTest
|
|||
Maps.<String, Object>newHashMap(),
|
||||
Lists.<String>newArrayList(),
|
||||
Lists.<String>newArrayList(),
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
0,
|
||||
0
|
||||
);
|
||||
|
|
|
@ -34,7 +34,7 @@ public class PeriodDropRuleTest
|
|||
private final static DataSegment.Builder builder = DataSegment.builder()
|
||||
.dataSource("test")
|
||||
.version(new DateTime("2012-12-31T01:00:00").toString())
|
||||
.shardSpec(new NoneShardSpec());
|
||||
.shardSpec(NoneShardSpec.instance());
|
||||
|
||||
@Test
|
||||
public void testAppliesToAll()
|
||||
|
|
|
@ -38,7 +38,7 @@ public class PeriodLoadRuleTest
|
|||
private final static DataSegment.Builder builder = DataSegment.builder()
|
||||
.dataSource("test")
|
||||
.version(new DateTime().toString())
|
||||
.shardSpec(new NoneShardSpec());
|
||||
.shardSpec(NoneShardSpec.instance());
|
||||
|
||||
@Test
|
||||
public void testAppliesToAll()
|
||||
|
|
|
@ -175,7 +175,7 @@ public class DruidJsonValidatorTest
|
|||
null,
|
||||
null,
|
||||
1,
|
||||
new NoneShardSpec(),
|
||||
NoneShardSpec.instance(),
|
||||
new IndexSpec(),
|
||||
null,
|
||||
0,
|
||||
|
|
Loading…
Reference in New Issue