Replace the deprecated usage of NoneShardSpec (#3166)

This commit is contained in:
Hyukjin Kwon 2016-06-26 02:27:25 +09:00 committed by Gian Merlino
parent 4cc39b2ee7
commit 45f553fc28
65 changed files with 91 additions and 91 deletions

View File

@ -43,7 +43,7 @@ public class DataSegmentPusherUtilTest
loadSpec,
Arrays.asList("dim1", "dim2"),
Arrays.asList("met1", "met2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
null,
1
);

View File

@ -100,7 +100,7 @@ public class DataSegmentTest
loadSpec,
Arrays.asList("dim1", "dim2"),
Arrays.asList("met1", "met2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
TEST_VERSION,
1
);
@ -152,7 +152,7 @@ public class DataSegmentTest
.dataSource("foo")
.interval(new Interval("2012-01-01/2012-01-02"))
.version(new DateTime("2012-01-01T11:22:33.444Z").toString())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.build();
Assert.assertEquals(

View File

@ -11,8 +11,8 @@ public class NoneShardSpecTest
@Test
public void testEqualsAndHashCode()
{
final ShardSpec one = new NoneShardSpec();
final ShardSpec two = new NoneShardSpec();
final ShardSpec one = NoneShardSpec.instance();
final ShardSpec two = NoneShardSpec.instance();
Assert.assertEquals(one, two);
Assert.assertEquals(one.hashCode(), two.hashCode());
}

View File

@ -383,4 +383,4 @@ public class SelectBenchmark
}
}
}
}
}

View File

@ -49,7 +49,7 @@ public class AzureDataSegmentKillerTest extends EasyMockSupport
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
);

View File

@ -56,7 +56,7 @@ public class AzureDataSegmentPullerTest extends EasyMockSupport
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
);

View File

@ -60,7 +60,7 @@ public class AzureDataSegmentPusherTest extends EasyMockSupport
ImmutableMap.<String, Object>of("containerName", containerName, "blobPath", blobPath),
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
);
@ -99,7 +99,7 @@ public class AzureDataSegmentPusherTest extends EasyMockSupport
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
size
);

View File

@ -79,7 +79,7 @@ public class CloudFilesDataSegmentPusherTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
size
);

View File

@ -122,7 +122,7 @@ public class HdfsDataSegmentKillerTest
),
ImmutableList.of("product"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
12334
);

View File

@ -66,7 +66,7 @@ public class HdfsDataSegmentPusherTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
size
);

View File

@ -53,7 +53,7 @@ public class S3DataSegmentMoverTest
),
ImmutableList.of("dim1", "dim1"),
ImmutableList.of("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
);
@ -128,7 +128,7 @@ public class S3DataSegmentMoverTest
),
ImmutableList.of("dim1", "dim1"),
ImmutableList.of("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
), ImmutableMap.<String, Object>of("bucket", "DOES NOT EXIST", "baseKey", "baseKey"));
@ -151,7 +151,7 @@ public class S3DataSegmentMoverTest
),
ImmutableList.of("dim1", "dim1"),
ImmutableList.of("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1
), ImmutableMap.<String, Object>of("bucket", "DOES NOT EXIST", "baseKey", "baseKey2"));

View File

@ -73,7 +73,7 @@ public class S3DataSegmentPusherTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
size
);

View File

@ -171,7 +171,7 @@ public class DetermineHashedPartitionsJob implements Jobby
List<HadoopyShardSpec> actualSpecs = Lists.newArrayListWithExpectedSize(numberOfShards);
if (numberOfShards == 1) {
actualSpecs.add(new HadoopyShardSpec(new NoneShardSpec(), shardCount++));
actualSpecs.add(new HadoopyShardSpec(NoneShardSpec.instance(), shardCount++));
} else {
for (int i = 0; i < numberOfShards; ++i) {
actualSpecs.add(

View File

@ -645,7 +645,7 @@ public class DeterminePartitionsJob implements Jobby
final ShardSpec shardSpec;
if (currentDimPartitions.partitions.isEmpty()) {
shardSpec = new NoneShardSpec();
shardSpec = NoneShardSpec.instance();
} else {
if (currentDimPartition.rows < config.getTargetPartitionSize() * SHARD_COMBINE_THRESHOLD) {
// Combine with previous shard

View File

@ -80,7 +80,7 @@ public class HadoopDruidDetermineConfigurationJob implements Jobby
shardSpecs.put(bucket, specs);
log.info("DateTime[%s], spec[%s]", bucket, specs);
} else {
final HadoopyShardSpec spec = new HadoopyShardSpec(new NoneShardSpec(), shardCount++);
final HadoopyShardSpec spec = new HadoopyShardSpec(NoneShardSpec.instance(), shardCount++);
shardSpecs.put(bucket, Lists.newArrayList(spec));
log.info("DateTime[%s], spec[%s]", bucket, spec);
}

View File

@ -73,7 +73,7 @@ public class HadoopIngestionSpecUpdateDatasourcePathSpecSegmentsTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
2
);

View File

@ -73,7 +73,7 @@ public class DatasourceInputFormatTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
2
)
@ -89,7 +89,7 @@ public class DatasourceInputFormatTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
11
)
@ -105,7 +105,7 @@ public class DatasourceInputFormatTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
4
)

View File

@ -53,7 +53,7 @@ public class DatasourceInputSplitTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
12334
),

View File

@ -44,7 +44,7 @@ public class WindowedDataSegmentTest
),
ImmutableList.of("host"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
2
);

View File

@ -96,7 +96,7 @@ public class DatasourcePathSpecTest
),
ImmutableList.of("product"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
12334
)
@ -112,7 +112,7 @@ public class DatasourcePathSpecTest
),
ImmutableList.of("product"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
12335
)

View File

@ -214,7 +214,7 @@ public class IndexTask extends AbstractFixedIntervalTask
shardSpecs.add(new HashBasedNumberedShardSpec(i, numShards, null, jsonMapper));
}
} else {
shardSpecs = ImmutableList.<ShardSpec>of(new NoneShardSpec());
shardSpecs = ImmutableList.<ShardSpec>of(NoneShardSpec.instance());
}
}
for (final ShardSpec shardSpec : shardSpecs) {
@ -301,7 +301,7 @@ public class IndexTask extends AbstractFixedIntervalTask
final List<ShardSpec> shardSpecs = Lists.newArrayList();
if (numberOfShards == 1) {
shardSpecs.add(new NoneShardSpec());
shardSpecs.add(NoneShardSpec.instance());
} else {
for (int i = 0; i < numberOfShards; ++i) {
shardSpecs.add(new HashBasedNumberedShardSpec(i, numberOfShards, null, jsonMapper));

View File

@ -320,7 +320,7 @@ public abstract class MergeTaskBase extends AbstractFixedIntervalTask
.interval(mergedInterval)
.version(version)
.binaryVersion(IndexIO.CURRENT_VERSION_ID)
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.dimensions(Lists.newArrayList(mergedDimensions))
.metrics(Lists.newArrayList(mergedMetrics))
.build();

View File

@ -60,7 +60,7 @@ public class ConvertSegmentTaskTest
ImmutableMap.<String, Object>of(),
ImmutableList.<String>of(),
ImmutableList.<String>of(),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
102937
);

View File

@ -47,7 +47,7 @@ public class HadoopConverterTaskSerDeTest
private static final Map<String, Object> LOAD_SPEC = ImmutableMap.<String, Object>of("someKey", "someVal");
private static final List<String> DIMENSIONS = ImmutableList.of("dim1", "dim2");
private static final List<String> METRICS = ImmutableList.of("metric1", "metric2");
private static final ShardSpec SHARD_SPEC = new NoneShardSpec();
private static final ShardSpec SHARD_SPEC = NoneShardSpec.instance();
private static final int BINARY_VERSION = 34718;
private static final long SEGMENT_SIZE = 7483901348790L;
private static final IndexSpec INDEX_SPEC = new IndexSpec(new ConciseBitmapSerdeFactory(), "lz4", "lzf");

View File

@ -331,7 +331,7 @@ public class TaskSerdeTest
null,
null,
1,
new NoneShardSpec(),
NoneShardSpec.instance(),
indexSpec,
null,
0,
@ -487,7 +487,7 @@ public class TaskSerdeTest
ImmutableMap.<String, Object>of(),
ImmutableList.of("dim1", "dim2"),
ImmutableList.of("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
12345L
),
@ -511,7 +511,7 @@ public class TaskSerdeTest
ImmutableMap.<String, Object>of(),
ImmutableList.of("dim1", "dim2"),
ImmutableList.of("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
12345L
);

View File

@ -752,7 +752,7 @@ public class TaskLifecycleTest
.version("2011-04-6T16:52:46.119-05:00")
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build();

View File

@ -169,4 +169,4 @@ public class LongGenericColumnPartSerde implements ColumnPartSerde
}
};
}
}
}

View File

@ -163,7 +163,7 @@ public class MultiSegmentSelectQueryTest
interval.getStart(),
interval.getEnd(),
version,
new NoneShardSpec()
NoneShardSpec.instance()
);
}

View File

@ -74,4 +74,4 @@ public class CompressedVSizeIndexedV3SupplierTest extends CompressedVSizeIndexed
buffer, ByteOrder.nativeOrder()
);
}
}
}

View File

@ -445,7 +445,7 @@ public class SchemalessIndex
Ordering.natural().nullsFirst()
);
ShardSpec noneShardSpec = new NoneShardSpec();
ShardSpec noneShardSpec = NoneShardSpec.instance();
for (int i = 0; i < intervals.size(); i++) {
timeline.add(intervals.get(i), i, noneShardSpec.createChunk(filesToMap.get(i)));

View File

@ -165,4 +165,4 @@ public class CompressedIntsIndexedWriterTest
vals = new int[0];
checkSerializedSizeAndData(2);
}
}
}

View File

@ -217,4 +217,4 @@ public class CompressedVSizeIndexedV3WriterTest
vals = new ArrayList<>();
checkSerializedSizeAndData(1, 2);
}
}
}

View File

@ -161,4 +161,4 @@ public class CompressedVSizeIntsIndexedWriterTest
vals = new int[0];
checkSerializedSizeAndData(2);
}
}
}

View File

@ -108,4 +108,4 @@ public class VSizeIndexedIntsWriterTest
vals = new int[0];
checkSerializedSizeAndData();
}
}
}

View File

@ -96,4 +96,4 @@ public class OnheapIncrementalIndexTest
Assert.assertEquals(0, checkFailedCount.get());
}
}
}

View File

@ -45,7 +45,7 @@ public class RealtimeTuningConfig implements TuningConfig, AppenderatorConfig
private static final VersioningPolicy defaultVersioningPolicy = new IntervalStartVersioningPolicy();
private static final RejectionPolicyFactory defaultRejectionPolicyFactory = new ServerTimeRejectionPolicyFactory();
private static final int defaultMaxPendingPersists = 0;
private static final ShardSpec defaultShardSpec = new NoneShardSpec();
private static final ShardSpec defaultShardSpec = NoneShardSpec.instance();
private static final IndexSpec defaultIndexSpec = new IndexSpec();
private static final Boolean defaultBuildV9Directly = Boolean.FALSE;
private static final Boolean defaultReportParseExceptions = Boolean.FALSE;

View File

@ -358,7 +358,7 @@ public class BrokerServerViewTest extends CuratorTestBase
.version(version)
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build();

View File

@ -147,7 +147,7 @@ public class CachingClusteredClientFunctionalityTest {
.dataSource("test")
.interval(interval)
.version(version)
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.build(),
new TierSelectorStrategy() {
@Override

View File

@ -2581,7 +2581,7 @@ public class CachingClusteredClientTest
null,
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
null,
-1
);

View File

@ -353,7 +353,7 @@ public class CoordinatorServerViewTest extends CuratorTestBase
.version(version)
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build();

View File

@ -61,7 +61,7 @@ public class DataSegmentTest
loadSpec,
Arrays.asList("dim1", "dim2"),
Arrays.asList("met1", "met2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
1
);
@ -108,7 +108,7 @@ public class DataSegmentTest
.dataSource("foo")
.interval(new Interval("2012-01-01/2012-01-02"))
.version(new DateTime("2012-01-01T11:22:33.444Z").toString())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.build();
Assert.assertEquals(

View File

@ -110,7 +110,7 @@ public class DirectDruidClientTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
0L
),
@ -220,7 +220,7 @@ public class DirectDruidClientTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
0L
),
@ -287,7 +287,7 @@ public class DirectDruidClientTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
0L
);

View File

@ -50,7 +50,7 @@ public class ImmutableSegmentLoadInfoTest
null,
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
0, 0
), Sets.newHashSet(new DruidServerMetadata("a", "host", 10, "type", "tier", 1))
);

View File

@ -51,7 +51,7 @@ public class ServerSelectorTest
.version("v1")
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build(),
@ -82,7 +82,7 @@ public class ServerSelectorTest
))
.metrics(
ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build()

View File

@ -123,7 +123,7 @@ public class TierSelectorStrategyTest
com.google.common.collect.Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
0L
),

View File

@ -78,7 +78,7 @@ public class IndexerSQLMetadataStorageCoordinatorTest
ImmutableMap.<String, Object>of(),
ImmutableList.of("dim1"),
ImmutableList.of("m1"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
100
);

View File

@ -57,7 +57,7 @@ public class MetadataSegmentManagerTest
),
ImmutableList.of("dim1", "dim2", "dim3"),
ImmutableList.of("count", "value"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1234L
);
@ -73,7 +73,7 @@ public class MetadataSegmentManagerTest
),
ImmutableList.of("dim1", "dim2", "dim3"),
ImmutableList.of("count", "value"),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
1234L
);

View File

@ -69,7 +69,7 @@ public class RealtimeTuningConfigTest
Assert.assertEquals(0, config.getHandoffConditionTimeout());
Assert.assertEquals(new IndexSpec(), config.getIndexSpec());
Assert.assertEquals(new Period("PT10M"), config.getIntermediatePersistPeriod());
Assert.assertEquals(new NoneShardSpec(), config.getShardSpec());
Assert.assertEquals(NoneShardSpec.instance(), config.getShardSpec());
Assert.assertEquals(0, config.getMaxPendingPersists());
Assert.assertEquals(75000, config.getMaxRowsInMemory());
Assert.assertEquals(0, config.getMergeThreadPriority());
@ -111,7 +111,7 @@ public class RealtimeTuningConfigTest
Assert.assertEquals(100, config.getHandoffConditionTimeout());
Assert.assertEquals(new IndexSpec(), config.getIndexSpec());
Assert.assertEquals(new Period("PT1H"), config.getIntermediatePersistPeriod());
Assert.assertEquals(new NoneShardSpec(), config.getShardSpec());
Assert.assertEquals(NoneShardSpec.instance(), config.getShardSpec());
Assert.assertEquals(100, config.getMaxPendingPersists());
Assert.assertEquals(100, config.getMaxRowsInMemory());
Assert.assertEquals(100, config.getMergeThreadPriority());

View File

@ -111,7 +111,7 @@ public class LocalDataSegmentKillerTest
),
ImmutableList.of("product"),
ImmutableList.of("visited_sum", "unique_hosts"),
new NoneShardSpec(),
NoneShardSpec.instance(),
9,
12334
);

View File

@ -54,7 +54,7 @@ public class LocalDataSegmentPusherTest
null,
null,
null,
new NoneShardSpec(),
NoneShardSpec.instance(),
null,
-1
);

View File

@ -146,7 +146,7 @@ public class SegmentLoaderLocalCacheManagerTest
.version("2015-05-27T03:38:35.683Z")
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(0)
.build();

View File

@ -354,4 +354,4 @@ public class CoordinatorBasedSegmentHandoffNotifierTest
0, 0
);
}
}
}

View File

@ -52,7 +52,7 @@ public class SegmentChangeRequestDropTest
loadSpec,
Arrays.asList("dim1", "dim2"),
Arrays.asList("met1", "met2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
1
);

View File

@ -52,7 +52,7 @@ public class SegmentChangeRequestLoadTest
loadSpec,
Arrays.asList("dim1", "dim2"),
Arrays.asList("met1", "met2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
1
);

View File

@ -467,7 +467,7 @@ public class ServerManagerTest
ImmutableMap.<String, Object>of("version", version, "interval", interval),
Arrays.asList("dim1", "dim2", "dim3"),
Arrays.asList("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
123L
)
@ -489,7 +489,7 @@ public class ServerManagerTest
ImmutableMap.<String, Object>of("version", version, "interval", interval),
Arrays.asList("dim1", "dim2", "dim3"),
Arrays.asList("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
123L
)

View File

@ -418,7 +418,7 @@ public class ZkCoordinatorTest extends CuratorTestBase
ImmutableMap.<String, Object>of("version", version, "interval", interval, "cacheDir", infoDir),
Arrays.asList("dim1", "dim2", "dim3"),
Arrays.asList("metric1", "metric2"),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
123L
);

View File

@ -102,7 +102,7 @@ public class DruidCoordinatorBalancerProfiler
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
4L
)

View File

@ -78,7 +78,7 @@ public class DruidCoordinatorBalancerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
11L
);
@ -89,7 +89,7 @@ public class DruidCoordinatorBalancerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
7L
);
@ -100,7 +100,7 @@ public class DruidCoordinatorBalancerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
4L
);
@ -111,7 +111,7 @@ public class DruidCoordinatorBalancerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
8L
);

View File

@ -79,7 +79,7 @@ public class DruidCoordinatorRuleRunnerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
IndexIO.CURRENT_VERSION_ID,
1
)
@ -992,7 +992,7 @@ public class DruidCoordinatorRuleRunnerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
1,
0
);
@ -1140,7 +1140,7 @@ public class DruidCoordinatorRuleRunnerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
1,
0
);

View File

@ -351,7 +351,7 @@ public class LoadQueuePeonTest extends CuratorTestBase
.version("2015-05-27T03:38:35.683Z")
.dimensions(ImmutableList.<String>of())
.metrics(ImmutableList.<String>of())
.shardSpec(new NoneShardSpec())
.shardSpec(NoneShardSpec.instance())
.binaryVersion(9)
.size(1200)
.build();

View File

@ -83,7 +83,7 @@ public class ReservoirSegmentSamplerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
11L
);
@ -94,7 +94,7 @@ public class ReservoirSegmentSamplerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
7L
);
@ -105,7 +105,7 @@ public class ReservoirSegmentSamplerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
4L
);
@ -116,7 +116,7 @@ public class ReservoirSegmentSamplerTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
8L
);

View File

@ -94,7 +94,7 @@ public class LoadRuleTest
Maps.<String, Object>newHashMap(),
Lists.<String>newArrayList(),
Lists.<String>newArrayList(),
new NoneShardSpec(),
NoneShardSpec.instance(),
0,
0
);

View File

@ -34,7 +34,7 @@ public class PeriodDropRuleTest
private final static DataSegment.Builder builder = DataSegment.builder()
.dataSource("test")
.version(new DateTime("2012-12-31T01:00:00").toString())
.shardSpec(new NoneShardSpec());
.shardSpec(NoneShardSpec.instance());
@Test
public void testAppliesToAll()

View File

@ -38,7 +38,7 @@ public class PeriodLoadRuleTest
private final static DataSegment.Builder builder = DataSegment.builder()
.dataSource("test")
.version(new DateTime().toString())
.shardSpec(new NoneShardSpec());
.shardSpec(NoneShardSpec.instance());
@Test
public void testAppliesToAll()

View File

@ -175,7 +175,7 @@ public class DruidJsonValidatorTest
null,
null,
1,
new NoneShardSpec(),
NoneShardSpec.instance(),
new IndexSpec(),
null,
0,