mirror of https://github.com/apache/druid.git
Costbalancer Multithread
This commit is contained in:
parent
82fd35726f
commit
a8d8d1ea06
|
@ -25,6 +25,12 @@ import java.util.List;
|
||||||
|
|
||||||
public interface BalancerStrategy
|
public interface BalancerStrategy
|
||||||
{
|
{
|
||||||
|
public static final String COST = "cost";
|
||||||
|
|
||||||
|
public static final String COST_MULTI = "cost_multi";
|
||||||
|
|
||||||
|
public static final String RANDOM = "random";
|
||||||
|
|
||||||
public ServerHolder findNewSegmentHomeBalancer(final DataSegment proposalSegment, final List<ServerHolder> serverHolders);
|
public ServerHolder findNewSegmentHomeBalancer(final DataSegment proposalSegment, final List<ServerHolder> serverHolders);
|
||||||
|
|
||||||
public ServerHolder findNewSegmentHomeReplicator(final DataSegment proposalSegment, final List<ServerHolder> serverHolders);
|
public ServerHolder findNewSegmentHomeReplicator(final DataSegment proposalSegment, final List<ServerHolder> serverHolders);
|
||||||
|
|
|
@ -0,0 +1,273 @@
|
||||||
|
/*
|
||||||
|
* Druid - a distributed column store.
|
||||||
|
* Copyright (C) 2012, 2013 Metamarkets Group Inc.
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or
|
||||||
|
* modify it under the terms of the GNU General Public License
|
||||||
|
* as published by the Free Software Foundation; either version 2
|
||||||
|
* of the License, or (at your option) any later version.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program; if not, write to the Free Software
|
||||||
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.druid.server.coordinator;
|
||||||
|
|
||||||
|
import com.google.api.client.util.Lists;
|
||||||
|
import com.metamx.common.Pair;
|
||||||
|
import com.metamx.emitter.EmittingLogger;
|
||||||
|
import io.druid.timeline.DataSegment;
|
||||||
|
import org.joda.time.DateTime;
|
||||||
|
import org.joda.time.Interval;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.Callable;
|
||||||
|
import java.util.concurrent.ExecutionException;
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.Executors;
|
||||||
|
import java.util.concurrent.Future;
|
||||||
|
|
||||||
|
public class CostBalancerMultithreadStrategy implements BalancerStrategy
|
||||||
|
{
|
||||||
|
private static final EmittingLogger log = new EmittingLogger(CostBalancerMultithreadStrategy.class);
|
||||||
|
private static final int DAY_IN_MILLIS = 1000 * 60 * 60 * 24;
|
||||||
|
private static final int SEVEN_DAYS_IN_MILLIS = 7 * DAY_IN_MILLIS;
|
||||||
|
private static final int THIRTY_DAYS_IN_MILLIS = 30 * DAY_IN_MILLIS;
|
||||||
|
private final long referenceTimestampInMillis;
|
||||||
|
|
||||||
|
public CostBalancerMultithreadStrategy(DateTime referenceTimestamp)
|
||||||
|
{
|
||||||
|
this.referenceTimestampInMillis = referenceTimestamp.getMillis();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ServerHolder findNewSegmentHomeReplicator(
|
||||||
|
DataSegment proposalSegment, List<ServerHolder> serverHolders
|
||||||
|
)
|
||||||
|
{
|
||||||
|
ServerHolder holder = chooseBestServer(proposalSegment, serverHolders, false).rhs;
|
||||||
|
if (holder != null && !holder.isServingSegment(proposalSegment)) {
|
||||||
|
return holder;
|
||||||
|
}
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ServerHolder findNewSegmentHomeBalancer(
|
||||||
|
DataSegment proposalSegment, List<ServerHolder> serverHolders
|
||||||
|
)
|
||||||
|
{
|
||||||
|
return chooseBestServer(proposalSegment, serverHolders, true).rhs;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* For assignment, we want to move to the lowest cost server that isn't already serving the segment.
|
||||||
|
*
|
||||||
|
* @param proposalSegment A DataSegment that we are proposing to move.
|
||||||
|
* @param serverHolders An iterable of ServerHolders for a particular tier.
|
||||||
|
*
|
||||||
|
* @return A ServerHolder with the new home for a segment.
|
||||||
|
*/
|
||||||
|
|
||||||
|
private Pair<Double, ServerHolder> chooseBestServer(
|
||||||
|
final DataSegment proposalSegment,
|
||||||
|
final Iterable<ServerHolder> serverHolders,
|
||||||
|
final boolean includeCurrentServer
|
||||||
|
)
|
||||||
|
{
|
||||||
|
Pair<Double, ServerHolder> bestServer = Pair.of(Double.POSITIVE_INFINITY, null);
|
||||||
|
|
||||||
|
ExecutorService service = Executors.newCachedThreadPool();
|
||||||
|
List<Future<Pair<Double, ServerHolder>>> futures = Lists.newArrayList();
|
||||||
|
|
||||||
|
for (final ServerHolder server : serverHolders) {
|
||||||
|
futures.add(service.submit(new CostCalculator(server, proposalSegment, includeCurrentServer)));
|
||||||
|
}
|
||||||
|
|
||||||
|
for (Future<Pair<Double, ServerHolder>> f : futures) {
|
||||||
|
try {
|
||||||
|
Pair<Double, ServerHolder> server = f.get();
|
||||||
|
if (server.lhs < bestServer.lhs) {
|
||||||
|
bestServer = server;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
catch (InterruptedException e) {
|
||||||
|
e.printStackTrace();
|
||||||
|
}
|
||||||
|
catch (ExecutionException e) {
|
||||||
|
e.printStackTrace();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
service.shutdown();
|
||||||
|
return bestServer;
|
||||||
|
}
|
||||||
|
|
||||||
|
private final class CostCalculator implements Callable<Pair<Double, ServerHolder>>
|
||||||
|
{
|
||||||
|
private final ServerHolder server;
|
||||||
|
private final DataSegment proposalSegment;
|
||||||
|
private final boolean includeCurrentServer;
|
||||||
|
|
||||||
|
CostCalculator(final ServerHolder server, final DataSegment proposalSegment, final boolean includeCurrentServer)
|
||||||
|
{
|
||||||
|
this.server = server;
|
||||||
|
this.proposalSegment = proposalSegment;
|
||||||
|
this.includeCurrentServer = includeCurrentServer;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Pair<Double, ServerHolder> call() throws Exception
|
||||||
|
{
|
||||||
|
if (includeCurrentServer || !server.isServingSegment(proposalSegment)) {
|
||||||
|
final long proposalSegmentSize = proposalSegment.getSize();
|
||||||
|
/** Don't calculate cost if the server doesn't have enough space or is loading the segment */
|
||||||
|
if (proposalSegmentSize > server.getAvailableSize() || server.isLoadingSegment(proposalSegment)) {
|
||||||
|
return Pair.of(Double.POSITIVE_INFINITY, null);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** The contribution to the total cost of a given server by proposing to move the segment to that server is... */
|
||||||
|
double cost = 0f;
|
||||||
|
/** the sum of the costs of other (exclusive of the proposalSegment) segments on the server */
|
||||||
|
for (DataSegment segment : server.getServer().getSegments().values()) {
|
||||||
|
if (!proposalSegment.equals(segment)) {
|
||||||
|
cost += computeJointSegmentCosts(proposalSegment, segment);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/** plus the costs of segments that will be loaded */
|
||||||
|
for (DataSegment segment : server.getPeon().getSegmentsToLoad()) {
|
||||||
|
cost += computeJointSegmentCosts(proposalSegment, segment);
|
||||||
|
}
|
||||||
|
|
||||||
|
return Pair.of(cost, server);
|
||||||
|
}
|
||||||
|
return Pair.of(Double.POSITIVE_INFINITY, null);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This defines the unnormalized cost function between two segments. There is a base cost given by
|
||||||
|
* the minimum size of the two segments and additional penalties.
|
||||||
|
* recencyPenalty: it is more likely that recent segments will be queried together
|
||||||
|
* dataSourcePenalty: if two segments belong to the same data source, they are more likely to be involved
|
||||||
|
* in the same queries
|
||||||
|
* gapPenalty: it is more likely that segments close together in time will be queried together
|
||||||
|
*
|
||||||
|
* @param segment1 The first DataSegment.
|
||||||
|
* @param segment2 The second DataSegment.
|
||||||
|
*
|
||||||
|
* @return The joint cost of placing the two DataSegments together on one node.
|
||||||
|
*/
|
||||||
|
public double computeJointSegmentCosts(final DataSegment segment1, final DataSegment segment2)
|
||||||
|
{
|
||||||
|
final Interval gap = segment1.getInterval().gap(segment2.getInterval());
|
||||||
|
|
||||||
|
final double baseCost = Math.min(segment1.getSize(), segment2.getSize());
|
||||||
|
double recencyPenalty = 1;
|
||||||
|
double dataSourcePenalty = 1;
|
||||||
|
double gapPenalty = 1;
|
||||||
|
|
||||||
|
if (segment1.getDataSource().equals(segment2.getDataSource())) {
|
||||||
|
dataSourcePenalty = 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
double segment1diff = referenceTimestampInMillis - segment1.getInterval().getEndMillis();
|
||||||
|
double segment2diff = referenceTimestampInMillis - segment2.getInterval().getEndMillis();
|
||||||
|
if (segment1diff < SEVEN_DAYS_IN_MILLIS && segment2diff < SEVEN_DAYS_IN_MILLIS) {
|
||||||
|
recencyPenalty = (2 - segment1diff / SEVEN_DAYS_IN_MILLIS) * (2 - segment2diff / SEVEN_DAYS_IN_MILLIS);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** gap is null if the two segment intervals overlap or if they're adjacent */
|
||||||
|
if (gap == null) {
|
||||||
|
gapPenalty = 2;
|
||||||
|
} else {
|
||||||
|
long gapMillis = gap.toDurationMillis();
|
||||||
|
if (gapMillis < THIRTY_DAYS_IN_MILLIS) {
|
||||||
|
gapPenalty = 2 - gapMillis / THIRTY_DAYS_IN_MILLIS;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
final double cost = baseCost * recencyPenalty * dataSourcePenalty * gapPenalty;
|
||||||
|
|
||||||
|
return cost;
|
||||||
|
}
|
||||||
|
|
||||||
|
public BalancerSegmentHolder pickSegmentToMove(final List<ServerHolder> serverHolders)
|
||||||
|
{
|
||||||
|
ReservoirSegmentSampler sampler = new ReservoirSegmentSampler();
|
||||||
|
return sampler.getRandomBalancerSegmentHolder(serverHolders);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Calculates the initial cost of the Druid segment configuration.
|
||||||
|
*
|
||||||
|
* @param serverHolders A list of ServerHolders for a particular tier.
|
||||||
|
*
|
||||||
|
* @return The initial cost of the Druid tier.
|
||||||
|
*/
|
||||||
|
public double calculateInitialTotalCost(final List<ServerHolder> serverHolders)
|
||||||
|
{
|
||||||
|
double cost = 0;
|
||||||
|
for (ServerHolder server : serverHolders) {
|
||||||
|
DataSegment[] segments = server.getServer().getSegments().values().toArray(new DataSegment[]{});
|
||||||
|
for (int i = 0; i < segments.length; ++i) {
|
||||||
|
for (int j = i; j < segments.length; ++j) {
|
||||||
|
cost += computeJointSegmentCosts(segments[i], segments[j]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return cost;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Calculates the cost normalization. This is such that the normalized cost is lower bounded
|
||||||
|
* by 1 (e.g. when each segment gets its own historical node).
|
||||||
|
*
|
||||||
|
* @param serverHolders A list of ServerHolders for a particular tier.
|
||||||
|
*
|
||||||
|
* @return The normalization value (the sum of the diagonal entries in the
|
||||||
|
* pairwise cost matrix). This is the cost of a cluster if each
|
||||||
|
* segment were to get its own historical node.
|
||||||
|
*/
|
||||||
|
public double calculateNormalization(final List<ServerHolder> serverHolders)
|
||||||
|
{
|
||||||
|
double cost = 0;
|
||||||
|
for (ServerHolder server : serverHolders) {
|
||||||
|
for (DataSegment segment : server.getServer().getSegments().values()) {
|
||||||
|
cost += computeJointSegmentCosts(segment, segment);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return cost;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void emitStats(
|
||||||
|
String tier,
|
||||||
|
CoordinatorStats stats, List<ServerHolder> serverHolderList
|
||||||
|
)
|
||||||
|
{
|
||||||
|
final double initialTotalCost = calculateInitialTotalCost(serverHolderList);
|
||||||
|
final double normalization = calculateNormalization(serverHolderList);
|
||||||
|
final double normalizedInitialCost = initialTotalCost / normalization;
|
||||||
|
|
||||||
|
stats.addToTieredStat("initialCost", tier, (long) initialTotalCost);
|
||||||
|
stats.addToTieredStat("normalization", tier, (long) normalization);
|
||||||
|
stats.addToTieredStat("normalizedInitialCostTimesOneThousand", tier, (long) (normalizedInitialCost * 1000));
|
||||||
|
|
||||||
|
log.info(
|
||||||
|
"[%s]: Initial Total Cost: [%f], Normalization: [%f], Initial Normalized Cost: [%f]",
|
||||||
|
tier,
|
||||||
|
initialTotalCost,
|
||||||
|
normalization,
|
||||||
|
normalizedInitialCost
|
||||||
|
);
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,31 @@
|
||||||
|
/*
|
||||||
|
* Druid - a distributed column store.
|
||||||
|
* Copyright (C) 2012, 2013 Metamarkets Group Inc.
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or
|
||||||
|
* modify it under the terms of the GNU General Public License
|
||||||
|
* as published by the Free Software Foundation; either version 2
|
||||||
|
* of the License, or (at your option) any later version.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program; if not, write to the Free Software
|
||||||
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
||||||
|
*/
|
||||||
|
package io.druid.server.coordinator;
|
||||||
|
|
||||||
|
import org.joda.time.DateTime;
|
||||||
|
|
||||||
|
public class CostBalancerMultithreadStrategyFactory implements BalancerStrategyFactory
|
||||||
|
{
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public BalancerStrategy createBalancerStrategy(DateTime referenceTimestamp)
|
||||||
|
{
|
||||||
|
return new CostBalancerStrategy(referenceTimestamp);
|
||||||
|
}
|
||||||
|
}
|
|
@ -19,6 +19,9 @@
|
||||||
|
|
||||||
package io.druid.server.coordinator;
|
package io.druid.server.coordinator;
|
||||||
|
|
||||||
|
import static io.druid.server.coordinator.BalancerStrategy.COST;
|
||||||
|
import static io.druid.server.coordinator.BalancerStrategy.COST_MULTI;
|
||||||
|
import static io.druid.server.coordinator.BalancerStrategy.RANDOM;
|
||||||
import com.google.common.base.Function;
|
import com.google.common.base.Function;
|
||||||
import com.google.common.base.Predicate;
|
import com.google.common.base.Predicate;
|
||||||
import com.google.common.base.Throwables;
|
import com.google.common.base.Throwables;
|
||||||
|
@ -736,6 +739,14 @@ public class DruidCoordinator
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
BalancerStrategyFactory factory= null;
|
||||||
|
if (COST.equals(config.getCoordinatorBalancerStrategy()))
|
||||||
|
factory = new CostBalancerStrategyFactory();
|
||||||
|
else if (RANDOM.equals(config.getCoordinatorBalancerStrategy()))
|
||||||
|
factory= new RandomBalancerStrategyFactory();
|
||||||
|
else if (COST_MULTI.equals(config.getCoordinatorBalancerStrategy()))
|
||||||
|
factory = new CostBalancerMultithreadStrategyFactory();
|
||||||
|
|
||||||
// Do coordinator stuff.
|
// Do coordinator stuff.
|
||||||
DruidCoordinatorRuntimeParams params =
|
DruidCoordinatorRuntimeParams params =
|
||||||
DruidCoordinatorRuntimeParams.newBuilder()
|
DruidCoordinatorRuntimeParams.newBuilder()
|
||||||
|
@ -743,6 +754,7 @@ public class DruidCoordinator
|
||||||
.withDatasources(databaseSegmentManager.getInventory())
|
.withDatasources(databaseSegmentManager.getInventory())
|
||||||
.withDynamicConfigs(getDynamicConfigs())
|
.withDynamicConfigs(getDynamicConfigs())
|
||||||
.withEmitter(emitter)
|
.withEmitter(emitter)
|
||||||
|
.withBalancerStrategyFactory(factory)
|
||||||
.build();
|
.build();
|
||||||
|
|
||||||
for (DruidCoordinatorHelper helper : helpers) {
|
for (DruidCoordinatorHelper helper : helpers) {
|
||||||
|
|
|
@ -42,6 +42,10 @@ public abstract class DruidCoordinatorConfig
|
||||||
@Default("PT1800s")
|
@Default("PT1800s")
|
||||||
public abstract Duration getCoordinatorIndexingPeriod();
|
public abstract Duration getCoordinatorIndexingPeriod();
|
||||||
|
|
||||||
|
@Config("druid.coordinator.balancer.strategy")
|
||||||
|
@Default("cost")
|
||||||
|
public abstract String getCoordinatorBalancerStrategy();
|
||||||
|
|
||||||
@Config("druid.coordinator.merge.on")
|
@Config("druid.coordinator.merge.on")
|
||||||
public boolean isMergeSegments()
|
public boolean isMergeSegments()
|
||||||
{
|
{
|
||||||
|
|
|
@ -0,0 +1,86 @@
|
||||||
|
package io.druid.server.coordination;
|
||||||
|
|
||||||
|
import com.google.api.client.util.Lists;
|
||||||
|
import com.google.common.collect.Maps;
|
||||||
|
import io.druid.client.DruidServer;
|
||||||
|
import io.druid.server.coordinator.CostBalancerStrategy;
|
||||||
|
import io.druid.server.coordinator.LoadQueuePeonTester;
|
||||||
|
import io.druid.server.coordinator.ServerHolder;
|
||||||
|
import io.druid.timeline.DataSegment;
|
||||||
|
import junit.framework.Assert;
|
||||||
|
import org.easymock.EasyMock;
|
||||||
|
import org.joda.time.DateTime;
|
||||||
|
import org.joda.time.DateTimeZone;
|
||||||
|
import org.joda.time.Interval;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
public class CostBalancerStrategyTest
|
||||||
|
{
|
||||||
|
private final List<ServerHolder> serverHolderList = Lists.newArrayList();
|
||||||
|
private final Interval day = DateTime.now().toDateMidnight().toInterval();
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup(){
|
||||||
|
for (int i = 0 ; i < 10; i++){
|
||||||
|
LoadQueuePeonTester fromPeon = new LoadQueuePeonTester();
|
||||||
|
DruidServer druidServer = EasyMock.createMock(DruidServer.class);
|
||||||
|
EasyMock.expect(druidServer.getName()).andReturn("DruidServer_Name_" + i).anyTimes();
|
||||||
|
EasyMock.expect(druidServer.getCurrSize()).andReturn(3000L).anyTimes();
|
||||||
|
EasyMock.expect(druidServer.getMaxSize()).andReturn(10000000L).anyTimes();
|
||||||
|
|
||||||
|
EasyMock.expect(druidServer.getSegment(EasyMock.<String>anyObject())).andReturn(null).anyTimes();
|
||||||
|
Map<String, DataSegment> segments = Maps.newHashMap();
|
||||||
|
for (int j = 0; j < 100; j ++){
|
||||||
|
DataSegment segment = getSegment(j);
|
||||||
|
segments.put(segment.getIdentifier(), segment);
|
||||||
|
EasyMock.expect(druidServer.getSegment(segment.getIdentifier())).andReturn(segment).anyTimes();
|
||||||
|
}
|
||||||
|
EasyMock.expect(druidServer.getSegments()).andReturn(segments).anyTimes();
|
||||||
|
|
||||||
|
EasyMock.replay(druidServer);
|
||||||
|
serverHolderList.add(new ServerHolder(druidServer, fromPeon));
|
||||||
|
}
|
||||||
|
LoadQueuePeonTester fromPeon = new LoadQueuePeonTester();
|
||||||
|
DruidServer druidServer = EasyMock.createMock(DruidServer.class);
|
||||||
|
EasyMock.expect(druidServer.getName()).andReturn("BEST_SERVER").anyTimes();
|
||||||
|
EasyMock.expect(druidServer.getCurrSize()).andReturn(3000L).anyTimes();
|
||||||
|
EasyMock.expect(druidServer.getMaxSize()).andReturn(10000000L).anyTimes();
|
||||||
|
|
||||||
|
EasyMock.expect(druidServer.getSegment(EasyMock.<String>anyObject())).andReturn(null).anyTimes();
|
||||||
|
Map<String, DataSegment> segments = Maps.newHashMap();
|
||||||
|
for (int j = 0; j < 98; j ++){
|
||||||
|
DataSegment segment = getSegment(j);
|
||||||
|
segments.put(segment.getIdentifier(), segment);
|
||||||
|
EasyMock.expect(druidServer.getSegment(segment.getIdentifier())).andReturn(segment).anyTimes();
|
||||||
|
}
|
||||||
|
EasyMock.expect(druidServer.getSegments()).andReturn(segments).anyTimes();
|
||||||
|
|
||||||
|
EasyMock.replay(druidServer);
|
||||||
|
serverHolderList.add(new ServerHolder(druidServer, fromPeon));
|
||||||
|
}
|
||||||
|
|
||||||
|
private DataSegment getSegment(int index){
|
||||||
|
DataSegment segment = EasyMock.createMock(DataSegment.class);
|
||||||
|
EasyMock.expect(segment.getInterval()).andReturn(day).anyTimes();
|
||||||
|
EasyMock.expect(segment.getIdentifier()).andReturn("DUMMY_SEGID_" + index).anyTimes();
|
||||||
|
EasyMock.expect(segment.getDataSource()).andReturn("DUMMY").anyTimes();
|
||||||
|
EasyMock.expect(segment.getSize()).andReturn(index * 100L).anyTimes();
|
||||||
|
EasyMock.replay(segment);
|
||||||
|
return segment;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void test() throws InterruptedException {
|
||||||
|
DataSegment segment = getSegment(1000);
|
||||||
|
|
||||||
|
CostBalancerStrategy strategy = new CostBalancerStrategy(DateTime.now(DateTimeZone.UTC));
|
||||||
|
ServerHolder holder = strategy.findNewSegmentHomeReplicator(segment, serverHolderList);
|
||||||
|
Assert.assertNotNull("Should be able to find a place for new segment!!", holder);
|
||||||
|
Assert.assertEquals("Best Server should be BEST_SERVER", "BEST_SERVER", holder.getServer().getName());
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
Loading…
Reference in New Issue