SOLR-1095: Refactor code to standardize replica assignment

This commit is contained in:
Noble Paul 2017-06-30 15:49:40 +09:30
parent 0159d494f5
commit 196d84b9e0
7 changed files with 234 additions and 102 deletions

View File

@ -25,10 +25,14 @@ import java.util.LinkedHashMap;
import java.util.List; import java.util.List;
import java.util.Locale; import java.util.Locale;
import java.util.Map; import java.util.Map;
import java.util.Random;
import java.util.Set; import java.util.Set;
import java.util.function.Supplier;
import java.util.regex.Matcher; import java.util.regex.Matcher;
import java.util.regex.Pattern; import java.util.regex.Pattern;
import java.util.stream.Collectors;
import com.google.common.collect.ImmutableMap;
import org.apache.solr.client.solrj.cloud.autoscaling.Policy; import org.apache.solr.client.solrj.cloud.autoscaling.Policy;
import org.apache.solr.client.solrj.cloud.autoscaling.PolicyHelper; import org.apache.solr.client.solrj.cloud.autoscaling.PolicyHelper;
import org.apache.solr.client.solrj.impl.CloudSolrClient; import org.apache.solr.client.solrj.impl.CloudSolrClient;
@ -40,7 +44,9 @@ import org.apache.solr.common.SolrException;
import org.apache.solr.common.cloud.ClusterState; import org.apache.solr.common.cloud.ClusterState;
import org.apache.solr.common.cloud.DocCollection; import org.apache.solr.common.cloud.DocCollection;
import org.apache.solr.common.cloud.Replica; import org.apache.solr.common.cloud.Replica;
import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.cloud.Slice; import org.apache.solr.common.cloud.Slice;
import org.apache.solr.common.cloud.ZkNodeProps;
import org.apache.solr.common.cloud.ZkStateReader; import org.apache.solr.common.cloud.ZkStateReader;
import org.apache.solr.common.util.StrUtils; import org.apache.solr.common.util.StrUtils;
import org.apache.solr.common.util.Utils; import org.apache.solr.common.util.Utils;
@ -49,6 +55,11 @@ import org.apache.zookeeper.KeeperException;
import static java.util.Collections.singletonMap; import static java.util.Collections.singletonMap;
import static org.apache.solr.client.solrj.cloud.autoscaling.Policy.POLICY; import static org.apache.solr.client.solrj.cloud.autoscaling.Policy.POLICY;
import static org.apache.solr.cloud.OverseerCollectionMessageHandler.CREATE_NODE_SET;
import static org.apache.solr.cloud.OverseerCollectionMessageHandler.CREATE_NODE_SET_EMPTY;
import static org.apache.solr.cloud.OverseerCollectionMessageHandler.CREATE_NODE_SET_SHUFFLE;
import static org.apache.solr.cloud.OverseerCollectionMessageHandler.CREATE_NODE_SET_SHUFFLE_DEFAULT;
import static org.apache.solr.common.cloud.DocCollection.SNITCH;
import static org.apache.solr.common.cloud.ZkStateReader.CORE_NAME_PROP; import static org.apache.solr.common.cloud.ZkStateReader.CORE_NAME_PROP;
import static org.apache.solr.common.cloud.ZkStateReader.MAX_SHARDS_PER_NODE; import static org.apache.solr.common.cloud.ZkStateReader.MAX_SHARDS_PER_NODE;
import static org.apache.solr.common.cloud.ZkStateReader.SOLR_AUTOSCALING_CONF_PATH; import static org.apache.solr.common.cloud.ZkStateReader.SOLR_AUTOSCALING_CONF_PATH;
@ -141,6 +152,91 @@ public class Assign {
else return replicaName; else return replicaName;
} }
} }
public static List<String> getLiveOrLiveAndCreateNodeSetList(final Set<String> liveNodes, final ZkNodeProps message, final Random random) {
// TODO: add smarter options that look at the current number of cores per
// node?
// for now we just go random (except when createNodeSet and createNodeSet.shuffle=false are passed in)
List<String> nodeList;
final String createNodeSetStr = message.getStr(CREATE_NODE_SET);
final List<String> createNodeList = (createNodeSetStr == null) ? null : StrUtils.splitSmart((CREATE_NODE_SET_EMPTY.equals(createNodeSetStr) ? "" : createNodeSetStr), ",", true);
if (createNodeList != null) {
nodeList = new ArrayList<>(createNodeList);
nodeList.retainAll(liveNodes);
if (message.getBool(CREATE_NODE_SET_SHUFFLE, CREATE_NODE_SET_SHUFFLE_DEFAULT)) {
Collections.shuffle(nodeList, random);
}
} else {
nodeList = new ArrayList<>(liveNodes);
Collections.shuffle(nodeList, random);
}
return nodeList;
}
public static List<ReplicaPosition> identifyNodes(Supplier<CoreContainer> coreContainer,
ZkStateReader zkStateReader,
ClusterState clusterState,
List<String> nodeList,
String collectionName,
ZkNodeProps message,
List<String> shardNames,
int numNrtReplicas,
int numTlogReplicas,
int numPullReplicas) throws KeeperException, InterruptedException {
List<Map> rulesMap = (List) message.get("rule");
String policyName = message.getStr(POLICY);
Map autoScalingJson = Utils.getJson(zkStateReader.getZkClient(), SOLR_AUTOSCALING_CONF_PATH, true);
if (rulesMap == null && policyName == null) {
int i = 0;
List<ReplicaPosition> result = new ArrayList<>();
for (String aShard : shardNames) {
for (Map.Entry<Replica.Type, Integer> e : ImmutableMap.of(Replica.Type.NRT, numNrtReplicas,
Replica.Type.TLOG, numTlogReplicas,
Replica.Type.PULL, numPullReplicas
).entrySet()) {
for (int j = 0; j < e.getValue(); j++){
result.add(new ReplicaPosition(aShard, j, e.getKey(), nodeList.get(i % nodeList.size())));
i++;
}
}
}
return result;
} else {
if (numTlogReplicas + numPullReplicas != 0) {
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST,
Replica.Type.TLOG + " or " + Replica.Type.PULL + " replica types not supported with placement rules or cluster policies");
}
}
if (policyName != null || autoScalingJson.get(Policy.CLUSTER_POLICY) != null) {
return getPositionsUsingPolicy(collectionName,
shardNames, numNrtReplicas, policyName, zkStateReader, nodeList);
} else {
List<Rule> rules = new ArrayList<>();
for (Object map : rulesMap) rules.add(new Rule((Map) map));
Map<String, Integer> sharVsReplicaCount = new HashMap<>();
for (String shard : shardNames) sharVsReplicaCount.put(shard, numNrtReplicas);
ReplicaAssigner replicaAssigner = new ReplicaAssigner(rules,
sharVsReplicaCount,
(List<Map>) message.get(SNITCH),
new HashMap<>(),//this is a new collection. So, there are no nodes in any shard
nodeList,
coreContainer.get(),
clusterState);
Map<ReplicaPosition, String> nodeMappings = replicaAssigner.getNodeMappings();
return nodeMappings.entrySet().stream()
.map(e -> new ReplicaPosition(e.getKey().shard, e.getKey().index, e.getKey().type, e.getValue()))
.collect(Collectors.toList());
}
}
static class ReplicaCount { static class ReplicaCount {
public final String nodeName; public final String nodeName;
@ -191,21 +287,21 @@ public class Assign {
} }
List l = (List) coll.get(DocCollection.RULE); List l = (List) coll.get(DocCollection.RULE);
Map<ReplicaAssigner.Position, String> positions = null; List<ReplicaPosition> replicaPositions = null;
if (l != null) { if (l != null) {
positions = getNodesViaRules(clusterState, shard, numberOfNodes, cc, coll, createNodeList, l); replicaPositions = getNodesViaRules(clusterState, shard, numberOfNodes, cc, coll, createNodeList, l);
} }
String policyName = coll.getStr(POLICY); String policyName = coll.getStr(POLICY);
Map autoScalingJson = Utils.getJson(cc.getZkController().getZkClient(), SOLR_AUTOSCALING_CONF_PATH, true); Map autoScalingJson = Utils.getJson(cc.getZkController().getZkClient(), SOLR_AUTOSCALING_CONF_PATH, true);
if (policyName != null || autoScalingJson.get(Policy.CLUSTER_POLICY) != null) { if (policyName != null || autoScalingJson.get(Policy.CLUSTER_POLICY) != null) {
positions= Assign.getPositionsUsingPolicy(collectionName, Collections.singletonList(shard), numberOfNodes, replicaPositions = Assign.getPositionsUsingPolicy(collectionName, Collections.singletonList(shard), numberOfNodes,
policyName, cc.getZkController().getZkStateReader(), createNodeList); policyName, cc.getZkController().getZkStateReader(), createNodeList);
} }
if(positions != null){ if(replicaPositions != null){
List<ReplicaCount> repCounts = new ArrayList<>(); List<ReplicaCount> repCounts = new ArrayList<>();
for (String s : positions.values()) { for (ReplicaPosition p : replicaPositions) {
repCounts.add(new ReplicaCount(s)); repCounts.add(new ReplicaCount(p.node));
} }
return repCounts; return repCounts;
} }
@ -215,7 +311,8 @@ public class Assign {
return sortedNodeList; return sortedNodeList;
} }
public static Map<ReplicaAssigner.Position, String> getPositionsUsingPolicy(String collName, List<String> shardNames, int numReplicas,
public static List<ReplicaPosition> getPositionsUsingPolicy(String collName, List<String> shardNames, int numReplicas,
String policyName, ZkStateReader zkStateReader, String policyName, ZkStateReader zkStateReader,
List<String> nodesList) throws KeeperException, InterruptedException { List<String> nodesList) throws KeeperException, InterruptedException {
try (CloudSolrClient csc = new CloudSolrClient.Builder() try (CloudSolrClient csc = new CloudSolrClient.Builder()
@ -226,11 +323,11 @@ public class Assign {
Map<String, List<String>> locations = PolicyHelper.getReplicaLocations(collName, Map<String, List<String>> locations = PolicyHelper.getReplicaLocations(collName,
autoScalingJson, autoScalingJson,
clientDataProvider, singletonMap(collName, policyName), shardNames, numReplicas, nodesList); clientDataProvider, singletonMap(collName, policyName), shardNames, numReplicas, nodesList);
Map<ReplicaAssigner.Position, String> result = new HashMap<>(); List<ReplicaPosition> result = new ArrayList<>();
for (Map.Entry<String, List<String>> e : locations.entrySet()) { for (Map.Entry<String, List<String>> e : locations.entrySet()) {
List<String> value = e.getValue(); List<String> value = e.getValue();
for (int i = 0; i < value.size(); i++) { for (int i = 0; i < value.size(); i++) {
result.put(new ReplicaAssigner.Position(e.getKey(), i, Replica.Type.NRT), value.get(i)); result.add(new ReplicaPosition(e.getKey(), i, Replica.Type.NRT, value.get(i)));
} }
} }
return result; return result;
@ -239,7 +336,7 @@ public class Assign {
} }
} }
private static Map<ReplicaAssigner.Position, String> getNodesViaRules(ClusterState clusterState, String shard, int numberOfNodes, private static List<ReplicaPosition> getNodesViaRules(ClusterState clusterState, String shard, int numberOfNodes,
CoreContainer cc, DocCollection coll, List<String> createNodeList, List l) { CoreContainer cc, DocCollection coll, List<String> createNodeList, List l) {
ArrayList<Rule> rules = new ArrayList<>(); ArrayList<Rule> rules = new ArrayList<>();
for (Object o : l) rules.add(new Rule((Map) o)); for (Object o : l) rules.add(new Rule((Map) o));
@ -253,18 +350,18 @@ public class Assign {
n.put(replica.getNodeName(), ++count); n.put(replica.getNodeName(), ++count);
} }
} }
List snitches = (List) coll.get(DocCollection.SNITCH); List snitches = (List) coll.get(SNITCH);
List<String> nodesList = createNodeList == null ? List<String> nodesList = createNodeList == null ?
new ArrayList<>(clusterState.getLiveNodes()) : new ArrayList<>(clusterState.getLiveNodes()) :
createNodeList; createNodeList;
Map<ReplicaAssigner.Position, String> positions = new ReplicaAssigner( Map<ReplicaPosition, String> positions = new ReplicaAssigner(
rules, rules,
Collections.singletonMap(shard, numberOfNodes), Collections.singletonMap(shard, numberOfNodes),
snitches, snitches,
shardVsNodes, shardVsNodes,
nodesList, cc, clusterState).getNodeMappings(); nodesList, cc, clusterState).getNodeMappings();
return positions;// getReplicaCounts(positions); return positions.entrySet().stream().map(e -> e.getKey().setNode(e.getValue())).collect(Collectors.toList());// getReplicaCounts(positions);
} }
private static HashMap<String, ReplicaCount> getNodeNameVsShardCount(String collectionName, private static HashMap<String, ReplicaCount> getNodeNameVsShardCount(String collectionName,

View File

@ -31,7 +31,7 @@ import java.util.concurrent.TimeUnit;
import org.apache.solr.cloud.OverseerCollectionMessageHandler.Cmd; import org.apache.solr.cloud.OverseerCollectionMessageHandler.Cmd;
import org.apache.solr.cloud.overseer.ClusterStateMutator; import org.apache.solr.cloud.overseer.ClusterStateMutator;
import org.apache.solr.cloud.rule.ReplicaAssigner; import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.SolrException; import org.apache.solr.common.SolrException;
import org.apache.solr.common.SolrException.ErrorCode; import org.apache.solr.common.SolrException.ErrorCode;
import org.apache.solr.common.cloud.ClusterState; import org.apache.solr.common.cloud.ClusterState;
@ -131,12 +131,12 @@ public class CreateCollectionCmd implements Cmd {
// add our new cores to existing nodes serving the least number of cores // add our new cores to existing nodes serving the least number of cores
// but (for now) require that each core goes on a distinct node. // but (for now) require that each core goes on a distinct node.
final List<String> nodeList = OverseerCollectionMessageHandler.getLiveOrLiveAndCreateNodeSetList(clusterState.getLiveNodes(), message, RANDOM); final List<String> nodeList = Assign.getLiveOrLiveAndCreateNodeSetList(clusterState.getLiveNodes(), message, RANDOM);
Map<ReplicaAssigner.Position, String> positionVsNodes; List<ReplicaPosition> replicaPositions;
if (nodeList.isEmpty()) { if (nodeList.isEmpty()) {
log.warn("It is unusual to create a collection ("+collectionName+") without cores."); log.warn("It is unusual to create a collection ("+collectionName+") without cores.");
positionVsNodes = new HashMap<>(); replicaPositions = new ArrayList<>();
} else { } else {
int totalNumReplicas = numNrtReplicas + numTlogReplicas + numPullReplicas; int totalNumReplicas = numNrtReplicas + numTlogReplicas + numPullReplicas;
if (totalNumReplicas > nodeList.size()) { if (totalNumReplicas > nodeList.size()) {
@ -164,7 +164,9 @@ public class CreateCollectionCmd implements Cmd {
+ " shards to be created (higher than the allowed number)"); + " shards to be created (higher than the allowed number)");
} }
positionVsNodes = ocmh.identifyNodes(clusterState, nodeList, collectionName, message, shardNames, numNrtReplicas, numTlogReplicas, numPullReplicas); replicaPositions = Assign.identifyNodes(() -> ocmh.overseer.getZkController().getCoreContainer(),
ocmh.zkStateReader
, clusterState, nodeList, collectionName, message, shardNames, numNrtReplicas, numTlogReplicas, numPullReplicas);
} }
ZkStateReader zkStateReader = ocmh.zkStateReader; ZkStateReader zkStateReader = ocmh.zkStateReader;
@ -207,12 +209,11 @@ public class CreateCollectionCmd implements Cmd {
log.debug(formatString("Creating SolrCores for new collection {0}, shardNames {1} , nrtReplicas : {2}, tlogReplicas: {3}, pullReplicas: {4}", log.debug(formatString("Creating SolrCores for new collection {0}, shardNames {1} , nrtReplicas : {2}, tlogReplicas: {3}, pullReplicas: {4}",
collectionName, shardNames, numNrtReplicas, numTlogReplicas, numPullReplicas)); collectionName, shardNames, numNrtReplicas, numTlogReplicas, numPullReplicas));
Map<String,ShardRequest> coresToCreate = new LinkedHashMap<>(); Map<String,ShardRequest> coresToCreate = new LinkedHashMap<>();
for (Map.Entry<ReplicaAssigner.Position, String> e : positionVsNodes.entrySet()) { for (ReplicaPosition replicaPosition : replicaPositions) {
ReplicaAssigner.Position position = e.getKey(); String nodeName = replicaPosition.node;
String nodeName = e.getValue(); String coreName = Assign.buildCoreName(collectionName, replicaPosition.shard, replicaPosition.type, replicaPosition.index + 1);
String coreName = Assign.buildCoreName(collectionName, position.shard, position.type, position.index + 1);
log.debug(formatString("Creating core {0} as part of shard {1} of collection {2} on {3}" log.debug(formatString("Creating core {0} as part of shard {1} of collection {2} on {3}"
, coreName, position.shard, collectionName, nodeName)); , coreName, replicaPosition.shard, collectionName, nodeName));
String baseUrl = zkStateReader.getBaseUrlForNodeName(nodeName); String baseUrl = zkStateReader.getBaseUrlForNodeName(nodeName);
@ -222,11 +223,11 @@ public class CreateCollectionCmd implements Cmd {
ZkNodeProps props = new ZkNodeProps( ZkNodeProps props = new ZkNodeProps(
Overseer.QUEUE_OPERATION, ADDREPLICA.toString(), Overseer.QUEUE_OPERATION, ADDREPLICA.toString(),
ZkStateReader.COLLECTION_PROP, collectionName, ZkStateReader.COLLECTION_PROP, collectionName,
ZkStateReader.SHARD_ID_PROP, position.shard, ZkStateReader.SHARD_ID_PROP, replicaPosition.shard,
ZkStateReader.CORE_NAME_PROP, coreName, ZkStateReader.CORE_NAME_PROP, coreName,
ZkStateReader.STATE_PROP, Replica.State.DOWN.toString(), ZkStateReader.STATE_PROP, Replica.State.DOWN.toString(),
ZkStateReader.BASE_URL_PROP, baseUrl, ZkStateReader.BASE_URL_PROP, baseUrl,
ZkStateReader.REPLICA_TYPE, position.type.name()); ZkStateReader.REPLICA_TYPE, replicaPosition.type.name());
Overseer.getStateUpdateQueue(zkStateReader.getZkClient()).offer(Utils.toJSON(props)); Overseer.getStateUpdateQueue(zkStateReader.getZkClient()).offer(Utils.toJSON(props));
} }
@ -237,10 +238,10 @@ public class CreateCollectionCmd implements Cmd {
params.set(CoreAdminParams.NAME, coreName); params.set(CoreAdminParams.NAME, coreName);
params.set(COLL_CONF, configName); params.set(COLL_CONF, configName);
params.set(CoreAdminParams.COLLECTION, collectionName); params.set(CoreAdminParams.COLLECTION, collectionName);
params.set(CoreAdminParams.SHARD, position.shard); params.set(CoreAdminParams.SHARD, replicaPosition.shard);
params.set(ZkStateReader.NUM_SHARDS_PROP, numSlices); params.set(ZkStateReader.NUM_SHARDS_PROP, numSlices);
params.set(CoreAdminParams.NEW_COLLECTION, "true"); params.set(CoreAdminParams.NEW_COLLECTION, "true");
params.set(CoreAdminParams.REPLICA_TYPE, position.type.name()); params.set(CoreAdminParams.REPLICA_TYPE, replicaPosition.type.name());
if (async != null) { if (async != null) {
String coreAdminAsyncId = async + Math.abs(System.nanoTime()); String coreAdminAsyncId = async + Math.abs(System.nanoTime());

View File

@ -28,12 +28,13 @@ import java.util.LinkedHashMap;
import java.util.List; import java.util.List;
import java.util.Locale; import java.util.Locale;
import java.util.Map; import java.util.Map;
import java.util.Objects;
import java.util.Optional; import java.util.Optional;
import java.util.Properties; import java.util.Properties;
import java.util.Set; import java.util.Set;
import org.apache.solr.cloud.overseer.OverseerAction; import org.apache.solr.cloud.overseer.OverseerAction;
import org.apache.solr.cloud.rule.ReplicaAssigner; import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.SolrException; import org.apache.solr.common.SolrException;
import org.apache.solr.common.SolrException.ErrorCode; import org.apache.solr.common.SolrException.ErrorCode;
import org.apache.solr.common.cloud.ClusterState; import org.apache.solr.common.cloud.ClusterState;
@ -108,7 +109,7 @@ public class RestoreCmd implements OverseerCollectionMessageHandler.Cmd {
DocCollection backupCollectionState = backupMgr.readCollectionState(location, backupName, backupCollection); DocCollection backupCollectionState = backupMgr.readCollectionState(location, backupName, backupCollection);
// Get the Solr nodes to restore a collection. // Get the Solr nodes to restore a collection.
final List<String> nodeList = OverseerCollectionMessageHandler.getLiveOrLiveAndCreateNodeSetList( final List<String> nodeList = Assign.getLiveOrLiveAndCreateNodeSetList(
zkStateReader.getClusterState().getLiveNodes(), message, RANDOM); zkStateReader.getClusterState().getLiveNodes(), message, RANDOM);
int numShards = backupCollectionState.getActiveSlices().size(); int numShards = backupCollectionState.getActiveSlices().size();
@ -213,8 +214,11 @@ public class RestoreCmd implements OverseerCollectionMessageHandler.Cmd {
List<String> sliceNames = new ArrayList<>(); List<String> sliceNames = new ArrayList<>();
restoreCollection.getSlices().forEach(x -> sliceNames.add(x.getName())); restoreCollection.getSlices().forEach(x -> sliceNames.add(x.getName()));
Map<ReplicaAssigner.Position, String> positionVsNodes = ocmh.identifyNodes(clusterState, nodeList, List<ReplicaPosition> replicaPositions = Assign.identifyNodes(() -> ocmh.overseer.getZkController().getCoreContainer(),
restoreCollectionName, message, sliceNames, numNrtReplicas, numTlogReplicas, numPullReplicas); ocmh.zkStateReader, clusterState,
nodeList, restoreCollectionName,
message, sliceNames,
numNrtReplicas, numTlogReplicas, numPullReplicas);
//Create one replica per shard and copy backed up data to it //Create one replica per shard and copy backed up data to it
for (Slice slice : restoreCollection.getSlices()) { for (Slice slice : restoreCollection.getSlices()) {
@ -235,12 +239,11 @@ public class RestoreCmd implements OverseerCollectionMessageHandler.Cmd {
// Get the first node matching the shard to restore in // Get the first node matching the shard to restore in
String node; String node;
for (Map.Entry<ReplicaAssigner.Position, String> pvn : positionVsNodes.entrySet()) { for (ReplicaPosition replicaPosition : replicaPositions) {
ReplicaAssigner.Position position = pvn.getKey(); if (Objects.equals(replicaPosition.shard, slice.getName())) {
if (position.shard == slice.getName()) { node = replicaPosition.node;
node = pvn.getValue();
propMap.put(CoreAdminParams.NODE, node); propMap.put(CoreAdminParams.NODE, node);
positionVsNodes.remove(position); replicaPositions.remove(replicaPosition);
break; break;
} }
} }
@ -319,12 +322,11 @@ public class RestoreCmd implements OverseerCollectionMessageHandler.Cmd {
// Get the first node matching the shard to restore in // Get the first node matching the shard to restore in
String node; String node;
for (Map.Entry<ReplicaAssigner.Position, String> pvn : positionVsNodes.entrySet()) { for (ReplicaPosition replicaPosition : replicaPositions) {
ReplicaAssigner.Position position = pvn.getKey(); if (Objects.equals(replicaPosition.shard, slice.getName())) {
if (position.shard == slice.getName()) { node = replicaPosition.node;
node = pvn.getValue();
propMap.put(CoreAdminParams.NODE, node); propMap.put(CoreAdminParams.NODE, node);
positionVsNodes.remove(position); replicaPositions.remove(replicaPosition);
break; break;
} }
} }

View File

@ -30,7 +30,7 @@ import java.util.Set;
import org.apache.solr.client.solrj.request.CoreAdminRequest; import org.apache.solr.client.solrj.request.CoreAdminRequest;
import org.apache.solr.cloud.OverseerCollectionMessageHandler.Cmd; import org.apache.solr.cloud.OverseerCollectionMessageHandler.Cmd;
import org.apache.solr.cloud.overseer.OverseerAction; import org.apache.solr.cloud.overseer.OverseerAction;
import org.apache.solr.cloud.rule.ReplicaAssigner; import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.SolrException; import org.apache.solr.common.SolrException;
import org.apache.solr.common.cloud.ClusterState; import org.apache.solr.common.cloud.ClusterState;
import org.apache.solr.common.cloud.CompositeIdRouter; import org.apache.solr.common.cloud.CompositeIdRouter;
@ -381,7 +381,8 @@ public class SplitShardCmd implements Cmd {
// TODO: change this to handle sharding a slice into > 2 sub-shards. // TODO: change this to handle sharding a slice into > 2 sub-shards.
Map<ReplicaAssigner.Position, String> nodeMap = ocmh.identifyNodes(clusterState, List<ReplicaPosition> replicaPositions = Assign.identifyNodes(() -> ocmh.overseer.getZkController().getCoreContainer(),
ocmh.zkStateReader, clusterState,
new ArrayList<>(clusterState.getLiveNodes()), new ArrayList<>(clusterState.getLiveNodes()),
collectionName, collectionName,
new ZkNodeProps(collection.getProperties()), new ZkNodeProps(collection.getProperties()),
@ -389,10 +390,10 @@ public class SplitShardCmd implements Cmd {
List<Map<String, Object>> replicas = new ArrayList<>((repFactor - 1) * 2); List<Map<String, Object>> replicas = new ArrayList<>((repFactor - 1) * 2);
for (Map.Entry<ReplicaAssigner.Position, String> entry : nodeMap.entrySet()) { for (ReplicaPosition replicaPosition : replicaPositions) {
String sliceName = entry.getKey().shard; String sliceName = replicaPosition.shard;
String subShardNodeName = entry.getValue(); String subShardNodeName = replicaPosition.node;
String shardName = collectionName + "_" + sliceName + "_replica" + (entry.getKey().index); String shardName = collectionName + "_" + sliceName + "_replica" + (replicaPosition.index);
log.info("Creating replica shard " + shardName + " as part of slice " + sliceName + " of collection " log.info("Creating replica shard " + shardName + " as part of slice " + sliceName + " of collection "
+ collectionName + " on " + subShardNodeName); + collectionName + " on " + subShardNodeName);

View File

@ -33,6 +33,7 @@ import java.util.concurrent.atomic.AtomicReference;
import org.apache.solr.common.SolrException; import org.apache.solr.common.SolrException;
import org.apache.solr.common.cloud.ClusterState; import org.apache.solr.common.cloud.ClusterState;
import org.apache.solr.common.cloud.DocCollection; import org.apache.solr.common.cloud.DocCollection;
import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.cloud.Replica; import org.apache.solr.common.cloud.Replica;
import org.apache.solr.common.cloud.Slice; import org.apache.solr.common.cloud.Slice;
import org.apache.solr.common.cloud.rule.ImplicitSnitch; import org.apache.solr.common.cloud.rule.ImplicitSnitch;
@ -59,31 +60,6 @@ public class ReplicaAssigner {
private Map<String, AtomicInteger> nodeVsCores = new HashMap<>(); private Map<String, AtomicInteger> nodeVsCores = new HashMap<>();
public static class Position implements Comparable<Position> {
public final String shard;
public final int index;
public final Replica.Type type;
public Position(String shard, int replicaIdx, Replica.Type type) {
this.shard = shard;
this.index = replicaIdx;
this.type = type;
}
@Override
public int compareTo(Position that) {
//this is to ensure that we try one replica from each shard first instead of
// all replicas from same shard
return that.index > index ? -1 : that.index == index ? 0 : 1;
}
@Override
public String toString() {
return shard + ":" + index;
}
}
/** /**
* @param shardVsReplicaCount shard names vs no:of replicas required for each of those shards * @param shardVsReplicaCount shard names vs no:of replicas required for each of those shards
* @param snitches snitches details * @param snitches snitches details
@ -128,8 +104,8 @@ public class ReplicaAssigner {
* For each shard return a new set of nodes where the replicas need to be created satisfying * For each shard return a new set of nodes where the replicas need to be created satisfying
* the specified rule * the specified rule
*/ */
public Map<Position, String> getNodeMappings() { public Map<ReplicaPosition, String> getNodeMappings() {
Map<Position, String> result = getNodeMappings0(); Map<ReplicaPosition, String> result = getNodeMappings0();
if (result == null) { if (result == null) {
String msg = "Could not identify nodes matching the rules " + rules; String msg = "Could not identify nodes matching the rules " + rules;
if (!failedNodes.isEmpty()) { if (!failedNodes.isEmpty()) {
@ -149,7 +125,7 @@ public class ReplicaAssigner {
} }
Map<Position, String> getNodeMappings0() { Map<ReplicaPosition, String> getNodeMappings0() {
List<String> shardNames = new ArrayList<>(shardVsReplicaCount.keySet()); List<String> shardNames = new ArrayList<>(shardVsReplicaCount.keySet());
int[] shardOrder = new int[shardNames.size()]; int[] shardOrder = new int[shardNames.size()];
for (int i = 0; i < shardNames.size(); i++) shardOrder[i] = i; for (int i = 0; i < shardNames.size(); i++) shardOrder[i] = i;
@ -168,14 +144,14 @@ public class ReplicaAssigner {
} }
} }
Map<Position, String> result = tryAllPermutations(shardNames, shardOrder, nonWildCardShardRules, false); Map<ReplicaPosition, String> result = tryAllPermutations(shardNames, shardOrder, nonWildCardShardRules, false);
if (result == null && hasFuzzyRules) { if (result == null && hasFuzzyRules) {
result = tryAllPermutations(shardNames, shardOrder, nonWildCardShardRules, true); result = tryAllPermutations(shardNames, shardOrder, nonWildCardShardRules, true);
} }
return result; return result;
} }
private Map<Position, String> tryAllPermutations(List<String> shardNames, private Map<ReplicaPosition, String> tryAllPermutations(List<String> shardNames,
int[] shardOrder, int[] shardOrder,
int nonWildCardShardRules, int nonWildCardShardRules,
boolean fuzzyPhase) { boolean fuzzyPhase) {
@ -187,16 +163,16 @@ public class ReplicaAssigner {
for (; shardPermutations.hasNext(); ) { for (; shardPermutations.hasNext(); ) {
int[] p = shardPermutations.next(); int[] p = shardPermutations.next();
List<Position> positions = new ArrayList<>(); List<ReplicaPosition> replicaPositions = new ArrayList<>();
for (int pos : p) { for (int pos : p) {
for (int j = 0; j < shardVsReplicaCount.get(shardNames.get(pos)); j++) { for (int j = 0; j < shardVsReplicaCount.get(shardNames.get(pos)); j++) {
positions.add(new Position(shardNames.get(pos), j, Replica.Type.NRT)); replicaPositions.add(new ReplicaPosition(shardNames.get(pos), j, Replica.Type.NRT));
} }
} }
Collections.sort(positions); Collections.sort(replicaPositions);
for (Iterator<int[]> it = permutations(rules.size()); it.hasNext(); ) { for (Iterator<int[]> it = permutations(rules.size()); it.hasNext(); ) {
int[] permutation = it.next(); int[] permutation = it.next();
Map<Position, String> result = tryAPermutationOfRules(permutation, positions, fuzzyPhase); Map<ReplicaPosition, String> result = tryAPermutationOfRules(permutation, replicaPositions, fuzzyPhase);
if (result != null) return result; if (result != null) return result;
} }
} }
@ -205,9 +181,9 @@ public class ReplicaAssigner {
} }
private Map<Position, String> tryAPermutationOfRules(int[] rulePermutation, List<Position> positions, boolean fuzzyPhase) { private Map<ReplicaPosition, String> tryAPermutationOfRules(int[] rulePermutation, List<ReplicaPosition> replicaPositions, boolean fuzzyPhase) {
Map<String, Map<String, Object>> nodeVsTagsCopy = getDeepCopy(nodeVsTags, 2); Map<String, Map<String, Object>> nodeVsTagsCopy = getDeepCopy(nodeVsTags, 2);
Map<Position, String> result = new LinkedHashMap<>(); Map<ReplicaPosition, String> result = new LinkedHashMap<>();
int startPosition = 0; int startPosition = 0;
Map<String, Map<String, Integer>> copyOfCurrentState = getDeepCopy(shardVsNodes, 2); Map<String, Map<String, Integer>> copyOfCurrentState = getDeepCopy(shardVsNodes, 2);
List<String> sortedLiveNodes = new ArrayList<>(this.participatingLiveNodes); List<String> sortedLiveNodes = new ArrayList<>(this.participatingLiveNodes);
@ -232,7 +208,7 @@ public class ReplicaAssigner {
return result1; return result1;
}); });
forEachPosition: forEachPosition:
for (Position position : positions) { for (ReplicaPosition replicaPosition : replicaPositions) {
//trying to assign a node by verifying each rule in this rulePermutation //trying to assign a node by verifying each rule in this rulePermutation
forEachNode: forEachNode:
for (int j = 0; j < sortedLiveNodes.size(); j++) { for (int j = 0; j < sortedLiveNodes.size(); j++) {
@ -242,16 +218,16 @@ public class ReplicaAssigner {
Rule rule = rules.get(rulePermutation[i]); Rule rule = rules.get(rulePermutation[i]);
//trying to assign a replica into this node in this shard //trying to assign a replica into this node in this shard
Rule.MatchStatus status = rule.tryAssignNodeToShard(liveNode, Rule.MatchStatus status = rule.tryAssignNodeToShard(liveNode,
copyOfCurrentState, nodeVsTagsCopy, position.shard, fuzzyPhase ? FUZZY_ASSIGN : ASSIGN); copyOfCurrentState, nodeVsTagsCopy, replicaPosition.shard, fuzzyPhase ? FUZZY_ASSIGN : ASSIGN);
if (status == Rule.MatchStatus.CANNOT_ASSIGN_FAIL) { if (status == Rule.MatchStatus.CANNOT_ASSIGN_FAIL) {
continue forEachNode;//try another node for this position continue forEachNode;//try another node for this position
} }
} }
//We have reached this far means this node can be applied to this position //We have reached this far means this node can be applied to this position
//and all rules are fine. So let us change the currentState //and all rules are fine. So let us change the currentState
result.put(position, liveNode); result.put(replicaPosition, liveNode);
Map<String, Integer> nodeNames = copyOfCurrentState.get(position.shard); Map<String, Integer> nodeNames = copyOfCurrentState.get(replicaPosition.shard);
if (nodeNames == null) copyOfCurrentState.put(position.shard, nodeNames = new HashMap<>()); if (nodeNames == null) copyOfCurrentState.put(replicaPosition.shard, nodeNames = new HashMap<>());
Integer n = nodeNames.get(liveNode); Integer n = nodeNames.get(liveNode);
n = n == null ? 1 : n + 1; n = n == null ? 1 : n + 1;
nodeNames.put(liveNode, n); nodeNames.put(liveNode, n);
@ -267,11 +243,11 @@ public class ReplicaAssigner {
return null; return null;
} }
if (positions.size() > result.size()) { if (replicaPositions.size() > result.size()) {
return null; return null;
} }
for (Map.Entry<Position, String> e : result.entrySet()) { for (Map.Entry<ReplicaPosition, String> e : result.entrySet()) {
for (int i = 0; i < rulePermutation.length; i++) { for (int i = 0; i < rulePermutation.length; i++) {
Rule rule = rules.get(rulePermutation[i]); Rule rule = rules.get(rulePermutation[i]);
Rule.MatchStatus matchStatus = rule.tryAssignNodeToShard(e.getValue(), Rule.MatchStatus matchStatus = rule.tryAssignNodeToShard(e.getValue(),

View File

@ -0,0 +1,55 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.solr.common.cloud;
public class ReplicaPosition implements Comparable<ReplicaPosition> {
public final String shard;
public final int index;
public final Replica.Type type;
public String node;
public ReplicaPosition(String shard, int replicaIdx, Replica.Type type) {
this.shard = shard;
this.index = replicaIdx;
this.type = type;
}
public ReplicaPosition(String shard, int replicaIdx, Replica.Type type, String node) {
this.shard = shard;
this.index = replicaIdx;
this.type = type;
this.node = node;
}
@Override
public int compareTo(ReplicaPosition that) {
//this is to ensure that we try one replica from each shard first instead of
// all replicas from same shard
return that.index > index ? -1 : that.index == index ? 0 : 1;
}
@Override
public String toString() {
return shard + ":" + index;
}
public ReplicaPosition setNode(String node) {
this.node = node;
return this;
}
}

View File

@ -28,7 +28,7 @@ import java.util.Set;
import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableList;
import org.apache.solr.SolrTestCaseJ4; import org.apache.solr.SolrTestCaseJ4;
import org.apache.solr.cloud.rule.ReplicaAssigner.Position; import org.apache.solr.common.cloud.ReplicaPosition;
import org.apache.solr.common.cloud.ZkStateReader; import org.apache.solr.common.cloud.ZkStateReader;
import org.apache.solr.common.cloud.rule.Snitch; import org.apache.solr.common.cloud.rule.Snitch;
import org.apache.solr.common.cloud.rule.SnitchContext; import org.apache.solr.common.cloud.rule.SnitchContext;
@ -73,7 +73,7 @@ public class RuleEngineTest extends SolrTestCaseJ4{
"'replica':'1',shard:'*','node':'*'}," + "'replica':'1',shard:'*','node':'*'}," +
" {'freedisk':'>1'}]"); " {'freedisk':'>1'}]");
Map<Position, String> mapping = new ReplicaAssigner( Map<ReplicaPosition, String> mapping = new ReplicaAssigner(
rules, rules,
shardVsReplicaCount, singletonList(MockSnitch.class.getName()), shardVsReplicaCount, singletonList(MockSnitch.class.getName()),
new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null ).getNodeMappings(); new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null ).getNodeMappings();
@ -147,7 +147,7 @@ public class RuleEngineTest extends SolrTestCaseJ4{
"{node:'!127.0.0.1:49947_'}," + "{node:'!127.0.0.1:49947_'}," +
"{freedisk:'>1'}]"); "{freedisk:'>1'}]");
Map shardVsReplicaCount = makeMap("shard1", 2, "shard2", 2); Map shardVsReplicaCount = makeMap("shard1", 2, "shard2", 2);
Map<Position, String> mapping = new ReplicaAssigner( Map<ReplicaPosition, String> mapping = new ReplicaAssigner(
rules, rules,
shardVsReplicaCount, singletonList(MockSnitch.class.getName()), shardVsReplicaCount, singletonList(MockSnitch.class.getName()),
new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null).getNodeMappings(); new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null).getNodeMappings();
@ -236,7 +236,7 @@ public class RuleEngineTest extends SolrTestCaseJ4{
"node5:80", makeMap("rack", "182") "node5:80", makeMap("rack", "182")
); );
MockSnitch.nodeVsTags = nodeVsTags; MockSnitch.nodeVsTags = nodeVsTags;
Map<Position, String> mapping = new ReplicaAssigner( Map<ReplicaPosition, String> mapping = new ReplicaAssigner(
rules, rules,
shardVsReplicaCount, singletonList(MockSnitch.class.getName()), shardVsReplicaCount, singletonList(MockSnitch.class.getName()),
new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null).getNodeMappings0(); new HashMap(), new ArrayList<>(MockSnitch.nodeVsTags.keySet()), null, null).getNodeMappings0();