Avoid sending duplicate remote failed shard requests (#31313)
Today if a write replication request fails, we will send a shard-failed message to the master node to fail that replica. However, if there are many ongoing write replication requests and the master node is busy, we might overwhelm the cluster and the master node with many shard-failed requests. This commit tries to minimize the shard-failed requests in the above scenario by caching the ongoing shard-failed requests. This issue was discussed at https://discuss.elastic.co/t/half-dead-node-lead-to-cluster-hang/113658/25.
This commit is contained in:
parent
d9a6d69a0d
commit
2a8381d3fa
|
@ -25,10 +25,10 @@ import org.elasticsearch.ElasticsearchException;
|
||||||
import org.elasticsearch.ExceptionsHelper;
|
import org.elasticsearch.ExceptionsHelper;
|
||||||
import org.elasticsearch.Version;
|
import org.elasticsearch.Version;
|
||||||
import org.elasticsearch.cluster.ClusterChangedEvent;
|
import org.elasticsearch.cluster.ClusterChangedEvent;
|
||||||
import org.elasticsearch.cluster.ClusterStateTaskExecutor;
|
|
||||||
import org.elasticsearch.cluster.ClusterState;
|
import org.elasticsearch.cluster.ClusterState;
|
||||||
import org.elasticsearch.cluster.ClusterStateObserver;
|
import org.elasticsearch.cluster.ClusterStateObserver;
|
||||||
import org.elasticsearch.cluster.ClusterStateTaskConfig;
|
import org.elasticsearch.cluster.ClusterStateTaskConfig;
|
||||||
|
import org.elasticsearch.cluster.ClusterStateTaskExecutor;
|
||||||
import org.elasticsearch.cluster.ClusterStateTaskListener;
|
import org.elasticsearch.cluster.ClusterStateTaskListener;
|
||||||
import org.elasticsearch.cluster.MasterNodeChangePredicate;
|
import org.elasticsearch.cluster.MasterNodeChangePredicate;
|
||||||
import org.elasticsearch.cluster.NotMasterException;
|
import org.elasticsearch.cluster.NotMasterException;
|
||||||
|
@ -48,6 +48,7 @@ import org.elasticsearch.common.io.stream.StreamInput;
|
||||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||||
import org.elasticsearch.common.settings.Settings;
|
import org.elasticsearch.common.settings.Settings;
|
||||||
import org.elasticsearch.common.unit.TimeValue;
|
import org.elasticsearch.common.unit.TimeValue;
|
||||||
|
import org.elasticsearch.common.util.concurrent.ConcurrentCollections;
|
||||||
import org.elasticsearch.discovery.Discovery;
|
import org.elasticsearch.discovery.Discovery;
|
||||||
import org.elasticsearch.index.shard.ShardId;
|
import org.elasticsearch.index.shard.ShardId;
|
||||||
import org.elasticsearch.node.NodeClosedException;
|
import org.elasticsearch.node.NodeClosedException;
|
||||||
|
@ -68,7 +69,9 @@ import java.util.ArrayList;
|
||||||
import java.util.HashSet;
|
import java.util.HashSet;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Locale;
|
import java.util.Locale;
|
||||||
|
import java.util.Objects;
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
|
import java.util.concurrent.ConcurrentMap;
|
||||||
import java.util.function.Predicate;
|
import java.util.function.Predicate;
|
||||||
|
|
||||||
public class ShardStateAction extends AbstractComponent {
|
public class ShardStateAction extends AbstractComponent {
|
||||||
|
@ -80,6 +83,10 @@ public class ShardStateAction extends AbstractComponent {
|
||||||
private final ClusterService clusterService;
|
private final ClusterService clusterService;
|
||||||
private final ThreadPool threadPool;
|
private final ThreadPool threadPool;
|
||||||
|
|
||||||
|
// a list of shards that failed during replication
|
||||||
|
// we keep track of these shards in order to avoid sending duplicate failed shard requests for a single failing shard.
|
||||||
|
private final ConcurrentMap<FailedShardEntry, CompositeListener> remoteFailedShardsCache = ConcurrentCollections.newConcurrentMap();
|
||||||
|
|
||||||
@Inject
|
@Inject
|
||||||
public ShardStateAction(Settings settings, ClusterService clusterService, TransportService transportService,
|
public ShardStateAction(Settings settings, ClusterService clusterService, TransportService transportService,
|
||||||
AllocationService allocationService, RoutingService routingService, ThreadPool threadPool) {
|
AllocationService allocationService, RoutingService routingService, ThreadPool threadPool) {
|
||||||
|
@ -146,8 +153,35 @@ public class ShardStateAction extends AbstractComponent {
|
||||||
*/
|
*/
|
||||||
public void remoteShardFailed(final ShardId shardId, String allocationId, long primaryTerm, boolean markAsStale, final String message, @Nullable final Exception failure, Listener listener) {
|
public void remoteShardFailed(final ShardId shardId, String allocationId, long primaryTerm, boolean markAsStale, final String message, @Nullable final Exception failure, Listener listener) {
|
||||||
assert primaryTerm > 0L : "primary term should be strictly positive";
|
assert primaryTerm > 0L : "primary term should be strictly positive";
|
||||||
FailedShardEntry shardEntry = new FailedShardEntry(shardId, allocationId, primaryTerm, message, failure, markAsStale);
|
final FailedShardEntry shardEntry = new FailedShardEntry(shardId, allocationId, primaryTerm, message, failure, markAsStale);
|
||||||
sendShardAction(SHARD_FAILED_ACTION_NAME, clusterService.state(), shardEntry, listener);
|
final CompositeListener compositeListener = new CompositeListener(listener);
|
||||||
|
final CompositeListener existingListener = remoteFailedShardsCache.putIfAbsent(shardEntry, compositeListener);
|
||||||
|
if (existingListener == null) {
|
||||||
|
sendShardAction(SHARD_FAILED_ACTION_NAME, clusterService.state(), shardEntry, new Listener() {
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
try {
|
||||||
|
compositeListener.onSuccess();
|
||||||
|
} finally {
|
||||||
|
remoteFailedShardsCache.remove(shardEntry);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception e) {
|
||||||
|
try {
|
||||||
|
compositeListener.onFailure(e);
|
||||||
|
} finally {
|
||||||
|
remoteFailedShardsCache.remove(shardEntry);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
existingListener.addListener(listener);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int remoteShardFailedCacheSize() {
|
||||||
|
return remoteFailedShardsCache.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -414,6 +448,23 @@ public class ShardStateAction extends AbstractComponent {
|
||||||
components.add("markAsStale [" + markAsStale + "]");
|
components.add("markAsStale [" + markAsStale + "]");
|
||||||
return String.join(", ", components);
|
return String.join(", ", components);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean equals(Object o) {
|
||||||
|
if (this == o) return true;
|
||||||
|
if (o == null || getClass() != o.getClass()) return false;
|
||||||
|
FailedShardEntry that = (FailedShardEntry) o;
|
||||||
|
// Exclude message and exception from equals and hashCode
|
||||||
|
return Objects.equals(this.shardId, that.shardId) &&
|
||||||
|
Objects.equals(this.allocationId, that.allocationId) &&
|
||||||
|
primaryTerm == that.primaryTerm &&
|
||||||
|
markAsStale == that.markAsStale;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int hashCode() {
|
||||||
|
return Objects.hash(shardId, allocationId, primaryTerm, markAsStale);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public void shardStarted(final ShardRouting shardRouting, final String message, Listener listener) {
|
public void shardStarted(final ShardRouting shardRouting, final String message, Listener listener) {
|
||||||
|
@ -585,6 +636,72 @@ public class ShardStateAction extends AbstractComponent {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A composite listener that allows registering multiple listeners dynamically.
|
||||||
|
*/
|
||||||
|
static final class CompositeListener implements Listener {
|
||||||
|
private boolean isNotified = false;
|
||||||
|
private Exception failure = null;
|
||||||
|
private final List<Listener> listeners = new ArrayList<>();
|
||||||
|
|
||||||
|
CompositeListener(Listener listener) {
|
||||||
|
listeners.add(listener);
|
||||||
|
}
|
||||||
|
|
||||||
|
void addListener(Listener listener) {
|
||||||
|
final boolean ready;
|
||||||
|
synchronized (this) {
|
||||||
|
ready = this.isNotified;
|
||||||
|
if (ready == false) {
|
||||||
|
listeners.add(listener);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (ready) {
|
||||||
|
if (failure != null) {
|
||||||
|
listener.onFailure(failure);
|
||||||
|
} else {
|
||||||
|
listener.onSuccess();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void onCompleted(Exception failure) {
|
||||||
|
synchronized (this) {
|
||||||
|
this.failure = failure;
|
||||||
|
this.isNotified = true;
|
||||||
|
}
|
||||||
|
RuntimeException firstException = null;
|
||||||
|
for (Listener listener : listeners) {
|
||||||
|
try {
|
||||||
|
if (failure != null) {
|
||||||
|
listener.onFailure(failure);
|
||||||
|
} else {
|
||||||
|
listener.onSuccess();
|
||||||
|
}
|
||||||
|
} catch (RuntimeException innerEx) {
|
||||||
|
if (firstException == null) {
|
||||||
|
firstException = innerEx;
|
||||||
|
} else {
|
||||||
|
firstException.addSuppressed(innerEx);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (firstException != null) {
|
||||||
|
throw firstException;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
onCompleted(null);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception failure) {
|
||||||
|
onCompleted(failure);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
public static class NoLongerPrimaryShardException extends ElasticsearchException {
|
public static class NoLongerPrimaryShardException extends ElasticsearchException {
|
||||||
|
|
||||||
public NoLongerPrimaryShardException(ShardId shardId, String msg) {
|
public NoLongerPrimaryShardException(ShardId shardId, String msg) {
|
||||||
|
|
|
@ -22,11 +22,11 @@ package org.elasticsearch.cluster.action.shard;
|
||||||
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
||||||
import org.apache.lucene.index.CorruptIndexException;
|
import org.apache.lucene.index.CorruptIndexException;
|
||||||
import org.elasticsearch.Version;
|
import org.elasticsearch.Version;
|
||||||
import org.elasticsearch.cluster.action.shard.ShardStateAction.FailedShardEntry;
|
|
||||||
import org.elasticsearch.cluster.ClusterName;
|
import org.elasticsearch.cluster.ClusterName;
|
||||||
import org.elasticsearch.cluster.ClusterState;
|
import org.elasticsearch.cluster.ClusterState;
|
||||||
import org.elasticsearch.cluster.ClusterStateTaskExecutor;
|
import org.elasticsearch.cluster.ClusterStateTaskExecutor;
|
||||||
import org.elasticsearch.cluster.ESAllocationTestCase;
|
import org.elasticsearch.cluster.ESAllocationTestCase;
|
||||||
|
import org.elasticsearch.cluster.action.shard.ShardStateAction.FailedShardEntry;
|
||||||
import org.elasticsearch.cluster.metadata.IndexMetaData;
|
import org.elasticsearch.cluster.metadata.IndexMetaData;
|
||||||
import org.elasticsearch.cluster.metadata.MetaData;
|
import org.elasticsearch.cluster.metadata.MetaData;
|
||||||
import org.elasticsearch.cluster.node.DiscoveryNodes;
|
import org.elasticsearch.cluster.node.DiscoveryNodes;
|
||||||
|
@ -43,6 +43,7 @@ import org.elasticsearch.cluster.routing.allocation.FailedShard;
|
||||||
import org.elasticsearch.cluster.routing.allocation.StaleShard;
|
import org.elasticsearch.cluster.routing.allocation.StaleShard;
|
||||||
import org.elasticsearch.cluster.routing.allocation.decider.ClusterRebalanceAllocationDecider;
|
import org.elasticsearch.cluster.routing.allocation.decider.ClusterRebalanceAllocationDecider;
|
||||||
import org.elasticsearch.common.UUIDs;
|
import org.elasticsearch.common.UUIDs;
|
||||||
|
import org.elasticsearch.common.collect.Tuple;
|
||||||
import org.elasticsearch.common.settings.Settings;
|
import org.elasticsearch.common.settings.Settings;
|
||||||
import org.elasticsearch.common.util.set.Sets;
|
import org.elasticsearch.common.util.set.Sets;
|
||||||
import org.elasticsearch.index.Index;
|
import org.elasticsearch.index.Index;
|
||||||
|
@ -53,9 +54,7 @@ import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
import java.util.function.Function;
|
|
||||||
import java.util.stream.Collectors;
|
import java.util.stream.Collectors;
|
||||||
import java.util.stream.IntStream;
|
import java.util.stream.IntStream;
|
||||||
|
|
||||||
|
@ -131,10 +130,15 @@ public class ShardFailedClusterStateTaskExecutorTests extends ESAllocationTestCa
|
||||||
tasks.addAll(failingTasks);
|
tasks.addAll(failingTasks);
|
||||||
tasks.addAll(nonExistentTasks);
|
tasks.addAll(nonExistentTasks);
|
||||||
ClusterStateTaskExecutor.ClusterTasksResult<FailedShardEntry> result = failingExecutor.execute(currentState, tasks);
|
ClusterStateTaskExecutor.ClusterTasksResult<FailedShardEntry> result = failingExecutor.execute(currentState, tasks);
|
||||||
Map<FailedShardEntry, ClusterStateTaskExecutor.TaskResult> taskResultMap =
|
List<Tuple<FailedShardEntry, ClusterStateTaskExecutor.TaskResult>> taskResultList = new ArrayList<>();
|
||||||
failingTasks.stream().collect(Collectors.toMap(Function.identity(), task -> ClusterStateTaskExecutor.TaskResult.failure(new RuntimeException("simulated applyFailedShards failure"))));
|
for (FailedShardEntry failingTask : failingTasks) {
|
||||||
taskResultMap.putAll(nonExistentTasks.stream().collect(Collectors.toMap(Function.identity(), task -> ClusterStateTaskExecutor.TaskResult.success())));
|
taskResultList.add(Tuple.tuple(failingTask,
|
||||||
assertTaskResults(taskResultMap, result, currentState, false);
|
ClusterStateTaskExecutor.TaskResult.failure(new RuntimeException("simulated applyFailedShards failure"))));
|
||||||
|
}
|
||||||
|
for (FailedShardEntry nonExistentTask : nonExistentTasks) {
|
||||||
|
taskResultList.add(Tuple.tuple(nonExistentTask, ClusterStateTaskExecutor.TaskResult.success()));
|
||||||
|
}
|
||||||
|
assertTaskResults(taskResultList, result, currentState, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
public void testIllegalShardFailureRequests() throws Exception {
|
public void testIllegalShardFailureRequests() throws Exception {
|
||||||
|
@ -147,14 +151,14 @@ public class ShardFailedClusterStateTaskExecutorTests extends ESAllocationTestCa
|
||||||
tasks.add(new FailedShardEntry(failingTask.shardId, failingTask.allocationId,
|
tasks.add(new FailedShardEntry(failingTask.shardId, failingTask.allocationId,
|
||||||
randomIntBetween(1, (int) primaryTerm - 1), failingTask.message, failingTask.failure, randomBoolean()));
|
randomIntBetween(1, (int) primaryTerm - 1), failingTask.message, failingTask.failure, randomBoolean()));
|
||||||
}
|
}
|
||||||
Map<FailedShardEntry, ClusterStateTaskExecutor.TaskResult> taskResultMap =
|
List<Tuple<FailedShardEntry, ClusterStateTaskExecutor.TaskResult>> taskResultList = tasks.stream()
|
||||||
tasks.stream().collect(Collectors.toMap(
|
.map(task -> Tuple.tuple(task, ClusterStateTaskExecutor.TaskResult.failure(
|
||||||
Function.identity(),
|
new ShardStateAction.NoLongerPrimaryShardException(task.shardId, "primary term ["
|
||||||
task -> ClusterStateTaskExecutor.TaskResult.failure(new ShardStateAction.NoLongerPrimaryShardException(task.shardId,
|
+ task.primaryTerm + "] did not match current primary term ["
|
||||||
"primary term [" + task.primaryTerm + "] did not match current primary term [" +
|
+ currentState.metaData().index(task.shardId.getIndex()).primaryTerm(task.shardId.id()) + "]"))))
|
||||||
currentState.metaData().index(task.shardId.getIndex()).primaryTerm(task.shardId.id()) + "]"))));
|
.collect(Collectors.toList());
|
||||||
ClusterStateTaskExecutor.ClusterTasksResult<FailedShardEntry> result = executor.execute(currentState, tasks);
|
ClusterStateTaskExecutor.ClusterTasksResult<FailedShardEntry> result = executor.execute(currentState, tasks);
|
||||||
assertTaskResults(taskResultMap, result, currentState, false);
|
assertTaskResults(taskResultList, result, currentState, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
public void testMarkAsStaleWhenFailingShard() throws Exception {
|
public void testMarkAsStaleWhenFailingShard() throws Exception {
|
||||||
|
@ -251,44 +255,44 @@ public class ShardFailedClusterStateTaskExecutorTests extends ESAllocationTestCa
|
||||||
ClusterState clusterState,
|
ClusterState clusterState,
|
||||||
boolean clusterStateChanged
|
boolean clusterStateChanged
|
||||||
) {
|
) {
|
||||||
Map<ShardStateAction.FailedShardEntry, ClusterStateTaskExecutor.TaskResult> taskResultMap =
|
List<Tuple<FailedShardEntry, ClusterStateTaskExecutor.TaskResult>> taskResultList = tasks.stream()
|
||||||
tasks.stream().collect(Collectors.toMap(Function.identity(), task -> ClusterStateTaskExecutor.TaskResult.success()));
|
.map(t -> Tuple.tuple(t, ClusterStateTaskExecutor.TaskResult.success())).collect(Collectors.toList());
|
||||||
assertTaskResults(taskResultMap, result, clusterState, clusterStateChanged);
|
assertTaskResults(taskResultList, result, clusterState, clusterStateChanged);
|
||||||
}
|
}
|
||||||
|
|
||||||
private static void assertTaskResults(
|
private static void assertTaskResults(
|
||||||
Map<ShardStateAction.FailedShardEntry, ClusterStateTaskExecutor.TaskResult> taskResultMap,
|
List<Tuple<ShardStateAction.FailedShardEntry, ClusterStateTaskExecutor.TaskResult>> taskResultList,
|
||||||
ClusterStateTaskExecutor.ClusterTasksResult<ShardStateAction.FailedShardEntry> result,
|
ClusterStateTaskExecutor.ClusterTasksResult<ShardStateAction.FailedShardEntry> result,
|
||||||
ClusterState clusterState,
|
ClusterState clusterState,
|
||||||
boolean clusterStateChanged
|
boolean clusterStateChanged
|
||||||
) {
|
) {
|
||||||
// there should be as many task results as tasks
|
// there should be as many task results as tasks
|
||||||
assertEquals(taskResultMap.size(), result.executionResults.size());
|
assertEquals(taskResultList.size(), result.executionResults.size());
|
||||||
|
|
||||||
for (Map.Entry<ShardStateAction.FailedShardEntry, ClusterStateTaskExecutor.TaskResult> entry : taskResultMap.entrySet()) {
|
for (Tuple<FailedShardEntry, ClusterStateTaskExecutor.TaskResult> entry : taskResultList) {
|
||||||
// every task should have a corresponding task result
|
// every task should have a corresponding task result
|
||||||
assertTrue(result.executionResults.containsKey(entry.getKey()));
|
assertTrue(result.executionResults.containsKey(entry.v1()));
|
||||||
|
|
||||||
// the task results are as expected
|
// the task results are as expected
|
||||||
assertEquals(entry.getKey().toString(), entry.getValue().isSuccess(), result.executionResults.get(entry.getKey()).isSuccess());
|
assertEquals(entry.v1().toString(), entry.v2().isSuccess(), result.executionResults.get(entry.v1()).isSuccess());
|
||||||
}
|
}
|
||||||
|
|
||||||
List<ShardRouting> shards = clusterState.getRoutingTable().allShards();
|
List<ShardRouting> shards = clusterState.getRoutingTable().allShards();
|
||||||
for (Map.Entry<ShardStateAction.FailedShardEntry, ClusterStateTaskExecutor.TaskResult> entry : taskResultMap.entrySet()) {
|
for (Tuple<FailedShardEntry, ClusterStateTaskExecutor.TaskResult> entry : taskResultList) {
|
||||||
if (entry.getValue().isSuccess()) {
|
if (entry.v2().isSuccess()) {
|
||||||
// the shard was successfully failed and so should not be in the routing table
|
// the shard was successfully failed and so should not be in the routing table
|
||||||
for (ShardRouting shard : shards) {
|
for (ShardRouting shard : shards) {
|
||||||
if (shard.assignedToNode()) {
|
if (shard.assignedToNode()) {
|
||||||
assertFalse("entry key " + entry.getKey() + ", shard routing " + shard,
|
assertFalse("entry key " + entry.v1() + ", shard routing " + shard,
|
||||||
entry.getKey().getShardId().equals(shard.shardId()) &&
|
entry.v1().getShardId().equals(shard.shardId()) &&
|
||||||
entry.getKey().getAllocationId().equals(shard.allocationId().getId()));
|
entry.v1().getAllocationId().equals(shard.allocationId().getId()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// check we saw the expected failure
|
// check we saw the expected failure
|
||||||
ClusterStateTaskExecutor.TaskResult actualResult = result.executionResults.get(entry.getKey());
|
ClusterStateTaskExecutor.TaskResult actualResult = result.executionResults.get(entry.v1());
|
||||||
assertThat(actualResult.getFailure(), instanceOf(entry.getValue().getFailure().getClass()));
|
assertThat(actualResult.getFailure(), instanceOf(entry.v2().getFailure().getClass()));
|
||||||
assertThat(actualResult.getFailure().getMessage(), equalTo(entry.getValue().getFailure().getMessage()));
|
assertThat(actualResult.getFailure().getMessage(), equalTo(entry.v2().getFailure().getMessage()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -59,9 +59,10 @@ import org.junit.Before;
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.UUID;
|
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
|
import java.util.UUID;
|
||||||
import java.util.concurrent.CountDownLatch;
|
import java.util.concurrent.CountDownLatch;
|
||||||
|
import java.util.concurrent.Phaser;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
import java.util.concurrent.atomic.AtomicBoolean;
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
import java.util.concurrent.atomic.AtomicInteger;
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
@ -73,6 +74,8 @@ import static org.elasticsearch.test.ClusterServiceUtils.createClusterService;
|
||||||
import static org.elasticsearch.test.ClusterServiceUtils.setState;
|
import static org.elasticsearch.test.ClusterServiceUtils.setState;
|
||||||
import static org.hamcrest.CoreMatchers.equalTo;
|
import static org.hamcrest.CoreMatchers.equalTo;
|
||||||
import static org.hamcrest.CoreMatchers.instanceOf;
|
import static org.hamcrest.CoreMatchers.instanceOf;
|
||||||
|
import static org.hamcrest.CoreMatchers.sameInstance;
|
||||||
|
import static org.hamcrest.Matchers.arrayWithSize;
|
||||||
import static org.hamcrest.Matchers.greaterThanOrEqualTo;
|
import static org.hamcrest.Matchers.greaterThanOrEqualTo;
|
||||||
import static org.hamcrest.Matchers.is;
|
import static org.hamcrest.Matchers.is;
|
||||||
import static org.hamcrest.Matchers.nullValue;
|
import static org.hamcrest.Matchers.nullValue;
|
||||||
|
@ -138,6 +141,7 @@ public class ShardStateActionTests extends ESTestCase {
|
||||||
clusterService.close();
|
clusterService.close();
|
||||||
transportService.close();
|
transportService.close();
|
||||||
super.tearDown();
|
super.tearDown();
|
||||||
|
assertThat(shardStateAction.remoteShardFailedCacheSize(), equalTo(0));
|
||||||
}
|
}
|
||||||
|
|
||||||
@AfterClass
|
@AfterClass
|
||||||
|
@ -381,6 +385,89 @@ public class ShardStateActionTests extends ESTestCase {
|
||||||
assertThat(failure.get().getMessage(), equalTo(catastrophicError.getMessage()));
|
assertThat(failure.get().getMessage(), equalTo(catastrophicError.getMessage()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void testCacheRemoteShardFailed() throws Exception {
|
||||||
|
final String index = "test";
|
||||||
|
setState(clusterService, ClusterStateCreationUtils.stateWithActivePrimary(index, true, randomInt(5)));
|
||||||
|
ShardRouting failedShard = getRandomShardRouting(index);
|
||||||
|
boolean markAsStale = randomBoolean();
|
||||||
|
int numListeners = between(1, 100);
|
||||||
|
CountDownLatch latch = new CountDownLatch(numListeners);
|
||||||
|
long primaryTerm = randomLongBetween(1, Long.MAX_VALUE);
|
||||||
|
for (int i = 0; i < numListeners; i++) {
|
||||||
|
shardStateAction.remoteShardFailed(failedShard.shardId(), failedShard.allocationId().getId(),
|
||||||
|
primaryTerm, markAsStale, "test", getSimulatedFailure(), new ShardStateAction.Listener() {
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
latch.countDown();
|
||||||
|
}
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception e) {
|
||||||
|
latch.countDown();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
CapturingTransport.CapturedRequest[] capturedRequests = transport.getCapturedRequestsAndClear();
|
||||||
|
assertThat(capturedRequests, arrayWithSize(1));
|
||||||
|
transport.handleResponse(capturedRequests[0].requestId, TransportResponse.Empty.INSTANCE);
|
||||||
|
latch.await();
|
||||||
|
assertThat(transport.capturedRequests(), arrayWithSize(0));
|
||||||
|
}
|
||||||
|
|
||||||
|
public void testRemoteShardFailedConcurrently() throws Exception {
|
||||||
|
final String index = "test";
|
||||||
|
final AtomicBoolean shutdown = new AtomicBoolean(false);
|
||||||
|
setState(clusterService, ClusterStateCreationUtils.stateWithActivePrimary(index, true, randomInt(5)));
|
||||||
|
ShardRouting[] failedShards = new ShardRouting[between(1, 5)];
|
||||||
|
for (int i = 0; i < failedShards.length; i++) {
|
||||||
|
failedShards[i] = getRandomShardRouting(index);
|
||||||
|
}
|
||||||
|
Thread[] clientThreads = new Thread[between(1, 6)];
|
||||||
|
int iterationsPerThread = scaledRandomIntBetween(50, 500);
|
||||||
|
Phaser barrier = new Phaser(clientThreads.length + 2); // one for master thread, one for the main thread
|
||||||
|
Thread masterThread = new Thread(() -> {
|
||||||
|
barrier.arriveAndAwaitAdvance();
|
||||||
|
while (shutdown.get() == false) {
|
||||||
|
for (CapturingTransport.CapturedRequest request : transport.getCapturedRequestsAndClear()) {
|
||||||
|
if (randomBoolean()) {
|
||||||
|
transport.handleResponse(request.requestId, TransportResponse.Empty.INSTANCE);
|
||||||
|
} else {
|
||||||
|
transport.handleRemoteError(request.requestId, randomFrom(getSimulatedFailure()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
masterThread.start();
|
||||||
|
|
||||||
|
AtomicInteger notifiedResponses = new AtomicInteger();
|
||||||
|
for (int t = 0; t < clientThreads.length; t++) {
|
||||||
|
clientThreads[t] = new Thread(() -> {
|
||||||
|
barrier.arriveAndAwaitAdvance();
|
||||||
|
for (int i = 0; i < iterationsPerThread; i++) {
|
||||||
|
ShardRouting failedShard = randomFrom(failedShards);
|
||||||
|
shardStateAction.remoteShardFailed(failedShard.shardId(), failedShard.allocationId().getId(),
|
||||||
|
randomLongBetween(1, Long.MAX_VALUE), randomBoolean(), "test", getSimulatedFailure(), new ShardStateAction.Listener() {
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
notifiedResponses.incrementAndGet();
|
||||||
|
}
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception e) {
|
||||||
|
notifiedResponses.incrementAndGet();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
clientThreads[t].start();
|
||||||
|
}
|
||||||
|
barrier.arriveAndAwaitAdvance();
|
||||||
|
for (Thread t : clientThreads) {
|
||||||
|
t.join();
|
||||||
|
}
|
||||||
|
assertBusy(() -> assertThat(notifiedResponses.get(), equalTo(clientThreads.length * iterationsPerThread)));
|
||||||
|
shutdown.set(true);
|
||||||
|
masterThread.join();
|
||||||
|
}
|
||||||
|
|
||||||
private ShardRouting getRandomShardRouting(String index) {
|
private ShardRouting getRandomShardRouting(String index) {
|
||||||
IndexRoutingTable indexRoutingTable = clusterService.state().routingTable().index(index);
|
IndexRoutingTable indexRoutingTable = clusterService.state().routingTable().index(index);
|
||||||
ShardsIterator shardsIterator = indexRoutingTable.randomAllActiveShardsIt();
|
ShardsIterator shardsIterator = indexRoutingTable.randomAllActiveShardsIt();
|
||||||
|
@ -452,4 +539,61 @@ public class ShardStateActionTests extends ESTestCase {
|
||||||
return out.bytes();
|
return out.bytes();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void testCompositeListener() throws Exception {
|
||||||
|
AtomicInteger successCount = new AtomicInteger();
|
||||||
|
AtomicInteger failureCount = new AtomicInteger();
|
||||||
|
Exception failure = randomBoolean() ? getSimulatedFailure() : null;
|
||||||
|
ShardStateAction.CompositeListener compositeListener = new ShardStateAction.CompositeListener(new ShardStateAction.Listener() {
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
successCount.incrementAndGet();
|
||||||
|
}
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception e) {
|
||||||
|
assertThat(e, sameInstance(failure));
|
||||||
|
failureCount.incrementAndGet();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
int iterationsPerThread = scaledRandomIntBetween(100, 1000);
|
||||||
|
Thread[] threads = new Thread[between(1, 4)];
|
||||||
|
Phaser barrier = new Phaser(threads.length + 1);
|
||||||
|
for (int i = 0; i < threads.length; i++) {
|
||||||
|
threads[i] = new Thread(() -> {
|
||||||
|
barrier.arriveAndAwaitAdvance();
|
||||||
|
for (int n = 0; n < iterationsPerThread; n++) {
|
||||||
|
compositeListener.addListener(new ShardStateAction.Listener() {
|
||||||
|
@Override
|
||||||
|
public void onSuccess() {
|
||||||
|
successCount.incrementAndGet();
|
||||||
|
}
|
||||||
|
@Override
|
||||||
|
public void onFailure(Exception e) {
|
||||||
|
assertThat(e, sameInstance(failure));
|
||||||
|
failureCount.incrementAndGet();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
threads[i].start();
|
||||||
|
}
|
||||||
|
barrier.arriveAndAwaitAdvance();
|
||||||
|
if (failure != null) {
|
||||||
|
compositeListener.onFailure(failure);
|
||||||
|
} else {
|
||||||
|
compositeListener.onSuccess();
|
||||||
|
}
|
||||||
|
for (Thread t : threads) {
|
||||||
|
t.join();
|
||||||
|
}
|
||||||
|
assertBusy(() -> {
|
||||||
|
if (failure != null) {
|
||||||
|
assertThat(successCount.get(), equalTo(0));
|
||||||
|
assertThat(failureCount.get(), equalTo(threads.length*iterationsPerThread + 1));
|
||||||
|
} else {
|
||||||
|
assertThat(successCount.get(), equalTo(threads.length*iterationsPerThread + 1));
|
||||||
|
assertThat(failureCount.get(), equalTo(0));
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue