[Docs] "The the" is a great band, but ... (#26644)
Removing several occurrences of this typo in the docs and javadocs, seems to be a common mistake. Corrections turn up once in a while in PRs, better to correct some of this in one sweep.
This commit is contained in:
parent
b6c6effa2a
commit
c7c6443b10
|
@ -60,7 +60,7 @@ public abstract class DocWriteResponse extends ReplicationResponse implements Wr
|
||||||
private static final String FORCED_REFRESH = "forced_refresh";
|
private static final String FORCED_REFRESH = "forced_refresh";
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* An enum that represents the the results of CRUD operations, primarily used to communicate the type of
|
* An enum that represents the results of CRUD operations, primarily used to communicate the type of
|
||||||
* operation that occurred.
|
* operation that occurred.
|
||||||
*/
|
*/
|
||||||
public enum Result implements Writeable {
|
public enum Result implements Writeable {
|
||||||
|
|
|
@ -110,7 +110,7 @@ public class ListTasksResponse extends BaseTasksResponse implements ToXContentOb
|
||||||
// we found parent in the list of tasks - add it to the parent list
|
// we found parent in the list of tasks - add it to the parent list
|
||||||
parentTask.addGroup(taskGroup);
|
parentTask.addGroup(taskGroup);
|
||||||
} else {
|
} else {
|
||||||
// we got zombie or the parent was filtered out - add it to the the top task list
|
// we got zombie or the parent was filtered out - add it to the top task list
|
||||||
topLevelTasks.add(taskGroup);
|
topLevelTasks.add(taskGroup);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -64,7 +64,7 @@ public final class AutoCreateIndex {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Should the index be auto created?
|
* Should the index be auto created?
|
||||||
* @throws IndexNotFoundException if the the index doesn't exist and shouldn't be auto created
|
* @throws IndexNotFoundException if the index doesn't exist and shouldn't be auto created
|
||||||
*/
|
*/
|
||||||
public boolean shouldAutoCreate(String index, ClusterState state) {
|
public boolean shouldAutoCreate(String index, ClusterState state) {
|
||||||
if (resolver.hasIndexOrAlias(index, state)) {
|
if (resolver.hasIndexOrAlias(index, state)) {
|
||||||
|
|
|
@ -343,7 +343,7 @@ public class ReplicationOperation<
|
||||||
public interface Replicas<RequestT extends ReplicationRequest<RequestT>> {
|
public interface Replicas<RequestT extends ReplicationRequest<RequestT>> {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Performs the the specified request on the specified replica.
|
* Performs the specified request on the specified replica.
|
||||||
*
|
*
|
||||||
* @param replica the shard this request should be executed on
|
* @param replica the shard this request should be executed on
|
||||||
* @param replicaRequest the operation to perform
|
* @param replicaRequest the operation to perform
|
||||||
|
|
|
@ -270,7 +270,7 @@ public class DiscoveryNode implements Writeable, ToXContentFragment {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The unique ephemeral id of the node. Ephemeral ids are meant to be attached the the life span
|
* The unique ephemeral id of the node. Ephemeral ids are meant to be attached the life span
|
||||||
* of a node process. When ever a node is restarted, it's ephemeral id is required to change (while it's {@link #getId()}
|
* of a node process. When ever a node is restarted, it's ephemeral id is required to change (while it's {@link #getId()}
|
||||||
* will be read from the data folder and will remain the same across restarts). Since all node attributes and addresses
|
* will be read from the data folder and will remain the same across restarts). Since all node attributes and addresses
|
||||||
* are maintained during the life span of a node process, we can (and are) using the ephemeralId in
|
* are maintained during the life span of a node process, we can (and are) using the ephemeralId in
|
||||||
|
|
|
@ -44,9 +44,7 @@ import java.util.List;
|
||||||
import java.util.Locale;
|
import java.util.Locale;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.Optional;
|
import java.util.Optional;
|
||||||
import java.util.OptionalDouble;
|
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
import java.util.stream.Collectors;
|
|
||||||
|
|
||||||
import static java.util.Collections.emptyMap;
|
import static java.util.Collections.emptyMap;
|
||||||
|
|
||||||
|
@ -402,7 +400,7 @@ public class IndexShardRoutingTable implements Iterable<ShardRouting> {
|
||||||
@Override
|
@Override
|
||||||
public int compare(ShardRouting s1, ShardRouting s2) {
|
public int compare(ShardRouting s1, ShardRouting s2) {
|
||||||
if (s1.currentNodeId().equals(s2.currentNodeId())) {
|
if (s1.currentNodeId().equals(s2.currentNodeId())) {
|
||||||
// these shards on the the same node
|
// these shards on the same node
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
Double shard1rank = nodeRanks.get(s1.currentNodeId());
|
Double shard1rank = nodeRanks.get(s1.currentNodeId());
|
||||||
|
|
|
@ -39,7 +39,7 @@ public final class XContentParserUtils {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Makes sure that current token is of type {@link XContentParser.Token#FIELD_NAME} and the the field name is equal to the provided one
|
* Makes sure that current token is of type {@link XContentParser.Token#FIELD_NAME} and the field name is equal to the provided one
|
||||||
* @throws ParsingException if the token is not of type {@link XContentParser.Token#FIELD_NAME} or is not equal to the given field name
|
* @throws ParsingException if the token is not of type {@link XContentParser.Token#FIELD_NAME} or is not equal to the given field name
|
||||||
*/
|
*/
|
||||||
public static void ensureFieldName(XContentParser parser, Token token, String fieldName) throws IOException {
|
public static void ensureFieldName(XContentParser parser, Token token, String fieldName) throws IOException {
|
||||||
|
|
|
@ -34,7 +34,7 @@ class VersionValue implements Accountable {
|
||||||
|
|
||||||
/** the seq number of the operation that last changed the associated uuid */
|
/** the seq number of the operation that last changed the associated uuid */
|
||||||
final long seqNo;
|
final long seqNo;
|
||||||
/** the the term of the operation that last changed the associated uuid */
|
/** the term of the operation that last changed the associated uuid */
|
||||||
final long term;
|
final long term;
|
||||||
|
|
||||||
VersionValue(long version, long seqNo, long term) {
|
VersionValue(long version, long seqNo, long term) {
|
||||||
|
|
|
@ -22,7 +22,6 @@ package org.elasticsearch.index.seqno;
|
||||||
import org.elasticsearch.common.io.stream.StreamInput;
|
import org.elasticsearch.common.io.stream.StreamInput;
|
||||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||||
import org.elasticsearch.common.io.stream.Writeable;
|
import org.elasticsearch.common.io.stream.Writeable;
|
||||||
import org.elasticsearch.common.xcontent.ToXContent.Params;
|
|
||||||
import org.elasticsearch.common.xcontent.ToXContentFragment;
|
import org.elasticsearch.common.xcontent.ToXContentFragment;
|
||||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||||
|
|
||||||
|
@ -42,7 +41,7 @@ public class SeqNoStats implements ToXContentFragment, Writeable {
|
||||||
public SeqNoStats(long maxSeqNo, long localCheckpoint, long globalCheckpoint) {
|
public SeqNoStats(long maxSeqNo, long localCheckpoint, long globalCheckpoint) {
|
||||||
assert localCheckpoint <= maxSeqNo:
|
assert localCheckpoint <= maxSeqNo:
|
||||||
"local checkpoint [" + localCheckpoint + "] is above maximum seq no [" + maxSeqNo + "]";
|
"local checkpoint [" + localCheckpoint + "] is above maximum seq no [" + maxSeqNo + "]";
|
||||||
// note that the the global checkpoint can be higher from both maxSeqNo and localCheckpoint
|
// note that the global checkpoint can be higher from both maxSeqNo and localCheckpoint
|
||||||
// as we use this stats object to describe lucene commits as well as live statistic.
|
// as we use this stats object to describe lucene commits as well as live statistic.
|
||||||
this.maxSeqNo = maxSeqNo;
|
this.maxSeqNo = maxSeqNo;
|
||||||
this.localCheckpoint = localCheckpoint;
|
this.localCheckpoint = localCheckpoint;
|
||||||
|
|
|
@ -609,6 +609,7 @@ public class IndexShard extends AbstractIndexShardComponent implements IndicesCl
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public IndexShardState state() {
|
public IndexShardState state() {
|
||||||
return state;
|
return state;
|
||||||
}
|
}
|
||||||
|
@ -1346,6 +1347,7 @@ public class IndexShard extends AbstractIndexShardComponent implements IndicesCl
|
||||||
* Returns the current {@link RecoveryState} if this shard is recovering or has been recovering.
|
* Returns the current {@link RecoveryState} if this shard is recovering or has been recovering.
|
||||||
* Returns null if the recovery has not yet started or shard was not recovered (created via an API).
|
* Returns null if the recovery has not yet started or shard was not recovered (created via an API).
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public RecoveryState recoveryState() {
|
public RecoveryState recoveryState() {
|
||||||
return this.recoveryState;
|
return this.recoveryState;
|
||||||
}
|
}
|
||||||
|
@ -1757,7 +1759,7 @@ public class IndexShard extends AbstractIndexShardComponent implements IndicesCl
|
||||||
* case that the global checkpoint update from the primary is ahead of the local checkpoint on this shard. In this case, we
|
* case that the global checkpoint update from the primary is ahead of the local checkpoint on this shard. In this case, we
|
||||||
* ignore the global checkpoint update. This can happen if we are in the translog stage of recovery. Prior to this, the engine
|
* ignore the global checkpoint update. This can happen if we are in the translog stage of recovery. Prior to this, the engine
|
||||||
* is not opened and this shard will not receive global checkpoint updates, and after this the shard will be contributing to
|
* is not opened and this shard will not receive global checkpoint updates, and after this the shard will be contributing to
|
||||||
* calculations of the the global checkpoint. However, we can not assert that we are in the translog stage of recovery here as
|
* calculations of the global checkpoint. However, we can not assert that we are in the translog stage of recovery here as
|
||||||
* while the global checkpoint update may have emanated from the primary when we were in that state, we could subsequently move
|
* while the global checkpoint update may have emanated from the primary when we were in that state, we could subsequently move
|
||||||
* to recovery finalization, or even finished recovery before the update arrives here.
|
* to recovery finalization, or even finished recovery before the update arrives here.
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -183,7 +183,7 @@ final class IndexShardOperationPermits implements Closeable {
|
||||||
* - blockOperations can be called on a recovery thread which can be expected to be interrupted when recovery is cancelled;
|
* - blockOperations can be called on a recovery thread which can be expected to be interrupted when recovery is cancelled;
|
||||||
* interruptions are bad here as permit acquisition will throw an interrupted exception which will be swallowed by
|
* interruptions are bad here as permit acquisition will throw an interrupted exception which will be swallowed by
|
||||||
* the threaded action listener if the queue of the thread pool on which it submits is full
|
* the threaded action listener if the queue of the thread pool on which it submits is full
|
||||||
* - if a permit is acquired and the queue of the thread pool which the the threaded action listener uses is full, the
|
* - if a permit is acquired and the queue of the thread pool which the threaded action listener uses is full, the
|
||||||
* onFailure handler is executed on the calling thread; this should not be the recovery thread as it would delay the
|
* onFailure handler is executed on the calling thread; this should not be the recovery thread as it would delay the
|
||||||
* recovery
|
* recovery
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -272,7 +272,7 @@ public class ScriptService extends AbstractComponent implements Closeable, Clust
|
||||||
|
|
||||||
if (type == ScriptType.STORED) {
|
if (type == ScriptType.STORED) {
|
||||||
// * lang and options will both be null when looking up a stored script,
|
// * lang and options will both be null when looking up a stored script,
|
||||||
// so we must get the source to retrieve the them before checking if the
|
// so we must get the source to retrieve them before checking if the
|
||||||
// context is supported
|
// context is supported
|
||||||
// * a stored script must be pulled from the cluster state every time in case
|
// * a stored script must be pulled from the cluster state every time in case
|
||||||
// the script has been updated since the last compilation
|
// the script has been updated since the last compilation
|
||||||
|
|
|
@ -144,7 +144,7 @@ public class TransportService extends AbstractLifecycleComponent {
|
||||||
/**
|
/**
|
||||||
* Build the service.
|
* Build the service.
|
||||||
*
|
*
|
||||||
* @param clusterSettings if non null the the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
* @param clusterSettings if non null, the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
||||||
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
||||||
*/
|
*/
|
||||||
public TransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor transportInterceptor,
|
public TransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor transportInterceptor,
|
||||||
|
|
|
@ -20,6 +20,7 @@
|
||||||
package org.elasticsearch.index.engine;
|
package org.elasticsearch.index.engine;
|
||||||
|
|
||||||
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
||||||
|
|
||||||
import org.apache.logging.log4j.Level;
|
import org.apache.logging.log4j.Level;
|
||||||
import org.apache.logging.log4j.LogManager;
|
import org.apache.logging.log4j.LogManager;
|
||||||
import org.apache.logging.log4j.Logger;
|
import org.apache.logging.log4j.Logger;
|
||||||
|
@ -2848,7 +2849,7 @@ public class InternalEngineTests extends ESTestCase {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests that when the the close method returns the engine is actually guaranteed to have cleaned up and that resources are closed
|
* Tests that when the close method returns the engine is actually guaranteed to have cleaned up and that resources are closed
|
||||||
*/
|
*/
|
||||||
public void testConcurrentEngineClosed() throws BrokenBarrierException, InterruptedException {
|
public void testConcurrentEngineClosed() throws BrokenBarrierException, InterruptedException {
|
||||||
Thread[] closingThreads = new Thread[3];
|
Thread[] closingThreads = new Thread[3];
|
||||||
|
|
|
@ -517,7 +517,7 @@ public class GlobalCheckpointTrackerTests extends ESTestCase {
|
||||||
/**
|
/**
|
||||||
* If we do not update the global checkpoint in {@link GlobalCheckpointTracker#markAllocationIdAsInSync(String, long)} after adding the
|
* If we do not update the global checkpoint in {@link GlobalCheckpointTracker#markAllocationIdAsInSync(String, long)} after adding the
|
||||||
* allocation ID to the in-sync set and removing it from pending, the local checkpoint update that freed the thread waiting for the
|
* allocation ID to the in-sync set and removing it from pending, the local checkpoint update that freed the thread waiting for the
|
||||||
* local checkpoint to advance could miss updating the global checkpoint in a race if the the waiting thread did not add the allocation
|
* local checkpoint to advance could miss updating the global checkpoint in a race if the waiting thread did not add the allocation
|
||||||
* ID to the in-sync set and remove it from the pending set before the local checkpoint updating thread executed the global checkpoint
|
* ID to the in-sync set and remove it from the pending set before the local checkpoint updating thread executed the global checkpoint
|
||||||
* update. This test fails without an additional call to {@link GlobalCheckpointTracker#updateGlobalCheckpointOnPrimary()} after
|
* update. This test fails without an additional call to {@link GlobalCheckpointTracker#updateGlobalCheckpointOnPrimary()} after
|
||||||
* removing the allocation ID from the pending set in {@link GlobalCheckpointTracker#markAllocationIdAsInSync(String, long)} (even if a
|
* removing the allocation ID from the pending set in {@link GlobalCheckpointTracker#markAllocationIdAsInSync(String, long)} (even if a
|
||||||
|
|
|
@ -372,7 +372,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested Filter gets rewritten
|
* Test the nested Filter gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedRewrites() throws IOException {
|
public void testNestedRewrites() throws IOException {
|
||||||
FieldSortBuilder sortBuilder = new FieldSortBuilder(MAPPED_STRING_FIELDNAME);
|
FieldSortBuilder sortBuilder = new FieldSortBuilder(MAPPED_STRING_FIELDNAME);
|
||||||
|
@ -389,7 +389,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested sort gets rewritten
|
* Test the nested sort gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedSortRewrites() throws IOException {
|
public void testNestedSortRewrites() throws IOException {
|
||||||
FieldSortBuilder sortBuilder = new FieldSortBuilder(MAPPED_STRING_FIELDNAME);
|
FieldSortBuilder sortBuilder = new FieldSortBuilder(MAPPED_STRING_FIELDNAME);
|
||||||
|
|
|
@ -588,7 +588,7 @@ public class GeoDistanceSortBuilderTests extends AbstractSortTestCase<GeoDistanc
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested Filter gets rewritten
|
* Test the nested Filter gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedRewrites() throws IOException {
|
public void testNestedRewrites() throws IOException {
|
||||||
GeoDistanceSortBuilder sortBuilder = new GeoDistanceSortBuilder("fieldName", 0.0, 0.0);
|
GeoDistanceSortBuilder sortBuilder = new GeoDistanceSortBuilder("fieldName", 0.0, 0.0);
|
||||||
|
@ -605,7 +605,7 @@ public class GeoDistanceSortBuilderTests extends AbstractSortTestCase<GeoDistanc
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested sort gets rewritten
|
* Test the nested sort gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedSortRewrites() throws IOException {
|
public void testNestedSortRewrites() throws IOException {
|
||||||
GeoDistanceSortBuilder sortBuilder = new GeoDistanceSortBuilder("fieldName", 0.0, 0.0);
|
GeoDistanceSortBuilder sortBuilder = new GeoDistanceSortBuilder("fieldName", 0.0, 0.0);
|
||||||
|
|
|
@ -31,11 +31,11 @@ import org.elasticsearch.index.fielddata.IndexFieldData.XFieldComparatorSource.N
|
||||||
import org.elasticsearch.index.fielddata.fieldcomparator.BytesRefFieldComparatorSource;
|
import org.elasticsearch.index.fielddata.fieldcomparator.BytesRefFieldComparatorSource;
|
||||||
import org.elasticsearch.index.fielddata.fieldcomparator.DoubleValuesComparatorSource;
|
import org.elasticsearch.index.fielddata.fieldcomparator.DoubleValuesComparatorSource;
|
||||||
import org.elasticsearch.index.mapper.TypeFieldMapper;
|
import org.elasticsearch.index.mapper.TypeFieldMapper;
|
||||||
import org.elasticsearch.index.query.QueryBuilders;
|
|
||||||
import org.elasticsearch.index.query.QueryShardContext;
|
|
||||||
import org.elasticsearch.index.query.MatchNoneQueryBuilder;
|
import org.elasticsearch.index.query.MatchNoneQueryBuilder;
|
||||||
import org.elasticsearch.index.query.QueryBuilder;
|
import org.elasticsearch.index.query.QueryBuilder;
|
||||||
|
import org.elasticsearch.index.query.QueryBuilders;
|
||||||
import org.elasticsearch.index.query.QueryRewriteContext;
|
import org.elasticsearch.index.query.QueryRewriteContext;
|
||||||
|
import org.elasticsearch.index.query.QueryShardContext;
|
||||||
import org.elasticsearch.index.query.RangeQueryBuilder;
|
import org.elasticsearch.index.query.RangeQueryBuilder;
|
||||||
import org.elasticsearch.script.Script;
|
import org.elasticsearch.script.Script;
|
||||||
import org.elasticsearch.script.ScriptType;
|
import org.elasticsearch.script.ScriptType;
|
||||||
|
@ -360,7 +360,7 @@ public class ScriptSortBuilderTests extends AbstractSortTestCase<ScriptSortBuild
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested Filter gets rewritten
|
* Test the nested Filter gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedRewrites() throws IOException {
|
public void testNestedRewrites() throws IOException {
|
||||||
ScriptSortBuilder sortBuilder = new ScriptSortBuilder(mockScript("something"), ScriptSortType.STRING);
|
ScriptSortBuilder sortBuilder = new ScriptSortBuilder(mockScript("something"), ScriptSortType.STRING);
|
||||||
|
@ -377,7 +377,7 @@ public class ScriptSortBuilderTests extends AbstractSortTestCase<ScriptSortBuild
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the the nested sort gets rewritten
|
* Test the nested sort gets rewritten
|
||||||
*/
|
*/
|
||||||
public void testNestedSortRewrites() throws IOException {
|
public void testNestedSortRewrites() throws IOException {
|
||||||
ScriptSortBuilder sortBuilder = new ScriptSortBuilder(mockScript("something"), ScriptSortType.STRING);
|
ScriptSortBuilder sortBuilder = new ScriptSortBuilder(mockScript("something"), ScriptSortType.STRING);
|
||||||
|
|
|
@ -660,7 +660,7 @@ Note that def types will be assumed to be of the boolean type. Any def type eva
|
||||||
*Examples:*
|
*Examples:*
|
||||||
[source,Java]
|
[source,Java]
|
||||||
----
|
----
|
||||||
boolean x = !false; // declares the boolean variable x and sets it to the the opposite of the false value
|
boolean x = !false; // declares the boolean variable x and sets it to the opposite of the false value
|
||||||
boolean y = !x; // declares the boolean variable y and sets it to the opposite of the boolean variable x
|
boolean y = !x; // declares the boolean variable y and sets it to the opposite of the boolean variable x
|
||||||
def z = !y; // declares the def variable z and sets it to the opposite of the boolean variable y
|
def z = !y; // declares the def variable z and sets it to the opposite of the boolean variable y
|
||||||
----
|
----
|
||||||
|
|
|
@ -239,7 +239,7 @@ The number of documents that were successfully deleted.
|
||||||
|
|
||||||
`batches`::
|
`batches`::
|
||||||
|
|
||||||
The number of scroll responses pulled back by the the delete by query.
|
The number of scroll responses pulled back by the delete by query.
|
||||||
|
|
||||||
`version_conflicts`::
|
`version_conflicts`::
|
||||||
|
|
||||||
|
|
|
@ -616,7 +616,7 @@ The number of documents that were successfully created.
|
||||||
|
|
||||||
`batches`::
|
`batches`::
|
||||||
|
|
||||||
The number of scroll responses pulled back by the the reindex.
|
The number of scroll responses pulled back by the reindex.
|
||||||
|
|
||||||
`version_conflicts`::
|
`version_conflicts`::
|
||||||
|
|
||||||
|
|
|
@ -290,7 +290,7 @@ The number of documents that were successfully updated.
|
||||||
|
|
||||||
`batches`::
|
`batches`::
|
||||||
|
|
||||||
The number of scroll responses pulled back by the the update by query.
|
The number of scroll responses pulled back by the update by query.
|
||||||
|
|
||||||
`version_conflicts`::
|
`version_conflicts`::
|
||||||
|
|
||||||
|
|
|
@ -2012,7 +2012,7 @@ into:
|
||||||
// NOTCONSOLE
|
// NOTCONSOLE
|
||||||
|
|
||||||
If there is already a `bar` field nested under `foo` then
|
If there is already a `bar` field nested under `foo` then
|
||||||
this processor merges the the `foo.bar` field into it. If the field is
|
this processor merges the `foo.bar` field into it. If the field is
|
||||||
a scalar value then it will turn that field into an array field.
|
a scalar value then it will turn that field into an array field.
|
||||||
|
|
||||||
For example, the following document:
|
For example, the following document:
|
||||||
|
@ -2054,7 +2054,7 @@ Consider the following document:
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
// NOTCONSOLE
|
// NOTCONSOLE
|
||||||
|
|
||||||
Then the the `foo` needs to be renamed first before the `dot_expander`
|
Then the `foo` needs to be renamed first before the `dot_expander`
|
||||||
processor is applied. So in order for the `foo.bar` field to properly
|
processor is applied. So in order for the `foo.bar` field to properly
|
||||||
be expanded into the `bar` field under the `foo` field the following
|
be expanded into the `bar` field under the `foo` field the following
|
||||||
pipeline should be used:
|
pipeline should be used:
|
||||||
|
|
|
@ -21,7 +21,7 @@ include-tagged::{plugins-examples-dir}/script-expert-scoring/src/main/java/org/e
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
|
||||||
You can execute the script by specifying its `lang` as `expert_scripts`, and the name
|
You can execute the script by specifying its `lang` as `expert_scripts`, and the name
|
||||||
of the script as the the script source:
|
of the script as the script source:
|
||||||
|
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
|
|
|
@ -291,7 +291,7 @@ GET /my-index/_search
|
||||||
|
|
||||||
<1> The version is optional, but useful in certain cases. We can ensure that we are trying to percolate
|
<1> The version is optional, but useful in certain cases. We can ensure that we are trying to percolate
|
||||||
the document we just have indexed. A change may be made after we have indexed, and if that is the
|
the document we just have indexed. A change may be made after we have indexed, and if that is the
|
||||||
case the then the search request would fail with a version conflict error.
|
case the search request would fail with a version conflict error.
|
||||||
|
|
||||||
The search response returned is identical as in the previous example.
|
The search response returned is identical as in the previous example.
|
||||||
|
|
||||||
|
|
|
@ -9,7 +9,7 @@ To upgrade using a zip or compressed tarball:
|
||||||
.. Extract the zip or tarball to a _new_ directory. This is critical if you
|
.. Extract the zip or tarball to a _new_ directory. This is critical if you
|
||||||
are not using external `config` and `data` directories.
|
are not using external `config` and `data` directories.
|
||||||
|
|
||||||
.. Set the the `ES_PATH_CONF` environment variable to specify the location of
|
.. Set the `ES_PATH_CONF` environment variable to specify the location of
|
||||||
your external `config` directory and `jvm.options` file. If you are not
|
your external `config` directory and `jvm.options` file. If you are not
|
||||||
using an external `config` directory, copy your old configuration
|
using an external `config` directory, copy your old configuration
|
||||||
over to the new installation.
|
over to the new installation.
|
||||||
|
|
|
@ -110,7 +110,7 @@ public class TribeIntegrationTests extends ESIntegTestCase {
|
||||||
private static final Predicate<InternalTestCluster> CLUSTER2_ONLY = c -> c.getClusterName().equals(cluster2.getClusterName());
|
private static final Predicate<InternalTestCluster> CLUSTER2_ONLY = c -> c.getClusterName().equals(cluster2.getClusterName());
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A predicate that is used to select the the two remote clusters
|
* A predicate that is used to select the two remote clusters
|
||||||
**/
|
**/
|
||||||
private static final Predicate<InternalTestCluster> ALL = c -> true;
|
private static final Predicate<InternalTestCluster> ALL = c -> true;
|
||||||
|
|
||||||
|
|
|
@ -65,7 +65,7 @@ public interface GceInstancesService {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* cloud.gce.max_wait: How long exponential backoff should retry before definitely failing.
|
* cloud.gce.max_wait: How long exponential backoff should retry before definitely failing.
|
||||||
* It's a total time since the the initial call is made.
|
* It's a total time since the initial call is made.
|
||||||
* A negative value will retry indefinitely. Defaults to `-1s` (retry indefinitely).
|
* A negative value will retry indefinitely. Defaults to `-1s` (retry indefinitely).
|
||||||
*/
|
*/
|
||||||
Setting<TimeValue> MAX_WAIT_SETTING =
|
Setting<TimeValue> MAX_WAIT_SETTING =
|
||||||
|
|
|
@ -231,7 +231,7 @@ public class GeoIpProcessorFactoryTests extends ESTestCase {
|
||||||
Files.copy(new ByteArrayInputStream(StreamsUtils.copyToBytesFromClasspath("/GeoLite2-Country.mmdb.gz")),
|
Files.copy(new ByteArrayInputStream(StreamsUtils.copyToBytesFromClasspath("/GeoLite2-Country.mmdb.gz")),
|
||||||
geoIpConfigDir.resolve("GeoLite2-Country.mmdb.gz"));
|
geoIpConfigDir.resolve("GeoLite2-Country.mmdb.gz"));
|
||||||
|
|
||||||
// Loading another database reader instances, because otherwise we can't test lazy loading as the the
|
// Loading another database reader instances, because otherwise we can't test lazy loading as the
|
||||||
// database readers used at class level are reused between tests. (we want to keep that otherwise running this
|
// database readers used at class level are reused between tests. (we want to keep that otherwise running this
|
||||||
// test will take roughly 4 times more time)
|
// test will take roughly 4 times more time)
|
||||||
Map<String, DatabaseReaderLazyLoader> databaseReaders =
|
Map<String, DatabaseReaderLazyLoader> databaseReaders =
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
# This file contains some utilities to test the the .deb/.rpm
|
# This file contains some utilities to test the .deb/.rpm
|
||||||
# packages and the SysV/Systemd scripts.
|
# packages and the SysV/Systemd scripts.
|
||||||
|
|
||||||
# WARNING: This testing file must be executed as root and can
|
# WARNING: This testing file must be executed as root and can
|
||||||
|
|
|
@ -20,6 +20,7 @@
|
||||||
package org.elasticsearch.test.transport;
|
package org.elasticsearch.test.transport;
|
||||||
|
|
||||||
import com.carrotsearch.randomizedtesting.SysGlobals;
|
import com.carrotsearch.randomizedtesting.SysGlobals;
|
||||||
|
|
||||||
import org.elasticsearch.Version;
|
import org.elasticsearch.Version;
|
||||||
import org.elasticsearch.cluster.ClusterModule;
|
import org.elasticsearch.cluster.ClusterModule;
|
||||||
import org.elasticsearch.cluster.node.DiscoveryNode;
|
import org.elasticsearch.cluster.node.DiscoveryNode;
|
||||||
|
@ -126,7 +127,7 @@ public final class MockTransportService extends TransportService {
|
||||||
/**
|
/**
|
||||||
* Build the service.
|
* Build the service.
|
||||||
*
|
*
|
||||||
* @param clusterSettings if non null the the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
* @param clusterSettings if non null the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
||||||
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
||||||
*/
|
*/
|
||||||
public MockTransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor interceptor,
|
public MockTransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor interceptor,
|
||||||
|
@ -139,7 +140,7 @@ public final class MockTransportService extends TransportService {
|
||||||
/**
|
/**
|
||||||
* Build the service.
|
* Build the service.
|
||||||
*
|
*
|
||||||
* @param clusterSettings if non null the the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
* @param clusterSettings if non null the {@linkplain TransportService} will register with the {@link ClusterSettings} for settings
|
||||||
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
* updates for {@link #TRACE_LOG_EXCLUDE_SETTING} and {@link #TRACE_LOG_INCLUDE_SETTING}.
|
||||||
*/
|
*/
|
||||||
public MockTransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor interceptor,
|
public MockTransportService(Settings settings, Transport transport, ThreadPool threadPool, TransportInterceptor interceptor,
|
||||||
|
@ -345,7 +346,7 @@ public final class MockTransportService extends TransportService {
|
||||||
final long startTime = System.currentTimeMillis();
|
final long startTime = System.currentTimeMillis();
|
||||||
|
|
||||||
addDelegate(transportAddress, new ClearableTransport(original) {
|
addDelegate(transportAddress, new ClearableTransport(original) {
|
||||||
private final Queue<Runnable> requestsToSendWhenCleared = new LinkedBlockingDeque<Runnable>();
|
private final Queue<Runnable> requestsToSendWhenCleared = new LinkedBlockingDeque<>();
|
||||||
private boolean cleared = false;
|
private boolean cleared = false;
|
||||||
|
|
||||||
TimeValue getDelay() {
|
TimeValue getDelay() {
|
||||||
|
|
Loading…
Reference in New Issue