Fix a variety of typos and misspelled words (#32792)

This commit is contained in:
Kazuhiro Sera 2018-10-04 02:11:39 +09:00 committed by David Turner
parent ee21067a41
commit d45fe43a68
101 changed files with 153 additions and 153 deletions

View File

@ -320,7 +320,7 @@ have to test Elasticsearch.
#### Configurations
Gradle organizes dependencies and build artifacts into "configurations" and
allows you to use these configurations arbitrarilly. Here are some of the most
allows you to use these configurations arbitrarily. Here are some of the most
common configurations in our build and how we use them:
<dl>

View File

@ -250,7 +250,7 @@ Pass arbitrary jvm arguments.
Running backwards compatibility tests is disabled by default since it
requires a release version of elasticsearch to be present on the test system.
To run backwards compatibilty tests untar or unzip a release and run the tests
To run backwards compatibility tests untar or unzip a release and run the tests
with the following command:
---------------------------------------------------------------------------

View File

@ -122,7 +122,7 @@ class VersionCollection {
if (isReleased(version) == false) {
// caveat 1 - This should only ever contain 2 non released branches in flight. An example is 6.x is frozen,
// and 6.2 is cut but not yet released there is some simple logic to make sure that in the case of more than 2,
// it will bail. The order is that the minor snapshot is fufilled first, and then the staged minor snapshot
// it will bail. The order is that the minor snapshot is fulfilled first, and then the staged minor snapshot
if (nextMinorSnapshot == null) {
// it has not been set yet
nextMinorSnapshot = replaceAsSnapshot(version)

View File

@ -72,7 +72,7 @@ public class RestTestsFromSnippetsTask extends SnippetsTask {
/**
* Root directory containing all the files generated by this task. It is
* contained withing testRoot.
* contained within testRoot.
*/
File outputRoot() {
return new File(testRoot, '/rest-api-spec/test')

View File

@ -337,7 +337,7 @@ class NodeInfo {
case 'deb':
return new File(baseDir, "${distro}-extracted/etc/elasticsearch")
default:
throw new InvalidUserDataException("Unkown distribution: ${distro}")
throw new InvalidUserDataException("Unknown distribution: ${distro}")
}
}
}

View File

@ -22,7 +22,7 @@ task sample {
// dependsOn buildResources.outputDir
// for now it's just
dependsOn buildResources
// we have to refference it at configuration time in order to be picked up
// we have to reference it at configuration time in order to be picked up
ext.checkstyle_suppressions = buildResources.copy('checkstyle_suppressions.xml')
doLast {
println "This task is using ${file(checkstyle_suppressions)}"

View File

@ -215,7 +215,7 @@ public class Detector implements ToXContentObject {
}
/**
* Excludes frequently-occuring metrics from the analysis;
* Excludes frequently-occurring metrics from the analysis;
* can apply to 'by' field, 'over' field, or both
*
* @return the value that the user set

View File

@ -228,7 +228,7 @@ final class JvmOptionsParser {
// no range is present, apply the JVM option to the specified major version only
upper = lower;
} else if (end == null) {
// a range of the form \\d+- is present, apply the JVM option to all major versions larger than the specifed one
// a range of the form \\d+- is present, apply the JVM option to all major versions larger than the specified one
upper = Integer.MAX_VALUE;
} else {
// a range of the form \\d+-\\d+ is present, apply the JVM option to the specified range of major versions

View File

@ -307,7 +307,7 @@ You can also customize the response consumer used to buffer the asynchronous
responses. The default consumer will buffer up to 100MB of response on the
JVM heap. If the response is larger then the request will fail. You could,
for example, lower the maximum size which might be useful if you are running
in a heap constrained environment like the exmaple above.
in a heap constrained environment like the example above.
Once you've created the singleton you can use it when making requests:

View File

@ -3,7 +3,7 @@
Painless has a strict whitelist for methods and classes to ensure all
painless scripts are secure. Most of these methods are exposed directly
from the Java Runtime Enviroment (JRE) while others are part of
from the Java Runtime Environment (JRE) while others are part of
Elasticsearch or Painless itself. Below is a list of all available
classes grouped with their respected methods. Clicking on the method
name takes you to the documentation for that specific method. Methods

View File

@ -32,7 +32,7 @@ PUT _snapshot/my_hdfs_repository
"type": "hdfs",
"settings": {
"uri": "hdfs://namenode:8020/",
"path": "elasticsearch/respositories/my_hdfs_repository",
"path": "elasticsearch/repositories/my_hdfs_repository",
"conf.dfs.client.read.shortcircuit": "true"
}
}
@ -149,7 +149,7 @@ PUT _snapshot/my_hdfs_repository
"type": "hdfs",
"settings": {
"uri": "hdfs://namenode:8020/",
"path": "/user/elasticsearch/respositories/my_hdfs_repository",
"path": "/user/elasticsearch/repositories/my_hdfs_repository",
"security.principal": "elasticsearch@REALM"
}
}
@ -167,7 +167,7 @@ PUT _snapshot/my_hdfs_repository
"type": "hdfs",
"settings": {
"uri": "hdfs://namenode:8020/",
"path": "/user/elasticsearch/respositories/my_hdfs_repository",
"path": "/user/elasticsearch/repositories/my_hdfs_repository",
"security.principal": "elasticsearch/_HOST@REALM"
}
}

View File

@ -72,7 +72,7 @@ parameter or in the `filename` field in an input YAML file.
You can optionally provide IP addresses or DNS names for each instance. If
neither IP addresses nor DNS names are specified, the Elastic stack products
cannot perform hostname verification and you might need to configure the
`verfication_mode` security setting to `certificate` only. For more information
`verification_mode` security setting to `certificate` only. For more information
about this setting, see <<security-settings>>.
All certificates that are generated by this command are signed by a CA. You can

View File

@ -207,7 +207,7 @@ The `forecasts_stats` object shows statistics about forecasts. It has the follow
(object) Counts per forecast status, for example: {"finished" : 2}.
NOTE: `memory_bytes`, `records`, `processing_time_ms` and `status` require at least 1 forecast, otherwise
these fields are ommitted.
these fields are omitted.
[float]
[[ml-stats-node]]

View File

@ -45,7 +45,7 @@ For more information, see
==== Examples
Imagine we have an index named `sensor-1` full of raw data. We know that the data will grow over time, so there
will be a `sensor-2`, `sensor-3`, etc. Let's create a Rollup job that targets the index pattern `sensor-*` to accomodate
will be a `sensor-2`, `sensor-3`, etc. Let's create a Rollup job that targets the index pattern `sensor-*` to accommodate
this future scaling:
[source,js]

View File

@ -44,7 +44,7 @@ from `artifacts.elastic.co/maven` by adding it to the repositories list:
=== Setup
The driver main class is `org.elasticsearch.xpack.sql.jdbc.jdbc.JdbcDriver`.
Note the driver implements the JDBC 4.0 +Service Provider+ mechanism meaning it is registerd automatically
Note the driver implements the JDBC 4.0 +Service Provider+ mechanism meaning it is registered automatically
as long as its available in the classpath.
Once registered, the driver understands the following syntax as an URL:

View File

@ -143,7 +143,7 @@
* described by later documentation.
* <p>
* Storebable nodes have three methods for writing -- setup, load, and store. These methods
* are used in conjuction with a parent node aware of the storeable node (lhs) that has a node
* are used in conjunction with a parent node aware of the storeable node (lhs) that has a node
* representing a value to store (rhs). The setup method is always once called before a store
* to give storeable nodes a chance to write any prefixes they may have and any values such as
* array indices before the store happens. Load is called on a storeable node that must also
@ -152,7 +152,7 @@
* Sub nodes are partial nodes that require a parent to work correctly. These nodes can really
* represent anything the parent node would like to split up into logical pieces and don't really
* have any distinct set of rules. The currently existing subnodes all have ANode as a super class
* somewhere in their class heirachy so the parent node can defer some analysis and writing to
* somewhere in their class hierarchy so the parent node can defer some analysis and writing to
* the sub node.
*/
package org.elasticsearch.painless.node;

View File

@ -434,7 +434,7 @@ public class PainlessDocGenerator {
if (classPackage.startsWith("org.apache.lucene")) {
return "lucene-core";
}
throw new IllegalArgumentException("Unrecognized packge: " + classPackage);
throw new IllegalArgumentException("Unrecognized package: " + classPackage);
}
private static void emitGeneratedWarning(PrintStream stream) {

View File

@ -83,7 +83,7 @@ public class PainlessExecuteRequestTests extends AbstractStreamableXContentTestC
QueryBuilder query = randomBoolean() ? new MatchAllQueryBuilder() : null;
// TODO: pass down XContextType to createTestInstance() method.
// otherwise the document itself is different causing test failures.
// This should be done in a seperate change as the test instance is created before xcontent type is randomly picked and
// This should be done in a separate change as the test instance is created before xcontent type is randomly picked and
// all the createTestInstance() methods need to be changed, which will make this a big chnage
// BytesReference doc = randomBoolean() ? new BytesArray("{}") : null;
BytesReference doc = null;

View File

@ -42,8 +42,8 @@ import java.util.stream.Collectors;
/**
* Returns the results for a {@link RankEvalRequest}.<br>
* The repsonse contains a detailed section for each evaluation query in the request and
* possible failures that happened when executin individual queries.
* The response contains a detailed section for each evaluation query in the request and
* possible failures that happened when execution individual queries.
**/
public class RankEvalResponse extends ActionResponse implements ToXContentObject {

View File

@ -481,7 +481,7 @@ public class AsyncBulkByScrollActionTests extends ESTestCase {
/**
* Execute a bulk retry test case. The total number of failures is random and the number of retries attempted is set to
* testRequest.getMaxRetries and controled by the failWithRejection parameter.
* testRequest.getMaxRetries and controlled by the failWithRejection parameter.
*/
private void bulkRetryTestCase(boolean failWithRejection) throws Exception {
int totalFailures = randomIntBetween(1, testRequest.getMaxRetries());

View File

@ -122,7 +122,7 @@ public class CancelTests extends ReindexTestCase {
logger.debug("waiting for updates to be blocked");
boolean blocked = awaitBusy(
() -> ALLOWED_OPERATIONS.hasQueuedThreads() && ALLOWED_OPERATIONS.availablePermits() == 0,
1, TimeUnit.MINUTES); // 10 seconds is usually fine but on heavilly loaded machines this can wake a while
1, TimeUnit.MINUTES); // 10 seconds is usually fine but on heavily loaded machines this can take a while
assertTrue("updates blocked", blocked);
// Status should show the task running

View File

@ -60,7 +60,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
String index = "foo";
String type = "mytype";
String[] equilavent = {"I WİLL USE TURKİSH CASING", "ı will use turkish casıng"};
String[] equivalent = {"I WİLL USE TURKİSH CASING", "ı will use turkish casıng"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -75,8 +75,8 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
// both values should collate to same value
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equilavent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[1] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equivalent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[1] + "\"}", XContentType.JSON)
);
// searching for either of the terms should return both results since they collate to the same value
@ -85,7 +85,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
.types(type)
.source(new SearchSourceBuilder()
.fetchSource(false)
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equilavent[0] : equilavent[1]))
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equivalent[0] : equivalent[1]))
.sort("collate")
.sort("_id", SortOrder.DESC) // secondary sort should kick in because both will collate to same value
);
@ -100,7 +100,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
String index = "foo";
String type = "mytype";
String[] equilavent = {"a", "C", "a", "B"};
String[] equivalent = {"a", "C", "a", "B"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -114,9 +114,9 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
// everything should be indexed fine, no exceptions
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":[\"" + equilavent[0] + "\", \""
+ equilavent[1] + "\"]}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[2] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":[\"" + equivalent[0] + "\", \""
+ equivalent[1] + "\"]}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[2] + "\"}", XContentType.JSON)
);
// using sort mode = max, values B and C will be used for the sort
@ -161,7 +161,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
String index = "foo";
String type = "mytype";
String[] equilavent = {"I W\u0049\u0307LL USE TURKİSH CASING", "ı will use turkish casıng"};
String[] equivalent = {"I W\u0049\u0307LL USE TURKİSH CASING", "ı will use turkish casıng"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -176,8 +176,8 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
assertAcked(client().admin().indices().prepareCreate(index).addMapping(type, builder));
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equilavent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[1] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equivalent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[1] + "\"}", XContentType.JSON)
);
// searching for either of the terms should return both results since they collate to the same value
@ -186,7 +186,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
.types(type)
.source(new SearchSourceBuilder()
.fetchSource(false)
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equilavent[0] : equilavent[1]))
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equivalent[0] : equivalent[1]))
.sort("collate")
.sort("_id", SortOrder.DESC) // secondary sort should kick in because both will collate to same value
);
@ -204,7 +204,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
String index = "foo";
String type = "mytype";
String[] equilavent = {"TESTING", "testing"};
String[] equivalent = {"TESTING", "testing"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -219,8 +219,8 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
assertAcked(client().admin().indices().prepareCreate(index).addMapping(type, builder));
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equilavent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[1] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equivalent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[1] + "\"}", XContentType.JSON)
);
SearchRequest request = new SearchRequest()
@ -228,7 +228,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
.types(type)
.source(new SearchSourceBuilder()
.fetchSource(false)
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equilavent[0] : equilavent[1]))
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equivalent[0] : equivalent[1]))
.sort("collate")
.sort("_id", SortOrder.DESC) // secondary sort should kick in because both will collate to same value
);
@ -247,7 +247,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
String index = "foo";
String type = "mytype";
String[] equilavent = {"foo-bar", "foo bar"};
String[] equivalent = {"foo-bar", "foo bar"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -262,8 +262,8 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
assertAcked(client().admin().indices().prepareCreate(index).addMapping(type, builder));
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equilavent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[1] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equivalent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[1] + "\"}", XContentType.JSON)
);
SearchRequest request = new SearchRequest()
@ -271,7 +271,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
.types(type)
.source(new SearchSourceBuilder()
.fetchSource(false)
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equilavent[0] : equilavent[1]))
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equivalent[0] : equivalent[1]))
.sort("collate")
.sort("_id", SortOrder.DESC) // secondary sort should kick in because both will collate to same value
);
@ -467,7 +467,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
RuleBasedCollator tailoredCollator = new RuleBasedCollator(baseCollator.getRules() + DIN5007_2_tailorings);
String tailoredRules = tailoredCollator.getRules();
String[] equilavent = {"Töne", "Toene"};
String[] equivalent = {"Töne", "Toene"};
XContentBuilder builder = jsonBuilder()
.startObject().startObject("properties")
@ -481,8 +481,8 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
assertAcked(client().admin().indices().prepareCreate(index).addMapping(type, builder));
indexRandom(true,
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equilavent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equilavent[1] + "\"}", XContentType.JSON)
client().prepareIndex(index, type, "1").setSource("{\"collate\":\"" + equivalent[0] + "\"}", XContentType.JSON),
client().prepareIndex(index, type, "2").setSource("{\"collate\":\"" + equivalent[1] + "\"}", XContentType.JSON)
);
SearchRequest request = new SearchRequest()
@ -490,7 +490,7 @@ public class ICUCollationKeywordFieldMapperIT extends ESIntegTestCase {
.types(type)
.source(new SearchSourceBuilder()
.fetchSource(false)
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equilavent[0] : equilavent[1]))
.query(QueryBuilders.termQuery("collate", randomBoolean() ? equivalent[0] : equivalent[1]))
.sort("collate", SortOrder.ASC)
.sort("_id", SortOrder.DESC) // secondary sort should kick in because both will collate to same value
);

View File

@ -1,7 +1,7 @@
# tests that the custom suggester works
# the issue that prompted serializing Suggestion as a registered named writeable was not revealed until
# a user found that it would fail when reducing suggestions in a multi node envrionment
# a user found that it would fail when reducing suggestions in a multi node environment
# https://github.com/elastic/elasticsearch/issues/26585
"test custom suggester":
- do:

View File

@ -748,7 +748,7 @@ os_parsers:
# possibility of false positive when different marketing names share same NT kernel
# e.g. windows server 2003 and windows xp
# lots of ua strings have Windows NT 4.1 !?!?!?!? !?!? !? !????!?! !!! ??? !?!?! ?
# (very) roughly ordered in terms of frequency of occurence of regex (win xp currently most frequent, etc)
# (very) roughly ordered in terms of frequency of occurrence of regex (win xp currently most frequent, etc)
##########
# ie mobile desktop mode
@ -2848,7 +2848,7 @@ device_parsers:
device_replacement: 'Micromax $1'
brand_replacement: 'Micromax'
model_replacement: '$1'
# be carefull here with Acer e.g. A500
# be careful here with Acer e.g. A500
- regex: '; *(A\d{2}|A[12]\d{2}|A90S|A110Q) Build'
regex_flag: 'i'
device_replacement: 'Micromax $1'

View File

@ -1,5 +1,5 @@
---
"Create things in the cluster state that we'll validate are there after the ugprade":
"Create things in the cluster state that we'll validate are there after the upgrade":
- do:
snapshot.create_repository:
repository: my_repo

View File

@ -190,7 +190,7 @@ setup() {
@test "[SYSTEMD] start Elasticsearch with custom JVM options" {
assert_file_exist $ESENVFILE
# The custom config directory is not under /tmp or /var/tmp because
# systemd's private temp directory functionaly means different
# systemd's private temp directory functionally means different
# processes can have different views of what's in these directories
local temp=`mktemp -p /etc -d`
cp "$ESCONFIG"/elasticsearch.yml "$temp"

View File

@ -97,7 +97,7 @@ fi
rm -rf "$ESPLUGINS"
# The custom plugins directory is not under /tmp or /var/tmp because
# systemd's private temp directory functionaly means different
# systemd's private temp directory functionally means different
# processes can have different views of what's in these directories
local es_plugins=$(mktemp -p /var -d -t 'plugins.XXXX')
chown -R elasticsearch:elasticsearch "$es_plugins"

View File

@ -556,7 +556,7 @@ run_elasticsearch_tests() {
move_config() {
local oldConfig="$ESCONFIG"
# The custom config directory is not under /tmp or /var/tmp because
# systemd's private temp directory functionaly means different
# systemd's private temp directory functionally means different
# processes can have different views of what's in these directories
export ESCONFIG="${1:-$(mktemp -p /etc -d -t 'config.XXXX')}"
echo "Moving configuration directory from $oldConfig to $ESCONFIG"

View File

@ -328,7 +328,7 @@ public final class SearchPhaseController extends AbstractComponent {
SearchPhaseResult searchResultProvider = resultsLookup.apply(shardDoc.shardIndex);
if (searchResultProvider == null) {
// this can happen if we are hitting a shard failure during the fetch phase
// in this case we referenced the shard result via teh ScoreDoc but never got a
// in this case we referenced the shard result via the ScoreDoc but never got a
// result from fetch.
// TODO it would be nice to assert this in the future
continue;
@ -380,7 +380,7 @@ public final class SearchPhaseController extends AbstractComponent {
SearchPhaseResult fetchResultProvider = resultsLookup.apply(shardDoc.shardIndex);
if (fetchResultProvider == null) {
// this can happen if we are hitting a shard failure during the fetch phase
// in this case we referenced the shard result via teh ScoreDoc but never got a
// in this case we referenced the shard result via the ScoreDoc but never got a
// result from fetch.
// TODO it would be nice to assert this in the future
continue;

View File

@ -84,7 +84,7 @@ final class Natives {
*/
static String getShortPathName(final String path) {
if (!JNA_AVAILABLE) {
logger.warn("cannot obtain short path for [{}] because JNA is not avilable", path);
logger.warn("cannot obtain short path for [{}] because JNA is not available", path);
return path;
}
return JNANatives.getShortPathName(path);
@ -123,7 +123,7 @@ final class Natives {
static void trySetMaxSizeVirtualMemory() {
if (!JNA_AVAILABLE) {
logger.warn("cannot getrlimit RLIMIT_AS beacuse JNA is not available");
logger.warn("cannot getrlimit RLIMIT_AS because JNA is not available");
return;
}
JNANatives.trySetMaxSizeVirtualMemory();

View File

@ -367,7 +367,7 @@ public class IndexNameExpressionResolver extends AbstractComponent {
}
aliases.add(alias);
} else {
// If not, we have a non required alias for this index - no futher checking needed
// If not, we have a non required alias for this index - no further checking needed
return null;
}
}

View File

@ -135,11 +135,11 @@ public class MetaDataIndexUpgradeService extends AbstractComponent {
private void checkMappingsCompatibility(IndexMetaData indexMetaData) {
try {
// We cannot instantiate real analysis server or similiarity service at this point because the node
// We cannot instantiate real analysis server or similarity service at this point because the node
// might not have been started yet. However, we don't really need real analyzers or similarities at
// this stage - so we can fake it using constant maps accepting every key.
// This is ok because all used similarities and analyzers for this index were known before the upgrade.
// Missing analyzers and similarities plugin will still trigger the apropriate error during the
// Missing analyzers and similarities plugin will still trigger the appropriate error during the
// actual upgrade.
IndexSettings indexSettings = new IndexSettings(indexMetaData, this.settings);

View File

@ -200,7 +200,7 @@ public class TemplateUpgradeService extends AbstractComponent implements Cluster
if (anyUpgradeFailed.get()) {
logger.info("Templates were partially upgraded to version {}", Version.CURRENT);
} else {
logger.info("Templates were upgraded successfuly to version {}", Version.CURRENT);
logger.info("Templates were upgraded successfully to version {}", Version.CURRENT);
}
// Check upgraders are satisfied after the update completed. If they still
// report that changes are required, this might indicate a bug or that something

View File

@ -265,7 +265,7 @@ public class IndexShardRoutingTable implements Iterable<ShardRouting> {
/**
* Returns an iterator over active and initializing shards, ordered by the adaptive replica
* selection forumla. Making sure though that its random within the active shards of the same
* selection formula. Making sure though that its random within the active shards of the same
* (or missing) rank, and initializing shards are the last to iterate through.
*/
public ShardIterator activeInitializingShardsRankedIt(@Nullable ResponseCollectorService collector,

View File

@ -51,7 +51,7 @@ public interface DateMathParser {
*
* Examples are
*
* <code>2014-11-18||-2y</code> substracts two years from the input date
* <code>2014-11-18||-2y</code> subtracts two years from the input date
* <code>now/m</code> rounds the current time to minute granularity
*
* Supported rounding units are

View File

@ -41,7 +41,7 @@ public class PreConfiguredCharFilter extends PreConfiguredAnalysisComponent<Char
}
/**
* Create a pre-configured char filter that may not vary at all, provide access to the elasticsearch verison
* Create a pre-configured char filter that may not vary at all, provide access to the elasticsearch version
*/
public static PreConfiguredCharFilter singletonWithVersion(String name, boolean useFilterForMultitermQueries,
BiFunction<Reader, org.elasticsearch.Version, Reader> create) {

View File

@ -150,7 +150,7 @@ public abstract class AbstractBulkByScrollRequest<Self extends AbstractBulkByScr
e = addValidationError("stored_fields is not supported in this context", e);
}
if (maxRetries < 0) {
e = addValidationError("retries cannnot be negative", e);
e = addValidationError("retries cannot be negative", e);
}
if (false == (size == -1 || size > 0)) {
e = addValidationError(

View File

@ -791,7 +791,7 @@ public class IndexShard extends AbstractIndexShardComponent implements IndicesCl
+ "]";
ensureWriteAllowed(origin);
// When there is a single type, the unique identifier is only composed of the _id,
// so there is no way to differenciate foo#1 from bar#1. This is especially an issue
// so there is no way to differentiate foo#1 from bar#1. This is especially an issue
// if a user first deletes foo#1 and then indexes bar#1: since we do not encode the
// _type in the uid it might look like we are reindexing the same document, which
// would fail if bar#1 is indexed with a lower version than foo#1 was deleted with.

View File

@ -1047,7 +1047,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref
}
/**
* returns the history uuid the store points at, or null if not existant.
* returns the history uuid the store points at, or null if nonexistent.
*/
public String getHistoryUUID() {
return commitUserData.get(Engine.HISTORY_UUID_KEY);

View File

@ -85,12 +85,12 @@ public class TranslogStats implements Streamable, ToXContentFragment {
return numberOfOperations;
}
/** the size of the generations in the translog that weren't yet to comitted to lucene */
/** the size of the generations in the translog that weren't yet to committed to lucene */
public long getUncommittedSizeInBytes() {
return uncommittedSizeInBytes;
}
/** the number of operations in generations of the translog that weren't yet to comitted to lucene */
/** the number of operations in generations of the translog that weren't yet to committed to lucene */
public int getUncommittedOperations() {
return uncommittedOperations;
}

View File

@ -50,7 +50,7 @@ public class FilterAggregatorFactory extends AggregatorFactory<FilterAggregatorF
* Returns the {@link Weight} for this filter aggregation, creating it if
* necessary. This is done lazily so that the {@link Weight} is only created
* if the aggregation collects documents reducing the overhead of the
* aggregation in teh case where no documents are collected.
* aggregation in the case where no documents are collected.
*
* Note that as aggregations are initialsed and executed in a serial manner,
* no concurrency considerations are necessary here.

View File

@ -53,7 +53,7 @@ public final class FragmentBuilderHelper {
if (!fragInfo.getSubInfos().isEmpty() && containsBrokenAnalysis(fieldType.indexAnalyzer())) {
/* This is a special case where broken analysis like WDF is used for term-vector creation at index-time
* which can potentially mess up the offsets. To prevent a SAIIOBException we need to resort
* the fragments based on their offsets rather than using soley the positions as it is done in
* the fragments based on their offsets rather than using solely the positions as it is done in
* the FastVectorHighlighter. Yet, this is really a lucene problem and should be fixed in lucene rather
* than in this hack... aka. "we are are working on in!" */
final List<SubInfo> subInfos = fragInfo.getSubInfos();

View File

@ -150,7 +150,7 @@ public interface Transport extends LifecycleComponent {
}
/**
* This class represents a response context that encapsulates the actual response handler, the action and the conneciton it was
* This class represents a response context that encapsulates the actual response handler, the action and the connection it was
* executed on.
*/
final class ResponseContext<T extends TransportResponse> {

View File

@ -149,7 +149,7 @@ public class BulkRequestTests extends ESTestCase {
IllegalArgumentException exc = expectThrows(IllegalArgumentException.class,
() -> bulkRequest.add(bulkAction.getBytes(StandardCharsets.UTF_8), 0, bulkAction.length(), null, null, XContentType.JSON));
assertThat(exc.getMessage(),
containsString("Malformed action/metadata line [5], expected a simple value for field [_unkown] but found [START_ARRAY]"));
containsString("Malformed action/metadata line [5], expected a simple value for field [_unknown] but found [START_ARRAY]"));
}
public void testSimpleBulk8() throws Exception {

View File

@ -1020,7 +1020,7 @@ public class TransportReplicationActionTests extends ESTestCase {
// publish a new state (same as the old state with the version incremented)
setState(clusterService, stateWithNodes);
// Assert that the request was retried, this time successfull
// Assert that the request was retried, this time successful
assertTrue("action should have been successfully called on retry but was not", calledSuccessfully.get());
transportService.stop();
}

View File

@ -599,7 +599,7 @@ public class OperationRoutingTests extends ESTestCase{
collector.addNodeStatistics("node_1", 4, TimeValue.timeValueMillis(300).nanos(), TimeValue.timeValueMillis(250).nanos());
groupIterator = opRouting.searchShards(state, indexNames, null, null, collector, outstandingRequests);
shardChoice = groupIterator.get(0).nextOrNull();
// finally, node 2 is choosen instead
// finally, node 2 is chosen instead
assertThat(shardChoice.currentNodeId(), equalTo("node_2"));
IOUtils.close(clusterService);

View File

@ -63,7 +63,7 @@ public class FilterAllocationDeciderTests extends ESAllocationTestCase {
"node2").build());
RoutingTable routingTable = state.routingTable();
// we can initally only allocate on node2
// we can initially only allocate on node2
assertEquals(routingTable.index("idx").shard(0).shards().get(0).state(), INITIALIZING);
assertEquals(routingTable.index("idx").shard(0).shards().get(0).currentNodeId(), "node2");
routingTable = service.applyFailedShard(state, routingTable.index("idx").shard(0).shards().get(0), randomBoolean()).routingTable();

View File

@ -37,7 +37,7 @@ public class MultiFieldCopyToMapperTests extends ESTestCase {
public void testExceptionForCopyToInMultiFields() throws IOException {
XContentBuilder mapping = createMappinmgWithCopyToInMultiField();
// first check that for newer versions we throw exception if copy_to is found withing multi field
// first check that for newer versions we throw exception if copy_to is found within multi field
MapperService mapperService = MapperTestUtils.newMapperService(xContentRegistry(), createTempDir(), Settings.EMPTY, "test");
try {
mapperService.parse("type", new CompressedXContent(Strings.toString(mapping)), true);

View File

@ -579,7 +579,7 @@ public class IndexShardIT extends ESSingleNodeTestCase {
// Generate a couple of segments
client().prepareIndex("test", "_doc", "1").setSource("{\"foo\":\"" + randomAlphaOfLength(100) + "\"}", XContentType.JSON)
.setRefreshPolicy(IMMEDIATE).get();
// Use routing so 2 documents are guarenteed to be on the same shard
// Use routing so 2 documents are guaranteed to be on the same shard
String routing = randomAlphaOfLength(5);
client().prepareIndex("test", "_doc", "2").setSource("{\"foo\":\"" + randomAlphaOfLength(100) + "\"}", XContentType.JSON)
.setRefreshPolicy(IMMEDIATE).setRouting(routing).get();

View File

@ -2323,7 +2323,7 @@ public class TranslogTests extends ESTestCase {
@Override
void deleteReaderFiles(TranslogReader reader) {
if (fail.fail()) {
// simulate going OOM and dieing just at the wrong moment.
// simulate going OOM and dying just at the wrong moment.
throw new RuntimeException("simulated");
} else {
super.deleteReaderFiles(reader);

View File

@ -69,7 +69,7 @@ public class ProcessProbeTests extends ESTestCase {
ProcessStats.Mem mem = stats.getMem();
assertNotNull(mem);
// Commited total virtual memory can return -1 if not supported, let's see which platforms fail
// Committed total virtual memory can return -1 if not supported, let's see which platforms fail
assertThat(mem.getTotalVirtual().getBytes(), greaterThan(0L));
}
}

View File

@ -113,7 +113,7 @@ public class InternalScriptedMetricTests extends InternalAggregationTestCase<Int
*/
@Override
protected ScriptService mockScriptService() {
// mock script always retuns the size of the input aggs list as result
// mock script always returns the size of the input aggs list as result
@SuppressWarnings("unchecked")
MockScriptEngine scriptEngine = new MockScriptEngine(MockScriptEngine.NAME,
Collections.singletonMap(REDUCE_SCRIPT_NAME, script -> ((List<Object>) script.get("states")).size()),

View File

@ -170,7 +170,7 @@ public class InternalTopHitsTests extends InternalAggregationTestCase<InternalTo
case STRING_VAL:
return new BytesRef(randomAlphaOfLength(5));
default:
throw new UnsupportedOperationException("Unkown SortField.Type: " + type);
throw new UnsupportedOperationException("Unknown SortField.Type: " + type);
}
}

View File

@ -159,7 +159,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
}
/**
* Test that missing values get transfered correctly to the SortField
* Test that missing values get transferred correctly to the SortField
*/
public void testBuildSortFieldMissingValue() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();
@ -190,7 +190,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
}
/**
* Test that the sort builder order gets transfered correctly to the SortField
* Test that the sort builder order gets transferred correctly to the SortField
*/
public void testBuildSortFieldOrder() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();
@ -214,7 +214,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
}
/**
* Test that the sort builder mode gets transfered correctly to the SortField
* Test that the sort builder mode gets transferred correctly to the SortField
*/
public void testMultiValueMode() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();
@ -249,7 +249,7 @@ public class FieldSortBuilderTests extends AbstractSortTestCase<FieldSortBuilder
comparatorSource = (XFieldComparatorSource) sortField.getComparatorSource();
assertEquals(MultiValueMode.MEDIAN, comparatorSource.sortMode());
// sort mode should also be set by build() implicitely to MIN or MAX if not set explicitely on builder
// sort mode should also be set by build() implicitly to MIN or MAX if not set explicitly on builder
sortBuilder = new FieldSortBuilder("value");
sortField = sortBuilder.build(shardContextMock).field;
assertThat(sortField, instanceOf(SortedNumericSortField.class));

View File

@ -445,7 +445,7 @@ public class GeoDistanceSortBuilderTests extends AbstractSortTestCase<GeoDistanc
}
/**
* Test that the sort builder order gets transfered correctly to the SortField
* Test that the sort builder order gets transferred correctly to the SortField
*/
public void testBuildSortFieldOrder() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();
@ -460,7 +460,7 @@ public class GeoDistanceSortBuilderTests extends AbstractSortTestCase<GeoDistanc
}
/**
* Test that the sort builder mode gets transfered correctly to the SortField
* Test that the sort builder mode gets transferred correctly to the SortField
*/
public void testMultiValueMode() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();

View File

@ -267,7 +267,7 @@ public class ScriptSortBuilderTests extends AbstractSortTestCase<ScriptSortBuild
}
/**
* Test that the sort builder mode gets transfered correctly to the SortField
* Test that the sort builder mode gets transferred correctly to the SortField
*/
public void testMultiValueMode() throws IOException {
QueryShardContext shardContextMock = createMockShardContext();

View File

@ -2,5 +2,5 @@
{"field1": "value0"}
{"index": {"_index": "test", "_type": "doc", "_id": 1}}
{"field1": "value1"}
{"index": {"_index": "test", "_type": "doc", "_id": 2, "_unkown": ["foo", "bar"]}}
{"index": {"_index": "test", "_type": "doc", "_id": 2, "_unknown": ["foo", "bar"]}}
{"field1": "value2"}

View File

@ -2,7 +2,7 @@
"person":{
"dynamic_templates":[
{
"tempalte_1":{
"template_1":{
"match":"multi*",
"mapping":{
"type":"{dynamic_type}",

View File

@ -43,7 +43,7 @@ public class TestThreadInfoPatternConverterTests extends ESTestCase {
// Test threads get the test name
assertEquals(getTestName(), threadInfo(Thread.currentThread().getName()));
// Suite initalization gets "suite"
// Suite initialization gets "suite"
assertEquals("suite", suiteInfo);
// And stuff that doesn't match anything gets wrapped in [] so we can see it

View File

@ -482,7 +482,7 @@ public class DoSectionTests extends AbstractClientYamlTestFragmentParserTestCase
" type: test_type\n" +
"warnings:\n" +
" - some test warning they are typically pretty long\n" +
" - some other test warning somtimes they have [in] them"
" - some other test warning sometimes they have [in] them"
);
DoSection doSection = DoSection.parse(parser);
@ -496,7 +496,7 @@ public class DoSectionTests extends AbstractClientYamlTestFragmentParserTestCase
assertThat(doSection.getApiCallSection().getBodies().size(), equalTo(0));
assertThat(doSection.getExpectedWarningHeaders(), equalTo(Arrays.asList(
"some test warning they are typically pretty long",
"some other test warning somtimes they have [in] them")));
"some other test warning sometimes they have [in] them")));
parser = createParser(YamlXContent.yamlXContent,
"indices.get_field_mapping:\n" +

View File

@ -99,7 +99,7 @@ public class InternalTestClusterTests extends ESTestCase {
/**
* a set of settings that are expected to have different values betweem clusters, even they have been initialized with the same
* base settins.
* base settings.
*/
static final Set<String> clusterUniqueSettings = new HashSet<>();

View File

@ -61,7 +61,7 @@ configuring the `http` attachment type, you must specify the request URL. The
`reporting` attachment type is a special type to include PDF rendered dashboards
from kibana. This type is consistently polling the kibana app if the dashboard
rendering is done, preventing long running HTTP connections, that are potentially
killed by firewalls or load balancers inbetween.
killed by firewalls or load balancers in-between.
[source,js]
--------------------------------------------------

View File

@ -35,7 +35,7 @@ import java.util.List;
* <p>
* Optionally, each hop can contain a "guiding query" that further limits the set of documents considered.
* In our weblog example above we might choose to constrain the second hop to only look at log records that
* had a reponse code of 404.
* had a response code of 404.
* </p>
* <p>
* If absent, the list of {@link VertexRequest}s is inherited from the prior Hop's list to avoid repeating

View File

@ -123,7 +123,7 @@ public class XPackPlugin extends XPackClientPlugin implements ScriptPlugin, Exte
//private final Environment env;
protected boolean transportClientMode;
protected final Licensing licensing;
// These should not be directly accessed as they cannot be overriden in tests. Please use the getters so they can be overridden.
// These should not be directly accessed as they cannot be overridden in tests. Please use the getters so they can be overridden.
private static final SetOnce<XPackLicenseState> licenseState = new SetOnce<>();
private static final SetOnce<SSLService> sslService = new SetOnce<>();
private static final SetOnce<LicenseService> licenseService = new SetOnce<>();

View File

@ -385,7 +385,7 @@ public class Detector implements ToXContentObject, Writeable {
}
/**
* Excludes frequently-occuring metrics from the analysis;
* Excludes frequently-occurring metrics from the analysis;
* can apply to 'by' field, 'over' field, or both
*
* @return the value that the user set

View File

@ -106,7 +106,7 @@ public abstract class WatchRecord implements ToXContentObject {
}
if (executionResult.conditionResult().met()) {
final Collection<ActionWrapperResult> values = executionResult.actionsResults().values();
// acknowledged as state wins because the user had explicitely set this, where as throttled may happen due to execution
// acknowledged as state wins because the user had explicitly set this, where as throttled may happen due to execution
if (values.stream().anyMatch((r) -> r.action().status() == Action.Result.Status.ACKNOWLEDGED)) {
return ExecutionState.ACKNOWLEDGED;
} else if (values.stream().anyMatch((r) -> r.action().status() == Action.Result.Status.THROTTLED)) {

View File

@ -193,7 +193,7 @@ public class TransportGraphExploreAction extends HandledTransportAction<GraphExp
// A single sample pool of docs is built at the root of the aggs tree.
// For quality's sake it might have made more sense to sample top docs
// for each of the terms from the previous hop (e.g. an initial query for "beatles"
// may have seperate doc-sample pools for significant root terms "john", "paul", "yoko" etc)
// may have separate doc-sample pools for significant root terms "john", "paul", "yoko" etc)
// but I found this dramatically slowed down execution - each pool typically had different docs which
// each had non-overlapping sets of terms that needed frequencies looking up for significant terms.
// A common sample pool reduces the specialization that can be given to each root term but

View File

@ -11,7 +11,7 @@ import org.elasticsearch.test.ESSingleNodeTestCase;
import org.elasticsearch.xpack.core.ml.MachineLearningField;
/**
* An extention to {@link ESSingleNodeTestCase} that adds node settings specifically needed for ML test cases.
* An extension to {@link ESSingleNodeTestCase} that adds node settings specifically needed for ML test cases.
*/
public abstract class MlSingleNodeTestCase extends ESSingleNodeTestCase {

View File

@ -184,7 +184,7 @@ public class BasicDistributedJobsIT extends BaseMlIntegTestCase {
@TestLogging("org.elasticsearch.xpack.persistent:TRACE,org.elasticsearch.cluster.service:DEBUG,org.elasticsearch.xpack.ml.action:DEBUG")
public void testDedicatedMlNode() throws Exception {
internalCluster().ensureAtMostNumDataNodes(0);
// start 2 non ml node that will never get a job allocated. (but ml apis are accessable from this node)
// start 2 non ml node that will never get a job allocated. (but ml apis are accessible from this node)
internalCluster().startNode(Settings.builder().put(MachineLearning.ML_ENABLED.getKey(), false));
internalCluster().startNode(Settings.builder().put(MachineLearning.ML_ENABLED.getKey(), false));
// start ml node

View File

@ -262,7 +262,7 @@ public class RollupJobTask extends AllocatedPersistentTask implements SchedulerE
updatePersistentTaskState(state,
ActionListener.wrap(
(task) -> {
logger.debug("Succesfully updated state for rollup job [" + job.getConfig().getId() + "] to ["
logger.debug("Successfully updated state for rollup job [" + job.getConfig().getId() + "] to ["
+ state.getIndexerState() + "][" + state.getPosition() + "]");
listener.onResponse(new StartRollupJobAction.Response(true));
},
@ -308,7 +308,7 @@ public class RollupJobTask extends AllocatedPersistentTask implements SchedulerE
updatePersistentTaskState(state,
ActionListener.wrap(
(task) -> {
logger.debug("Succesfully updated state for rollup job [" + job.getConfig().getId()
logger.debug("Successfully updated state for rollup job [" + job.getConfig().getId()
+ "] to [" + state.getIndexerState() + "]");
listener.onResponse(new StopRollupJobAction.Response(true));
},

View File

@ -202,7 +202,7 @@ public class RollupJobTaskTests extends ESTestCase {
} else if (c == 1) {
assertThat(((RollupJobStatus) taskState).getIndexerState(), equalTo(IndexerState.STOPPED));
} else {
fail("Should not have updated persistent statuse > 2 times");
fail("Should not have updated persistent statuses > 2 times");
}
listener.onResponse(new PersistentTasksCustomMetaData.PersistentTask<>("foo", RollupField.TASK_NAME, job, 1,
new PersistentTasksCustomMetaData.Assignment("foo", "foo")));
@ -688,7 +688,7 @@ public class RollupJobTaskTests extends ESTestCase {
} else if (c == 2) {
assertThat(((RollupJobStatus) taskState).getIndexerState(), equalTo(IndexerState.STOPPED));
} else {
fail("Should not have updated persistent statuse > 3 times");
fail("Should not have updated persistent statuses > 3 times");
}
listener.onResponse(new PersistentTasksCustomMetaData.PersistentTask<>("foo", RollupField.TASK_NAME, job, 1,
new PersistentTasksCustomMetaData.Assignment("foo", "foo")));

View File

@ -115,7 +115,7 @@ public class LdapSessionFactory extends SessionFactory {
* Securely escapes the username and inserts it into the template using MessageFormat
*
* @param username username to insert into the DN template. Any commas, equals or plus will be escaped.
* @return DN (distinquished name) build from the template.
* @return DN (distinguished name) build from the template.
*/
String buildDnFromTemplate(String username, String template) {
//this value must be escaped to avoid manipulation of the template DN.

View File

@ -151,7 +151,7 @@ public interface UserRoleMapper {
* {@link ExpressionModel} class can take a custom {@code Predicate} that tests whether the data in the model
* matches the {@link FieldExpression.FieldValue value} in the expression.
*
* The string constructor parameter may or may not actaully parse as a DN - the "dn" field <em>should</em>
* The string constructor parameter may or may not actually parse as a DN - the "dn" field <em>should</em>
* always be a DN, however groups will be a DN if they're from an LDAP/AD realm, but often won't be for a SAML realm.
*
* Because the {@link FieldExpression.FieldValue} might be a pattern ({@link CharacterRunAutomaton automaton}),

View File

@ -53,7 +53,7 @@ public interface ServerTransportFilter {
throws IOException;
/**
* The server trasnport filter that should be used in nodes as it ensures that an incoming
* The server transport filter that should be used in nodes as it ensures that an incoming
* request is properly authenticated and authorized
*/
class NodeProfile implements ServerTransportFilter {

View File

@ -31,7 +31,7 @@ import static org.hamcrest.Matchers.not;
/**
* This test ensures, that the plugin template upgrader can add and remove
* templates when started within security, as this requires certain
* system priviliges
* system privileges
*/
@ClusterScope(maxNumDataNodes = 1, scope = Scope.SUITE, numClientNodes = 0)
public class TemplateUpgraderTests extends SecurityIntegTestCase {

View File

@ -40,7 +40,7 @@ public class LdapMetaDataResolverTests extends ESTestCase {
new Attribute("cn", "Clint Barton"),
new Attribute("uid", "hawkeye"),
new Attribute("email", "clint.barton@shield.gov"),
new Attribute("memberOf", "cn=staff,ou=groups,dc=exmaple,dc=com", "cn=admin,ou=groups,dc=exmaple,dc=com")
new Attribute("memberOf", "cn=staff,ou=groups,dc=example,dc=com", "cn=admin,ou=groups,dc=example,dc=com")
);
final Map<String, Object> map = resolve(attributes);
assertThat(map.size(), equalTo(2));

View File

@ -1636,7 +1636,7 @@ public class SamlAuthenticatorTests extends SamlTestCase {
/*
Permutation 7 - Mangle the contents of the response to be
<Response>
<Extentions>
<Extensions>
<ForgedAssertion><?ForgedAssertion>
<LegitimateAssertion>
<LegitimateAssertionSignature></LegitimateAssertionSignature>
@ -1645,16 +1645,16 @@ public class SamlAuthenticatorTests extends SamlTestCase {
*/
final Element response = (Element) legitimateDocument.
getElementsByTagNameNS(SAML20P_NS, "Response").item(0);
final Element extentions = legitimateDocument.createElement("Extensions");
final Element extensions = legitimateDocument.createElement("Extensions");
final Element assertion = (Element) legitimateDocument.
getElementsByTagNameNS(SAML20_NS, "Assertion").item(0);
response.insertBefore(extentions, assertion);
response.insertBefore(extensions, assertion);
final Element forgedAssertion = (Element) assertion.cloneNode(true);
forgedAssertion.setAttribute("ID", "_forged_assertion_id");
final Element forgedSignature = (Element) forgedAssertion.
getElementsByTagNameNS("http://www.w3.org/2000/09/xmldsig#", "Signature").item(0);
forgedAssertion.removeChild(forgedSignature);
extentions.appendChild(forgedAssertion);
extensions.appendChild(forgedAssertion);
final SamlToken forgedToken = token(SamlUtils.toString((legitimateDocument.getDocumentElement())));
final ElasticsearchSecurityException exception = expectSamlException(() -> authenticator.authenticate(forgedToken));
assertThat(exception.getMessage(), containsString("Failed to parse SAML"));

View File

@ -121,7 +121,7 @@ public class IpFilteringUpdateTests extends SecurityIntegTestCase {
}
}
// issue #762, occured because in the above test we use HTTP and transport
// issue #762, occurred because in the above test we use HTTP and transport
public void testThatDisablingIpFilterWorksAsExpected() throws Exception {
Settings settings = Settings.builder()
.put("xpack.security.transport.filter.deny", "127.0.0.8")

View File

@ -23,7 +23,7 @@ role3:
cluster: ALL
indices: '*': ALL
# invalid role indices privilegs
# invalid role indices privileges
role4:
cluster: ALL
indices:

View File

@ -43,7 +43,7 @@ import javax.sql.DataSource;
* This class tries to cater to both audiences - use the legacy, Writer way if needed though strive to use the
* proper typical approach, that of specifying intention and output (file) in the URL.
*
* For this reason the {@link System#out} and {@link System#err} are being refered in this class though are used only
* For this reason the {@link System#out} and {@link System#err} are being referred in this class though are used only
* when needed.
*/
public final class Debug {

View File

@ -23,7 +23,7 @@ public class FunctionDefinition {
private final List<String> aliases;
private final Class<? extends Function> clazz;
/**
* Is this a datetime function comaptible with {@code EXTRACT}.
* Is this a datetime function compatible with {@code EXTRACT}.
*/
private final boolean datetime;
private final Builder builder;
@ -60,7 +60,7 @@ public class FunctionDefinition {
}
/**
* Is this a datetime function comaptible with {@code EXTRACT}.
* Is this a datetime function compatible with {@code EXTRACT}.
*/
boolean datetime() {
return datetime;

View File

@ -15,7 +15,7 @@ abstract class StringFunctionUtils {
*
* @param s the original String
* @param start starting position for the substring within the original string. 0-based index position
* @param length length in characters of the substracted substring
* @param length length in characters of the subtracted substring
* @return the resulting String
*/
static String substring(String s, int start, int length) {

View File

@ -291,7 +291,7 @@ abstract class ExpressionBuilder extends IdentifierBuilder {
}
return new Neg(source(ctx.operator), value);
default:
throw new ParsingException(loc, "Unknown arithemtic {}", ctx.operator.getText());
throw new ParsingException(loc, "Unknown arithmetic {}", ctx.operator.getText());
}
}
@ -314,7 +314,7 @@ abstract class ExpressionBuilder extends IdentifierBuilder {
case SqlBaseParser.MINUS:
return new Sub(loc, left, right);
default:
throw new ParsingException(loc, "Unknown arithemtic {}", ctx.operator.getText());
throw new ParsingException(loc, "Unknown arithmetic {}", ctx.operator.getText());
}
}

View File

@ -172,7 +172,7 @@ setup:
"job_id":"datafeeds-crud-1",
"indexes":["index-foo"],
"types":["type-bar"],
"query":{"match_all_mispelled":{}}
"query":{"match_all_misspelled":{}}
}
---

View File

@ -34,7 +34,7 @@ setup:
# read the monitoring indices.
- do:
xpack.security.put_role:
name: "unkown_agent_role"
name: "unknown_agent_role"
body: >
{
"cluster": ["monitor"],
@ -51,7 +51,7 @@ setup:
body: >
{
"password": "s3krit",
"roles" : [ "unkown_agent_role" ]
"roles" : [ "unknown_agent_role" ]
}
---
@ -70,7 +70,7 @@ teardown:
ignore: 404
- do:
xpack.security.delete_role:
name: "unkown_agent_role"
name: "unknown_agent_role"
ignore: 404
---

View File

@ -177,7 +177,7 @@ setup:
- is_true: acknowledged
---
"Test delete non-existant job":
"Test delete non-existent job":
- do:
catch: /the task with id does_not_exist doesn't exist/

View File

@ -37,7 +37,7 @@ setup:
}
---
"Test start non-existant job":
"Test start nonexistent job":
- do:
catch: /Task for Rollup Job \[does_not_exist\] not found/

View File

@ -37,7 +37,7 @@ setup:
}
---
"Test stop non-existant job":
"Test stop nonexistent job":
- do:
catch: /Task for Rollup Job \[does_not_exist\] not found/

View File

@ -91,7 +91,7 @@ teardown:
ignore: 404
---
"Test shared index seperating user by using DLS":
"Test shared index separating user by using DLS":
- do:
headers:
Authorization: "Basic am9lOngtcGFjay10ZXN0LXBhc3N3b3Jk"

View File

@ -12,7 +12,7 @@ teardown:
ignore: 404
---
"Ensure that ack status is reset after unsuccesful execution":
"Ensure that ack status is reset after unsuccessful execution":
- do:
xpack.watcher.put_watch:

View File

@ -47,7 +47,7 @@ public class IndexUpgradeIT extends IndexUpgradeIntegTestCase {
public void testIndexUpgradeInfoLicense() throws Exception {
// This test disables all licenses and generates a new one using dev private key
// in non-snapshot builds we are using produciton public key for license verification
// in non-snapshot builds we are using production public key for license verification
// which makes this test to fail
assumeTrue("License is only valid when tested against snapshot/test keys", Build.CURRENT.isSnapshot());
assertAcked(client().admin().indices().prepareCreate("test").get());

View File

@ -143,7 +143,7 @@ final class WatcherIndexingListener extends AbstractComponent implements Indexin
*
* @param shardId The shard id object of the document being processed
* @param index The index operation
* @param ex The exception occured during indexing
* @param ex The exception occurred during indexing
*/
@Override
public void postIndex(ShardId shardId, Engine.Index index, Exception ex) {

View File

@ -122,7 +122,7 @@ public class ExecutionService extends AbstractComponent {
/**
* Pause the execution of the watcher executor, and empty the state.
* Pausing means, that no new watch executions will be done unless this pausing is explicitely unset.
* Pausing means, that no new watch executions will be done unless this pausing is explicitly unset.
* This is important when watcher is stopped, so that scheduled watches do not accidentally get executed.
* This should not be used when we need to reload watcher based on some cluster state changes, then just calling
* {@link #clearExecutionsAndQueue()} is the way to go
@ -338,7 +338,7 @@ public class ExecutionService extends AbstractComponent {
public void updateWatchStatus(Watch watch) throws IOException {
// at the moment we store the status together with the watch,
// so we just need to update the watch itself
// we do not want to update the status.state field, as it might have been deactivated inbetween
// we do not want to update the status.state field, as it might have been deactivated in-between
Map<String, String> parameters = MapBuilder.<String, String>newMapBuilder()
.put(Watch.INCLUDE_STATUS_KEY, "true")
.put(WatchStatus.INCLUDE_STATE, "false")

View File

@ -117,7 +117,7 @@ public class SlackActionTests extends ESTestCase {
hasError = true;
break;
case 1:
when(response.status()).thenReturn(randomIntBetween(300, 600)); // error reponse
when(response.status()).thenReturn(randomIntBetween(300, 600)); // error response
messages.add(SentMessages.SentMessage.responded(randomAlphaOfLength(10), message, request, response));
hasError = true;
break;

View File

@ -522,7 +522,7 @@ public class HttpClientTests extends ESTestCase {
});
HttpRequest request = HttpRequest.builder("localhost", serverSocket.getLocalPort()).path("/").build();
expectThrows(ClientProtocolException.class, () -> httpClient.execute(request));
assertThat("A server side exception occured, but shouldn't", hasExceptionHappened.get(), is(nullValue()));
assertThat("A server side exception occurred, but shouldn't", hasExceptionHappened.get(), is(nullValue()));
} finally {
terminate(executor);
}

View File

@ -83,7 +83,7 @@ public class DeleteWatchTests extends AbstractWatcherIntegrationTestCase {
// watch has been executed successfully
String state = ObjectPath.eval("state", source);
assertThat(state, is("executed"));
// no exception occured
// no exception occurred
assertThat(source, not(hasKey("exception")));
}
}

View File

@ -105,7 +105,7 @@ public class ExecuteWatchTests extends AbstractWatcherIntegrationTestCase {
}
if (mode.force()) {
// since we're forcing, lets ack the action, such that it'd suppoed to be throttled
// since we're forcing, lets ack the action, such that it'd supposed to be throttled
// but forcing will ignore the throttling
// lets wait for the watch to be ackable

View File

@ -62,7 +62,7 @@ teardown:
ignore: 404
---
"Test shared index seperating user by using DLS role query with user's username":
"Test shared index separating user by using DLS role query with user's username":
- do:
xpack.security.put_role:
name: "small_companies_role"
@ -130,7 +130,7 @@ teardown:
- match: { hits.hits.0._source.user.username: john}
---
"Test shared index seperating user by using DLS role query with user's metadata":
"Test shared index separating user by using DLS role query with user's metadata":
- do:
xpack.security.put_role:
name: "small_companies_role"

View File

@ -213,7 +213,7 @@ public class RestSqlSecurityIT extends SqlSecurityTestCase {
/**
* Test the hijacking a scroll fails. This test is only implemented for
* REST because it is the only API where it is simple to hijack a scroll.
* It should excercise the same code as the other APIs but if we were truly
* It should exercise the same code as the other APIs but if we were truly
* paranoid we'd hack together something to test the others as well.
*/
public void testHijackScrollFails() throws Exception {

View File

@ -264,7 +264,7 @@ public abstract class SqlSecurityTestCase extends ESRestTestCase {
createAuditLogAsserter()
.expectSqlCompositeAction("test_admin", "test")
/* Scrolling doesn't have to access the index again, at least not through sql.
* If we asserted query and scroll logs then we would see the scoll. */
* If we asserted query and scroll logs then we would see the scroll. */
.expect(true, SQL_ACTION_NAME, "test_admin", empty())
.expect(true, SQL_ACTION_NAME, "test_admin", empty())
.expectSqlCompositeAction("only_a", "test")