Fixed typos (#20843)

This commit is contained in:
Pascal Borreli 2016-10-10 21:51:47 +01:00 committed by Lee Hinman
parent 44d960ea85
commit fcb01deb34
19 changed files with 24 additions and 24 deletions

View File

@ -63,7 +63,7 @@ Response:
} }
-------------------------------------------------- --------------------------------------------------
<1> 1000 documents were sampled in total becase we asked for a maximum of 200 from an index with 5 shards. The cost of performing the nested significant_terms aggregation was therefore limited rather than unbounded. <1> 1000 documents were sampled in total because we asked for a maximum of 200 from an index with 5 shards. The cost of performing the nested significant_terms aggregation was therefore limited rather than unbounded.
<2> The results of the significant_terms aggregation are not skewed by any single over-active Twitter user because we asked for a maximum of one tweet from any one user in our sample. <2> The results of the significant_terms aggregation are not skewed by any single over-active Twitter user because we asked for a maximum of one tweet from any one user in our sample.
@ -92,7 +92,7 @@ Controlling diversity using a field:
{ {
"aggs" : { "aggs" : {
"sample" : { "sample" : {
"diverisfied_sampler" : { "diversified_sampler" : {
"field" : "author", "field" : "author",
"max_docs_per_value" : 3 "max_docs_per_value" : 3
} }

View File

@ -499,7 +499,7 @@ TIP: for indexed scripts replace the `file` parameter with an `id` parameter.
"aggs" : { "aggs" : {
"genres" : { "genres" : {
"terms" : { "terms" : {
"field" : "gendre", "field" : "gender",
"script" : { "script" : {
"inline" : "'Genre: ' +_value" "inline" : "'Genre: ' +_value"
"lang" : "painless" "lang" : "painless"

View File

@ -2,7 +2,7 @@
=== Lowercase Tokenizer === Lowercase Tokenizer
The `lowercase` toknenizer, like the The `lowercase` tokenizer, like the
<<analysis-letter-tokenizer, `letter` tokenizer>> breaks text into terms <<analysis-letter-tokenizer, `letter` tokenizer>> breaks text into terms
whenever it encounters a character which is not a letter, but it also whenever it encounters a character which is not a letter, but it also
lowecases all terms. It is functionally equivalent to the lowecases all terms. It is functionally equivalent to the

View File

@ -5,7 +5,7 @@
=== Use bulk requests === Use bulk requests
Bulk requests will yield much better performance than single-document index Bulk requests will yield much better performance than single-document index
requests. In order to know the optimal size of a bulk request, you shoud run requests. In order to know the optimal size of a bulk request, you should run
a benchmark on a single node with a single shard. First try to index 100 a benchmark on a single node with a single shard. First try to index 100
documents at once, then 200, then 400, etc. doubling the number of documents documents at once, then 200, then 400, etc. doubling the number of documents
in a bulk request in every benchmark run. When the indexing speed starts to in a bulk request in every benchmark run. When the indexing speed starts to
@ -32,7 +32,7 @@ When it happens, you should pause indexing a bit before trying again, ideally
with randomized exponential backoff. with randomized exponential backoff.
Similarly to sizing bulk requests, only testing can tell what the optimal Similarly to sizing bulk requests, only testing can tell what the optimal
number of workers is. This can be tested by progressivily increasing the number of workers is. This can be tested by progressively increasing the
number of workers until either I/O or CPU is saturated on the cluster. number of workers until either I/O or CPU is saturated on the cluster.
[float] [float]
@ -58,7 +58,7 @@ original values.
=== Disable swapping === Disable swapping
You should make sure that the operating system is not swapping out the java You should make sure that the operating system is not swapping out the java
process by <<setup-configuration-memory,disabling swappping>>. process by <<setup-configuration-memory,disabling swapping>>.
[float] [float]
=== Give memory to the filesystem cache === Give memory to the filesystem cache

View File

@ -79,7 +79,7 @@ By default, elasticsearch completely relies on the operating system file system
cache for caching I/O operations. It is possible to set `index.store.preload` cache for caching I/O operations. It is possible to set `index.store.preload`
in order to tell the operating system to load the content of hot index in order to tell the operating system to load the content of hot index
files into memory upon opening. This setting accept a comma-separated list of files into memory upon opening. This setting accept a comma-separated list of
files extensions: all files whose extenion is in the list will be pre-loaded files extensions: all files whose extension is in the list will be pre-loaded
upon opening. This can be useful to improve search performance of an index, upon opening. This can be useful to improve search performance of an index,
especially when the host operating system is restarted, since this causes the especially when the host operating system is restarted, since this causes the
file system cache to be trashed. However note that this may slow down the file system cache to be trashed. However note that this may slow down the

View File

@ -660,7 +660,7 @@ A node will not start if either of these plugins are not available.
The <<ingest-stats,node stats API>> can be used to fetch ingest usage statistics, globally and on a per The <<ingest-stats,node stats API>> can be used to fetch ingest usage statistics, globally and on a per
pipeline basis. Useful to find out which pipelines are used the most or spent the most time on preprocessing. pipeline basis. Useful to find out which pipelines are used the most or spent the most time on preprocessing.
[[append-procesesor]] [[append-processor]]
=== Append Processor === Append Processor
Appends one or more values to an existing array if the field already exists and it is an array. Appends one or more values to an existing array if the field already exists and it is an array.
Converts a scalar to an array and appends one or more values to it if the field exists and it is a scalar. Converts a scalar to an array and appends one or more values to it if the field exists and it is a scalar.

View File

@ -353,7 +353,7 @@ remove it as described in the previous section.
===== Time-series ===== Time-series
When doing time series analysis with elastisearch, it is common to have many When doing time series analysis with elasticsearch, it is common to have many
numeric fields that you will often aggregate on but never filter on. In such a numeric fields that you will often aggregate on but never filter on. In such a
case, you could disable indexing on those fields to save disk space and also case, you could disable indexing on those fields to save disk space and also
maybe gain some indexing speed: maybe gain some indexing speed:

View File

@ -3,5 +3,5 @@
==== Mapper attachments plugin ==== Mapper attachments plugin
* The mapper attachments plugin has been depecated in elasticsearch 5.0 and is now removed. * The mapper attachments plugin has been deprecated in elasticsearch 5.0 and is now removed.
You can use {plugins}/ingest-attachment.html[ingest attachment plugin] instead. You can use {plugins}/ingest-attachment.html[ingest attachment plugin] instead.

View File

@ -4,6 +4,6 @@
==== Unquoted JSON ==== Unquoted JSON
In previous versions of Elasticsearch, JSON documents were allowed to contain unquoted field names. In previous versions of Elasticsearch, JSON documents were allowed to contain unquoted field names.
This feature was removed in the 5.x series, but a backwards-compability layer was added via the This feature was removed in the 5.x series, but a backwards-compatibility layer was added via the
system property `elasticsearch.json.allow_unquoted_field_names`. This backwards-compability layer system property `elasticsearch.json.allow_unquoted_field_names`. This backwards-compatibility layer
has been removed in Elasticsearch 6.0.0. has been removed in Elasticsearch 6.0.0.

View File

@ -287,7 +287,7 @@ machine. In production, however, it is recommended to run only one node of Elast
By default, Elasticsearch is configured to prevent more than one node from sharing the same data By default, Elasticsearch is configured to prevent more than one node from sharing the same data
path. To allow for more than one node (e.g., on your development machine), use the setting path. To allow for more than one node (e.g., on your development machine), use the setting
`node.max_local_storage_nodes` and set this to a positve integer larger than one. `node.max_local_storage_nodes` and set this to a positive integer larger than one.
WARNING: Never run different node types (i.e. master, data) from the same data directory. This can WARNING: Never run different node types (i.e. master, data) from the same data directory. This can
lead to unexpected data loss. lead to unexpected data loss.

View File

@ -13,7 +13,7 @@ the same pattern:
} }
------------------------------------- -------------------------------------
<1> The language the script is written in, which defaults to `painless`. <1> The language the script is written in, which defaults to `painless`.
<2> The script itself which may be specfied as `inline`, `id`, or `file`. <2> The script itself which may be specified as `inline`, `id`, or `file`.
<3> Any named parameters that should be passed into the script. <3> Any named parameters that should be passed into the script.
For example, the following script is used in a search request to return a For example, the following script is used in a search request to return a

View File

@ -277,7 +277,7 @@ GET /_snapshot/my_backup/_all
----------------------------------- -----------------------------------
// CONSOLE // CONSOLE
The command fails if some of the snapshots are unavailable. The boolean parameter `ignore_unvailable` can be used to The command fails if some of the snapshots are unavailable. The boolean parameter `ignore_unavailable` can be used to
return all snapshots that are currently available. return all snapshots that are currently available.
A currently running snapshot can be retrieved using the following command: A currently running snapshot can be retrieved using the following command:

View File

@ -318,7 +318,7 @@ In the above example, the field is a <<geo-point,`geo_point`>> and origin can be
math (for example `now-1h`) is supported for origin. math (for example `now-1h`) is supported for origin.
`scale`:: `scale`::
Required for all types. Defines the distance from origin + offest at which the computed Required for all types. Defines the distance from origin + offset at which the computed
score will equal `decay` parameter. For geo fields: Can be defined as number+unit (1km, 12m,...). score will equal `decay` parameter. For geo fields: Can be defined as number+unit (1km, 12m,...).
Default unit is meters. For date fields: Can to be defined as a number+unit ("1h", "10d",...). Default unit is meters. For date fields: Can to be defined as a number+unit ("1h", "10d",...).
Default unit is milliseconds. For numeric field: Any number. Default unit is milliseconds. For numeric field: Any number.

View File

@ -101,7 +101,7 @@ process has unlimited address space and is enforced only on Linux. To
pass the maximum size virtual memory check, you must configure your pass the maximum size virtual memory check, you must configure your
system to allow the Elasticsearch process the ability to have unlimited system to allow the Elasticsearch process the ability to have unlimited
address space. This can be done via `/etc/security/limits.conf` using address space. This can be done via `/etc/security/limits.conf` using
the `as` setting to `unlimited` (note that you might have to increaes the `as` setting to `unlimited` (note that you might have to increase
the limits for the `root` user too). the limits for the `root` user too).
=== Maximum map count check === Maximum map count check

View File

@ -70,7 +70,7 @@ environments, otherwise you might end up with nodes joining the wrong cluster.
[[node.name]] [[node.name]]
=== `node.name` === `node.name`
By default, Elasticsearch will take the 7 first charachter of the randomly generated uuid used as the node id. By default, Elasticsearch will take the 7 first character of the randomly generated uuid used as the node id.
Note that the node id is persisted and does not change when a node restarts and therefore the default node name Note that the node id is persisted and does not change when a node restarts and therefore the default node name
will also not change. will also not change.

View File

@ -1,7 +1,7 @@
==== SysV `init` vs `systemd` ==== SysV `init` vs `systemd`
Elasticsearch is not started automatically after installation. How to start Elasticsearch is not started automatically after installation. How to start
and stop Elasticsearch depends on whether your sytem uses SysV `init` or and stop Elasticsearch depends on whether your system uses SysV `init` or
`systemd` (used by newer distributions). You can tell which is being used by `systemd` (used by newer distributions). You can tell which is being used by
running this command: running this command:

View File

@ -21,7 +21,7 @@ endif::[]
ifeval::["{release-state}"!="unreleased"] ifeval::["{release-state}"!="unreleased"]
The `.zip` archive for Elastisearch v{version} can be downloaded and installed as follows: The `.zip` archive for Elasticsearch v{version} can be downloaded and installed as follows:
["source","sh",subs="attributes"] ["source","sh",subs="attributes"]
@ -49,7 +49,7 @@ endif::[]
ifeval::["{release-state}"!="unreleased"] ifeval::["{release-state}"!="unreleased"]
The `.tar.gz` archive for Elastisearch v{version} can be downloaded and installed as follows: The `.tar.gz` archive for Elasticsearch v{version} can be downloaded and installed as follows:
["source","sh",subs="attributes"] ["source","sh",subs="attributes"]
-------------------------------------------- --------------------------------------------

View File

@ -67,7 +67,7 @@ ES_JAVA_OPTS="-Xms4000m -Xmx4000m" ./bin/elasticsearch <2>
<2> Set the minimum and maximum heap size to 4000 MB. <2> Set the minimum and maximum heap size to 4000 MB.
NOTE: Configuring the heap for the <<windows-service,Windows service>> NOTE: Configuring the heap for the <<windows-service,Windows service>>
is different than the above. The values initiallly populated for the is different than the above. The values initially populated for the
Windows service can be configured as above but are different after the Windows service can be configured as above but are different after the
service has been installed. Consult the service has been installed. Consult the
<<windows-service,Windows service documentation>> for additional <<windows-service,Windows service documentation>> for additional

View File

@ -253,7 +253,7 @@ Usually, you would combine assertions and matchers in your test like this
[source,java] [source,java]
---------------------------- ----------------------------
SearchResponse seearchResponse = client().prepareSearch() ...; SearchResponse searchResponse = client().prepareSearch() ...;
assertHitCount(searchResponse, 4); assertHitCount(searchResponse, 4);
assertFirstHit(searchResponse, hasId("4")); assertFirstHit(searchResponse, hasId("4"));
assertSearchHits(searchResponse, "1", "2", "3", "4"); assertSearchHits(searchResponse, "1", "2", "3", "4");