mirror of
https://github.com/honeymoose/OpenSearch.git
synced 2025-02-17 10:25:15 +00:00
Fixed typos (#20843)
This commit is contained in:
parent
44d960ea85
commit
fcb01deb34
@ -63,7 +63,7 @@ Response:
|
|||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
|
||||||
<1> 1000 documents were sampled in total becase we asked for a maximum of 200 from an index with 5 shards. The cost of performing the nested significant_terms aggregation was therefore limited rather than unbounded.
|
<1> 1000 documents were sampled in total because we asked for a maximum of 200 from an index with 5 shards. The cost of performing the nested significant_terms aggregation was therefore limited rather than unbounded.
|
||||||
<2> The results of the significant_terms aggregation are not skewed by any single over-active Twitter user because we asked for a maximum of one tweet from any one user in our sample.
|
<2> The results of the significant_terms aggregation are not skewed by any single over-active Twitter user because we asked for a maximum of one tweet from any one user in our sample.
|
||||||
|
|
||||||
|
|
||||||
@ -92,7 +92,7 @@ Controlling diversity using a field:
|
|||||||
{
|
{
|
||||||
"aggs" : {
|
"aggs" : {
|
||||||
"sample" : {
|
"sample" : {
|
||||||
"diverisfied_sampler" : {
|
"diversified_sampler" : {
|
||||||
"field" : "author",
|
"field" : "author",
|
||||||
"max_docs_per_value" : 3
|
"max_docs_per_value" : 3
|
||||||
}
|
}
|
||||||
|
@ -499,7 +499,7 @@ TIP: for indexed scripts replace the `file` parameter with an `id` parameter.
|
|||||||
"aggs" : {
|
"aggs" : {
|
||||||
"genres" : {
|
"genres" : {
|
||||||
"terms" : {
|
"terms" : {
|
||||||
"field" : "gendre",
|
"field" : "gender",
|
||||||
"script" : {
|
"script" : {
|
||||||
"inline" : "'Genre: ' +_value"
|
"inline" : "'Genre: ' +_value"
|
||||||
"lang" : "painless"
|
"lang" : "painless"
|
||||||
|
@ -2,7 +2,7 @@
|
|||||||
=== Lowercase Tokenizer
|
=== Lowercase Tokenizer
|
||||||
|
|
||||||
|
|
||||||
The `lowercase` toknenizer, like the
|
The `lowercase` tokenizer, like the
|
||||||
<<analysis-letter-tokenizer, `letter` tokenizer>> breaks text into terms
|
<<analysis-letter-tokenizer, `letter` tokenizer>> breaks text into terms
|
||||||
whenever it encounters a character which is not a letter, but it also
|
whenever it encounters a character which is not a letter, but it also
|
||||||
lowecases all terms. It is functionally equivalent to the
|
lowecases all terms. It is functionally equivalent to the
|
||||||
|
@ -5,7 +5,7 @@
|
|||||||
=== Use bulk requests
|
=== Use bulk requests
|
||||||
|
|
||||||
Bulk requests will yield much better performance than single-document index
|
Bulk requests will yield much better performance than single-document index
|
||||||
requests. In order to know the optimal size of a bulk request, you shoud run
|
requests. In order to know the optimal size of a bulk request, you should run
|
||||||
a benchmark on a single node with a single shard. First try to index 100
|
a benchmark on a single node with a single shard. First try to index 100
|
||||||
documents at once, then 200, then 400, etc. doubling the number of documents
|
documents at once, then 200, then 400, etc. doubling the number of documents
|
||||||
in a bulk request in every benchmark run. When the indexing speed starts to
|
in a bulk request in every benchmark run. When the indexing speed starts to
|
||||||
@ -32,7 +32,7 @@ When it happens, you should pause indexing a bit before trying again, ideally
|
|||||||
with randomized exponential backoff.
|
with randomized exponential backoff.
|
||||||
|
|
||||||
Similarly to sizing bulk requests, only testing can tell what the optimal
|
Similarly to sizing bulk requests, only testing can tell what the optimal
|
||||||
number of workers is. This can be tested by progressivily increasing the
|
number of workers is. This can be tested by progressively increasing the
|
||||||
number of workers until either I/O or CPU is saturated on the cluster.
|
number of workers until either I/O or CPU is saturated on the cluster.
|
||||||
|
|
||||||
[float]
|
[float]
|
||||||
@ -58,7 +58,7 @@ original values.
|
|||||||
=== Disable swapping
|
=== Disable swapping
|
||||||
|
|
||||||
You should make sure that the operating system is not swapping out the java
|
You should make sure that the operating system is not swapping out the java
|
||||||
process by <<setup-configuration-memory,disabling swappping>>.
|
process by <<setup-configuration-memory,disabling swapping>>.
|
||||||
|
|
||||||
[float]
|
[float]
|
||||||
=== Give memory to the filesystem cache
|
=== Give memory to the filesystem cache
|
||||||
|
@ -79,7 +79,7 @@ By default, elasticsearch completely relies on the operating system file system
|
|||||||
cache for caching I/O operations. It is possible to set `index.store.preload`
|
cache for caching I/O operations. It is possible to set `index.store.preload`
|
||||||
in order to tell the operating system to load the content of hot index
|
in order to tell the operating system to load the content of hot index
|
||||||
files into memory upon opening. This setting accept a comma-separated list of
|
files into memory upon opening. This setting accept a comma-separated list of
|
||||||
files extensions: all files whose extenion is in the list will be pre-loaded
|
files extensions: all files whose extension is in the list will be pre-loaded
|
||||||
upon opening. This can be useful to improve search performance of an index,
|
upon opening. This can be useful to improve search performance of an index,
|
||||||
especially when the host operating system is restarted, since this causes the
|
especially when the host operating system is restarted, since this causes the
|
||||||
file system cache to be trashed. However note that this may slow down the
|
file system cache to be trashed. However note that this may slow down the
|
||||||
|
@ -660,7 +660,7 @@ A node will not start if either of these plugins are not available.
|
|||||||
The <<ingest-stats,node stats API>> can be used to fetch ingest usage statistics, globally and on a per
|
The <<ingest-stats,node stats API>> can be used to fetch ingest usage statistics, globally and on a per
|
||||||
pipeline basis. Useful to find out which pipelines are used the most or spent the most time on preprocessing.
|
pipeline basis. Useful to find out which pipelines are used the most or spent the most time on preprocessing.
|
||||||
|
|
||||||
[[append-procesesor]]
|
[[append-processor]]
|
||||||
=== Append Processor
|
=== Append Processor
|
||||||
Appends one or more values to an existing array if the field already exists and it is an array.
|
Appends one or more values to an existing array if the field already exists and it is an array.
|
||||||
Converts a scalar to an array and appends one or more values to it if the field exists and it is a scalar.
|
Converts a scalar to an array and appends one or more values to it if the field exists and it is a scalar.
|
||||||
|
@ -353,7 +353,7 @@ remove it as described in the previous section.
|
|||||||
|
|
||||||
===== Time-series
|
===== Time-series
|
||||||
|
|
||||||
When doing time series analysis with elastisearch, it is common to have many
|
When doing time series analysis with elasticsearch, it is common to have many
|
||||||
numeric fields that you will often aggregate on but never filter on. In such a
|
numeric fields that you will often aggregate on but never filter on. In such a
|
||||||
case, you could disable indexing on those fields to save disk space and also
|
case, you could disable indexing on those fields to save disk space and also
|
||||||
maybe gain some indexing speed:
|
maybe gain some indexing speed:
|
||||||
|
@ -3,5 +3,5 @@
|
|||||||
|
|
||||||
==== Mapper attachments plugin
|
==== Mapper attachments plugin
|
||||||
|
|
||||||
* The mapper attachments plugin has been depecated in elasticsearch 5.0 and is now removed.
|
* The mapper attachments plugin has been deprecated in elasticsearch 5.0 and is now removed.
|
||||||
You can use {plugins}/ingest-attachment.html[ingest attachment plugin] instead.
|
You can use {plugins}/ingest-attachment.html[ingest attachment plugin] instead.
|
||||||
|
@ -4,6 +4,6 @@
|
|||||||
==== Unquoted JSON
|
==== Unquoted JSON
|
||||||
|
|
||||||
In previous versions of Elasticsearch, JSON documents were allowed to contain unquoted field names.
|
In previous versions of Elasticsearch, JSON documents were allowed to contain unquoted field names.
|
||||||
This feature was removed in the 5.x series, but a backwards-compability layer was added via the
|
This feature was removed in the 5.x series, but a backwards-compatibility layer was added via the
|
||||||
system property `elasticsearch.json.allow_unquoted_field_names`. This backwards-compability layer
|
system property `elasticsearch.json.allow_unquoted_field_names`. This backwards-compatibility layer
|
||||||
has been removed in Elasticsearch 6.0.0.
|
has been removed in Elasticsearch 6.0.0.
|
||||||
|
@ -287,7 +287,7 @@ machine. In production, however, it is recommended to run only one node of Elast
|
|||||||
|
|
||||||
By default, Elasticsearch is configured to prevent more than one node from sharing the same data
|
By default, Elasticsearch is configured to prevent more than one node from sharing the same data
|
||||||
path. To allow for more than one node (e.g., on your development machine), use the setting
|
path. To allow for more than one node (e.g., on your development machine), use the setting
|
||||||
`node.max_local_storage_nodes` and set this to a positve integer larger than one.
|
`node.max_local_storage_nodes` and set this to a positive integer larger than one.
|
||||||
|
|
||||||
WARNING: Never run different node types (i.e. master, data) from the same data directory. This can
|
WARNING: Never run different node types (i.e. master, data) from the same data directory. This can
|
||||||
lead to unexpected data loss.
|
lead to unexpected data loss.
|
||||||
|
@ -13,7 +13,7 @@ the same pattern:
|
|||||||
}
|
}
|
||||||
-------------------------------------
|
-------------------------------------
|
||||||
<1> The language the script is written in, which defaults to `painless`.
|
<1> The language the script is written in, which defaults to `painless`.
|
||||||
<2> The script itself which may be specfied as `inline`, `id`, or `file`.
|
<2> The script itself which may be specified as `inline`, `id`, or `file`.
|
||||||
<3> Any named parameters that should be passed into the script.
|
<3> Any named parameters that should be passed into the script.
|
||||||
|
|
||||||
For example, the following script is used in a search request to return a
|
For example, the following script is used in a search request to return a
|
||||||
|
@ -277,7 +277,7 @@ GET /_snapshot/my_backup/_all
|
|||||||
-----------------------------------
|
-----------------------------------
|
||||||
// CONSOLE
|
// CONSOLE
|
||||||
|
|
||||||
The command fails if some of the snapshots are unavailable. The boolean parameter `ignore_unvailable` can be used to
|
The command fails if some of the snapshots are unavailable. The boolean parameter `ignore_unavailable` can be used to
|
||||||
return all snapshots that are currently available.
|
return all snapshots that are currently available.
|
||||||
|
|
||||||
A currently running snapshot can be retrieved using the following command:
|
A currently running snapshot can be retrieved using the following command:
|
||||||
|
@ -318,7 +318,7 @@ In the above example, the field is a <<geo-point,`geo_point`>> and origin can be
|
|||||||
math (for example `now-1h`) is supported for origin.
|
math (for example `now-1h`) is supported for origin.
|
||||||
|
|
||||||
`scale`::
|
`scale`::
|
||||||
Required for all types. Defines the distance from origin + offest at which the computed
|
Required for all types. Defines the distance from origin + offset at which the computed
|
||||||
score will equal `decay` parameter. For geo fields: Can be defined as number+unit (1km, 12m,...).
|
score will equal `decay` parameter. For geo fields: Can be defined as number+unit (1km, 12m,...).
|
||||||
Default unit is meters. For date fields: Can to be defined as a number+unit ("1h", "10d",...).
|
Default unit is meters. For date fields: Can to be defined as a number+unit ("1h", "10d",...).
|
||||||
Default unit is milliseconds. For numeric field: Any number.
|
Default unit is milliseconds. For numeric field: Any number.
|
||||||
|
@ -101,7 +101,7 @@ process has unlimited address space and is enforced only on Linux. To
|
|||||||
pass the maximum size virtual memory check, you must configure your
|
pass the maximum size virtual memory check, you must configure your
|
||||||
system to allow the Elasticsearch process the ability to have unlimited
|
system to allow the Elasticsearch process the ability to have unlimited
|
||||||
address space. This can be done via `/etc/security/limits.conf` using
|
address space. This can be done via `/etc/security/limits.conf` using
|
||||||
the `as` setting to `unlimited` (note that you might have to increaes
|
the `as` setting to `unlimited` (note that you might have to increase
|
||||||
the limits for the `root` user too).
|
the limits for the `root` user too).
|
||||||
|
|
||||||
=== Maximum map count check
|
=== Maximum map count check
|
||||||
|
@ -70,7 +70,7 @@ environments, otherwise you might end up with nodes joining the wrong cluster.
|
|||||||
[[node.name]]
|
[[node.name]]
|
||||||
=== `node.name`
|
=== `node.name`
|
||||||
|
|
||||||
By default, Elasticsearch will take the 7 first charachter of the randomly generated uuid used as the node id.
|
By default, Elasticsearch will take the 7 first character of the randomly generated uuid used as the node id.
|
||||||
Note that the node id is persisted and does not change when a node restarts and therefore the default node name
|
Note that the node id is persisted and does not change when a node restarts and therefore the default node name
|
||||||
will also not change.
|
will also not change.
|
||||||
|
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
==== SysV `init` vs `systemd`
|
==== SysV `init` vs `systemd`
|
||||||
|
|
||||||
Elasticsearch is not started automatically after installation. How to start
|
Elasticsearch is not started automatically after installation. How to start
|
||||||
and stop Elasticsearch depends on whether your sytem uses SysV `init` or
|
and stop Elasticsearch depends on whether your system uses SysV `init` or
|
||||||
`systemd` (used by newer distributions). You can tell which is being used by
|
`systemd` (used by newer distributions). You can tell which is being used by
|
||||||
running this command:
|
running this command:
|
||||||
|
|
||||||
|
@ -21,7 +21,7 @@ endif::[]
|
|||||||
|
|
||||||
ifeval::["{release-state}"!="unreleased"]
|
ifeval::["{release-state}"!="unreleased"]
|
||||||
|
|
||||||
The `.zip` archive for Elastisearch v{version} can be downloaded and installed as follows:
|
The `.zip` archive for Elasticsearch v{version} can be downloaded and installed as follows:
|
||||||
|
|
||||||
|
|
||||||
["source","sh",subs="attributes"]
|
["source","sh",subs="attributes"]
|
||||||
@ -49,7 +49,7 @@ endif::[]
|
|||||||
|
|
||||||
ifeval::["{release-state}"!="unreleased"]
|
ifeval::["{release-state}"!="unreleased"]
|
||||||
|
|
||||||
The `.tar.gz` archive for Elastisearch v{version} can be downloaded and installed as follows:
|
The `.tar.gz` archive for Elasticsearch v{version} can be downloaded and installed as follows:
|
||||||
|
|
||||||
["source","sh",subs="attributes"]
|
["source","sh",subs="attributes"]
|
||||||
--------------------------------------------
|
--------------------------------------------
|
||||||
|
@ -67,7 +67,7 @@ ES_JAVA_OPTS="-Xms4000m -Xmx4000m" ./bin/elasticsearch <2>
|
|||||||
<2> Set the minimum and maximum heap size to 4000 MB.
|
<2> Set the minimum and maximum heap size to 4000 MB.
|
||||||
|
|
||||||
NOTE: Configuring the heap for the <<windows-service,Windows service>>
|
NOTE: Configuring the heap for the <<windows-service,Windows service>>
|
||||||
is different than the above. The values initiallly populated for the
|
is different than the above. The values initially populated for the
|
||||||
Windows service can be configured as above but are different after the
|
Windows service can be configured as above but are different after the
|
||||||
service has been installed. Consult the
|
service has been installed. Consult the
|
||||||
<<windows-service,Windows service documentation>> for additional
|
<<windows-service,Windows service documentation>> for additional
|
||||||
|
@ -253,7 +253,7 @@ Usually, you would combine assertions and matchers in your test like this
|
|||||||
|
|
||||||
[source,java]
|
[source,java]
|
||||||
----------------------------
|
----------------------------
|
||||||
SearchResponse seearchResponse = client().prepareSearch() ...;
|
SearchResponse searchResponse = client().prepareSearch() ...;
|
||||||
assertHitCount(searchResponse, 4);
|
assertHitCount(searchResponse, 4);
|
||||||
assertFirstHit(searchResponse, hasId("4"));
|
assertFirstHit(searchResponse, hasId("4"));
|
||||||
assertSearchHits(searchResponse, "1", "2", "3", "4");
|
assertSearchHits(searchResponse, "1", "2", "3", "4");
|
||||||
|
Loading…
x
Reference in New Issue
Block a user