Docs: Updated the experimental annotations in the docs as follows:
* Removed the docs for `index.compound_format` and `index.compound_on_flush` - these are expert settings which should probably be removed (see https://github.com/elastic/elasticsearch/issues/10778) * Removed the docs for `index.index_concurrency` - another expert setting * Labelled the segments verbose output as experimental * Marked the `compression`, `precision_threshold` and `rehash` options as experimental in the cardinality and percentile aggs * Improved the experimental text on `significant_terms`, `execution_hint` in the terms agg, and `terminate_after` param on count and search * Removed the experimental flag on the `geobounds` agg * Marked the settings in the `merge` and `store` modules as experimental, rather than the modules themselves Closes #10782
This commit is contained in:
parent
f1a0e2216a
commit
37ed61807f
|
@ -15,29 +15,6 @@ all the relevant modules settings can be provided when creating an index
|
||||||
There are specific index level settings that are not associated with any
|
There are specific index level settings that are not associated with any
|
||||||
specific module. These include:
|
specific module. These include:
|
||||||
|
|
||||||
[[index-compound-format]]`index.compound_format`::
|
|
||||||
|
|
||||||
experimental[]
|
|
||||||
Should the compound file format be used (boolean setting).
|
|
||||||
The compound format was created to reduce the number of open
|
|
||||||
file handles when using file based storage. However, by default it is set
|
|
||||||
to `false` as the non-compound format gives better performance. It is important
|
|
||||||
that OS is configured to give Elasticsearch ``enough'' file handles.
|
|
||||||
See <<file-descriptors>>.
|
|
||||||
+
|
|
||||||
Alternatively, `compound_format` can be set to a number between `0` and
|
|
||||||
`1`, where `0` means `false`, `1` means `true` and a number inbetween
|
|
||||||
represents a percentage: if the merged segment is less than this
|
|
||||||
percentage of the total index, then it is written in compound format,
|
|
||||||
otherwise it is written in non-compound format.
|
|
||||||
|
|
||||||
[[index-compound-on-flush]]`index.compound_on_flush`::
|
|
||||||
|
|
||||||
experimental[]
|
|
||||||
Should a new segment (create by indexing, not by merging) be written
|
|
||||||
in compound format or non-compound format? Defaults to `true`.
|
|
||||||
This is a dynamic setting.
|
|
||||||
|
|
||||||
`index.refresh_interval`::
|
`index.refresh_interval`::
|
||||||
A time setting controlling how often the
|
A time setting controlling how often the
|
||||||
refresh operation will be executed. Defaults to `1s`. Can be set to `-1`
|
refresh operation will be executed. Defaults to `1s`. Can be set to `-1`
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
[[index-modules-merge]]
|
[[index-modules-merge]]
|
||||||
== Merge
|
== Merge
|
||||||
|
|
||||||
experimental[]
|
experimental[All of the settings exposed in the `merge` module are expert only and may be removed in the future]
|
||||||
|
|
||||||
A shard in elasticsearch is a Lucene index, and a Lucene index is broken
|
A shard in elasticsearch is a Lucene index, and a Lucene index is broken
|
||||||
down into segments. Segments are internal storage elements in the index
|
down into segments. Segments are internal storage elements in the index
|
||||||
|
@ -72,12 +72,6 @@ This policy has the following settings:
|
||||||
Higher values favor selecting merges that reclaim deletions. A value of
|
Higher values favor selecting merges that reclaim deletions. A value of
|
||||||
`0.0` means deletions don't impact merge selection. Defaults to `2.0`.
|
`0.0` means deletions don't impact merge selection. Defaults to `2.0`.
|
||||||
|
|
||||||
`index.compound_format`::
|
|
||||||
|
|
||||||
Should the index be stored in compound format or not. Defaults to `false`.
|
|
||||||
See <<index-compound-format,`index.compound_format`>> in
|
|
||||||
<<index-modules-settings>>.
|
|
||||||
|
|
||||||
For normal merging, this policy first computes a "budget" of how many
|
For normal merging, this policy first computes a "budget" of how many
|
||||||
segments are allowed to be in the index. If the index is over-budget,
|
segments are allowed to be in the index. If the index is over-budget,
|
||||||
then the policy sorts segments by decreasing size (proportionally considering percent
|
then the policy sorts segments by decreasing size (proportionally considering percent
|
||||||
|
|
|
@ -1,8 +1,6 @@
|
||||||
[[index-modules-store]]
|
[[index-modules-store]]
|
||||||
== Store
|
== Store
|
||||||
|
|
||||||
experimental[]
|
|
||||||
|
|
||||||
The store module allows you to control how index data is stored.
|
The store module allows you to control how index data is stored.
|
||||||
|
|
||||||
The index can either be stored in-memory (no persistence) or on-disk
|
The index can either be stored in-memory (no persistence) or on-disk
|
||||||
|
@ -20,6 +18,7 @@ heap space* using the "Memory" (see below) storage type. It translates
|
||||||
to the fact that there is no need for extra large JVM heaps (with their
|
to the fact that there is no need for extra large JVM heaps (with their
|
||||||
own consequences) for storing the index in memory.
|
own consequences) for storing the index in memory.
|
||||||
|
|
||||||
|
experimental[All of the settings exposed in the `store` module are expert only and may be removed in the future]
|
||||||
|
|
||||||
[float]
|
[float]
|
||||||
[[file-system]]
|
[[file-system]]
|
||||||
|
|
|
@ -78,7 +78,7 @@ compound:: Whether the segment is stored in a compound file. When true, this
|
||||||
|
|
||||||
To add additional information that can be used for debugging, use the `verbose` flag.
|
To add additional information that can be used for debugging, use the `verbose` flag.
|
||||||
|
|
||||||
NOTE: The format of additional verbose information is experimental and can change at any time.
|
experimental[The format of the additional verbose information is experimental and can change at any time]
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
|
|
@ -61,9 +61,6 @@ settings API:
|
||||||
`index.refresh_interval`::
|
`index.refresh_interval`::
|
||||||
The async refresh interval of a shard.
|
The async refresh interval of a shard.
|
||||||
|
|
||||||
`index.index_concurrency`::
|
|
||||||
experimental[] Defaults to `8`.
|
|
||||||
|
|
||||||
`index.translog.flush_threshold_ops`::
|
`index.translog.flush_threshold_ops`::
|
||||||
When to flush based on operations.
|
When to flush based on operations.
|
||||||
|
|
||||||
|
@ -151,14 +148,6 @@ settings API:
|
||||||
`index.translog.fs.type`::
|
`index.translog.fs.type`::
|
||||||
experimental[] Either `simple` or `buffered` (default).
|
experimental[] Either `simple` or `buffered` (default).
|
||||||
|
|
||||||
`index.compound_format`::
|
|
||||||
experimental[] See <<index-compound-format,`index.compound_format`>> in
|
|
||||||
<<index-modules-settings>>.
|
|
||||||
|
|
||||||
`index.compound_on_flush`::
|
|
||||||
experimental[] See <<index-compound-on-flush,`index.compound_on_flush>> in
|
|
||||||
<<index-modules-settings>>.
|
|
||||||
|
|
||||||
<<index-modules-slowlog>>::
|
<<index-modules-slowlog>>::
|
||||||
All the settings for slow log.
|
All the settings for slow log.
|
||||||
|
|
||||||
|
|
|
@ -424,10 +424,7 @@ automatically loaded.
|
||||||
[float]
|
[float]
|
||||||
=== Lucene Expressions Scripts
|
=== Lucene Expressions Scripts
|
||||||
|
|
||||||
[WARNING]
|
experimental[The Lucene expressions module is undergoing significant development and the exposed functionality is likely to change in the future]
|
||||||
========================
|
|
||||||
This feature is *experimental* and subject to change in future versions.
|
|
||||||
========================
|
|
||||||
|
|
||||||
Lucene's expressions module provides a mechanism to compile a
|
Lucene's expressions module provides a mechanism to compile a
|
||||||
`javascript` expression to bytecode. This allows very fast execution,
|
`javascript` expression to bytecode. This allows very fast execution,
|
||||||
|
|
|
@ -3,7 +3,7 @@
|
||||||
|
|
||||||
An aggregation that returns interesting or unusual occurrences of terms in a set.
|
An aggregation that returns interesting or unusual occurrences of terms in a set.
|
||||||
|
|
||||||
experimental[]
|
experimental[The `significant_terms` aggregation can be very heavy when run on large indices. Work is in progress to provide more lightweight sampling techniques. As a result, the API for this feature may change in non-backwards compatible ways]
|
||||||
|
|
||||||
.Example use cases:
|
.Example use cases:
|
||||||
* Suggesting "H5N1" when users search for "bird flu" in text
|
* Suggesting "H5N1" when users search for "bird flu" in text
|
||||||
|
|
|
@ -613,7 +613,7 @@ this would typically be too costly in terms of RAM.
|
||||||
[[search-aggregations-bucket-terms-aggregation-execution-hint]]
|
[[search-aggregations-bucket-terms-aggregation-execution-hint]]
|
||||||
==== Execution hint
|
==== Execution hint
|
||||||
|
|
||||||
experimental[]
|
experimental[The automated execution optimization is experimental, so this parameter is provided temporarily as a way to override the default behaviour]
|
||||||
|
|
||||||
There are different mechanisms by which terms aggregations can be executed:
|
There are different mechanisms by which terms aggregations can be executed:
|
||||||
|
|
||||||
|
|
|
@ -23,10 +23,10 @@ match a query:
|
||||||
|
|
||||||
==== Precision control
|
==== Precision control
|
||||||
|
|
||||||
experimental[]
|
|
||||||
|
|
||||||
This aggregation also supports the `precision_threshold` and `rehash` options:
|
This aggregation also supports the `precision_threshold` and `rehash` options:
|
||||||
|
|
||||||
|
experimental[The `precision_threshold` and `rehash` options are specific to the current internal implementation of the `cardinality` agg, which may change in the future]
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
{
|
{
|
||||||
|
@ -42,14 +42,14 @@ This aggregation also supports the `precision_threshold` and `rehash` options:
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
|
||||||
<1> experimental[] The `precision_threshold` options allows to trade memory for accuracy, and
|
<1> The `precision_threshold` options allows to trade memory for accuracy, and
|
||||||
defines a unique count below which counts are expected to be close to
|
defines a unique count below which counts are expected to be close to
|
||||||
accurate. Above this value, counts might become a bit more fuzzy. The maximum
|
accurate. Above this value, counts might become a bit more fuzzy. The maximum
|
||||||
supported value is 40000, thresholds above this number will have the same
|
supported value is 40000, thresholds above this number will have the same
|
||||||
effect as a threshold of 40000.
|
effect as a threshold of 40000.
|
||||||
Default value depends on the number of parent aggregations that multiple
|
Default value depends on the number of parent aggregations that multiple
|
||||||
create buckets (such as terms or histograms).
|
create buckets (such as terms or histograms).
|
||||||
<2> experimental[] If you computed a hash on client-side, stored it into your documents and want
|
<2> If you computed a hash on client-side, stored it into your documents and want
|
||||||
Elasticsearch to use them to compute counts using this hash function without
|
Elasticsearch to use them to compute counts using this hash function without
|
||||||
rehashing values, it is possible to specify `rehash: false`. Default value is
|
rehashing values, it is possible to specify `rehash: false`. Default value is
|
||||||
`true`. Please note that the hash must be indexed as a long when `rehash` is
|
`true`. Please note that the hash must be indexed as a long when `rehash` is
|
||||||
|
|
|
@ -1,8 +1,6 @@
|
||||||
[[search-aggregations-metrics-geobounds-aggregation]]
|
[[search-aggregations-metrics-geobounds-aggregation]]
|
||||||
=== Geo Bounds Aggregation
|
=== Geo Bounds Aggregation
|
||||||
|
|
||||||
experimental[]
|
|
||||||
|
|
||||||
A metric aggregation that computes the bounding box containing all geo_point values for a field.
|
A metric aggregation that computes the bounding box containing all geo_point values for a field.
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -155,7 +155,7 @@ it. It would not be the case on more skewed distributions.
|
||||||
[[search-aggregations-metrics-percentile-aggregation-compression]]
|
[[search-aggregations-metrics-percentile-aggregation-compression]]
|
||||||
==== Compression
|
==== Compression
|
||||||
|
|
||||||
experimental[]
|
experimental[The `compression` parameter is specific to the current internal implementation of percentiles, and may change in the future]
|
||||||
|
|
||||||
Approximate algorithms must balance memory utilization with estimation accuracy.
|
Approximate algorithms must balance memory utilization with estimation accuracy.
|
||||||
This balance can be controlled using a `compression` parameter:
|
This balance can be controlled using a `compression` parameter:
|
||||||
|
|
|
@ -64,7 +64,8 @@ query.
|
||||||
|default_operator |The default operator to be used, can be `AND` or
|
|default_operator |The default operator to be used, can be `AND` or
|
||||||
`OR`. Defaults to `OR`.
|
`OR`. Defaults to `OR`.
|
||||||
|
|
||||||
|terminate_after |experimental[] The maximum count for each shard, upon
|
|terminate_after |experimental[The API for this feature may change in the future]
|
||||||
|
The maximum count for each shard, upon
|
||||||
reaching which the query execution will terminate early.
|
reaching which the query execution will terminate early.
|
||||||
If set, the response will have a boolean field `terminated_early` to
|
If set, the response will have a boolean field `terminated_early` to
|
||||||
indicate whether the query execution has actually terminated_early.
|
indicate whether the query execution has actually terminated_early.
|
||||||
|
|
|
@ -77,7 +77,8 @@ And here is a sample response:
|
||||||
|
|
||||||
`terminate_after`::
|
`terminate_after`::
|
||||||
|
|
||||||
experimental[] The maximum number of documents to collect for each shard,
|
experimental[The API for this feature may change in the future]
|
||||||
|
The maximum number of documents to collect for each shard,
|
||||||
upon reaching which the query execution will terminate early. If set, the
|
upon reaching which the query execution will terminate early. If set, the
|
||||||
response will have a boolean field `terminated_early` to indicate whether
|
response will have a boolean field `terminated_early` to indicate whether
|
||||||
the query execution has actually terminated_early. Defaults to no
|
the query execution has actually terminated_early. Defaults to no
|
||||||
|
|
|
@ -82,7 +82,8 @@ scores and return them as part of each hit.
|
||||||
within the specified time value and bail with the hits accumulated up to
|
within the specified time value and bail with the hits accumulated up to
|
||||||
that point when expired. Defaults to no timeout.
|
that point when expired. Defaults to no timeout.
|
||||||
|
|
||||||
|`terminate_after` |experimental[] The maximum number of documents to collect for
|
|`terminate_after` |experimental[The API for this feature may change in the future]
|
||||||
|
The maximum number of documents to collect for
|
||||||
each shard, upon reaching which the query execution will terminate early.
|
each shard, upon reaching which the query execution will terminate early.
|
||||||
If set, the response will have a boolean field `terminated_early` to
|
If set, the response will have a boolean field `terminated_early` to
|
||||||
indicate whether the query execution has actually terminated_early.
|
indicate whether the query execution has actually terminated_early.
|
||||||
|
|
Loading…
Reference in New Issue