* [ML][Inference] document new settings * [DOCS] Minor edits
This commit is contained in:
parent
62a3154d0e
commit
d068818b16
|
@ -50,6 +50,25 @@ IMPORTANT: If you want to use {ml} features in your cluster, you must have
|
|||
`xpack.ml.enabled` set to `true` on all master-eligible nodes. This is the
|
||||
default behavior.
|
||||
|
||||
`xpack.ml.inference_model.cache_size`::
|
||||
The maximum inference cache size allowed. The inference cache exists in the JVM
|
||||
heap on each ingest node. The cache affords faster processing times for the
|
||||
`inference` processor. The value can be a static byte sized value (i.e. "2gb")
|
||||
or a percentage of total allocated heap. The default is "40%".
|
||||
|
||||
`xpack.ml.inference_model.time_to_live`::
|
||||
The time to live (TTL) for models in the inference model cache. The TTL is
|
||||
calculated from last access. The `inference` processor attempts to load the
|
||||
model from cache. If the `inference` processor does not receive any documents
|
||||
for the duration of the TTL, the referenced model is flagged for eviction from
|
||||
the cache. If a document is processed later, the model is again loaded into the
|
||||
cache. Defaults to `5m`.
|
||||
|
||||
`xpack.ml.max_inference_processors` (<<cluster-update-settings,Dynamic>>)::
|
||||
The total number of `inference` type processors allowed across all ingest
|
||||
pipelines. Once the limit is reached, adding an `inference` processor to
|
||||
a pipeline is disallowed. Defaults to `50`.
|
||||
|
||||
`xpack.ml.max_machine_memory_percent` (<<cluster-update-settings,Dynamic>>)::
|
||||
The maximum percentage of the machine's memory that {ml} may use for running
|
||||
analytics processes. (These processes are separate to the {es} JVM.) Defaults to
|
||||
|
|
Loading…
Reference in New Issue