2019-12-24 13:22:05 -05:00
|
|
|
|
tag::aggregations[]
|
|
|
|
|
If set, the {dfeed} performs aggregation searches. Support for aggregations is
|
2020-01-22 11:08:31 -05:00
|
|
|
|
limited and should be used only with low cardinality data. For more information,
|
2019-12-24 13:22:05 -05:00
|
|
|
|
see
|
|
|
|
|
{ml-docs}/ml-configuring-aggregation.html[Aggregating data for faster performance].
|
|
|
|
|
end::aggregations[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::allow-lazy-open[]
|
|
|
|
|
Advanced configuration option. Specifies whether this job can open when there is
|
|
|
|
|
insufficient {ml} node capacity for it to be immediately assigned to a node. The
|
|
|
|
|
default value is `false`; if a {ml} node with capacity to run the job cannot
|
|
|
|
|
immediately be found, the <<ml-open-job,open {anomaly-jobs} API>> returns an
|
|
|
|
|
error. However, this is also subject to the cluster-wide
|
|
|
|
|
`xpack.ml.max_lazy_ml_nodes` setting; see <<advanced-ml-settings>>. If this
|
|
|
|
|
option is set to `true`, the <<ml-open-job,open {anomaly-jobs} API>> does not
|
|
|
|
|
return an error and the job waits in the `opening` state until sufficient {ml}
|
|
|
|
|
node capacity is available.
|
|
|
|
|
end::allow-lazy-open[]
|
|
|
|
|
|
|
|
|
|
tag::allow-lazy-start[]
|
|
|
|
|
Whether this job should be allowed to start when there is insufficient {ml} node
|
|
|
|
|
capacity for it to be immediately assigned to a node. The default is `false`,
|
|
|
|
|
which means that the <<start-dfanalytics>> will return an error if a {ml} node
|
|
|
|
|
with capacity to run the job cannot immediately be found. (However, this is also
|
|
|
|
|
subject to the cluster-wide `xpack.ml.max_lazy_ml_nodes` setting - see
|
|
|
|
|
<<advanced-ml-settings>>.) If this option is set to `true` then the
|
|
|
|
|
<<start-dfanalytics>> will not return an error, and the job will wait in the
|
|
|
|
|
`starting` state until sufficient {ml} node capacity is available.
|
|
|
|
|
end::allow-lazy-start[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::allow-no-datafeeds[]
|
|
|
|
|
Specifies what to do when the request:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* Contains wildcard expressions and there are no {dfeeds} that match.
|
|
|
|
|
* Contains the `_all` string or no identifiers and there are no matches.
|
|
|
|
|
* Contains wildcard expressions and there are only partial matches.
|
|
|
|
|
|
|
|
|
|
The default value is `true`, which returns an empty `datafeeds` array when
|
|
|
|
|
there are no matches and the subset of results when there are partial matches.
|
|
|
|
|
If this parameter is `false`, the request returns a `404` status code when there
|
|
|
|
|
are no matches or only partial matches.
|
|
|
|
|
--
|
|
|
|
|
end::allow-no-datafeeds[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::allow-no-jobs[]
|
|
|
|
|
Specifies what to do when the request:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* Contains wildcard expressions and there are no jobs that match.
|
|
|
|
|
* Contains the `_all` string or no identifiers and there are no matches.
|
|
|
|
|
* Contains wildcard expressions and there are only partial matches.
|
|
|
|
|
|
|
|
|
|
The default value is `true`, which returns an empty `jobs` array
|
|
|
|
|
when there are no matches and the subset of results when there are partial
|
|
|
|
|
matches. If this parameter is `false`, the request returns a `404` status code
|
|
|
|
|
when there are no matches or only partial matches.
|
|
|
|
|
--
|
|
|
|
|
end::allow-no-jobs[]
|
|
|
|
|
|
|
|
|
|
tag::allow-no-match[]
|
|
|
|
|
Specifies what to do when the request:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* Contains wildcard expressions and there are no {dfanalytics-jobs} that match.
|
|
|
|
|
* Contains the `_all` string or no identifiers and there are no matches.
|
|
|
|
|
* Contains wildcard expressions and there are only partial matches.
|
|
|
|
|
|
|
|
|
|
The default value is `true`, which returns an empty `data_frame_analytics` array
|
|
|
|
|
when there are no matches and the subset of results when there are partial
|
|
|
|
|
matches. If this parameter is `false`, the request returns a `404` status code
|
|
|
|
|
when there are no matches or only partial matches.
|
|
|
|
|
--
|
|
|
|
|
end::allow-no-match[]
|
|
|
|
|
|
|
|
|
|
tag::analysis[]
|
|
|
|
|
Defines the type of {dfanalytics} you want to perform on your source index. For
|
|
|
|
|
example: `outlier_detection`. See <<ml-dfa-analysis-objects>>.
|
|
|
|
|
end::analysis[]
|
|
|
|
|
|
|
|
|
|
tag::analysis-config[]
|
2019-12-24 13:22:05 -05:00
|
|
|
|
The analysis configuration, which specifies how to analyze the data. After you
|
|
|
|
|
create a job, you cannot change the analysis configuration; all the properties
|
|
|
|
|
are informational.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::analysis-config[]
|
|
|
|
|
|
|
|
|
|
tag::analysis-limits[]
|
|
|
|
|
Limits can be applied for the resources required to hold the mathematical models
|
|
|
|
|
in memory. These limits are approximate and can be set per job. They do not
|
2019-12-24 13:22:05 -05:00
|
|
|
|
control the memory used by other processes, for example the {es} Java processes.
|
|
|
|
|
If necessary, you can increase the limits after the job is created.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::analysis-limits[]
|
|
|
|
|
|
|
|
|
|
tag::analyzed-fields[]
|
|
|
|
|
Specify `includes` and/or `excludes` patterns to select which fields will be
|
2020-01-24 07:31:44 -05:00
|
|
|
|
included in the analysis. The patterns specified in `excludes` are applied last,
|
|
|
|
|
therefore `excludes` takes precedence. In other words, if the same field is
|
|
|
|
|
specified in both `includes` and `excludes`, then the field will not be included
|
|
|
|
|
in the analysis.
|
2020-01-09 10:21:35 -05:00
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
The supported fields for each type of analysis are as follows:
|
|
|
|
|
|
|
|
|
|
* {oldetection-cap} requires numeric or boolean data to analyze. The algorithms
|
|
|
|
|
don't support missing values therefore fields that have data types other than
|
|
|
|
|
numeric or boolean are ignored. Documents where included fields contain missing
|
|
|
|
|
values, null values, or an array are also ignored. Therefore the `dest` index
|
|
|
|
|
may contain documents that don't have an {olscore}.
|
|
|
|
|
* {regression-cap} supports fields that are numeric, `boolean`, `text`,
|
|
|
|
|
`keyword`, and `ip`. It is also tolerant of missing values. Fields that are
|
|
|
|
|
supported are included in the analysis, other fields are ignored. Documents
|
|
|
|
|
where included fields contain an array with two or more values are also
|
|
|
|
|
ignored. Documents in the `dest` index that don’t contain a results field are
|
|
|
|
|
not included in the {reganalysis}.
|
|
|
|
|
* {classification-cap} supports fields that are numeric, `boolean`, `text`,
|
|
|
|
|
`keyword`, and `ip`. It is also tolerant of missing values. Fields that are
|
|
|
|
|
supported are included in the analysis, other fields are ignored. Documents
|
|
|
|
|
where included fields contain an array with two or more values are also ignored.
|
|
|
|
|
Documents in the `dest` index that don’t contain a results field are not
|
|
|
|
|
included in the {classanalysis}. {classanalysis-cap} can be improved by mapping
|
|
|
|
|
ordinal variable values to a single number. For example, in case of age ranges,
|
|
|
|
|
you can model the values as "0-14" = 0, "15-24" = 1, "25-34" = 2, and so on.
|
|
|
|
|
If `analyzed_fields` is not set, only the relevant fields will be included. For
|
|
|
|
|
example, all the numeric fields for {oldetection}. For more information about
|
|
|
|
|
field selection, see <<explain-dfanalytics>>.
|
|
|
|
|
--
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::analyzed-fields[]
|
|
|
|
|
|
2020-01-09 10:21:35 -05:00
|
|
|
|
tag::analyzed-fields-excludes[]
|
|
|
|
|
An array of strings that defines the fields that will be excluded from the
|
|
|
|
|
analysis. You do not need to add fields with unsupported data types to
|
|
|
|
|
`excludes`, these fields are excluded from the analysis automatically.
|
|
|
|
|
end::analyzed-fields-excludes[]
|
|
|
|
|
|
|
|
|
|
tag::analyzed-fields-includes[]
|
2020-03-03 10:25:24 -05:00
|
|
|
|
An array of strings that defines the fields that will be included in the
|
|
|
|
|
analysis.
|
2020-01-09 10:21:35 -05:00
|
|
|
|
end::analyzed-fields-includes[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::assignment-explanation-anomaly-jobs[]
|
|
|
|
|
For open {anomaly-jobs} only, contains messages relating to the selection
|
|
|
|
|
of a node to run the job.
|
|
|
|
|
end::assignment-explanation-anomaly-jobs[]
|
|
|
|
|
|
|
|
|
|
tag::assignment-explanation-datafeeds[]
|
2020-03-05 10:25:43 -05:00
|
|
|
|
For started {dfeeds} only, contains messages relating to the selection of a
|
|
|
|
|
node.
|
2020-03-02 10:28:55 -05:00
|
|
|
|
end::assignment-explanation-datafeeds[]
|
2020-02-26 12:20:36 -05:00
|
|
|
|
|
2020-02-26 05:09:37 -05:00
|
|
|
|
tag::assignment-explanation-dfanalytics[]
|
|
|
|
|
Contains messages relating to the selection of a node.
|
|
|
|
|
end::assignment-explanation-dfanalytics[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::background-persist-interval[]
|
|
|
|
|
Advanced configuration option. The time between each periodic persistence of the
|
|
|
|
|
model. The default value is a randomized value between 3 to 4 hours, which
|
|
|
|
|
avoids all jobs persisting at exactly the same time. The smallest allowed value
|
|
|
|
|
is 1 hour.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
TIP: For very large models (several GB), persistence could take 10-20 minutes,
|
|
|
|
|
so do not set the `background_persist_interval` value too low.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::background-persist-interval[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::bucket-allocation-failures-count[]
|
|
|
|
|
The number of buckets for which new entities in incoming data were not processed
|
|
|
|
|
due to insufficient model memory. This situation is also signified by a
|
|
|
|
|
`hard_limit: memory_status` property value.
|
|
|
|
|
end::bucket-allocation-failures-count[]
|
|
|
|
|
|
2020-02-26 12:20:36 -05:00
|
|
|
|
tag::bucket-count[]
|
|
|
|
|
The number of buckets processed.
|
|
|
|
|
end::bucket-count[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::bucket-count-anomaly-jobs[]
|
|
|
|
|
The number of bucket results produced by the job.
|
|
|
|
|
end::bucket-count-anomaly-jobs[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::bucket-span[]
|
|
|
|
|
The size of the interval that the analysis is aggregated into, typically between
|
2020-01-22 11:08:31 -05:00
|
|
|
|
`5m` and `1h`. The default value is `5m`. If the {anomaly-job} uses a {dfeed}
|
|
|
|
|
with {ml-docs}/ml-configuring-aggregation.html[aggregations], this value must be
|
|
|
|
|
divisible by the interval of the date histogram aggregation. For more
|
|
|
|
|
information, see {ml-docs}/ml-buckets.html[Buckets].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::bucket-span[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::bucket-span-results[]
|
|
|
|
|
The length of the bucket in seconds. This value matches the `bucket_span`
|
|
|
|
|
that is specified in the job.
|
|
|
|
|
end::bucket-span-results[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::bucket-time-exponential-average[]
|
|
|
|
|
Exponential moving average of all bucket processing times, in milliseconds.
|
|
|
|
|
end::bucket-time-exponential-average[]
|
|
|
|
|
|
|
|
|
|
tag::bucket-time-exponential-average-hour[]
|
|
|
|
|
Exponentially-weighted moving average of bucket processing times
|
|
|
|
|
calculated in a 1 hour time window, in milliseconds.
|
|
|
|
|
end::bucket-time-exponential-average-hour[]
|
|
|
|
|
|
|
|
|
|
tag::bucket-time-maximum[]
|
|
|
|
|
Maximum among all bucket processing times, in milliseconds.
|
|
|
|
|
end::bucket-time-maximum[]
|
|
|
|
|
|
|
|
|
|
tag::bucket-time-minimum[]
|
|
|
|
|
Minimum among all bucket processing times, in milliseconds.
|
|
|
|
|
end::bucket-time-minimum[]
|
|
|
|
|
|
|
|
|
|
tag::bucket-time-total[]
|
|
|
|
|
Sum of all bucket processing times, in milliseconds.
|
|
|
|
|
end::bucket-time-total[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::by-field-name[]
|
|
|
|
|
The field used to split the data. In particular, this property is used for
|
|
|
|
|
analyzing the splits with respect to their own history. It is used for finding
|
|
|
|
|
unusual values in the context of the split.
|
|
|
|
|
end::by-field-name[]
|
|
|
|
|
|
2020-01-02 14:23:26 -05:00
|
|
|
|
tag::calendar-id[]
|
|
|
|
|
A string that uniquely identifies a calendar.
|
|
|
|
|
end::calendar-id[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::categorization-analyzer[]
|
|
|
|
|
If `categorization_field_name` is specified, you can also define the analyzer
|
|
|
|
|
that is used to interpret the categorization field. This property cannot be used
|
|
|
|
|
at the same time as `categorization_filters`. The categorization analyzer
|
2019-12-27 16:30:26 -05:00
|
|
|
|
specifies how the categorization field is interpreted by the categorization
|
2019-12-13 05:48:21 -05:00
|
|
|
|
process. The syntax is very similar to that used to define the `analyzer` in the
|
|
|
|
|
<<indices-analyze,Analyze endpoint>>. For more information, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-categories.html[Categorizing log messages].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
The `categorization_analyzer` field can be specified either as a string or as an
|
|
|
|
|
object. If it is a string it must refer to a
|
|
|
|
|
<<analysis-analyzers,built-in analyzer>> or one added by another plugin. If it
|
|
|
|
|
is an object it has the following properties:
|
|
|
|
|
--
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`analysis_config`.`categorization_analyzer`.`char_filter`::::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(array of strings or objects)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=char-filter]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`analysis_config`.`categorization_analyzer`.`tokenizer`::::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(string or object)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=tokenizer]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`analysis_config`.`categorization_analyzer`.`filter`::::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(array of strings or objects)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=filter]
|
|
|
|
|
end::categorization-analyzer[]
|
|
|
|
|
|
|
|
|
|
tag::categorization-examples-limit[]
|
|
|
|
|
The maximum number of examples stored per category in memory and in the results
|
2019-12-27 16:30:26 -05:00
|
|
|
|
data store. The default value is `4`. If you increase this value, more examples
|
2019-12-13 05:48:21 -05:00
|
|
|
|
are available, however it requires that you have more storage available. If you
|
|
|
|
|
set this value to `0`, no examples are stored.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: The `categorization_examples_limit` only applies to analysis that uses
|
|
|
|
|
categorization. For more information, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-categories.html[Categorizing log messages].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::categorization-examples-limit[]
|
|
|
|
|
|
|
|
|
|
tag::categorization-field-name[]
|
|
|
|
|
If this property is specified, the values of the specified field will be
|
|
|
|
|
categorized. The resulting categories must be used in a detector by setting
|
|
|
|
|
`by_field_name`, `over_field_name`, or `partition_field_name` to the keyword
|
|
|
|
|
`mlcategory`. For more information, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-categories.html[Categorizing log messages].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::categorization-field-name[]
|
|
|
|
|
|
|
|
|
|
tag::categorization-filters[]
|
|
|
|
|
If `categorization_field_name` is specified, you can also define optional
|
|
|
|
|
filters. This property expects an array of regular expressions. The expressions
|
|
|
|
|
are used to filter out matching sequences from the categorization field values.
|
|
|
|
|
You can use this functionality to fine tune the categorization by excluding
|
|
|
|
|
sequences from consideration when categories are defined. For example, you can
|
|
|
|
|
exclude SQL statements that appear in your log files. For more information, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-categories.html[Categorizing log messages]. This
|
2019-12-13 05:48:21 -05:00
|
|
|
|
property cannot be used at the same time as `categorization_analyzer`. If you
|
|
|
|
|
only want to define simple regular expression filters that are applied prior to
|
|
|
|
|
tokenization, setting this property is the easiest method. If you also want to
|
|
|
|
|
customize the tokenizer or post-tokenization filtering, use the
|
|
|
|
|
`categorization_analyzer` property instead and include the filters as
|
|
|
|
|
`pattern_replace` character filters. The effect is exactly the same.
|
|
|
|
|
end::categorization-filters[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::categorization-status[]
|
|
|
|
|
The status of categorization for the job. Contains one of the following values:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* `ok`: Categorization is performing acceptably well (or not being used at all).
|
|
|
|
|
* `warn`: Categorization is detecting a distribution of categories that suggests
|
|
|
|
|
the input data is inappropriate for categorization. Problems could be that there
|
|
|
|
|
is only one category, more than 90% of categories are rare, the number of
|
|
|
|
|
categories is greater than 50% of the number of categorized documents, there are
|
|
|
|
|
no frequently matched categories, or more than 50% of categories are dead.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::categorization-status[]
|
|
|
|
|
|
|
|
|
|
tag::categorized-doc-count[]
|
|
|
|
|
The number of documents that have had a field categorized.
|
|
|
|
|
end::categorized-doc-count[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::char-filter[]
|
|
|
|
|
One or more <<analysis-charfilters,character filters>>. In addition to the
|
|
|
|
|
built-in character filters, other plugins can provide more character filters.
|
|
|
|
|
This property is optional. If it is not specified, no character filters are
|
|
|
|
|
applied prior to categorization. If you are customizing some other aspect of the
|
|
|
|
|
analyzer and you need to achieve the equivalent of `categorization_filters`
|
|
|
|
|
(which are not permitted when some other aspect of the analyzer is customized),
|
|
|
|
|
add them here as
|
|
|
|
|
<<analysis-pattern-replace-charfilter,pattern replace character filters>>.
|
|
|
|
|
end::char-filter[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::chunking-config[]
|
2020-03-05 10:25:43 -05:00
|
|
|
|
{dfeeds-cap} might be required to search over long time periods, for several
|
|
|
|
|
months or years. This search is split into time chunks in order to ensure the
|
|
|
|
|
load on {es} is managed. Chunking configuration controls how the size of these
|
|
|
|
|
time chunks are calculated and is an advanced configuration option.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
A chunking configuration object has the following properties:
|
|
|
|
|
|
|
|
|
|
`chunking_config`.`mode`:::
|
|
|
|
|
(string)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=mode]
|
|
|
|
|
|
|
|
|
|
`chunking_config`.`time_span`:::
|
|
|
|
|
(<<time-units,time units>>)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=time-span]
|
|
|
|
|
end::chunking-config[]
|
|
|
|
|
|
2020-03-13 13:35:51 -04:00
|
|
|
|
tag::class-assignment-objective[]
|
|
|
|
|
Defines the objective to optimize when assigning class labels. Available
|
|
|
|
|
objectives are `maximize_accuracy` and `maximize_minimum_recall`. When maximizing
|
|
|
|
|
accuracy class labels are chosen to maximize the number of correct predictions.
|
|
|
|
|
When maximizing minimum recall labels are chosen to maximize the minimum recall
|
|
|
|
|
for any class. Defaults to maximize_minimum_recall.
|
|
|
|
|
end::class-assignment-objective[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::compute-feature-influence[]
|
|
|
|
|
If `true`, the feature influence calculation is enabled. Defaults to `true`.
|
|
|
|
|
end::compute-feature-influence[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules[]
|
|
|
|
|
An array of custom rule objects, which enable you to customize the way detectors
|
|
|
|
|
operate. For example, a rule may dictate to the detector conditions under which
|
|
|
|
|
results should be skipped. For more examples, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-detector-custom-rules.html[Customizing detectors with custom rules].
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::custom-rules[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-actions[]
|
|
|
|
|
The set of actions to be triggered when the rule applies. If
|
2019-12-13 05:48:21 -05:00
|
|
|
|
more than one action is specified the effects of all actions are combined. The
|
|
|
|
|
available actions include:
|
|
|
|
|
|
|
|
|
|
* `skip_result`: The result will not be created. This is the default value.
|
|
|
|
|
Unless you also specify `skip_model_update`, the model will be updated as usual
|
|
|
|
|
with the corresponding series value.
|
|
|
|
|
* `skip_model_update`: The value for that series will not be used to update the
|
|
|
|
|
model. Unless you also specify `skip_result`, the results will be created as
|
|
|
|
|
usual. This action is suitable when certain values are expected to be
|
|
|
|
|
consistently anomalous and they affect the model in a way that negatively
|
|
|
|
|
impacts the rest of the results.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::custom-rules-actions[]
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::custom-rules-scope[]
|
|
|
|
|
An optional scope of series where the rule applies. A rule must either
|
2019-12-13 05:48:21 -05:00
|
|
|
|
have a non-empty scope or at least one condition. By default, the scope includes
|
|
|
|
|
all series. Scoping is allowed for any of the fields that are also specified in
|
|
|
|
|
`by_field_name`, `over_field_name`, or `partition_field_name`. To add a scope
|
|
|
|
|
for a field, add the field name as a key in the scope object and set its value
|
|
|
|
|
to an object with the following properties:
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::custom-rules-scope[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-scope-filter-id[]
|
|
|
|
|
The id of the filter to be used.
|
|
|
|
|
end::custom-rules-scope-filter-id[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-scope-filter-type[]
|
|
|
|
|
Either `include` (the rule applies for values in the filter) or `exclude` (the
|
|
|
|
|
rule applies for values not in the filter). Defaults to `include`.
|
|
|
|
|
end::custom-rules-scope-filter-type[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-conditions[]
|
|
|
|
|
An optional array of numeric conditions when the rule applies. A rule must
|
|
|
|
|
either have a non-empty scope or at least one condition. Multiple conditions are
|
2020-03-05 10:25:43 -05:00
|
|
|
|
combined together with a logical `AND`. A condition has the following
|
|
|
|
|
properties:
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::custom-rules-conditions[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-conditions-applies-to[]
|
|
|
|
|
Specifies the result property to which the condition applies. The available
|
|
|
|
|
options are `actual`, `typical`, `diff_from_typical`, `time`. If your detector
|
|
|
|
|
uses `lat_long`, `metric`, `rare`, or `freq_rare` functions, you can only
|
|
|
|
|
specify conditions that apply to `time`.
|
|
|
|
|
end::custom-rules-conditions-applies-to[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-conditions-operator[]
|
|
|
|
|
Specifies the condition operator. The available options are `gt` (greater than),
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`gte` (greater than or equals), `lt` (less than) and `lte` (less than or
|
|
|
|
|
equals).
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::custom-rules-conditions-operator[]
|
|
|
|
|
|
|
|
|
|
tag::custom-rules-conditions-value[]
|
|
|
|
|
The value that is compared against the `applies_to` field using the `operator`.
|
|
|
|
|
end::custom-rules-conditions-value[]
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
|
|
|
|
tag::custom-settings[]
|
|
|
|
|
Advanced configuration option. Contains custom meta data about the job. For
|
|
|
|
|
example, it can contain custom URL information as shown in
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-configuring-url.html[Adding custom URLs to {ml} results].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::custom-settings[]
|
|
|
|
|
|
|
|
|
|
tag::data-description[]
|
|
|
|
|
The data description defines the format of the input data when you send data to
|
|
|
|
|
the job by using the <<ml-post-data,post data>> API. Note that when configure
|
|
|
|
|
a {dfeed}, these properties are automatically set.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
When data is received via the <<ml-post-data,post data>> API, it is not stored
|
|
|
|
|
in {es}. Only the results for {anomaly-detect} are retained.
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`data_description`.`format`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(string) Only `JSON` format is supported at this time.
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`data_description`.`time_field`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(string) The name of the field that contains the timestamp.
|
|
|
|
|
The default value is `time`.
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`data_description`.`time_format`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(string)
|
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=time-format]
|
|
|
|
|
--
|
|
|
|
|
end::data-description[]
|
|
|
|
|
|
|
|
|
|
tag::data-frame-analytics[]
|
|
|
|
|
An array of {dfanalytics-job} resources, which are sorted by the `id` value in
|
|
|
|
|
ascending order.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`analysis`:::
|
|
|
|
|
(object) The type of analysis that is performed on the `source`.
|
|
|
|
|
|
|
|
|
|
`analyzed_fields`:::
|
|
|
|
|
(object) Contains `includes` and/or `excludes` patterns that select which fields
|
|
|
|
|
are included in the analysis.
|
|
|
|
|
|
|
|
|
|
`analyzed_fields`.`excludes`:::
|
|
|
|
|
(Optional, array) An array of strings that defines the fields that are excluded
|
|
|
|
|
from the analysis.
|
|
|
|
|
|
|
|
|
|
`analyzed_fields`.`includes`:::
|
|
|
|
|
(Optional, array) An array of strings that defines the fields that are included
|
|
|
|
|
in the analysis.
|
|
|
|
|
|
|
|
|
|
`dest`:::
|
|
|
|
|
(string) The destination configuration of the analysis.
|
|
|
|
|
|
|
|
|
|
`dest`.`index`:::
|
|
|
|
|
(string) The _destination index_ that stores the results of the
|
|
|
|
|
{dfanalytics-job}.
|
|
|
|
|
|
|
|
|
|
`dest`.`results_field`:::
|
|
|
|
|
(string) The name of the field that stores the results of the analysis. Defaults
|
|
|
|
|
to `ml`.
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
`id`:::
|
|
|
|
|
(string) The unique identifier of the {dfanalytics-job}.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`model_memory_limit`:::
|
|
|
|
|
(string) The `model_memory_limit` that has been set to the {dfanalytics-job}.
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
`source`:::
|
|
|
|
|
(object) The configuration of how the analysis data is sourced. It has an
|
|
|
|
|
`index` parameter and optionally a `query` and a `_source`.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`source`.`index`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(array) Index or indices on which to perform the analysis. It can be a single
|
|
|
|
|
index or index pattern as well as an array of indices or patterns.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`source`.`query`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(object) The query that has been specified for the {dfanalytics-job}. The {es}
|
|
|
|
|
query domain-specific language (<<query-dsl,DSL>>). This value corresponds to
|
|
|
|
|
the query object in an {es} search POST body. By default, this property has the
|
|
|
|
|
following value: `{"match_all": {}}`.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`source`.`_source`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(object) Contains the specified `includes` and/or `excludes` patterns that
|
|
|
|
|
select which fields are present in the destination. Fields that are excluded
|
|
|
|
|
cannot be included in the analysis.
|
2020-03-05 10:25:43 -05:00
|
|
|
|
|
|
|
|
|
`source`.`_source`.`excludes`:::
|
|
|
|
|
(array) An array of strings that defines the fields that are excluded from the
|
|
|
|
|
destination.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`source`.`_source`.`includes`:::
|
2019-12-13 05:48:21 -05:00
|
|
|
|
(array) An array of strings that defines the fields that are included in the
|
|
|
|
|
destination.
|
|
|
|
|
end::data-frame-analytics[]
|
|
|
|
|
|
|
|
|
|
tag::data-frame-analytics-stats[]
|
|
|
|
|
An array of statistics objects for {dfanalytics-jobs}, which are
|
|
|
|
|
sorted by the `id` value in ascending order.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`assignment_explanation`:::
|
|
|
|
|
(string)
|
|
|
|
|
For running jobs only, contains messages relating to the selection of a node to
|
|
|
|
|
run the job.
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
`id`:::
|
2020-03-05 10:25:43 -05:00
|
|
|
|
(string)
|
|
|
|
|
The unique identifier of the {dfanalytics-job}.
|
2020-02-29 06:03:40 -05:00
|
|
|
|
|
|
|
|
|
`memory_usage`:::
|
2020-03-05 10:25:43 -05:00
|
|
|
|
(Optional, object)
|
|
|
|
|
An object describing memory usage of the analytics. It is present only after the
|
|
|
|
|
job is started and memory usage is reported.
|
2020-02-29 06:03:40 -05:00
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`memory_usage`.`peak_usage_bytes`:::
|
|
|
|
|
(long)
|
|
|
|
|
The number of bytes used at the highest peak of memory usage.
|
2020-02-29 06:03:40 -05:00
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`memory_usage`.`timestamp`:::
|
|
|
|
|
(date)
|
|
|
|
|
The timestamp when memory usage was calculated.
|
2020-02-29 06:03:40 -05:00
|
|
|
|
|
2020-03-05 09:30:30 -05:00
|
|
|
|
`node`:::
|
|
|
|
|
(object)
|
|
|
|
|
Contains properties for the node that runs the job. This information is
|
|
|
|
|
available only for running jobs.
|
|
|
|
|
|
|
|
|
|
`node`.`attributes`:::
|
|
|
|
|
(object)
|
|
|
|
|
Lists node attributes such as `ml.machine_memory`, `ml.max_open_jobs`, and
|
|
|
|
|
`xpack.installed`.
|
|
|
|
|
|
|
|
|
|
`node`.`ephemeral_id`:::
|
|
|
|
|
(string)
|
|
|
|
|
The ephemeral id of the node.
|
|
|
|
|
|
|
|
|
|
`node`.`id`:::
|
|
|
|
|
(string)
|
|
|
|
|
The unique identifier of the node.
|
|
|
|
|
|
|
|
|
|
`node`.`name`:::
|
|
|
|
|
(string)
|
|
|
|
|
The node name.
|
|
|
|
|
|
|
|
|
|
`node`.`transport_address`:::
|
|
|
|
|
(string)
|
|
|
|
|
The host and port where transport HTTP connections are accepted.
|
|
|
|
|
|
2020-03-05 10:25:43 -05:00
|
|
|
|
`progress`:::
|
|
|
|
|
(array) The progress report of the {dfanalytics-job} by phase.
|
|
|
|
|
|
|
|
|
|
`progress`.`phase`:::
|
|
|
|
|
(string) Defines the phase of the {dfanalytics-job}. Possible phases:
|
|
|
|
|
`reindexing`, `loading_data`, `analyzing`, and `writing_results`.
|
|
|
|
|
|
|
|
|
|
`progress`.`progress_percent`:::
|
|
|
|
|
(integer) The progress that the {dfanalytics-job} has made expressed in
|
|
|
|
|
percentage.
|
|
|
|
|
|
|
|
|
|
`state`:::
|
|
|
|
|
(string) Current state of the {dfanalytics-job}.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::data-frame-analytics-stats[]
|
|
|
|
|
|
2019-12-18 03:10:12 -05:00
|
|
|
|
tag::datafeed-id[]
|
|
|
|
|
A numerical character string that uniquely identifies the
|
|
|
|
|
{dfeed}. This identifier can contain lowercase alphanumeric characters (a-z
|
|
|
|
|
and 0-9), hyphens, and underscores. It must start and end with alphanumeric
|
|
|
|
|
characters.
|
|
|
|
|
end::datafeed-id[]
|
|
|
|
|
|
|
|
|
|
tag::datafeed-id-wildcard[]
|
|
|
|
|
Identifier for the {dfeed}. It can be a {dfeed} identifier or a wildcard
|
|
|
|
|
expression.
|
|
|
|
|
end::datafeed-id-wildcard[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::dead-category-count[]
|
|
|
|
|
The number of categories created by categorization that will never be assigned
|
|
|
|
|
again because another category's definition makes it a superset of the dead
|
|
|
|
|
category. (Dead categories are a side effect of the way categorization has no
|
|
|
|
|
prior training.)
|
|
|
|
|
end::dead-category-count[]
|
|
|
|
|
|
2019-12-18 03:10:12 -05:00
|
|
|
|
tag::decompress-definition[]
|
2019-12-24 13:22:05 -05:00
|
|
|
|
Specifies whether the included model definition should be returned as a JSON map
|
|
|
|
|
(`true`) or in a custom compressed format (`false`). Defaults to `true`.
|
2019-12-18 03:10:12 -05:00
|
|
|
|
end::decompress-definition[]
|
|
|
|
|
|
|
|
|
|
tag::delayed-data-check-config[]
|
|
|
|
|
Specifies whether the {dfeed} checks for missing data and the size of the
|
|
|
|
|
window. For example: `{"enabled": true, "check_window": "1h"}`.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
The {dfeed} can optionally search over indices that have already been read in
|
2020-03-05 10:25:43 -05:00
|
|
|
|
an effort to determine whether any data has subsequently been added to the
|
|
|
|
|
index. If missing data is found, it is a good indication that the `query_delay`
|
|
|
|
|
option is set too low and the data is being indexed after the {dfeed} has passed
|
|
|
|
|
that moment in time. See
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-delayed-data-detection.html[Working with delayed data].
|
2019-12-18 03:10:12 -05:00
|
|
|
|
|
|
|
|
|
This check runs only on real-time {dfeeds}.
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`delayed_data_check_config`.`enabled`::
|
2019-12-18 03:10:12 -05:00
|
|
|
|
(boolean) Specifies whether the {dfeed} periodically checks for delayed data.
|
|
|
|
|
Defaults to `true`.
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
`delayed_data_check_config`.`check_window`::
|
2019-12-18 03:10:12 -05:00
|
|
|
|
(<<time-units,time units>>) The window of time that is searched for late data.
|
|
|
|
|
This window of time ends with the latest finalized bucket. It defaults to
|
|
|
|
|
`null`, which causes an appropriate `check_window` to be calculated when the
|
|
|
|
|
real-time {dfeed} runs. In particular, the default `check_window` span
|
|
|
|
|
calculation is based on the maximum of `2h` or `8 * bucket_span`.
|
|
|
|
|
--
|
|
|
|
|
end::delayed-data-check-config[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::dependent-variable[]
|
|
|
|
|
Defines which field of the document is to be predicted.
|
2019-11-06 07:40:27 -05:00
|
|
|
|
This parameter is supplied by field name and must match one of the fields in
|
|
|
|
|
the index being used to train. If this field is missing from a document, then
|
|
|
|
|
that document will not be used for training, but a prediction with the trained
|
|
|
|
|
model will be generated for it. It is also known as continuous target variable.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::dependent-variable[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::desc-results[]
|
|
|
|
|
If true, the results are sorted in descending order.
|
|
|
|
|
end::desc-results[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::description-dfa[]
|
|
|
|
|
A description of the job.
|
|
|
|
|
end::description-dfa[]
|
|
|
|
|
|
|
|
|
|
tag::dest[]
|
|
|
|
|
The destination configuration, consisting of `index` and
|
|
|
|
|
optionally `results_field` (`ml` by default).
|
|
|
|
|
|
|
|
|
|
`index`:::
|
|
|
|
|
(Required, string) Defines the _destination index_ to store the results of
|
|
|
|
|
the {dfanalytics-job}.
|
|
|
|
|
|
|
|
|
|
`results_field`:::
|
|
|
|
|
(Optional, string) Defines the name of the field in which to store the
|
|
|
|
|
results of the analysis. Default to `ml`.
|
|
|
|
|
end::dest[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::detector-description[]
|
|
|
|
|
A description of the detector. For example, `Low event rate`.
|
|
|
|
|
end::detector-description[]
|
|
|
|
|
|
|
|
|
|
tag::detector-field-name[]
|
|
|
|
|
The field that the detector uses in the function. If you use an event rate
|
|
|
|
|
function such as `count` or `rare`, do not specify this field.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: The `field_name` cannot contain double quotes or backslashes.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::detector-field-name[]
|
|
|
|
|
|
|
|
|
|
tag::detector-index[]
|
|
|
|
|
A unique identifier for the detector. This identifier is based on the order of
|
2020-01-09 11:34:34 -05:00
|
|
|
|
the detectors in the `analysis_config`, starting at zero.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::detector-index[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::earliest-record-timestamp[]
|
|
|
|
|
The timestamp of the earliest chronologically input document.
|
|
|
|
|
end::earliest-record-timestamp[]
|
|
|
|
|
|
|
|
|
|
tag::empty-bucket-count[]
|
|
|
|
|
The number of buckets which did not contain any data. If your data
|
|
|
|
|
contains many empty buckets, consider increasing your `bucket_span` or using
|
|
|
|
|
functions that are tolerant to gaps in data such as `mean`, `non_null_sum` or
|
|
|
|
|
`non_zero_count`.
|
|
|
|
|
end::empty-bucket-count[]
|
|
|
|
|
|
2019-11-06 07:40:27 -05:00
|
|
|
|
tag::eta[]
|
2020-01-09 10:21:35 -05:00
|
|
|
|
Advanced configuration option. The shrinkage applied to the weights. Smaller
|
|
|
|
|
values result in larger forests which have better generalization error. However,
|
|
|
|
|
the smaller the value the longer the training will take. For more information,
|
|
|
|
|
about shrinkage, see
|
|
|
|
|
https://en.wikipedia.org/wiki/Gradient_boosting#Shrinkage[this wiki article].
|
2019-11-06 07:40:27 -05:00
|
|
|
|
end::eta[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::exclude-frequent[]
|
|
|
|
|
Contains one of the following values: `all`, `none`, `by`, or `over`. If set,
|
|
|
|
|
frequent entities are excluded from influencing the anomaly results. Entities
|
|
|
|
|
can be considered frequent over time or frequent in a population. If you are
|
|
|
|
|
working with both over and by fields, then you can set `exclude_frequent` to
|
|
|
|
|
`all` for both fields, or to `by` or `over` for those specific fields.
|
|
|
|
|
end::exclude-frequent[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::exclude-interim-results[]
|
|
|
|
|
If `true`, the output excludes interim results. By default, interim results are
|
|
|
|
|
included.
|
|
|
|
|
end::exclude-interim-results[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::feature-bag-fraction[]
|
2020-01-09 10:21:35 -05:00
|
|
|
|
Advanced configuration option. Defines the fraction of features that will be
|
|
|
|
|
used when selecting a random bag for each candidate split.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::feature-bag-fraction[]
|
|
|
|
|
|
|
|
|
|
tag::feature-influence-threshold[]
|
|
|
|
|
The minimum {olscore} that a document needs to have in order to calculate its
|
|
|
|
|
{fiscore}. Value range: 0-1 (`0.1` by default).
|
|
|
|
|
end::feature-influence-threshold[]
|
|
|
|
|
|
|
|
|
|
tag::field-selection[]
|
|
|
|
|
An array of objects that explain selection for each field, sorted by
|
|
|
|
|
the field names. Each object in the array has the following properties:
|
|
|
|
|
|
|
|
|
|
`name`:::
|
|
|
|
|
(string) The field name.
|
|
|
|
|
|
|
|
|
|
`mapping_types`:::
|
|
|
|
|
(string) The mapping types of the field.
|
|
|
|
|
|
|
|
|
|
`is_included`:::
|
|
|
|
|
(boolean) Whether the field is selected to be included in the analysis.
|
|
|
|
|
|
|
|
|
|
`is_required`:::
|
|
|
|
|
(boolean) Whether the field is required.
|
|
|
|
|
|
|
|
|
|
`feature_type`:::
|
|
|
|
|
(string) The feature type of this field for the analysis. May be `categorical`
|
|
|
|
|
or `numerical`.
|
|
|
|
|
|
|
|
|
|
`reason`:::
|
|
|
|
|
(string) The reason a field is not selected to be included in the analysis.
|
|
|
|
|
end::field-selection[]
|
|
|
|
|
|
|
|
|
|
tag::filter[]
|
|
|
|
|
One or more <<analysis-tokenfilters,token filters>>. In addition to the built-in
|
|
|
|
|
token filters, other plugins can provide more token filters. This property is
|
|
|
|
|
optional. If it is not specified, no token filters are applied prior to
|
|
|
|
|
categorization.
|
|
|
|
|
end::filter[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2020-01-02 14:23:26 -05:00
|
|
|
|
tag::filter-id[]
|
|
|
|
|
A string that uniquely identifies a filter.
|
|
|
|
|
end::filter-id[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::forecast-total[]
|
|
|
|
|
The number of individual forecasts currently available for the job. A value of
|
|
|
|
|
`1` or more indicates that forecasts exist.
|
|
|
|
|
end::forecast-total[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::frequency[]
|
|
|
|
|
The interval at which scheduled queries are made while the {dfeed} runs in real
|
|
|
|
|
time. The default value is either the bucket span for short bucket spans, or,
|
|
|
|
|
for longer bucket spans, a sensible fraction of the bucket span. For example:
|
2020-01-22 11:08:31 -05:00
|
|
|
|
`150s`. When `frequency` is shorter than the bucket span, interim results for
|
|
|
|
|
the last (partial) bucket are written then eventually overwritten by the full
|
|
|
|
|
bucket results. If the {dfeed} uses aggregations, this value must be divisible
|
|
|
|
|
by the interval of the date histogram aggregation.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::frequency[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::frequent-category-count[]
|
|
|
|
|
The number of categories that match more than 1% of categorized documents.
|
|
|
|
|
end::frequent-category-count[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::from[]
|
|
|
|
|
Skips the specified number of {dfanalytics-jobs}. The default value is `0`.
|
|
|
|
|
end::from[]
|
|
|
|
|
|
|
|
|
|
tag::function[]
|
|
|
|
|
The analysis function that is used. For example, `count`, `rare`, `mean`, `min`,
|
|
|
|
|
`max`, and `sum`. For more information, see
|
2019-12-20 13:01:19 -05:00
|
|
|
|
{ml-docs}/ml-functions.html[Function reference].
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::function[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
|
|
|
|
tag::gamma[]
|
2020-01-09 10:21:35 -05:00
|
|
|
|
Advanced configuration option. Regularization parameter to prevent overfitting
|
|
|
|
|
on the training dataset. Multiplies a linear penalty associated with the size of
|
2019-11-06 07:40:27 -05:00
|
|
|
|
individual trees in the forest. The higher the value the more training will
|
2020-01-09 08:31:35 -05:00
|
|
|
|
prefer smaller trees. The smaller this parameter the larger individual trees
|
2019-11-06 07:40:27 -05:00
|
|
|
|
will be and the longer train will take.
|
|
|
|
|
end::gamma[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::groups[]
|
|
|
|
|
A list of job groups. A job can belong to no groups or many.
|
|
|
|
|
end::groups[]
|
|
|
|
|
|
2019-12-18 03:10:12 -05:00
|
|
|
|
tag::include-model-definition[]
|
|
|
|
|
Specifies if the model definition should be returned in the response. Defaults
|
|
|
|
|
to `false`. When `true`, only a single model must match the ID patterns
|
|
|
|
|
provided, otherwise a bad request is returned.
|
|
|
|
|
end::include-model-definition[]
|
|
|
|
|
|
|
|
|
|
tag::indices[]
|
|
|
|
|
An array of index names. Wildcards are supported. For example:
|
|
|
|
|
`["it_ops_metrics", "server*"]`.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: If any indices are in remote clusters then `cluster.remote.connect` must
|
|
|
|
|
not be set to `false` on any {ml} nodes.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::indices[]
|
|
|
|
|
|
2020-02-27 13:43:25 -05:00
|
|
|
|
tag::indices-options[]
|
2020-03-03 13:28:54 -05:00
|
|
|
|
Specifies index expansion options that are used during search.
|
|
|
|
|
+
|
|
|
|
|
--
|
2020-02-27 13:43:25 -05:00
|
|
|
|
For example:
|
|
|
|
|
```
|
|
|
|
|
{
|
|
|
|
|
"expand_wildcards": ["all"],
|
|
|
|
|
"ignore_unavailable": true,
|
|
|
|
|
"allow_no_indices": "false",
|
|
|
|
|
"ignore_throttled": true
|
|
|
|
|
}
|
|
|
|
|
```
|
2020-03-03 13:28:54 -05:00
|
|
|
|
For more information about these options, see <<multi-index>>.
|
|
|
|
|
--
|
2020-02-27 13:43:25 -05:00
|
|
|
|
end::indices-options[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::influencers[]
|
|
|
|
|
A comma separated list of influencer field names. Typically these can be the by,
|
|
|
|
|
over, or partition fields that are used in the detector configuration. You might
|
|
|
|
|
also want to use a field name that is not specifically named in a detector, but
|
|
|
|
|
is available as part of the input data. When you use multiple detectors, the use
|
2020-03-05 10:25:43 -05:00
|
|
|
|
of influencers is recommended as it aggregates results for each influencer
|
|
|
|
|
entity.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::influencers[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::input-bytes[]
|
|
|
|
|
The number of bytes of input data posted to the {anomaly-job}.
|
|
|
|
|
end::input-bytes[]
|
|
|
|
|
|
|
|
|
|
tag::input-field-count[]
|
|
|
|
|
The total number of fields in input documents posted to the {anomaly-job}. This
|
|
|
|
|
count includes fields that are not used in the analysis. However, be aware that
|
|
|
|
|
if you are using a {dfeed}, it extracts only the required fields from the
|
|
|
|
|
documents it retrieves before posting them to the job.
|
|
|
|
|
end::input-field-count[]
|
|
|
|
|
|
|
|
|
|
tag::input-record-count[]
|
|
|
|
|
The number of input documents posted to the {anomaly-job}.
|
|
|
|
|
end::input-record-count[]
|
|
|
|
|
|
|
|
|
|
tag::invalid-date-count[]
|
|
|
|
|
The number of input documents with either a missing date field or a date that
|
|
|
|
|
could not be parsed.
|
|
|
|
|
end::invalid-date-count[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::is-interim[]
|
|
|
|
|
If `true`, this is an interim result. In other words, the results are calculated
|
|
|
|
|
based on partial input data.
|
|
|
|
|
end::is-interim[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::job-id-anomaly-detection[]
|
|
|
|
|
Identifier for the {anomaly-job}.
|
|
|
|
|
end::job-id-anomaly-detection[]
|
|
|
|
|
|
|
|
|
|
tag::job-id-anomaly-detection-default[]
|
|
|
|
|
Identifier for the {anomaly-job}. It can be a job identifier, a group name, or a
|
|
|
|
|
wildcard expression. If you do not specify one of these options, the API returns
|
|
|
|
|
information for all {anomaly-jobs}.
|
|
|
|
|
end::job-id-anomaly-detection-default[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::job-id-anomaly-detection-define[]
|
|
|
|
|
Identifier for the {anomaly-job}. This identifier can contain lowercase
|
|
|
|
|
alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start
|
|
|
|
|
and end with alphanumeric characters.
|
|
|
|
|
end::job-id-anomaly-detection-define[]
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
|
|
|
|
tag::job-id-anomaly-detection-list[]
|
|
|
|
|
An identifier for the {anomaly-jobs}. It can be a job
|
|
|
|
|
identifier, a group name, or a comma-separated list of jobs or groups.
|
|
|
|
|
end::job-id-anomaly-detection-list[]
|
|
|
|
|
|
|
|
|
|
tag::job-id-anomaly-detection-wildcard[]
|
|
|
|
|
Identifier for the {anomaly-job}. It can be a job identifier, a group name, or a
|
|
|
|
|
wildcard expression.
|
|
|
|
|
end::job-id-anomaly-detection-wildcard[]
|
|
|
|
|
|
|
|
|
|
tag::job-id-anomaly-detection-wildcard-list[]
|
|
|
|
|
Identifier for the {anomaly-job}. It can be a job identifier, a group name, a
|
|
|
|
|
comma-separated list of jobs or groups, or a wildcard expression.
|
|
|
|
|
end::job-id-anomaly-detection-wildcard-list[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::job-id-data-frame-analytics[]
|
|
|
|
|
Identifier for the {dfanalytics-job}.
|
|
|
|
|
end::job-id-data-frame-analytics[]
|
|
|
|
|
|
|
|
|
|
tag::job-id-data-frame-analytics-default[]
|
|
|
|
|
Identifier for the {dfanalytics-job}. If you do not specify this option, the API
|
|
|
|
|
returns information for the first hundred {dfanalytics-jobs}.
|
|
|
|
|
end::job-id-data-frame-analytics-default[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::job-id-data-frame-analytics-define[]
|
|
|
|
|
Identifier for the {dfanalytics-job}. This identifier can contain lowercase
|
|
|
|
|
alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start
|
|
|
|
|
and end with alphanumeric characters.
|
|
|
|
|
end::job-id-data-frame-analytics-define[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::job-id-datafeed[]
|
|
|
|
|
The unique identifier for the job to which the {dfeed} sends data.
|
|
|
|
|
end::job-id-datafeed[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::jobs-stats-anomaly-detection[]
|
|
|
|
|
An array of {anomaly-job} statistics objects.
|
|
|
|
|
For more information, see <<ml-jobstats>>.
|
|
|
|
|
end::jobs-stats-anomaly-detection[]
|
|
|
|
|
|
|
|
|
|
tag::lambda[]
|
2020-01-09 10:21:35 -05:00
|
|
|
|
Advanced configuration option. Regularization parameter to prevent overfitting
|
|
|
|
|
on the training dataset. Multiplies an L2 regularisation term which applies to
|
|
|
|
|
leaf weights of the individual trees in the forest. The higher the value the
|
|
|
|
|
more training will attempt to keep leaf weights small. This makes the prediction
|
2019-11-06 07:40:27 -05:00
|
|
|
|
function smoother at the expense of potentially not being able to capture
|
|
|
|
|
relevant relationships between the features and the {depvar}. The smaller this
|
|
|
|
|
parameter the larger individual trees will be and the longer train will take.
|
|
|
|
|
end::lambda[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::last-data-time[]
|
|
|
|
|
The timestamp at which data was last analyzed, according to server time.
|
|
|
|
|
end::last-data-time[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::latency[]
|
|
|
|
|
The size of the window in which to expect data that is out of time order. The
|
|
|
|
|
default value is 0 (no latency). If you specify a non-zero value, it must be
|
|
|
|
|
greater than or equal to one second. For more information about time units, see
|
|
|
|
|
<<time-units>>.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: Latency is only applicable when you send data by using
|
|
|
|
|
the <<ml-post-data,post data>> API.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::latency[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::latest-empty-bucket-timestamp[]
|
|
|
|
|
The timestamp of the last bucket that did not contain any data.
|
|
|
|
|
end::latest-empty-bucket-timestamp[]
|
|
|
|
|
|
|
|
|
|
tag::latest-record-timestamp[]
|
|
|
|
|
The timestamp of the latest chronologically input document.
|
|
|
|
|
end::latest-record-timestamp[]
|
|
|
|
|
|
|
|
|
|
tag::latest-sparse-record-timestamp[]
|
|
|
|
|
The timestamp of the last bucket that was considered sparse.
|
|
|
|
|
end::latest-sparse-record-timestamp[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::max-empty-searches[]
|
|
|
|
|
If a real-time {dfeed} has never seen any data (including during any initial
|
|
|
|
|
training period) then it will automatically stop itself and close its associated
|
|
|
|
|
job after this many real-time searches that return no documents. In other words,
|
2020-03-05 10:25:43 -05:00
|
|
|
|
it will stop after `frequency` times `max_empty_searches` of real-time
|
|
|
|
|
operation. If not set then a {dfeed} with no end time that sees no data will
|
|
|
|
|
remain started until it is explicitly stopped. By default this setting is not
|
|
|
|
|
set.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::max-empty-searches[]
|
|
|
|
|
|
2020-03-11 06:45:27 -04:00
|
|
|
|
tag::max-trees[]
|
|
|
|
|
Advanced configuration option. Defines the maximum number of trees the forest is
|
2020-01-09 10:21:35 -05:00
|
|
|
|
allowed to contain. The maximum value is 2000.
|
2020-03-11 06:45:27 -04:00
|
|
|
|
end::max-trees[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::memory-estimation[]
|
|
|
|
|
An object containing the memory estimates. The object has the
|
|
|
|
|
following properties:
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
`expected_memory_without_disk`:::
|
|
|
|
|
(string) Estimated memory usage under the assumption that the whole
|
|
|
|
|
{dfanalytics} should happen in memory (i.e. without overflowing to disk).
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
`expected_memory_with_disk`:::
|
|
|
|
|
(string) Estimated memory usage under the assumption that overflowing to disk is
|
|
|
|
|
allowed during {dfanalytics}. `expected_memory_with_disk` is usually smaller
|
|
|
|
|
than `expected_memory_without_disk` as using disk allows to limit the main
|
|
|
|
|
memory needed to perform {dfanalytics}.
|
|
|
|
|
end::memory-estimation[]
|
2019-11-06 07:40:27 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::method[]
|
|
|
|
|
Sets the method that {oldetection} uses. If the method is not set {oldetection}
|
|
|
|
|
uses an ensemble of different methods and normalises and combines their
|
|
|
|
|
individual {olscores} to obtain the overall {olscore}. We recommend to use the
|
|
|
|
|
ensemble method. Available methods are `lof`, `ldof`, `distance_kth_nn`,
|
|
|
|
|
`distance_knn`.
|
|
|
|
|
end::method[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::missing-field-count[]
|
|
|
|
|
The number of input documents that are missing a field that the {anomaly-job} is
|
|
|
|
|
configured to analyze. Input documents with missing fields are still processed
|
|
|
|
|
because it is possible that not all fields are missing.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: If you are using {dfeeds} or posting data to the job in JSON format, a
|
|
|
|
|
high `missing_field_count` is often not an indication of data issues. It is not
|
|
|
|
|
necessarily a cause for concern.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::missing-field-count[]
|
|
|
|
|
|
2019-12-18 03:10:12 -05:00
|
|
|
|
tag::mode[]
|
|
|
|
|
There are three available modes:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* `auto`: The chunk size is dynamically calculated. This is the default and
|
|
|
|
|
recommended value.
|
|
|
|
|
* `manual`: Chunking is applied according to the specified `time_span`.
|
|
|
|
|
* `off`: No chunking is applied.
|
|
|
|
|
--
|
|
|
|
|
end::mode[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::model-bytes[]
|
|
|
|
|
The number of bytes of memory used by the models. This is the maximum value
|
|
|
|
|
since the last time the model was persisted. If the job is closed, this value
|
|
|
|
|
indicates the latest size.
|
|
|
|
|
end::model-bytes[]
|
|
|
|
|
|
|
|
|
|
tag::model-bytes-exceeded[]
|
|
|
|
|
The number of bytes over the high limit for memory usage at the last allocation
|
|
|
|
|
failure.
|
|
|
|
|
end::model-bytes-exceeded[]
|
|
|
|
|
|
2019-12-18 03:10:12 -05:00
|
|
|
|
tag::model-id[]
|
|
|
|
|
The unique identifier of the trained {infer} model.
|
|
|
|
|
end::model-id[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::model-memory-limit[]
|
|
|
|
|
The approximate maximum amount of memory resources that are required for
|
|
|
|
|
analytical processing. Once this limit is approached, data pruning becomes
|
|
|
|
|
more aggressive. Upon exceeding this limit, new entities are not modeled. The
|
|
|
|
|
default value for jobs created in version 6.1 and later is `1024mb`.
|
|
|
|
|
This value will need to be increased for jobs that are expected to analyze high
|
|
|
|
|
cardinality fields, but the default is set to a relatively small size to ensure
|
|
|
|
|
that high resource usage is a conscious decision. The default value for jobs
|
|
|
|
|
created in versions earlier than 6.1 is `4096mb`.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
If you specify a number instead of a string, the units are assumed to be MiB.
|
|
|
|
|
Specifying a string is recommended for clarity. If you specify a byte size unit
|
|
|
|
|
of `b` or `kb` and the number does not equate to a discrete number of megabytes,
|
|
|
|
|
it is rounded down to the closest MiB. The minimum valid value is 1 MiB. If you
|
|
|
|
|
specify a value less than 1 MiB, an error occurs. For more information about
|
|
|
|
|
supported byte size units, see <<byte-units>>.
|
|
|
|
|
|
|
|
|
|
If your `elasticsearch.yml` file contains an `xpack.ml.max_model_memory_limit`
|
|
|
|
|
setting, an error occurs when you try to create jobs that have
|
|
|
|
|
`model_memory_limit` values greater than that setting. For more information,
|
|
|
|
|
see <<ml-settings>>.
|
|
|
|
|
--
|
|
|
|
|
end::model-memory-limit[]
|
2019-12-10 08:29:19 -05:00
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::model-memory-limit-anomaly-jobs[]
|
|
|
|
|
The upper limit for model memory usage, checked on increasing values.
|
|
|
|
|
end::model-memory-limit-anomaly-jobs[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::model-memory-limit-dfa[]
|
|
|
|
|
The approximate maximum amount of memory resources that are permitted for
|
|
|
|
|
analytical processing. The default value for {dfanalytics-jobs} is `1gb`. If
|
|
|
|
|
your `elasticsearch.yml` file contains an `xpack.ml.max_model_memory_limit`
|
|
|
|
|
setting, an error occurs when you try to create {dfanalytics-jobs} that have
|
|
|
|
|
`model_memory_limit` values greater than that setting. For more information, see
|
|
|
|
|
<<ml-settings>>.
|
|
|
|
|
end::model-memory-limit-dfa[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::model-memory-status[]
|
|
|
|
|
The status of the mathematical models, which can have one of the following
|
|
|
|
|
values:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* `ok`: The models stayed below the configured value.
|
|
|
|
|
* `soft_limit`: The models used more than 60% of the configured memory limit
|
|
|
|
|
and older unused models will be pruned to free up space.
|
|
|
|
|
* `hard_limit`: The models used more space than the configured memory limit.
|
|
|
|
|
As a result, not all incoming data was processed.
|
|
|
|
|
--
|
|
|
|
|
end::model-memory-status[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::model-plot-config[]
|
|
|
|
|
This advanced configuration option stores model information along with the
|
|
|
|
|
results. It provides a more detailed view into {anomaly-detect}.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
WARNING: If you enable model plot it can add considerable overhead to the
|
|
|
|
|
performance of the system; it is not feasible for jobs with many entities.
|
|
|
|
|
|
|
|
|
|
Model plot provides a simplified and indicative view of the model and its
|
|
|
|
|
bounds. It does not display complex features such as multivariate correlations
|
|
|
|
|
or multimodal data. As such, anomalies may occasionally be reported which cannot
|
|
|
|
|
be seen in the model plot.
|
|
|
|
|
|
|
|
|
|
Model plot config can be configured when the job is created or updated later. It
|
|
|
|
|
must be disabled if performance issues are experienced.
|
|
|
|
|
--
|
|
|
|
|
end::model-plot-config[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::model-plot-config-enabled[]
|
|
|
|
|
If true, enables calculation and storage of the model bounds for each entity
|
|
|
|
|
that is being analyzed. By default, this is not enabled.
|
|
|
|
|
end::model-plot-config-enabled[]
|
|
|
|
|
|
|
|
|
|
tag::model-plot-config-terms[]
|
|
|
|
|
Limits data collection to this comma separated list of partition or by field
|
|
|
|
|
values. If terms are not specified or it is an empty string, no filtering is
|
|
|
|
|
applied. For example, "CPU,NetworkIn,DiskWrites". Wildcards are not supported.
|
|
|
|
|
Only the specified `terms` can be viewed when using the Single Metric Viewer.
|
|
|
|
|
end::model-plot-config-terms[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::model-snapshot-retention-days[]
|
2020-03-05 10:25:43 -05:00
|
|
|
|
Advanced configuration option. The period of time (in days) that model snapshots
|
|
|
|
|
are retained. Age is calculated relative to the timestamp of the newest model
|
|
|
|
|
snapshot. The default value is `1`, which means snapshots that are one day
|
|
|
|
|
(twenty-four hours) older than the newest snapshot are deleted.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::model-snapshot-retention-days[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::model-timestamp[]
|
|
|
|
|
The timestamp of the last record when the model stats were gathered.
|
|
|
|
|
end::model-timestamp[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::multivariate-by-fields[]
|
|
|
|
|
This functionality is reserved for internal use. It is not supported for use in
|
|
|
|
|
customer environments and is not subject to the support SLA of official GA
|
|
|
|
|
features.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
If set to `true`, the analysis will automatically find correlations between
|
|
|
|
|
metrics for a given `by` field value and report anomalies when those
|
|
|
|
|
correlations cease to hold. For example, suppose CPU and memory usage on host A
|
|
|
|
|
is usually highly correlated with the same metrics on host B. Perhaps this
|
|
|
|
|
correlation occurs because they are running a load-balanced application.
|
|
|
|
|
If you enable this property, then anomalies will be reported when, for example,
|
|
|
|
|
CPU usage on host A is high and the value of CPU usage on host B is low. That
|
|
|
|
|
is to say, you'll see an anomaly when the CPU of host A is unusual given
|
|
|
|
|
the CPU of host B.
|
|
|
|
|
|
|
|
|
|
NOTE: To use the `multivariate_by_fields` property, you must also specify
|
|
|
|
|
`by_field_name` in your detector.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::multivariate-by-fields[]
|
|
|
|
|
|
|
|
|
|
tag::n-neighbors[]
|
|
|
|
|
Defines the value for how many nearest neighbors each method of
|
|
|
|
|
{oldetection} will use to calculate its {olscore}. When the value is not set,
|
|
|
|
|
different values will be used for different ensemble members. This helps
|
|
|
|
|
improve diversity in the ensemble. Therefore, only override this if you are
|
|
|
|
|
confident that the value you choose is appropriate for the data set.
|
|
|
|
|
end::n-neighbors[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::node-address[]
|
|
|
|
|
The network address of the node.
|
|
|
|
|
end::node-address[]
|
|
|
|
|
|
|
|
|
|
tag::node-datafeeds[]
|
2020-02-26 12:20:36 -05:00
|
|
|
|
For started {dfeeds} only, this information pertains to the node upon which the
|
|
|
|
|
{dfeed} is started.
|
2020-03-02 10:28:55 -05:00
|
|
|
|
end::node-datafeeds[]
|
|
|
|
|
|
|
|
|
|
tag::node-ephemeral-id[]
|
|
|
|
|
The ephemeral ID of the node.
|
|
|
|
|
end::node-ephemeral-id[]
|
|
|
|
|
|
|
|
|
|
tag::node-id[]
|
|
|
|
|
The unique identifier of the node.
|
|
|
|
|
end::node-id[]
|
|
|
|
|
|
|
|
|
|
tag::node-jobs[]
|
|
|
|
|
Contains properties for the node that runs the job. This information is
|
|
|
|
|
available only for open jobs.
|
|
|
|
|
end::node-jobs[]
|
2020-02-26 12:20:36 -05:00
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::num-top-classes[]
|
|
|
|
|
Defines the number of categories for which the predicted
|
|
|
|
|
probabilities are reported. It must be non-negative. If it is greater than the
|
|
|
|
|
total number of categories (in the {version} version of the {stack}, it's two)
|
|
|
|
|
to predict then we will report all category probabilities. Defaults to 2.
|
|
|
|
|
end::num-top-classes[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::open-time[]
|
|
|
|
|
For open jobs only, the elapsed time for which the job has been open.
|
|
|
|
|
end::open-time[]
|
|
|
|
|
|
|
|
|
|
tag::out-of-order-timestamp-count[]
|
|
|
|
|
The number of input documents that are out of time sequence and outside
|
|
|
|
|
of the latency window. This information is applicable only when you provide data
|
|
|
|
|
to the {anomaly-job} by using the <<ml-post-data,post data API>>. These out of
|
|
|
|
|
order documents are discarded, since jobs require time series data to be in
|
|
|
|
|
ascending chronological order.
|
|
|
|
|
end::out-of-order-timestamp-count[]
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
|
|
|
|
tag::outlier-fraction[]
|
|
|
|
|
Sets the proportion of the data set that is assumed to be outlying prior to
|
|
|
|
|
{oldetection}. For example, 0.05 means it is assumed that 5% of values are real
|
|
|
|
|
outliers and 95% are inliers.
|
|
|
|
|
end::outlier-fraction[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::over-field-name[]
|
|
|
|
|
The field used to split the data. In particular, this property is used for
|
|
|
|
|
analyzing the splits with respect to the history of all splits. It is used for
|
|
|
|
|
finding unusual values in the population of all splits. For more information,
|
|
|
|
|
see {ml-docs}/ml-configuring-pop.html[Performing population analysis].
|
|
|
|
|
end::over-field-name[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::partition-field-name[]
|
|
|
|
|
The field used to segment the analysis. When you use this property, you have
|
|
|
|
|
completely independent baselines for each value of this field.
|
|
|
|
|
end::partition-field-name[]
|
|
|
|
|
|
|
|
|
|
tag::prediction-field-name[]
|
|
|
|
|
Defines the name of the prediction field in the results.
|
|
|
|
|
Defaults to `<dependent_variable>_prediction`.
|
|
|
|
|
end::prediction-field-name[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::processed-field-count[]
|
|
|
|
|
The total number of fields in all the documents that have been processed by the
|
|
|
|
|
{anomaly-job}. Only fields that are specified in the detector configuration
|
|
|
|
|
object contribute to this count. The timestamp is not included in this count.
|
|
|
|
|
end::processed-field-count[]
|
|
|
|
|
|
|
|
|
|
tag::processed-record-count[]
|
|
|
|
|
The number of input documents that have been processed by the {anomaly-job}.
|
|
|
|
|
This value includes documents with missing fields, since they are nonetheless
|
|
|
|
|
analyzed. If you use {dfeeds} and have aggregations in your search query, the
|
|
|
|
|
`processed_record_count` is the number of aggregation results processed, not the
|
|
|
|
|
number of {es} documents.
|
|
|
|
|
end::processed-record-count[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::query[]
|
|
|
|
|
The {es} query domain-specific language (DSL). This value corresponds to the
|
|
|
|
|
query object in an {es} search POST body. All the options that are supported by
|
|
|
|
|
{es} can be used, as this object is passed verbatim to {es}. By default, this
|
|
|
|
|
property has the following value: `{"match_all": {"boost": 1}}`.
|
|
|
|
|
end::query[]
|
|
|
|
|
|
|
|
|
|
tag::query-delay[]
|
|
|
|
|
The number of seconds behind real time that data is queried. For example, if
|
|
|
|
|
data from 10:04 a.m. might not be searchable in {es} until 10:06 a.m., set this
|
|
|
|
|
property to 120 seconds. The default value is randomly selected between `60s`
|
|
|
|
|
and `120s`. This randomness improves the query performance when there are
|
2020-01-22 11:08:31 -05:00
|
|
|
|
multiple jobs running on the same node. For more information, see
|
|
|
|
|
{ml-docs}/ml-delayed-data-detection.html[Handling delayed data].
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::query-delay[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::randomize-seed[]
|
|
|
|
|
Defines the seed to the random generator that is used to pick
|
2019-12-10 08:29:19 -05:00
|
|
|
|
which documents will be used for training. By default it is randomly generated.
|
|
|
|
|
Set it to a specific value to ensure the same documents are used for training
|
2019-12-13 05:48:21 -05:00
|
|
|
|
assuming other related parameters (e.g. `source`, `analyzed_fields`, etc.) are
|
|
|
|
|
the same.
|
|
|
|
|
end::randomize-seed[]
|
2019-12-10 08:29:19 -05:00
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::rare-category-count[]
|
|
|
|
|
The number of categories that match just one categorized document.
|
|
|
|
|
end::rare-category-count[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::renormalization-window-days[]
|
|
|
|
|
Advanced configuration option. The period over which adjustments to the score
|
|
|
|
|
are applied, as new data is seen. The default value is the longer of 30 days or
|
|
|
|
|
100 `bucket_spans`.
|
|
|
|
|
end::renormalization-window-days[]
|
|
|
|
|
|
|
|
|
|
tag::results-index-name[]
|
|
|
|
|
A text string that affects the name of the {ml} results index. The default value
|
|
|
|
|
is `shared`, which generates an index named `.ml-anomalies-shared`.
|
|
|
|
|
end::results-index-name[]
|
|
|
|
|
|
|
|
|
|
tag::results-retention-days[]
|
2020-03-05 10:25:43 -05:00
|
|
|
|
Advanced configuration option. The period of time (in days) that results are
|
|
|
|
|
retained. Age is calculated relative to the timestamp of the latest bucket
|
|
|
|
|
result. If this property has a non-null value, once per day at 00:30 (server
|
|
|
|
|
time), results that are the specified number of days older than the latest
|
2020-01-22 09:52:33 -05:00
|
|
|
|
bucket result are deleted from {es}. The default value is null, which means all
|
|
|
|
|
results are retained.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::results-retention-days[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::retain[]
|
|
|
|
|
If `true`, this snapshot will not be deleted during automatic cleanup of
|
|
|
|
|
snapshots older than `model_snapshot_retention_days`. However, this snapshot
|
|
|
|
|
will be deleted when the job is deleted. The default value is `false`.
|
|
|
|
|
end::retain[]
|
|
|
|
|
|
|
|
|
|
tag::script-fields[]
|
|
|
|
|
Specifies scripts that evaluate custom expressions and returns script fields to
|
|
|
|
|
the {dfeed}. The detector configuration objects in a job can contain functions
|
|
|
|
|
that use these script fields. For more information, see
|
|
|
|
|
{ml-docs}/ml-configuring-transform.html[Transforming data with script fields]
|
|
|
|
|
and <<request-body-search-script-fields,Script fields>>.
|
|
|
|
|
end::script-fields[]
|
|
|
|
|
|
|
|
|
|
tag::scroll-size[]
|
|
|
|
|
The `size` parameter that is used in {es} searches. The default value is `1000`.
|
|
|
|
|
end::scroll-size[]
|
|
|
|
|
|
2020-02-26 12:20:36 -05:00
|
|
|
|
tag::search-bucket-avg[]
|
|
|
|
|
The average search time per bucket, in milliseconds.
|
|
|
|
|
end::search-bucket-avg[]
|
|
|
|
|
|
|
|
|
|
tag::search-count[]
|
|
|
|
|
The number of searches run by the {dfeed}.
|
|
|
|
|
end::search-count[]
|
|
|
|
|
|
|
|
|
|
tag::search-exp-avg-hour[]
|
|
|
|
|
The exponential average search time per hour, in milliseconds.
|
|
|
|
|
end::search-exp-avg-hour[]
|
|
|
|
|
|
|
|
|
|
tag::search-time[]
|
|
|
|
|
The total time the {dfeed} spent searching, in milliseconds.
|
|
|
|
|
end::search-time[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::size[]
|
|
|
|
|
Specifies the maximum number of {dfanalytics-jobs} to obtain. The default value
|
|
|
|
|
is `100`.
|
|
|
|
|
end::size[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::snapshot-id[]
|
2019-12-31 13:53:05 -05:00
|
|
|
|
A numerical character string that uniquely identifies the model snapshot.
|
2019-12-24 13:22:05 -05:00
|
|
|
|
end::snapshot-id[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::source-put-dfa[]
|
|
|
|
|
The configuration of how to source the analysis data. It requires an
|
|
|
|
|
`index`. Optionally, `query` and `_source` may be specified.
|
|
|
|
|
|
|
|
|
|
`index`:::
|
|
|
|
|
(Required, string or array) Index or indices on which to perform the
|
|
|
|
|
analysis. It can be a single index or index pattern as well as an array of
|
|
|
|
|
indices or patterns.
|
2020-03-12 13:00:36 -04:00
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
WARNING: If your source indices contain documents with the same IDs, only the
|
|
|
|
|
document that is indexed last appears in the destination index.
|
|
|
|
|
--
|
2019-12-13 05:48:21 -05:00
|
|
|
|
|
|
|
|
|
`query`:::
|
|
|
|
|
(Optional, object) The {es} query domain-specific language
|
|
|
|
|
(<<query-dsl,DSL>>). This value corresponds to the query object in an {es}
|
|
|
|
|
search POST body. All the options that are supported by {es} can be used,
|
|
|
|
|
as this object is passed verbatim to {es}. By default, this property has
|
|
|
|
|
the following value: `{"match_all": {}}`.
|
|
|
|
|
|
|
|
|
|
`_source`:::
|
|
|
|
|
(Optional, object) Specify `includes` and/or `excludes` patterns to select
|
|
|
|
|
which fields will be present in the destination. Fields that are excluded
|
|
|
|
|
cannot be included in the analysis.
|
|
|
|
|
|
|
|
|
|
`includes`::::
|
|
|
|
|
(array) An array of strings that defines the fields that will be
|
|
|
|
|
included in the destination.
|
|
|
|
|
|
|
|
|
|
`excludes`::::
|
|
|
|
|
(array) An array of strings that defines the fields that will be
|
|
|
|
|
excluded from the destination.
|
|
|
|
|
end::source-put-dfa[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::sparse-bucket-count[]
|
|
|
|
|
The number of buckets that contained few data points compared to the expected
|
|
|
|
|
number of data points. If your data contains many sparse buckets, consider using
|
|
|
|
|
a longer `bucket_span`.
|
|
|
|
|
end::sparse-bucket-count[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::standardization-enabled[]
|
|
|
|
|
If `true`, then the following operation is performed on the columns before
|
|
|
|
|
computing outlier scores: (x_i - mean(x_i)) / sd(x_i). Defaults to `true`. For
|
|
|
|
|
more information, see
|
|
|
|
|
https://en.wikipedia.org/wiki/Feature_scaling#Standardization_(Z-score_Normalization)[this wiki page about standardization].
|
|
|
|
|
end::standardization-enabled[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::state-anomaly-job[]
|
|
|
|
|
The status of the {anomaly-job}, which can be one of the following values:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* `closed`: The job finished successfully with its model state persisted. The
|
|
|
|
|
job must be opened before it can accept further data.
|
|
|
|
|
* `closing`: The job close action is in progress and has not yet completed. A
|
|
|
|
|
closing job cannot accept further data.
|
|
|
|
|
* `failed`: The job did not finish successfully due to an error. This situation
|
2020-03-05 10:25:43 -05:00
|
|
|
|
can occur due to invalid input data, a fatal error occurring during the
|
|
|
|
|
analysis, or an external interaction such as the process being killed by the
|
|
|
|
|
Linux out of memory (OOM) killer. If the job had irrevocably failed, it must be
|
|
|
|
|
force closed and then deleted. If the {dfeed} can be corrected, the job can be
|
|
|
|
|
closed and then re-opened.
|
2020-03-02 10:28:55 -05:00
|
|
|
|
* `opened`: The job is available to receive and process data.
|
|
|
|
|
* `opening`: The job open action is in progress and has not yet completed.
|
|
|
|
|
--
|
|
|
|
|
end::state-anomaly-job[]
|
|
|
|
|
|
2020-02-26 12:20:36 -05:00
|
|
|
|
tag::state-datafeed[]
|
|
|
|
|
The status of the {dfeed}, which can be one of the following values:
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
* `started`: The {dfeed} is actively receiving data.
|
|
|
|
|
* `stopped`: The {dfeed} is stopped and will not receive data until it is
|
|
|
|
|
re-started.
|
|
|
|
|
--
|
|
|
|
|
end::state-datafeed[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::summary-count-field-name[]
|
|
|
|
|
If this property is specified, the data that is fed to the job is expected to be
|
|
|
|
|
pre-summarized. This property value is the name of the field that contains the
|
|
|
|
|
count of raw data points that have been summarized. The same
|
|
|
|
|
`summary_count_field_name` applies to all detectors in the job.
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: The `summary_count_field_name` property cannot be used with the `metric`
|
|
|
|
|
function.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::summary-count-field-name[]
|
|
|
|
|
|
2020-02-12 16:43:21 -05:00
|
|
|
|
tag::tags[]
|
|
|
|
|
A comma delimited string of tags. A {infer} model can have many tags, or none.
|
|
|
|
|
When supplied, only {infer} models that contain all the supplied tags are
|
|
|
|
|
returned.
|
|
|
|
|
end::tags[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::time-format[]
|
|
|
|
|
The time format, which can be `epoch`, `epoch_ms`, or a custom pattern. The
|
|
|
|
|
default value is `epoch`, which refers to UNIX or Epoch time (the number of
|
|
|
|
|
seconds since 1 Jan 1970). The value `epoch_ms` indicates that time is measured
|
|
|
|
|
in milliseconds since the epoch. The `epoch` and `epoch_ms` time formats accept
|
|
|
|
|
either integer or real values. +
|
|
|
|
|
+
|
|
|
|
|
--
|
|
|
|
|
NOTE: Custom patterns must conform to the Java `DateTimeFormatter` class.
|
|
|
|
|
When you use date-time formatting patterns, it is recommended that you provide
|
|
|
|
|
the full date, time and time zone. For example: `yyyy-MM-dd'T'HH:mm:ssX`.
|
|
|
|
|
If the pattern that you specify is not sufficient to produce a complete
|
|
|
|
|
timestamp, job creation fails.
|
|
|
|
|
|
|
|
|
|
--
|
|
|
|
|
end::time-format[]
|
|
|
|
|
|
2019-12-24 13:22:05 -05:00
|
|
|
|
tag::time-span[]
|
|
|
|
|
The time span that each search will be querying. This setting is only applicable
|
|
|
|
|
when the mode is set to `manual`. For example: `3h`.
|
|
|
|
|
end::time-span[]
|
|
|
|
|
|
|
|
|
|
tag::timeout-start[]
|
|
|
|
|
Controls the amount of time to wait until the {dfanalytics-job} starts. Defaults
|
|
|
|
|
to 20 seconds.
|
|
|
|
|
end::timeout-start[]
|
|
|
|
|
|
|
|
|
|
tag::timeout-stop[]
|
|
|
|
|
Controls the amount of time to wait until the {dfanalytics-job} stops. Defaults
|
|
|
|
|
to 20 seconds.
|
|
|
|
|
end::timeout-stop[]
|
|
|
|
|
|
|
|
|
|
tag::timestamp-results[]
|
|
|
|
|
The start time of the bucket for which these results were calculated.
|
|
|
|
|
end::timestamp-results[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::tokenizer[]
|
|
|
|
|
The name or definition of the <<analysis-tokenizers,tokenizer>> to use after
|
|
|
|
|
character filters are applied. This property is compulsory if
|
|
|
|
|
`categorization_analyzer` is specified as an object. Machine learning provides a
|
|
|
|
|
tokenizer called `ml_classic` that tokenizes in the same way as the
|
|
|
|
|
non-customizable tokenizer in older versions of the product. If you want to use
|
|
|
|
|
that tokenizer but change the character or token filters, specify
|
|
|
|
|
`"tokenizer": "ml_classic"` in your `categorization_analyzer`.
|
|
|
|
|
end::tokenizer[]
|
|
|
|
|
|
2020-03-02 10:28:55 -05:00
|
|
|
|
tag::total-by-field-count[]
|
|
|
|
|
The number of `by` field values that were analyzed by the models. This value is
|
|
|
|
|
cumulative for all detectors in the job.
|
|
|
|
|
end::total-by-field-count[]
|
|
|
|
|
|
|
|
|
|
tag::total-category-count[]
|
|
|
|
|
The number of categories created by categorization.
|
|
|
|
|
end::total-category-count[]
|
|
|
|
|
|
|
|
|
|
tag::total-over-field-count[]
|
|
|
|
|
The number of `over` field values that were analyzed by the models. This value
|
|
|
|
|
is cumulative for all detectors in the job.
|
|
|
|
|
end::total-over-field-count[]
|
|
|
|
|
|
|
|
|
|
tag::total-partition-field-count[]
|
|
|
|
|
The number of `partition` field values that were analyzed by the models. This
|
|
|
|
|
value is cumulative for all detectors in the job.
|
|
|
|
|
end::total-partition-field-count[]
|
|
|
|
|
|
2020-03-03 10:25:24 -05:00
|
|
|
|
tag::trained-model-configs[]
|
|
|
|
|
An array of trained model resources, which are sorted by the `model_id` value in
|
|
|
|
|
ascending order.
|
|
|
|
|
|
|
|
|
|
`model_id`:::
|
|
|
|
|
(string)
|
|
|
|
|
Idetifier for the trained model.
|
|
|
|
|
|
|
|
|
|
`created_by`:::
|
|
|
|
|
(string)
|
|
|
|
|
Information on the creator of the trained model.
|
|
|
|
|
|
|
|
|
|
`version`:::
|
|
|
|
|
(string)
|
|
|
|
|
The {es} version number in which the trained model was created.
|
|
|
|
|
|
|
|
|
|
`create_time`:::
|
|
|
|
|
(<<time-units,time units>>)
|
|
|
|
|
The time when the trained model was created.
|
|
|
|
|
|
|
|
|
|
`tags`:::
|
|
|
|
|
(string)
|
|
|
|
|
A comma delimited string of tags. A {infer} model can have many tags, or none.
|
|
|
|
|
|
|
|
|
|
`metadata`:::
|
|
|
|
|
(object)
|
|
|
|
|
An object containing metadata about the trained model. For example, models
|
|
|
|
|
created by {dfanalytics} contain an `analysis_config` and an `input`
|
|
|
|
|
object.
|
|
|
|
|
|
|
|
|
|
`estimated_heap_memory_usage_bytes`:::
|
|
|
|
|
(integer)
|
|
|
|
|
The estimated heap usage in bytes to keep the trained model in memory.
|
|
|
|
|
|
|
|
|
|
`estimated_operations`:::
|
|
|
|
|
(integer)
|
|
|
|
|
The estimated number of operations to use the trained model.
|
|
|
|
|
|
|
|
|
|
`license_level`:::
|
|
|
|
|
(string)
|
|
|
|
|
The license level of the trained model.
|
2020-03-11 13:49:39 -04:00
|
|
|
|
|
|
|
|
|
`default_field_map` :::
|
|
|
|
|
(object)
|
|
|
|
|
A string to string object that contains the default field map to use
|
|
|
|
|
when inferring against the model. For example, data frame analytics
|
|
|
|
|
may train the model on a specific multi-field `foo.keyword`.
|
|
|
|
|
The analytics job would then supply a default field map entry for
|
|
|
|
|
`"foo" : "foo.keyword"`.
|
|
|
|
|
|
|
|
|
|
Any field map described in the inference configuration takes precedence.
|
|
|
|
|
|
2020-03-03 10:25:24 -05:00
|
|
|
|
end::trained-model-configs[]
|
|
|
|
|
|
2019-12-13 05:48:21 -05:00
|
|
|
|
tag::training-percent[]
|
|
|
|
|
Defines what percentage of the eligible documents that will
|
|
|
|
|
be used for training. Documents that are ignored by the analysis (for example
|
2020-01-03 08:49:43 -05:00
|
|
|
|
those that contain arrays with more than one value) won’t be included in the
|
|
|
|
|
calculation for used percentage. Defaults to `100`.
|
2019-12-13 05:48:21 -05:00
|
|
|
|
end::training-percent[]
|
2019-12-10 08:29:19 -05:00
|
|
|
|
|
|
|
|
|
tag::use-null[]
|
|
|
|
|
Defines whether a new series is used as the null series when there is no value
|
|
|
|
|
for the by or partition fields. The default value is `false`.
|
|
|
|
|
end::use-null[]
|