2017-06-19 21:23:58 -04:00
|
|
|
[role="xpack"]
|
2018-08-31 19:49:24 -04:00
|
|
|
[testenv="platinum"]
|
2017-04-04 18:26:39 -04:00
|
|
|
[[ml-get-job-stats]]
|
2019-07-30 13:52:23 -04:00
|
|
|
=== Get {anomaly-job} statistics API
|
2017-12-14 13:52:49 -05:00
|
|
|
++++
|
2018-12-20 13:23:28 -05:00
|
|
|
<titleabbrev>Get job statistics</titleabbrev>
|
2017-12-14 13:52:49 -05:00
|
|
|
++++
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-07-30 13:52:23 -04:00
|
|
|
Retrieves usage information for {anomaly-jobs}.
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-06-27 12:42:47 -04:00
|
|
|
[[ml-get-job-stats-request]]
|
|
|
|
==== {api-request-title}
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2018-12-07 15:34:11 -05:00
|
|
|
`GET _ml/anomaly_detectors/<job_id>/_stats`
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2018-12-07 15:34:11 -05:00
|
|
|
`GET _ml/anomaly_detectors/<job_id>,<job_id>/_stats` +
|
2017-08-09 11:30:21 -04:00
|
|
|
|
2018-12-07 15:34:11 -05:00
|
|
|
`GET _ml/anomaly_detectors/_stats` +
|
2017-08-09 11:30:21 -04:00
|
|
|
|
2019-06-27 16:58:42 -04:00
|
|
|
`GET _ml/anomaly_detectors/_all/_stats`
|
|
|
|
|
|
|
|
[[ml-get-job-stats-prereqs]]
|
|
|
|
==== {api-prereq-title}
|
|
|
|
|
|
|
|
* If the {es} {security-features} are enabled, you must have `monitor_ml`,
|
|
|
|
`monitor`, `manage_ml`, or `manage` cluster privileges to use this API. See
|
2019-10-07 18:23:19 -04:00
|
|
|
<<security-privileges>>.
|
2017-08-09 11:30:21 -04:00
|
|
|
|
2019-06-27 12:42:47 -04:00
|
|
|
[[ml-get-job-stats-desc]]
|
|
|
|
==== {api-description-title}
|
2017-08-09 11:30:21 -04:00
|
|
|
|
2019-07-30 13:52:23 -04:00
|
|
|
You can get statistics for multiple {anomaly-jobs} in a single API request by
|
|
|
|
using a group name, a comma-separated list of jobs, or a wildcard expression.
|
|
|
|
You can get statistics for all {anomaly-jobs} by using `_all`, by specifying `*`
|
|
|
|
as the `<job_id>`, or by omitting the `<job_id>`.
|
2017-08-09 11:30:21 -04:00
|
|
|
|
2019-01-17 13:47:15 -05:00
|
|
|
IMPORTANT: This API returns a maximum of 10,000 jobs.
|
|
|
|
|
2019-06-27 12:42:47 -04:00
|
|
|
[[ml-get-job-stats-path-parms]]
|
|
|
|
==== {api-path-parms-title}
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-07-12 11:26:31 -04:00
|
|
|
`<job_id>`::
|
2019-12-27 16:30:26 -05:00
|
|
|
(Optional, string)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=job-id-anomaly-detection-default]
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-07-30 08:22:14 -04:00
|
|
|
[[ml-get-job-stats-query-parms]]
|
|
|
|
==== {api-query-parms-title}
|
|
|
|
|
|
|
|
`allow_no_jobs`::
|
2019-12-27 16:30:26 -05:00
|
|
|
(Optional, boolean)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=allow-no-jobs]
|
2019-07-30 08:22:14 -04:00
|
|
|
|
2019-06-27 12:42:47 -04:00
|
|
|
[[ml-get-job-stats-results]]
|
|
|
|
==== {api-response-body-title}
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-12-30 17:55:36 -05:00
|
|
|
The API returns the following information about the operational progress of a
|
|
|
|
job:
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-12-30 17:55:36 -05:00
|
|
|
`assignment_explanation`::
|
|
|
|
(string) For open jobs only, contains messages relating to the selection of a
|
|
|
|
node to run the job.
|
|
|
|
|
|
|
|
[[datacounts]]`data_counts`::
|
|
|
|
(object) An object that describes the quantity of input to the job and any
|
|
|
|
related error counts. The `data_count` values are cumulative for the lifetime of
|
|
|
|
a job. If a model snapshot is reverted or old results are deleted, the job
|
|
|
|
counts are not reset.
|
|
|
|
|
|
|
|
`data_counts`.`bucket_count`:::
|
|
|
|
(long) The number of bucket results produced by the job.
|
|
|
|
|
|
|
|
`data_counts`.`earliest_record_timestamp`:::
|
|
|
|
(date) The timestamp of the earliest chronologically input document.
|
|
|
|
|
|
|
|
`data_counts`.`empty_bucket_count`:::
|
|
|
|
(long) The number of buckets which did not contain any data. If your data
|
|
|
|
contains many empty buckets, consider increasing your `bucket_span` or using
|
|
|
|
functions that are tolerant to gaps in data such as `mean`, `non_null_sum` or
|
|
|
|
`non_zero_count`.
|
|
|
|
|
|
|
|
`data_counts`.`input_bytes`:::
|
|
|
|
(long) The number of raw bytes read by the job.
|
|
|
|
|
|
|
|
`data_counts`.`input_field_count`:::
|
|
|
|
(long) The total number of fields in input documents posted to the job. This
|
|
|
|
count includes fields that are not used in the analysis. However, be aware that
|
|
|
|
if you are using a {dfeed}, it extracts only the required fields from the
|
|
|
|
documents it retrieves before posting them to the job.
|
|
|
|
|
|
|
|
`data_counts`.`input_record_count`:::
|
|
|
|
(long) The number of data records read by the job.
|
|
|
|
|
|
|
|
`data_counts`.`invalid_date_count`:::
|
|
|
|
(long) The number of records with either a missing date field or a date that
|
|
|
|
could not be parsed.
|
|
|
|
|
|
|
|
`data_counts`.`job_id`:::
|
|
|
|
(string)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=job-id-anomaly-detection]
|
|
|
|
|
|
|
|
`data_counts`.`last_data_time`:::
|
|
|
|
(date) The timestamp at which data was last analyzed, according to server time.
|
|
|
|
|
|
|
|
`data_counts`.`latest_empty_bucket_timestamp`:::
|
|
|
|
(date) The timestamp of the last bucket that did not contain any data.
|
|
|
|
|
|
|
|
`data_counts`.`latest_record_timestamp`:::
|
|
|
|
(date) The timestamp of the latest chronologically input document.
|
|
|
|
|
|
|
|
`data_counts`.`latest_sparse_bucket_timestamp`:::
|
|
|
|
(date) The timestamp of the last bucket that was considered sparse.
|
|
|
|
|
|
|
|
`data_counts`.`missing_field_count`:::
|
|
|
|
(long) The number of input documents that are missing a field that the job is
|
|
|
|
configured to analyze. Input documents with missing fields are still processed
|
|
|
|
because it is possible that not all fields are missing. The value of
|
|
|
|
`processed_record_count` includes this count.
|
|
|
|
+
|
|
|
|
--
|
|
|
|
NOTE: If you are using {dfeeds} or posting data to the job in JSON format, a
|
|
|
|
high `missing_field_count` is often not an indication of data issues. It is not
|
|
|
|
necessarily a cause for concern.
|
|
|
|
|
|
|
|
--
|
|
|
|
|
|
|
|
`data_counts`.`out_of_order_timestamp_count`:::
|
|
|
|
(long) The number of input documents that are out of time sequence and outside
|
|
|
|
of the latency window. This information is applicable only when you provide data
|
|
|
|
to the job by using the <<ml-post-data,post data API>>. These out of order
|
|
|
|
documents are discarded, since jobs require time series data to be in ascending
|
|
|
|
chronological order.
|
|
|
|
|
|
|
|
`data_counts`.`processed_field_count`:::
|
|
|
|
(long) The total number of fields in all the documents that have been processed
|
|
|
|
by the job. Only fields that are specified in the detector configuration object
|
|
|
|
contribute to this count. The time stamp is not included in this count.
|
|
|
|
|
|
|
|
`data_counts`.`processed_record_count`:::
|
|
|
|
(long) The number of input documents that have been processed by the job. This
|
|
|
|
value includes documents with missing fields, since they are nonetheless
|
|
|
|
analyzed. If you use {dfeeds} and have aggregations in your search query, the
|
|
|
|
`processed_record_count` will be the number of aggregation results processed,
|
|
|
|
not the number of {es} documents.
|
|
|
|
|
|
|
|
`data_counts`.`sparse_bucket_count`:::
|
|
|
|
(long) The number of buckets that contained few data points compared to the
|
|
|
|
expected number of data points. If your data contains many sparse buckets,
|
|
|
|
consider using a longer `bucket_span`.
|
|
|
|
|
|
|
|
[[forecastsstats]]`forecasts_stats`::
|
|
|
|
(object) An object that provides statistical information about forecasts
|
|
|
|
of this job. It has the following properties:
|
|
|
|
+
|
|
|
|
--
|
|
|
|
NOTE: `memory_bytes`, `records`, `processing_time_ms` and `status` require at
|
|
|
|
least one forecast. Otherwise, these fields are omitted.
|
|
|
|
|
|
|
|
--
|
|
|
|
|
|
|
|
`forecasts_stats`.`forecasted_jobs`:::
|
|
|
|
(long) The number of jobs that have at least one forecast.
|
|
|
|
|
|
|
|
`forecasts_stats`.`memory_bytes`:::
|
|
|
|
(object) Statistics about the memory usage: minimum, maximum, average and total.
|
|
|
|
|
|
|
|
`forecasts_stats`.`processing_time_ms`:::
|
|
|
|
(object) Statistics about the forecast runtime in milliseconds: minimum, maximum, average and total.
|
|
|
|
|
|
|
|
`forecasts_stats`.`records`:::
|
|
|
|
(object) Statistics about the number of forecast records: minimum, maximum,
|
|
|
|
average and total.
|
|
|
|
|
|
|
|
`forecasts_stats`.`status`:::
|
|
|
|
(object) Counts per forecast status, for example: `{"finished" : 2}`.
|
|
|
|
|
|
|
|
`forecasts_stats`.`total`:::
|
|
|
|
(long) The number of forecasts currently available for this model.
|
|
|
|
|
|
|
|
`job_id`::
|
|
|
|
(string)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=job-id-anomaly-detection]
|
|
|
|
|
|
|
|
[[modelsizestats]]`model_size_stats`::
|
|
|
|
(object) An object that provides information about the size and contents of the
|
|
|
|
model. It has the following properties:
|
|
|
|
|
|
|
|
`model_size_stats`.`bucket_allocation_failures_count`:::
|
|
|
|
(long) The number of buckets for which new entities in incoming data were not
|
|
|
|
processed due to insufficient model memory. This situation is also signified
|
|
|
|
by a `hard_limit: memory_status` property value.
|
|
|
|
|
|
|
|
`model_size_stats`.`job_id`:::
|
|
|
|
(string)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=job-id-anomaly-detection]
|
|
|
|
|
|
|
|
`model_size_stats`.`log_time`:::
|
|
|
|
(date) The timestamp of the `model_size_stats` according to server time.
|
|
|
|
|
|
|
|
`model_size_stats`.`memory_status`:::
|
|
|
|
(string) The status of the mathematical models. This property can have one of
|
|
|
|
the following values:
|
|
|
|
+
|
|
|
|
--
|
|
|
|
* `ok`: The models stayed below the configured value.
|
|
|
|
* `soft_limit`: The models used more than 60% of the configured memory limit and
|
|
|
|
older unused models will be pruned to free up space.
|
|
|
|
* `hard_limit`: The models used more space than the configured memory limit. As
|
|
|
|
a result, not all incoming data was processed.
|
|
|
|
--
|
|
|
|
|
|
|
|
`model_size_stats`.`model_bytes`:::
|
|
|
|
(long) The number of bytes of memory used by the models. This is the maximum
|
|
|
|
value since the last time the model was persisted. If the job is closed,
|
|
|
|
this value indicates the latest size.
|
|
|
|
|
|
|
|
`model_size_stats`.`model_bytes_exceeded`:::
|
|
|
|
(long) The number of bytes over the high limit for memory usage at the last
|
|
|
|
allocation failure.
|
|
|
|
|
|
|
|
`model_size_stats`.`model_bytes_memory_limit`:::
|
|
|
|
(long) The upper limit for memory usage, checked on increasing values.
|
|
|
|
|
|
|
|
`model_size_stats`.`result_type`:::
|
|
|
|
(string) For internal use. The type of result.
|
|
|
|
|
|
|
|
`model_size_stats`.`total_by_field_count`:::
|
|
|
|
(long) The number of `by` field values that were analyzed by the models.
|
|
|
|
+
|
|
|
|
--
|
|
|
|
NOTE: The `by` field values are counted separately for each detector and
|
|
|
|
partition.
|
|
|
|
|
|
|
|
--
|
|
|
|
|
|
|
|
`model_size_stats`.`total_over_field_count`:::
|
|
|
|
(long) The number of `over` field values that were analyzed by the models.
|
|
|
|
+
|
|
|
|
--
|
|
|
|
NOTE: The `over` field values are counted separately for each detector and
|
|
|
|
partition.
|
|
|
|
|
|
|
|
--
|
|
|
|
|
|
|
|
`model_size_stats`.`total_partition_field_count`:::
|
|
|
|
(long) The number of `partition` field values that were analyzed by the models.
|
|
|
|
|
|
|
|
`model_size_stats`.`timestamp`:::
|
|
|
|
(date) The timestamp of the `model_size_stats` according to the timestamp of the
|
|
|
|
data.
|
|
|
|
|
|
|
|
[[stats-node]]`node`::
|
|
|
|
(object) Contains properties for the node that runs the job. This information is
|
|
|
|
available only for open jobs.
|
|
|
|
|
|
|
|
`node`.`attributes`:::
|
|
|
|
(object) Lists node attributes. For example,
|
|
|
|
`{"ml.machine_memory": "17179869184"}`.
|
|
|
|
|
|
|
|
`node`.`ephemeral_id`:::
|
|
|
|
(string) The ephemeral id of the node.
|
|
|
|
|
|
|
|
`node`.`id`:::
|
|
|
|
(string) The unique identifier of the node.
|
|
|
|
|
|
|
|
`node`.`name`:::
|
|
|
|
(string) The node name.
|
|
|
|
|
|
|
|
`node`.`transport_address`:::
|
|
|
|
(string) The host and port where transport HTTP connections are accepted.
|
|
|
|
|
|
|
|
`open_time`::
|
|
|
|
(string) For open jobs only, the elapsed time for which the job has been open.
|
|
|
|
For example, `28746386s`.
|
|
|
|
|
|
|
|
`state`::
|
|
|
|
(string) The status of the job, which can be one of the following values:
|
|
|
|
+
|
|
|
|
--
|
|
|
|
* `closed`: The job finished successfully with its model state persisted. The
|
|
|
|
job must be opened before it can accept further data.
|
|
|
|
* `closing`: The job close action is in progress and has not yet completed. A
|
|
|
|
closing job cannot accept further data.
|
|
|
|
* `failed`: The job did not finish successfully due to an error. This situation
|
|
|
|
can occur due to invalid input data. If the job had irrevocably failed, it must
|
|
|
|
be force closed and then deleted. If the {dfeed} can be corrected, the job can
|
|
|
|
be closed and then re-opened.
|
|
|
|
* `opened`: The job is available to receive and process data.
|
|
|
|
* `opening`: The job open action is in progress and has not yet completed.
|
|
|
|
--
|
|
|
|
|
|
|
|
[[timingstats]]`timing_stats`::
|
|
|
|
(object) An object that provides statistical information about timing aspect of
|
|
|
|
this job. It has the following properties:
|
|
|
|
|
|
|
|
`timing_stats`.`average_bucket_processing_time_ms`:::
|
|
|
|
(double) Average of all bucket processing times in milliseconds.
|
|
|
|
|
|
|
|
`timing_stats`.`bucket_count`:::
|
|
|
|
(long) The number of buckets processed.
|
|
|
|
|
|
|
|
`timing_stats`.`exponential_average_bucket_processing_time_ms`:::
|
|
|
|
(double) Exponential moving average of all bucket processing times in
|
|
|
|
milliseconds.
|
|
|
|
|
|
|
|
`timing_stats`.`exponential_average_bucket_processing_time_per_hour_ms`:::
|
|
|
|
(double) Exponentially-weighted moving average of bucket processing times
|
|
|
|
calculated in a 1 hour time window.
|
|
|
|
|
|
|
|
`timing_stats`.`job_id`:::
|
|
|
|
(string)
|
|
|
|
include::{docdir}/ml/ml-shared.asciidoc[tag=job-id-anomaly-detection]
|
|
|
|
|
|
|
|
`timing_stats`.`maximum_bucket_processing_time_ms`:::
|
|
|
|
(double) Maximum among all bucket processing times in milliseconds.
|
|
|
|
|
|
|
|
`timing_stats`.`minimum_bucket_processing_time_ms`:::
|
|
|
|
(double) Minimum among all bucket processing times in milliseconds.
|
|
|
|
|
|
|
|
`timing_stats`.`total_bucket_processing_time_ms`:::
|
|
|
|
(double) Sum of all bucket processing times in milliseconds.
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-07-30 08:22:14 -04:00
|
|
|
[[ml-get-job-stats-response-codes]]
|
|
|
|
==== {api-response-codes-title}
|
|
|
|
|
|
|
|
`404` (Missing resources)::
|
|
|
|
If `allow_no_jobs` is `false`, this code indicates that there are no
|
|
|
|
resources that match the request or only partial matches for the request.
|
|
|
|
|
2019-06-27 12:42:47 -04:00
|
|
|
[[ml-get-job-stats-example]]
|
|
|
|
==== {api-examples-title}
|
2017-04-04 18:26:39 -04:00
|
|
|
|
2019-09-06 11:31:13 -04:00
|
|
|
[source,console]
|
2017-04-11 21:52:47 -04:00
|
|
|
--------------------------------------------------
|
2019-12-30 17:55:36 -05:00
|
|
|
GET _ml/anomaly_detectors/low_request_rate/_stats
|
2017-04-11 21:52:47 -04:00
|
|
|
--------------------------------------------------
|
2019-12-30 17:55:36 -05:00
|
|
|
// TEST[skip:Kibana sample data]
|
2017-04-11 21:52:47 -04:00
|
|
|
|
2017-05-08 09:53:04 -04:00
|
|
|
The API returns the following results:
|
2017-04-21 11:23:27 -04:00
|
|
|
[source,js]
|
2017-04-04 18:26:39 -04:00
|
|
|
----
|
|
|
|
{
|
2019-12-30 17:55:36 -05:00
|
|
|
"count" : 1,
|
|
|
|
"jobs" : [
|
2017-04-04 18:26:39 -04:00
|
|
|
{
|
2019-12-30 17:55:36 -05:00
|
|
|
"job_id" : "low_request_rate",
|
|
|
|
"data_counts" : {
|
|
|
|
"job_id" : "low_request_rate",
|
|
|
|
"processed_record_count" : 1216,
|
|
|
|
"processed_field_count" : 1216,
|
|
|
|
"input_bytes" : 51678,
|
|
|
|
"input_field_count" : 1216,
|
|
|
|
"invalid_date_count" : 0,
|
|
|
|
"missing_field_count" : 0,
|
|
|
|
"out_of_order_timestamp_count" : 0,
|
|
|
|
"empty_bucket_count" : 242,
|
|
|
|
"sparse_bucket_count" : 0,
|
|
|
|
"bucket_count" : 1457,
|
|
|
|
"earliest_record_timestamp" : 1575172659612,
|
|
|
|
"latest_record_timestamp" : 1580417369440,
|
|
|
|
"last_data_time" : 1576017595046,
|
|
|
|
"latest_empty_bucket_timestamp" : 1580356800000,
|
|
|
|
"input_record_count" : 1216
|
2017-04-04 18:26:39 -04:00
|
|
|
},
|
2019-12-30 17:55:36 -05:00
|
|
|
"model_size_stats" : {
|
|
|
|
"job_id" : "low_request_rate",
|
|
|
|
"result_type" : "model_size_stats",
|
|
|
|
"model_bytes" : 41480,
|
|
|
|
"model_bytes_exceeded" : 0,
|
|
|
|
"model_bytes_memory_limit" : 10485760,
|
|
|
|
"total_by_field_count" : 3,
|
|
|
|
"total_over_field_count" : 0,
|
|
|
|
"total_partition_field_count" : 2,
|
|
|
|
"bucket_allocation_failures_count" : 0,
|
|
|
|
"memory_status" : "ok",
|
|
|
|
"log_time" : 1576017596000,
|
|
|
|
"timestamp" : 1580410800000
|
2017-04-04 18:26:39 -04:00
|
|
|
},
|
2019-12-30 17:55:36 -05:00
|
|
|
"forecasts_stats" : {
|
|
|
|
"total" : 1,
|
|
|
|
"forecasted_jobs" : 1,
|
|
|
|
"memory_bytes" : {
|
|
|
|
"total" : 9179.0,
|
|
|
|
"min" : 9179.0,
|
|
|
|
"avg" : 9179.0,
|
|
|
|
"max" : 9179.0
|
|
|
|
},
|
|
|
|
"records" : {
|
|
|
|
"total" : 168.0,
|
|
|
|
"min" : 168.0,
|
|
|
|
"avg" : 168.0,
|
|
|
|
"max" : 168.0
|
|
|
|
},
|
|
|
|
"processing_time_ms" : {
|
|
|
|
"total" : 40.0,
|
|
|
|
"min" : 40.0,
|
|
|
|
"avg" : 40.0,
|
|
|
|
"max" : 40.0
|
|
|
|
},
|
|
|
|
"status" : {
|
|
|
|
"finished" : 1
|
|
|
|
}
|
|
|
|
},
|
|
|
|
"state" : "opened",
|
|
|
|
"node" : {
|
|
|
|
"id" : "7bmMXyWCRs-TuPfGJJ_yMw",
|
|
|
|
"name" : "node-0",
|
|
|
|
"ephemeral_id" : "hoXMLZB0RWKfR9UPPUCxXX",
|
|
|
|
"transport_address" : "127.0.0.1:9300",
|
|
|
|
"attributes" : {
|
|
|
|
"ml.machine_memory" : "17179869184",
|
|
|
|
"xpack.installed" : "true",
|
|
|
|
"ml.max_open_jobs" : "20"
|
|
|
|
}
|
|
|
|
},
|
|
|
|
"assignment_explanation" : "",
|
|
|
|
"open_time" : "13s",
|
|
|
|
"timing_stats" : {
|
|
|
|
"job_id" : "low_request_rate",
|
|
|
|
"bucket_count" : 1457,
|
|
|
|
"total_bucket_processing_time_ms" : 1094.000000000001,
|
|
|
|
"minimum_bucket_processing_time_ms" : 0.0,
|
|
|
|
"maximum_bucket_processing_time_ms" : 48.0,
|
|
|
|
"average_bucket_processing_time_ms" : 0.75085792724777,
|
|
|
|
"exponential_average_bucket_processing_time_ms" : 0.5571716855800993,
|
|
|
|
"exponential_average_bucket_processing_time_per_hour_ms" : 15.0
|
2019-06-17 02:58:26 -04:00
|
|
|
}
|
2017-04-04 18:26:39 -04:00
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
2019-12-30 17:55:36 -05:00
|
|
|
----
|