[DOCS] Updates terms in machine learning datafeed APIs (#44883)

This commit is contained in:
Lisa Cawley 2019-07-26 10:47:03 -07:00 committed by lcawl
parent d4b2d21339
commit cef375f883
7 changed files with 49 additions and 45 deletions

View File

@ -4,10 +4,12 @@
:response: AcknowledgedResponse :response: AcknowledgedResponse
-- --
[id="{upid}-delete-datafeed"] [id="{upid}-delete-datafeed"]
=== Delete Datafeed API === Delete datafeed API
Deletes an existing datafeed.
[id="{upid}-{api}-request"] [id="{upid}-{api}-request"]
==== Delete Datafeed Request ==== Delete datafeed request
A +{request}+ object requires a non-null `datafeedId` and can optionally set `force`. A +{request}+ object requires a non-null `datafeedId` and can optionally set `force`.
@ -15,18 +17,17 @@ A +{request}+ object requires a non-null `datafeedId` and can optionally set `fo
--------------------------------------------------- ---------------------------------------------------
include-tagged::{doc-tests-file}[{api}-request] include-tagged::{doc-tests-file}[{api}-request]
--------------------------------------------------- ---------------------------------------------------
<1> Use to forcefully delete a started datafeed; <1> Use to forcefully delete a started datafeed. This method is quicker than
this method is quicker than stopping and deleting the datafeed. stopping and deleting the datafeed. Defaults to `false`.
Defaults to `false`.
include::../execution.asciidoc[] include::../execution.asciidoc[]
[id="{upid}-{api}-response"] [id="{upid}-{api}-response"]
==== Delete Datafeed Response ==== Delete datafeed response
The returned +{response}+ object indicates the acknowledgement of the request: The returned +{response}+ object indicates the acknowledgement of the request:
["source","java",subs="attributes,callouts,macros"] ["source","java",subs="attributes,callouts,macros"]
--------------------------------------------------- ---------------------------------------------------
include-tagged::{doc-tests-file}[{api}-response] include-tagged::{doc-tests-file}[{api}-response]
--------------------------------------------------- ---------------------------------------------------
<1> `isAcknowledged` was the deletion request acknowledged or not <1> `isAcknowledged` was the deletion request acknowledged or not.

View File

@ -4,14 +4,13 @@
:response: PutDatafeedResponse :response: PutDatafeedResponse
-- --
[id="{upid}-{api}"] [id="{upid}-{api}"]
=== Put Datafeed API === Put datafeed API
The Put Datafeed API can be used to create a new {ml} datafeed Creates a new {ml} datafeed in the cluster. The API accepts a +{request}+ object
in the cluster. The API accepts a +{request}+ object
as a request and returns a +{response}+. as a request and returns a +{response}+.
[id="{upid}-{api}-request"] [id="{upid}-{api}-request"]
==== Put Datafeed Request ==== Put datafeed request
A +{request}+ requires the following argument: A +{request}+ requires the following argument:
@ -19,10 +18,10 @@ A +{request}+ requires the following argument:
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-request] include-tagged::{doc-tests-file}[{api}-request]
-------------------------------------------------- --------------------------------------------------
<1> The configuration of the {ml} datafeed to create <1> The configuration of the {ml} datafeed to create.
[id="{upid}-{api}-config"] [id="{upid}-{api}-config"]
==== Datafeed Configuration ==== Datafeed configuration
The `DatafeedConfig` object contains all the details about the {ml} datafeed The `DatafeedConfig` object contains all the details about the {ml} datafeed
configuration. configuration.
@ -33,10 +32,10 @@ A `DatafeedConfig` requires the following arguments:
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-config] include-tagged::{doc-tests-file}[{api}-config]
-------------------------------------------------- --------------------------------------------------
<1> The datafeed ID and the job ID <1> The datafeed ID and the {anomaly-job} ID.
<2> The indices that contain the data to retrieve and feed into the job <2> The indices that contain the data to retrieve and feed into the {anomaly-job}.
==== Optional Arguments ==== Optional arguments
The following arguments are optional: The following arguments are optional:
["source","java",subs="attributes,callouts,macros"] ["source","java",subs="attributes,callouts,macros"]
@ -49,7 +48,8 @@ include-tagged::{doc-tests-file}[{api}-config-set-chunking-config]
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-config-set-frequency] include-tagged::{doc-tests-file}[{api}-config-set-frequency]
-------------------------------------------------- --------------------------------------------------
<1> The interval at which scheduled queries are made while the datafeed runs in real time. <1> The interval at which scheduled queries are made while the datafeed runs in
real time.
["source","java",subs="attributes,callouts,macros"] ["source","java",subs="attributes,callouts,macros"]
-------------------------------------------------- --------------------------------------------------
@ -72,8 +72,9 @@ The window must be larger than the Job's bucket size, but smaller than 24 hours,
and span less than 10,000 buckets. and span less than 10,000 buckets.
Defaults to `null`, which causes an appropriate window span to be calculated when Defaults to `null`, which causes an appropriate window span to be calculated when
the datafeed runs. the datafeed runs.
The default `check_window` span calculation is the max between `2h` or `8 * bucket_span`. The default `check_window` span calculation is the max between `2h` or
To explicitly disable, pass `DelayedDataCheckConfig.disabledDelayedDataCheckConfig()`. `8 * bucket_span`. To explicitly disable, pass
`DelayedDataCheckConfig.disabledDelayedDataCheckConfig()`.
["source","java",subs="attributes,callouts,macros"] ["source","java",subs="attributes,callouts,macros"]
-------------------------------------------------- --------------------------------------------------
@ -101,4 +102,4 @@ default values:
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-response] include-tagged::{doc-tests-file}[{api}-response]
-------------------------------------------------- --------------------------------------------------
<1> The created datafeed <1> The created datafeed.

View File

@ -4,14 +4,13 @@
:response: StartDatafeedResponse :response: StartDatafeedResponse
-- --
[id="{upid}-{api}"] [id="{upid}-{api}"]
=== Start Datafeed API === Start datafeed API
The Start Datafeed API provides the ability to start a {ml} datafeed in the cluster. Starts a {ml} datafeed in the cluster. It accepts a +{request}+ object and
It accepts a +{request}+ object and responds responds with a +{response}+ object.
with a +{response}+ object.
[id="{upid}-{api}-request"] [id="{upid}-{api}-request"]
==== Start Datafeed Request ==== Start datafeed request
A +{request}+ object is created referencing a non-null `datafeedId`. A +{request}+ object is created referencing a non-null `datafeedId`.
All other fields are optional for the request. All other fields are optional for the request.
@ -20,9 +19,9 @@ All other fields are optional for the request.
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-request] include-tagged::{doc-tests-file}[{api}-request]
-------------------------------------------------- --------------------------------------------------
<1> Constructing a new request referencing an existing `datafeedId` <1> Constructing a new request referencing an existing `datafeedId`.
==== Optional Arguments ==== Optional arguments
The following arguments are optional. The following arguments are optional.

View File

@ -4,14 +4,13 @@
:response: PutDatafeedResponse :response: PutDatafeedResponse
-- --
[id="{upid}-{api}"] [id="{upid}-{api}"]
=== Update Datafeed API === Update datafeed API
The Update Datafeed API can be used to update a {ml} datafeed Updates a {ml} datafeed in the cluster. The API accepts a +{request}+ object
in the cluster. The API accepts a +{request}+ object
as a request and returns a +{response}+. as a request and returns a +{response}+.
[id="{upid}-{api}-request"] [id="{upid}-{api}-request"]
==== Update Datafeed Request ==== Update datafeed request
A +{request}+ requires the following argument: A +{request}+ requires the following argument:
@ -22,7 +21,7 @@ include-tagged::{doc-tests-file}[{api}-request]
<1> The updated configuration of the {ml} datafeed <1> The updated configuration of the {ml} datafeed
[id="{upid}-{api}-config"] [id="{upid}-{api}-config"]
==== Updated Datafeed Arguments ==== Updated datafeed arguments
A `DatafeedUpdate` requires an existing non-null `datafeedId` and A `DatafeedUpdate` requires an existing non-null `datafeedId` and
allows updating various settings. allows updating various settings.
@ -31,12 +30,15 @@ allows updating various settings.
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-config] include-tagged::{doc-tests-file}[{api}-config]
-------------------------------------------------- --------------------------------------------------
<1> Mandatory, non-null `datafeedId` referencing an existing {ml} datafeed <1> Mandatory, non-null `datafeedId` referencing an existing {ml} datafeed.
<2> Optional, set the datafeed Aggregations for data gathering <2> Optional, set the datafeed aggregations for data gathering.
<3> Optional, the indices that contain the data to retrieve and feed into the job <3> Optional, the indices that contain the data to retrieve and feed into the
{anomaly-job}.
<4> Optional, specifies how data searches are split into time chunks. <4> Optional, specifies how data searches are split into time chunks.
<5> Optional, the interval at which scheduled queries are made while the datafeed runs in real time. <5> Optional, the interval at which scheduled queries are made while the
<6> Optional, a query to filter the search results by. Defaults to the `match_all` query. datafeed runs in real time.
<6> Optional, a query to filter the search results by. Defaults to the
`match_all` query.
<7> Optional, the time interval behind real time that data is queried. <7> Optional, the time interval behind real time that data is queried.
<8> Optional, allows the use of script fields. <8> Optional, allows the use of script fields.
<9> Optional, the `size` parameter used in the searches. <9> Optional, the `size` parameter used in the searches.
@ -53,4 +55,4 @@ the updated {ml} datafeed if it has been successfully updated.
-------------------------------------------------- --------------------------------------------------
include-tagged::{doc-tests-file}[{api}-response] include-tagged::{doc-tests-file}[{api}-response]
-------------------------------------------------- --------------------------------------------------
<1> The updated datafeed <1> The updated datafeed.

View File

@ -18,7 +18,7 @@ Instantiates a {dfeed}.
[[ml-put-datafeed-prereqs]] [[ml-put-datafeed-prereqs]]
==== {api-prereq-title} ==== {api-prereq-title}
* You must create a job before you create a {dfeed}. * You must create an {anomaly-job} before you create a {dfeed}.
* If {es} {security-features} are enabled, you must have `manage_ml` or `manage` * If {es} {security-features} are enabled, you must have `manage_ml` or `manage`
cluster privileges to use this API. See cluster privileges to use this API. See
{stack-ov}/security-privileges.html[Security privileges]. {stack-ov}/security-privileges.html[Security privileges].
@ -26,7 +26,7 @@ cluster privileges to use this API. See
[[ml-put-datafeed-desc]] [[ml-put-datafeed-desc]]
==== {api-description-title} ==== {api-description-title}
You can associate only one {dfeed} to each job. You can associate only one {dfeed} to each {anomaly-job}.
[IMPORTANT] [IMPORTANT]
==== ====
@ -75,7 +75,7 @@ those same roles.
`job_id`:: `job_id`::
(Required, string) A numerical character string that uniquely identifies the (Required, string) A numerical character string that uniquely identifies the
job. {anomaly-job}.
`query`:: `query`::
(Optional, object) The {es} query domain-specific language (DSL). This value (Optional, object) The {es} query domain-specific language (DSL). This value

View File

@ -18,8 +18,8 @@ Starts one or more {dfeeds}.
[[ml-start-datafeed-prereqs]] [[ml-start-datafeed-prereqs]]
==== {api-prereq-title} ==== {api-prereq-title}
* Before you can start a {dfeed}, the job must be open. Otherwise, an error * Before you can start a {dfeed}, the {anomaly-job} must be open. Otherwise, an
occurs. error occurs.
* If {es} {security-features} are enabled, you must have `manage_ml` or `manage` * If {es} {security-features} are enabled, you must have `manage_ml` or `manage`
cluster privileges to use this API. See cluster privileges to use this API. See
{stack-ov}/security-privileges.html[Security privileges]. {stack-ov}/security-privileges.html[Security privileges].
@ -36,7 +36,8 @@ If you want to analyze from the beginning of a dataset, you can specify any date
earlier than that beginning date. earlier than that beginning date.
If you do not specify a start time and the {dfeed} is associated with a new If you do not specify a start time and the {dfeed} is associated with a new
job, the analysis starts from the earliest time for which data is available. {anomaly-job}, the analysis starts from the earliest time for which data is
available.
When you start a {dfeed}, you can also specify an end time. If you do so, the When you start a {dfeed}, you can also specify an end time. If you do so, the
job analyzes data from the start time until the end time, at which point the job analyzes data from the start time until the end time, at which point the

View File

@ -67,7 +67,7 @@ The following properties can be updated after the {dfeed} is created:
`job_id`:: `job_id`::
(Optional, string) A numerical character string that uniquely identifies the (Optional, string) A numerical character string that uniquely identifies the
job. {anomaly-job}.
`query`:: `query`::
(Optional, object) The {es} query domain-specific language (DSL). This value (Optional, object) The {es} query domain-specific language (DSL). This value