[DOCS] Fix ML post_data docs (elastic/x-pack-elasticsearch#2689)
It was pointed out in https://github.com/elastic/elasticsearch-net/pull/2856#discussion_r142830656 that our post_data docs incorrectly say that reset_start and reset_end are body parameters. In fact they are query parameters. There were also a number of other errors and ommissions on this page that I have attempted to correct. Original commit: elastic/x-pack-elasticsearch@c83decacc7
This commit is contained in:
parent
a4f7db4f66
commit
c84d69fde3
|
@ -7,17 +7,22 @@ The post data API enables you to send data to an anomaly detection job for analy
|
|||
|
||||
==== Request
|
||||
|
||||
`POST _xpack/ml/anomaly_detectors/<job_id>/_data --data-binary @<data-file.json>`
|
||||
`POST _xpack/ml/anomaly_detectors/<job_id>/_data`
|
||||
|
||||
|
||||
==== Description
|
||||
|
||||
The job must have a state of `open` to receive and process the data.
|
||||
|
||||
The data that you send to the job must use the JSON format.
|
||||
The data that you send to the job must use the JSON format. Multiple JSON
|
||||
documents can be sent, either adjacent with no separator in between them or
|
||||
whitespace separated. Newline delimited JSON (NDJSON) is a possible whitespace
|
||||
separated format, and for this the `Content-Type` header should be set to
|
||||
`application/x-ndjson`.
|
||||
|
||||
File sizes are limited to 100 Mb. If your file is larger, split it into multiple
|
||||
files and upload each one separately in sequential time order. When running in
|
||||
Upload sizes are limited to the Elasticsearch HTTP receive buffer size
|
||||
(default 100 Mb). If your data is larger, split it into multiple chunks
|
||||
and upload each one separately in sequential time order. When running in
|
||||
real time, it is generally recommended that you perform many small uploads,
|
||||
rather than queueing data to upload larger files.
|
||||
|
||||
|
@ -29,9 +34,8 @@ The following records will not be processed:
|
|||
|
||||
//TBD link to Working with Out of Order timeseries concept doc
|
||||
|
||||
IMPORTANT: Data can only be accepted from a single connection. Use a single
|
||||
connection synchronously to send data, close, flush, or delete a single job.
|
||||
It is not currently possible to post data to multiple jobs using wildcards
|
||||
IMPORTANT: For each job, data can only be accepted from a single connection at
|
||||
a time. It is not currently possible to post data to multiple jobs using wildcards
|
||||
or a comma-separated list.
|
||||
|
||||
|
||||
|
@ -41,7 +45,7 @@ or a comma-separated list.
|
|||
(string) Identifier for the job
|
||||
|
||||
|
||||
==== Request Body
|
||||
==== Query Parameters
|
||||
|
||||
`reset_start`::
|
||||
(string) Specifies the start of the bucket resetting range
|
||||
|
@ -50,6 +54,12 @@ or a comma-separated list.
|
|||
(string) Specifies the end of the bucket resetting range
|
||||
|
||||
|
||||
==== Request Body
|
||||
|
||||
A sequence of one or more JSON documents containing the data to be analyzed.
|
||||
Only whitespace characters are permitted in between the documents.
|
||||
|
||||
|
||||
==== Authorization
|
||||
|
||||
You must have `manage_ml`, or `manage` cluster privileges to use this API.
|
||||
|
@ -60,7 +70,7 @@ For more information, see
|
|||
|
||||
==== Examples
|
||||
|
||||
The following example posts data from the farequote.json file to the `farequote` job:
|
||||
The following example posts data from the it_ops_new_kpi.json file to the `it_ops_new_kpi` job:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
|
@ -69,8 +79,6 @@ $ curl -s -H "Content-type: application/json"
|
|||
--data-binary @it_ops_new_kpi.json
|
||||
--------------------------------------------------
|
||||
|
||||
//TBD: Create example of how to post a small data example in Kibana?
|
||||
|
||||
When the data is sent, you receive information about the operational progress of the job.
|
||||
For example:
|
||||
|
||||
|
|
Loading…
Reference in New Issue