[role="xpack"] [testenv="basic"] [[get-inference-stats]] === Get {infer} trained model statistics API [subs="attributes"] ++++ Get {infer} trained model stats ++++ Retrieves usage information for trained {infer} models. experimental[] [[ml-get-inference-stats-request]] ==== {api-request-title} `GET _ml/inference/_stats` + `GET _ml/inference/_all/_stats` + `GET _ml/inference//_stats` + `GET _ml/inference/,/_stats` + `GET _ml/inference/,/_stats` [[ml-get-inference-stats-prereq]] ==== {api-prereq-title} Required privileges which should be added to a custom role: * cluster: `monitor_ml` For more information, see <> and <>. [[ml-get-inference-stats-desc]] ==== {api-description-title} You can get usage information for multiple trained models in a single API request by using a comma-separated list of model IDs or a wildcard expression. [[ml-get-inference-stats-path-params]] ==== {api-path-parms-title} ``:: (Optional, string) include::{docdir}/ml/ml-shared.asciidoc[tag=model-id] [[ml-get-inference-stats-query-params]] ==== {api-query-parms-title} `allow_no_match`:: (Optional, boolean) include::{docdir}/ml/ml-shared.asciidoc[tag=allow-no-match] `from`:: (Optional, integer) include::{docdir}/ml/ml-shared.asciidoc[tag=from] `size`:: (Optional, integer) include::{docdir}/ml/ml-shared.asciidoc[tag=size] [[ml-get-inference-stats-response-codes]] ==== {api-response-codes-title} `404` (Missing resources):: If `allow_no_match` is `false`, this code indicates that there are no resources that match the request or only partial matches for the request. [[ml-get-inference-stats-example]] ==== {api-examples-title} The following example gets usage information for all the trained models: [source,console] -------------------------------------------------- GET _ml/inference/_stats -------------------------------------------------- // TEST[skip:TBD] The API returns the following results: [source,console-result] ---- { "count": 2, "trained_model_stats": [ { "model_id": "flight-delay-prediction-1574775339910", "pipeline_count": 0 }, { "model_id": "regression-job-one-1574775307356", "pipeline_count": 1, "ingest": { "total": { "count": 178, "time_in_millis": 8, "current": 0, "failed": 0 }, "pipelines": { "flight-delay": { "count": 178, "time_in_millis": 8, "current": 0, "failed": 0, "processors": [ { "inference": { "type": "inference", "stats": { "count": 178, "time_in_millis": 7, "current": 0, "failed": 0 } } } ] } } } } ] } ---- // NOTCONSOLE