OpenSearch/docs/reference/ml/anomaly-detection/ml-configuring-populations....

88 lines
3.3 KiB
Plaintext
Raw Normal View History

[role="xpack"]
[[ml-configuring-populations]]
= Performing population analysis
Entities or events in your data can be considered anomalous when:
* Their behavior changes over time, relative to their own previous behavior, or
* Their behavior is different than other entities in a specified population.
The latter method of detecting outliers is known as _population analysis_. The
{ml} analytics build a profile of what a "typical" user, machine, or other entity
does over a specified time period and then identify when one is behaving
abnormally compared to the population.
This type of analysis is most useful when the behavior of the population as a
whole is mostly homogeneous and you want to identify outliers. In general,
population analysis is not useful when members of the population inherently
have vastly different behavior. You can, however, segment your data into groups
that behave similarly and run these as separate jobs. For example, you can use a
query filter in the {dfeed} to segment your data or you can use the
`partition_field_name` to split the analysis for the different groups.
Population analysis scales well and has a lower resource footprint than
individual analysis of each series. For example, you can analyze populations
of hundreds of thousands or millions of entities.
To specify the population, use the `over_field_name` property. For example:
[source,console]
----------------------------------
PUT _ml/anomaly_detectors/population
{
"description" : "Population analysis",
"analysis_config" : {
"bucket_span":"15m",
"influencers": [
"clientip"
],
"detectors": [
{
"function": "mean",
"field_name": "bytes",
"over_field_name": "clientip" <1>
}
]
},
"data_description" : {
"time_field":"timestamp",
"time_format": "epoch_ms"
}
}
----------------------------------
// TEST[skip:needs-licence]
<1> This `over_field_name` property indicates that the metrics for each client (
as identified by their IP address) are analyzed relative to other clients
in each bucket.
If your data is stored in {es}, you can use the population job wizard in {kib}
to create an {anomaly-job} with these same properties. For example, if you add
the sample web logs in {kib}, you can use the following job settings in the
population job wizard:
[role="screenshot"]
image::images/ml-population-job.png["Job settings in the population job wizard]
After you open the job and start the {dfeed} or supply data to the job, you can
view the results in {kib}. For example, you can view the results in the
**Anomaly Explorer**:
[role="screenshot"]
image::images/ml-population-results.png["Population analysis results in the Anomaly Explorer"]
As in this case, the results are often quite sparse. There might be just a few
data points for the selected time period. Population analysis is particularly
useful when you have many entities and the data for specific entitles is sporadic
or sparse.
If you click on a section in the timeline or swim lanes, you can see more
details about the anomalies:
[role="screenshot"]
image::images/ml-population-anomaly.png["Anomaly details for a specific user"]
In this example, the client IP address `30.156.16.164` received a low volume of
bytes on the date and time shown. This event is anomalous because the mean is
three times lower than the expected behavior of the population.