2013-11-24 06:13:08 -05:00
[[search-aggregations-bucket-terms-aggregation]]
2014-05-12 19:35:58 -04:00
=== Terms Aggregation
2013-11-24 06:13:08 -05:00
A multi-bucket value source based aggregation where buckets are dynamically built - one per unique value.
2017-08-30 06:11:10 -04:00
//////////////////////////
[source,js]
--------------------------------------------------
2019-01-18 03:34:11 -05:00
PUT /products
2017-08-30 06:11:10 -04:00
{
"mappings": {
2019-01-18 03:34:11 -05:00
"properties": {
"genre": {
"type": "keyword"
},
"product": {
"type": "keyword"
2017-08-30 06:11:10 -04:00
}
}
}
}
2019-01-18 03:34:11 -05:00
POST /products/_bulk?refresh
2017-08-30 06:11:10 -04:00
{"index":{"_id":0}}
{"genre": "rock", "product": "Product A"}
{"index":{"_id":1}}
{"genre": "rock"}
{"index":{"_id":2}}
{"genre": "rock"}
{"index":{"_id":3}}
{"genre": "jazz", "product": "Product Z"}
{"index":{"_id":4}}
{"genre": "jazz"}
{"index":{"_id":5}}
{"genre": "electronic"}
{"index":{"_id":6}}
{"genre": "electronic"}
{"index":{"_id":7}}
{"genre": "electronic"}
{"index":{"_id":8}}
{"genre": "electronic"}
{"index":{"_id":9}}
{"genre": "electronic"}
{"index":{"_id":10}}
{"genre": "electronic"}
-------------------------------------------------
// NOTCONSOLE
// TESTSETUP
//////////////////////////
2013-11-24 06:13:08 -05:00
Example:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2019-03-05 10:17:01 -05:00
"terms" : { "field" : "genre" } <1>
2013-11-24 06:13:08 -05:00
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
// TEST[s/_search/_search\?filter_path=aggregations/]
2019-03-05 10:17:01 -05:00
<1> `terms` aggregation should be a field of type `keyword` or any other data type suitable for bucket aggregations. In order to use it with `text` you will need to enable
<<fielddata, fielddata>>.
2013-11-24 06:13:08 -05:00
Response:
[source,js]
--------------------------------------------------
{
...
"aggregations" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-10-23 13:19:36 -04:00
"doc_count_error_upper_bound": 0, <1>
"sum_other_doc_count": 0, <2>
"buckets" : [ <3>
2013-11-24 06:13:08 -05:00
{
2017-08-30 06:11:10 -04:00
"key" : "electronic",
"doc_count" : 6
2013-11-24 06:13:08 -05:00
},
{
2016-07-01 08:54:25 -04:00
"key" : "rock",
2017-08-30 06:11:10 -04:00
"doc_count" : 3
2016-07-01 08:54:25 -04:00
},
{
2017-08-30 06:11:10 -04:00
"key" : "jazz",
"doc_count" : 2
}
2013-11-24 06:13:08 -05:00
]
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// TESTRESPONSE[s/\.\.\.//]
2014-10-23 13:19:36 -04:00
<1> an upper bound of the error on the document counts for each term, see <<search-aggregations-bucket-terms-aggregation-approximate-counts,below>>
2017-11-29 03:44:25 -05:00
<2> when there are lots of unique terms, Elasticsearch only returns the top terms; this number is the sum of the document counts for all buckets that are not part of the response
2014-10-23 13:19:36 -04:00
<3> the list of the top buckets, the meaning of `top` being defined by the <<search-aggregations-bucket-terms-aggregation-order,order>>
2013-11-29 06:35:25 -05:00
By default, the `terms` aggregation will return the buckets for the top ten terms ordered by the `doc_count`. One can
change this default behaviour by setting the `size` parameter.
2013-11-24 06:13:08 -05:00
2018-02-02 07:42:27 -05:00
[[search-aggregations-bucket-terms-aggregation-size]]
2014-07-07 08:58:42 -04:00
==== Size
2013-11-29 06:35:25 -05:00
The `size` parameter can be set to define how many term buckets should be returned out of the overall terms list. By
default, the node coordinating the search process will request each shard to provide its own top `size` term buckets
2014-01-20 05:18:26 -05:00
and once all shards respond, it will reduce the results to the final list that will then be returned to the client.
2013-11-29 06:35:25 -05:00
This means that if the number of unique terms is greater than `size`, the returned list is slightly off and not accurate
(it could be that the term counts are slightly off and it could even be that a term that should have been in the top
2016-06-14 06:37:10 -04:00
size buckets was not returned).
2014-05-09 11:55:56 -04:00
2018-01-25 02:59:41 -05:00
NOTE: If you want to retrieve **all** terms or all combinations of terms in a nested `terms` aggregation
you should use the <<search-aggregations-bucket-composite-aggregation,Composite>> aggregation which
allows to paginate over all possible terms rather than setting a size greater than the cardinality of the field in the
`terms` aggregation. The `terms` aggregation is meant to return the `top` terms and does not allow pagination.
2014-10-23 13:19:36 -04:00
[[search-aggregations-bucket-terms-aggregation-approximate-counts]]
2014-07-07 08:58:42 -04:00
==== Document counts are approximate
2014-09-26 15:04:42 -04:00
As described above, the document counts (and the results of any sub aggregations) in the terms aggregation are not always
accurate. This is because each shard provides its own view of what the ordered list of terms should be and these are
2014-07-07 08:58:42 -04:00
combined to give a final view. Consider the following scenario:
2014-09-26 15:04:42 -04:00
A request is made to obtain the top 5 terms in the field product, ordered by descending document count from an index with
3 shards. In this case each shard is asked to give its top 5 terms.
2014-07-07 08:58:42 -04:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-07-07 08:58:42 -04:00
{
"aggs" : {
"products" : {
2014-09-26 15:04:42 -04:00
"terms" : {
2014-07-07 08:58:42 -04:00
"field" : "product",
"size" : 5
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
// TEST[s/_search/_search\?filter_path=aggregations/]
2014-07-07 08:58:42 -04:00
2014-09-26 15:04:42 -04:00
The terms for each of the three shards are shown below with their
2014-07-07 08:58:42 -04:00
respective document counts in brackets:
[width="100%",cols="^2,^2,^2,^2",options="header"]
|=========================================================
| | Shard A | Shard B | Shard C
2014-09-26 15:04:42 -04:00
| 1 | Product A (25) | Product A (30) | Product A (45)
| 2 | Product B (18) | Product B (25) | Product C (44)
| 3 | Product C (6) | Product F (17) | Product Z (36)
| 4 | Product D (3) | Product Z (16) | Product G (30)
| 5 | Product E (2) | Product G (15) | Product E (29)
| 6 | Product F (2) | Product H (14) | Product H (28)
| 7 | Product G (2) | Product I (10) | Product Q (2)
| 8 | Product H (2) | Product Q (6) | Product D (1)
| 9 | Product I (1) | Product J (8) |
| 10 | Product J (1) | Product C (4) |
2014-07-07 08:58:42 -04:00
|=========================================================
The shards will return their top 5 terms so the results from the shards will be:
[width="100%",cols="^2,^2,^2,^2",options="header"]
|=========================================================
| | Shard A | Shard B | Shard C
2014-09-26 15:04:42 -04:00
| 1 | Product A (25) | Product A (30) | Product A (45)
| 2 | Product B (18) | Product B (25) | Product C (44)
| 3 | Product C (6) | Product F (17) | Product Z (36)
| 4 | Product D (3) | Product Z (16) | Product G (30)
| 5 | Product E (2) | Product G (15) | Product E (29)
2014-07-07 08:58:42 -04:00
|=========================================================
2014-09-26 15:04:42 -04:00
Taking the top 5 results from each of the shards (as requested) and combining them to make a final top 5 list produces
2014-07-07 08:58:42 -04:00
the following:
[width="40%",cols="^2,^2"]
|=========================================================
2014-09-26 15:04:42 -04:00
| 1 | Product A (100)
| 2 | Product Z (52)
| 3 | Product C (50)
| 4 | Product G (45)
| 5 | Product B (43)
2014-07-07 08:58:42 -04:00
|=========================================================
2014-09-26 15:04:42 -04:00
Because Product A was returned from all shards we know that its document count value is accurate. Product C was only
returned by shards A and C so its document count is shown as 50 but this is not an accurate count. Product C exists on
shard B, but its count of 4 was not high enough to put Product C into the top 5 list for that shard. Product Z was also
returned only by 2 shards but the third shard does not contain the term. There is no way of knowing, at the point of
combining the results to produce the final list of terms, that there is an error in the document count for Product C and
not for Product Z. Product H has a document count of 44 across all 3 shards but was not included in the final list of
2014-07-07 08:58:42 -04:00
terms because it did not make it into the top five terms on any of the shards.
==== Shard Size
2013-11-29 06:35:25 -05:00
The higher the requested `size` is, the more accurate the results will be, but also, the more expensive it will be to
compute the final results (both due to bigger priority queues that are managed on a shard level and due to bigger data
2014-09-26 15:04:42 -04:00
transfers between the nodes and the client).
2013-11-29 06:35:25 -05:00
The `shard_size` parameter can be used to minimize the extra work that comes with bigger requested `size`. When defined,
it will determine how many terms the coordinating node will request from each shard. Once all the shards responded, the
coordinating node will then reduce them to a final result which will be based on the `size` parameter - this way,
one can increase the accuracy of the returned terms and avoid the overhead of streaming a big list of buckets back to
2016-06-14 06:37:10 -04:00
the client.
2014-05-09 11:55:56 -04:00
2013-11-29 06:35:25 -05:00
2017-11-29 03:44:25 -05:00
NOTE: `shard_size` cannot be smaller than `size` (as it doesn't make much sense). When it is, Elasticsearch will
2013-11-29 06:35:25 -05:00
override it and reset it to be equal to `size`.
2013-11-24 06:13:08 -05:00
Remove single shard optimization when suggesting shard_size (#37041)
When executing terms aggregations we set the shard_size, meaning the
number of buckets to collect on each shard, to a value that's higher than
the number of requested buckets, to guarantee some basic level of
precision. We have an optimization in place so that we leave shard_size
set to size whenever we are searching against a single shard, in which
case maximum precision is guaranteed by definition.
Such optimization requires us access to the total number of shards that
the search is executing against. In the context of cross-cluster search,
once we will introduce multiple reduction steps (one per cluster) each
cluster will only know the number of local shards, which is problematic
as we should only optimize if we are searching against a single shard in a
single cluster. It could be that we are searching against one shard per cluster
in which case the current code would optimize number of terms causing
a loss of precision.
While discussing how to address the CCS scenario, we decided that we do
not want to introduce further complexity caused by this single shard
optimization, as it benefits only a minority of cases, especially when
the benefits are not so great.
This commit removes the single shard optimization, meaning that we will
always have heuristic enabled on how many number of buckets to collect
on the shards, even when searching against a single shard.
This will cause more buckets to be collected when searching against a single
shard compared to before. If that becomes a problem for some users, they
can work around that by setting the shard_size equal to the size.
Relates to #32125
2019-01-02 11:45:49 -05:00
The default `shard_size` is `(size * 1.5 + 10)`.
2015-02-25 06:00:55 -05:00
2014-07-07 08:58:42 -04:00
==== Calculating Document Count Error
2017-08-30 06:11:10 -04:00
There are two error values which can be shown on the terms aggregation. The first gives a value for the aggregation as
2014-09-26 15:04:42 -04:00
a whole which represents the maximum potential document count for a term which did not make it into the final list of
2017-08-30 06:11:10 -04:00
terms. This is calculated as the sum of the document count from the last term returned from each shard. For the example
2014-09-26 15:04:42 -04:00
given above the value would be 46 (2 + 15 + 29). This means that in the worst case scenario a term which was not returned
2014-07-07 08:58:42 -04:00
could have the 4th highest document count.
[source,js]
--------------------------------------------------
{
...
"aggregations" : {
"products" : {
"doc_count_error_upper_bound" : 46,
2017-08-30 06:11:10 -04:00
"sum_other_doc_count" : 79,
2014-07-07 08:58:42 -04:00
"buckets" : [
{
"key" : "Product A",
"doc_count" : 100
},
{
"key" : "Product Z",
"doc_count" : 52
2017-08-30 06:11:10 -04:00
}
2014-07-07 08:58:42 -04:00
...
]
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// TESTRESPONSE[s/\.\.\.//]
// TESTRESPONSE[s/: (\-)?[0-9]+/: $body.$_path/]
2014-07-07 08:58:42 -04:00
2015-02-04 09:43:22 -05:00
==== Per bucket document count error
2017-08-30 06:11:10 -04:00
The second error value can be enabled by setting the `show_term_doc_count_error` parameter to true:
2014-07-07 08:58:42 -04:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-07-07 08:58:42 -04:00
{
2017-08-30 06:11:10 -04:00
"aggs" : {
"products" : {
"terms" : {
"field" : "product",
"size" : 5,
"show_term_doc_count_error": true
}
}
}
}
--------------------------------------------------
// CONSOLE
// TEST[s/_search/_search\?filter_path=aggregations/]
This shows an error value for each term returned by the aggregation which represents the 'worst case' error in the document count
and can be useful when deciding on a value for the `shard_size` parameter. This is calculated by summing the document counts for
the last term returned by all shards which did not return the term. In the example above the error in the document count for Product C
would be 15 as Shard B was the only shard not to return the term and the document count of the last term it did return was 15.
The actual document count of Product C was 54 so the document count was only actually off by 4 even though the worst case was that
it would be off by 15. Product A, however has an error of 0 for its document count, since every shard returned it we can be confident
that the count returned is accurate.
2014-07-07 08:58:42 -04:00
2017-08-30 06:11:10 -04:00
[source,js]
--------------------------------------------------
{
...
2014-07-07 08:58:42 -04:00
"aggregations" : {
"products" : {
"doc_count_error_upper_bound" : 46,
2017-08-30 06:11:10 -04:00
"sum_other_doc_count" : 79,
2014-07-07 08:58:42 -04:00
"buckets" : [
{
"key" : "Product A",
"doc_count" : 100,
"doc_count_error_upper_bound" : 0
},
{
"key" : "Product Z",
"doc_count" : 52,
"doc_count_error_upper_bound" : 2
2017-08-30 06:11:10 -04:00
}
2014-07-07 08:58:42 -04:00
...
]
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// TESTRESPONSE[s/\.\.\.//]
// TESTRESPONSE[s/: (\-)?[0-9]+/: $body.$_path/]
2014-07-07 08:58:42 -04:00
2014-09-26 15:04:42 -04:00
These errors can only be calculated in this way when the terms are ordered by descending document count. When the aggregation is
ordered by the terms values themselves (either ascending or descending) there is no error in the document count since if a shard
does not return a particular term which appears in the results from another shard, it must not have that term in its index. When the
aggregation is either sorted by a sub aggregation or in order of ascending document count, the error in the document counts cannot be
2014-07-07 08:58:42 -04:00
determined and is given a value of -1 to indicate this.
2014-10-23 13:19:36 -04:00
[[search-aggregations-bucket-terms-aggregation-order]]
2013-11-24 06:13:08 -05:00
==== Order
2013-11-29 06:35:25 -05:00
The order of the buckets can be customized by setting the `order` parameter. By default, the buckets are ordered by
2016-08-18 07:52:45 -04:00
their `doc_count` descending. It is possible to change this behaviour as documented below:
2016-06-17 11:14:01 -04:00
2016-08-18 07:52:45 -04:00
WARNING: Sorting by ascending `_count` or by sub aggregation is discouraged as it increases the
<<search-aggregations-bucket-terms-aggregation-approximate-counts,error>> on document counts.
It is fine when a single shard is queried, or when the field that is being aggregated was used
as a routing key at index time: in these cases results will be accurate since shards have disjoint
values. However otherwise, errors are unbounded. One particular case that could still be useful
is sorting by <<search-aggregations-metrics-min-aggregation,`min`>> or
<<search-aggregations-metrics-max-aggregation,`max`>> aggregation: counts will not be accurate
but at least the top buckets will be correctly picked.
Ordering the buckets by their doc `_count` in an ascending manner:
2013-11-24 06:13:08 -05:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "genre",
2016-06-17 11:14:01 -04:00
"order" : { "_count" : "asc" }
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-24 06:13:08 -05:00
Ordering the buckets alphabetically by their terms in an ascending manner:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "genre",
2017-05-11 13:06:26 -04:00
"order" : { "_key" : "asc" }
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-24 06:13:08 -05:00
2017-05-11 13:06:26 -04:00
deprecated[6.0.0, Use `_key` instead of `_term` to order buckets by their term]
2013-11-24 06:13:08 -05:00
Ordering the buckets by single value metrics sub-aggregation (identified by the aggregation name):
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "genre",
2016-12-07 10:59:15 -05:00
"order" : { "max_play_count" : "desc" }
2013-11-24 06:13:08 -05:00
},
"aggs" : {
2016-12-07 10:59:15 -05:00
"max_play_count" : { "max" : { "field" : "play_count" } }
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-24 06:13:08 -05:00
Ordering the buckets by multi value metrics sub-aggregation (identified by the aggregation name):
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "genre",
2016-12-07 10:59:15 -05:00
"order" : { "playback_stats.max" : "desc" }
2013-11-24 06:13:08 -05:00
},
"aggs" : {
2016-07-01 08:54:25 -04:00
"playback_stats" : { "stats" : { "field" : "play_count" } }
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-24 06:13:08 -05:00
2016-08-18 08:08:44 -04:00
[NOTE]
2016-08-18 07:52:45 -04:00
.Pipeline aggs cannot be used for sorting
=======================================
<<search-aggregations-pipeline,Pipeline aggregations>> are run during the
reduce phase after all other aggregations have already completed. For this
reason, they cannot be used for ordering.
=======================================
2016-04-07 04:37:26 -04:00
2014-02-27 10:58:28 -05:00
It is also possible to order the buckets based on a "deeper" aggregation in the hierarchy. This is supported as long
2014-10-16 12:46:36 -04:00
as the aggregations path are of a single-bucket type, where the last aggregation in the path may either be a single-bucket
2014-02-27 10:58:28 -05:00
one or a metrics one. If it's a single-bucket type, the order will be defined by the number of docs in the bucket (i.e. `doc_count`),
in case it's a metrics one, the same rules as above apply (where the path must indicate the metric name to sort by in case of
a multi-value metrics aggregation, and in case of a single-value metrics aggregation the sort will be applied on that value).
The path must be defined in the following form:
2016-08-29 17:33:25 -04:00
// https://en.wikipedia.org/wiki/Extended_Backus%E2%80%93Naur_Form
[source,ebnf]
2014-02-27 10:58:28 -05:00
--------------------------------------------------
2016-08-29 17:33:25 -04:00
AGG_SEPARATOR = '>' ;
METRIC_SEPARATOR = '.' ;
AGG_NAME = <the name of the aggregation> ;
METRIC = <the name of the metric (in case of multi-value metrics aggregation)> ;
PATH = <AGG_NAME> [ <AGG_SEPARATOR>, <AGG_NAME> ]* [ <METRIC_SEPARATOR>, <METRIC> ] ;
2014-02-27 10:58:28 -05:00
--------------------------------------------------
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-02-27 10:58:28 -05:00
{
"aggs" : {
"countries" : {
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "artist.country",
"order" : { "rock>playback_stats.avg" : "desc" }
2014-02-27 10:58:28 -05:00
},
"aggs" : {
2016-07-01 08:54:25 -04:00
"rock" : {
"filter" : { "term" : { "genre" : "rock" }},
2014-02-27 10:58:28 -05:00
"aggs" : {
2016-07-01 08:54:25 -04:00
"playback_stats" : { "stats" : { "field" : "play_count" }}
2014-02-27 10:58:28 -05:00
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-02-27 10:58:28 -05:00
2016-07-01 08:54:25 -04:00
The above will sort the artist's countries buckets based on the average play count among the rock songs.
2014-02-27 10:58:28 -05:00
2014-09-03 04:26:54 -04:00
Multiple criteria can be used to order the buckets by providing an array of order criteria such as the following:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-09-03 04:26:54 -04:00
{
"aggs" : {
"countries" : {
"terms" : {
2016-07-01 08:54:25 -04:00
"field" : "artist.country",
"order" : [ { "rock>playback_stats.avg" : "desc" }, { "_count" : "desc" } ]
2014-09-03 04:26:54 -04:00
},
"aggs" : {
2016-07-01 08:54:25 -04:00
"rock" : {
2017-08-30 06:11:10 -04:00
"filter" : { "term" : { "genre" : "rock" }},
2014-09-03 04:26:54 -04:00
"aggs" : {
2016-07-01 08:54:25 -04:00
"playback_stats" : { "stats" : { "field" : "play_count" }}
2014-09-03 04:26:54 -04:00
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-09-03 04:26:54 -04:00
2016-07-01 08:54:25 -04:00
The above will sort the artist's countries buckets based on the average play count among the rock songs and then by
2014-09-03 04:26:54 -04:00
their `doc_count` in descending order.
2014-09-26 15:04:42 -04:00
NOTE: In the event that two buckets share the same values for all order criteria the bucket's term value is used as a
2014-09-03 04:26:54 -04:00
tie-breaker in ascending alphabetical order to prevent non-deterministic ordering of buckets.
2014-09-26 15:04:42 -04:00
==== Minimum document count
2014-01-08 09:08:18 -05:00
It is possible to only return terms that match more than a configured number of hits using the `min_doc_count` option:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-01-08 09:08:18 -05:00
{
"aggs" : {
"tags" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2014-11-27 14:29:34 -05:00
"field" : "tags",
2014-01-08 09:08:18 -05:00
"min_doc_count": 10
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-01-08 09:08:18 -05:00
The above aggregation would only return tags which have been found in 10 hits or more. Default value is `1`.
2014-05-09 11:55:56 -04:00
Terms are collected and ordered on a shard level and merged with the terms collected from other shards in a second step. However, the shard does not have the information about the global document count available. The decision if a term is added to a candidate list depends only on the order computed on the shard using local shard frequencies. The `min_doc_count` criterion is only applied after merging local terms statistics of all shards. In a way the decision to add the term as a candidate is made without being very _certain_ about if the term will actually reach the required `min_doc_count`. This might cause many (globally) high frequent terms to be missing in the final result if low frequent terms populated the candidate lists. To avoid this, the `shard_size` parameter can be increased to allow more candidate terms on the shards. However, this increases memory consumption and network traffic.
2014-09-26 15:04:42 -04:00
`shard_min_doc_count` parameter
2014-05-09 11:55:56 -04:00
2014-09-10 09:57:57 -04:00
The parameter `shard_min_doc_count` regulates the _certainty_ a shard has if the term should actually be added to the candidate list or not with respect to the `min_doc_count`. Terms will only be considered if their local shard frequency within the set is higher than the `shard_min_doc_count`. If your dictionary contains many low frequent terms and you are not interested in those (for example misspellings), then you can set the `shard_min_doc_count` parameter to filter out candidate terms on a shard level that will with a reasonable certainty not reach the required `min_doc_count` even after merging the local counts. `shard_min_doc_count` is set to `0` per default and has no effect unless you explicitly set it.
2014-05-09 11:55:56 -04:00
2014-01-08 09:08:18 -05:00
NOTE: Setting `min_doc_count`=`0` will also return buckets for terms that didn't match any hit. However, some of
2015-04-20 18:57:57 -04:00
the returned terms which have a document count of zero might only belong to deleted documents or documents
from other types, so there is no warranty that a `match_all` query would find a positive document count for
those terms.
2014-01-08 09:08:18 -05:00
WARNING: When NOT sorting on `doc_count` descending, high values of `min_doc_count` may return a number of buckets
which is less than `size` because not enough data was gathered from the shards. Missing buckets can be
back by increasing `shard_size`.
2014-05-09 11:55:56 -04:00
Setting `shard_min_doc_count` too high will cause terms to be filtered out on a shard level. This value should be set much lower than `min_doc_count/#shards`.
2014-01-08 09:08:18 -05:00
2014-09-07 05:09:52 -04:00
[[search-aggregations-bucket-terms-aggregation-script]]
2013-11-24 06:13:08 -05:00
==== Script
Generating the terms using a script:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2016-06-14 06:37:10 -04:00
"script" : {
2017-06-09 11:29:25 -04:00
"source": "doc['genre'].value",
2016-05-11 16:07:08 -04:00
"lang": "painless"
}
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-24 06:13:08 -05:00
2017-05-17 17:42:25 -04:00
This will interpret the `script` parameter as an `inline` script with the default script language and no script parameters. To use a stored script use the following syntax:
2015-05-12 05:37:22 -04:00
2017-08-30 06:11:10 -04:00
//////////////////////////
[source,js]
--------------------------------------------------
POST /_scripts/my_script
{
"script": {
"lang": "painless",
"source": "doc[params.field].value"
}
}
--------------------------------------------------
// CONSOLE
//////////////////////////
2015-05-12 05:37:22 -04:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2015-05-12 05:37:22 -04:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2015-05-12 05:37:22 -04:00
"terms" : {
"script" : {
2017-06-09 11:29:25 -04:00
"id": "my_script",
2015-05-12 05:37:22 -04:00
"params": {
2016-07-01 08:54:25 -04:00
"field": "genre"
2015-05-12 05:37:22 -04:00
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
// TEST[continued]
2015-05-12 05:37:22 -04:00
2013-11-24 06:13:08 -05:00
==== Value Script
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-24 06:13:08 -05:00
{
"aggs" : {
2016-07-01 08:54:25 -04:00
"genres" : {
2014-05-12 19:35:58 -04:00
"terms" : {
2018-10-02 10:29:32 -04:00
"field" : "genre",
2016-06-14 06:37:10 -04:00
"script" : {
2017-08-30 06:11:10 -04:00
"source" : "'Genre: ' +_value",
2016-06-14 06:37:10 -04:00
"lang" : "painless"
2016-05-11 16:07:08 -04:00
}
2013-11-24 06:13:08 -05:00
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-29 04:05:12 -05:00
==== Filtering Values
It is possible to filter the values for which buckets will be created. This can be done using the `include` and
2016-11-17 10:32:59 -05:00
`exclude` parameters which are based on regular expression strings or arrays of exact values. Additionally,
`include` clauses can filter using `partition` expressions.
===== Filtering Values with regular expressions
2013-11-29 04:05:12 -05:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-29 04:05:12 -05:00
{
"aggs" : {
"tags" : {
"terms" : {
"field" : "tags",
"include" : ".*sport.*",
"exclude" : "water_.*"
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-29 04:05:12 -05:00
In the above example, buckets will be created for all the tags that has the word `sport` in them, except those starting
2018-07-31 08:33:23 -04:00
with `water_` (so the tag `water_sports` will not be aggregated). The `include` regular expression will determine what
2013-11-29 04:05:12 -05:00
values are "allowed" to be aggregated, while the `exclude` determines the values that should not be aggregated. When
both are defined, the `exclude` has precedence, meaning, the `include` is evaluated first and only then the `exclude`.
2015-04-03 12:04:22 -04:00
The syntax is the same as <<regexp-syntax,regexp queries>>.
2013-11-27 08:55:54 -05:00
2016-11-17 10:32:59 -05:00
===== Filtering Values with exact values
2014-09-01 09:18:42 -04:00
For matching based on exact values the `include` and `exclude` parameters can simply take an array of
strings that represent the terms as they are found in the index:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-09-01 09:18:42 -04:00
{
"aggs" : {
"JapaneseCars" : {
"terms" : {
"field" : "make",
"include" : ["mazda", "honda"]
}
},
"ActiveCarManufacturers" : {
"terms" : {
"field" : "make",
"exclude" : ["rover", "jensen"]
}
}
}
}
2014-09-26 15:04:42 -04:00
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-09-01 09:18:42 -04:00
2016-11-17 10:32:59 -05:00
===== Filtering Values with partitions
Sometimes there are too many unique terms to process in a single request/response pair so
it can be useful to break the analysis up into multiple requests.
This can be achieved by grouping the field's values into a number of partitions at query-time and processing
only one partition in each request.
Consider this request which is looking for accounts that have not logged any access recently:
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2016-11-17 10:32:59 -05:00
{
"size": 0,
"aggs": {
"expired_sessions": {
"terms": {
"field": "account_id",
"include": {
"partition": 0,
"num_partitions": 20
},
"size": 10000,
"order": {
"last_access": "asc"
}
},
"aggs": {
"last_access": {
"max": {
"field": "access_date"
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2016-11-17 10:32:59 -05:00
This request is finding the last logged access date for a subset of customer accounts because we
might want to expire some customer accounts who haven't been seen for a long while.
The `num_partitions` setting has requested that the unique account_ids are organized evenly into twenty
partitions (0 to 19). and the `partition` setting in this request filters to only consider account_ids falling
into partition 0. Subsequent requests should ask for partitions 1 then 2 etc to complete the expired-account analysis.
Note that the `size` setting for the number of results returned needs to be tuned with the `num_partitions`.
For this particular account-expiration example the process for balancing values for `size` and `num_partitions` would be as follows:
1. Use the `cardinality` aggregation to estimate the total number of unique account_id values
2. Pick a value for `num_partitions` to break the number from 1) up into more manageable chunks
3. Pick a `size` value for the number of responses we want from each partition
4. Run a test request
If we have a circuit-breaker error we are trying to do too much in one request and must increase `num_partitions`.
If the request was successful but the last account ID in the date-sorted test response was still an account we might want to
expire then we may be missing accounts of interest and have set our numbers too low. We must either
* increase the `size` parameter to return more results per partition (could be heavy on memory) or
* increase the `num_partitions` to consider less accounts per request (could increase overall processing time as we need to make more requests)
2017-11-29 03:44:25 -05:00
Ultimately this is a balancing act between managing the Elasticsearch resources required to process a single request and the volume
2016-11-17 10:32:59 -05:00
of requests that the client application must issue to complete a task.
2014-09-07 05:09:52 -04:00
==== Multi-field terms aggregation
The `terms` aggregation does not support collecting terms from multiple fields
in the same document. The reason is that the `terms` agg doesn't collect the
string term values themselves, but rather uses
<<search-aggregations-bucket-terms-aggregation-execution-hint,global ordinals>>
to produce a list of all of the unique values in the field. Global ordinals
results in an important performance boost which would not be possible across
multiple fields.
There are two approaches that you can use to perform a `terms` agg across
multiple fields:
<<search-aggregations-bucket-terms-aggregation-script,Script>>::
Use a script to retrieve terms from multiple fields. This disables the global
ordinals optimization and will be slower than collecting terms from a single
field, but it gives you the flexibility to implement this option at search
time.
<<copy-to,`copy_to` field>>::
If you know ahead of time that you want to collect the terms from two or more
fields, then use `copy_to` in your mapping to create a new dedicated field at
index time which contains the values from both fields. You can aggregate on
this single field, which will benefit from the global ordinals optimization.
2019-04-11 18:38:25 -04:00
[[search-aggregations-bucket-terms-aggregation-collect]]
2014-05-02 04:09:40 -04:00
==== Collect mode
2014-09-26 15:04:42 -04:00
Deferring calculation of child aggregations
2014-05-02 04:09:40 -04:00
For fields with many unique terms and a small number of required results it can be more efficient to delay the calculation
of child aggregations until the top parent-level aggs have been pruned. Ordinarily, all branches of the aggregation tree
2016-06-08 06:46:11 -04:00
are expanded in one depth-first pass and only then any pruning occurs.
In some scenarios this can be very wasteful and can hit memory constraints.
2014-09-26 15:04:42 -04:00
An example problem scenario is querying a movie database for the 10 most popular actors and their 5 most common co-stars:
2014-05-02 04:09:40 -04:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-05-02 04:09:40 -04:00
{
"aggs" : {
"actors" : {
"terms" : {
"field" : "actors",
"size" : 10
},
"aggs" : {
"costars" : {
"terms" : {
"field" : "actors",
"size" : 5
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-05-02 04:09:40 -04:00
2016-06-08 06:46:11 -04:00
Even though the number of actors may be comparatively small and we want only 50 result buckets there is a combinatorial explosion of buckets
during calculation - a single actor can produce n² buckets where n is the number of actors. The sane option would be to first determine
2014-05-02 04:09:40 -04:00
the 10 most popular actors and only then examine the top co-stars for these 10 actors. This alternative strategy is what we call the `breadth_first` collection
2016-06-08 06:46:11 -04:00
mode as opposed to the `depth_first` mode.
NOTE: The `breadth_first` is the default mode for fields with a cardinality bigger than the requested size or when the cardinality is unknown (numeric fields or scripts for instance).
It is possible to override the default heuristic and to provide a collect mode directly in the request:
2014-09-26 15:04:42 -04:00
2014-05-02 04:09:40 -04:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2014-05-02 04:09:40 -04:00
{
"aggs" : {
"actors" : {
"terms" : {
"field" : "actors",
"size" : 10,
2016-06-08 06:46:11 -04:00
"collect_mode" : "breadth_first" <1>
2014-05-02 04:09:40 -04:00
},
"aggs" : {
"costars" : {
"terms" : {
"field" : "actors",
"size" : 5
}
}
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2014-05-02 04:09:40 -04:00
2016-06-08 06:46:11 -04:00
<1> the possible values are `breadth_first` and `depth_first`
2014-05-02 04:09:40 -04:00
When using `breadth_first` mode the set of documents that fall into the uppermost buckets are
2014-09-26 15:04:42 -04:00
cached for subsequent replay so there is a memory overhead in doing this which is linear with the number of matching documents.
2014-05-02 04:09:40 -04:00
Note that the `order` parameter can still be used to refer to data from a child aggregation when using the `breadth_first` setting - the parent
aggregation understands that this child aggregation will need to be called first before any of the other child aggregations.
2016-05-04 04:55:24 -04:00
WARNING: Nested aggregations such as `top_hits` which require access to score information under an aggregation that uses the `breadth_first`
collection mode need to replay the query on the second pass but only for the documents belonging to the top buckets.
2014-05-02 04:09:40 -04:00
2014-09-07 05:09:52 -04:00
[[search-aggregations-bucket-terms-aggregation-execution-hint]]
2013-11-27 08:55:54 -05:00
==== Execution hint
2014-07-21 05:55:22 -04:00
There are different mechanisms by which terms aggregations can be executed:
- by using field values directly in order to aggregate data per-bucket (`map`)
2017-08-21 13:12:27 -04:00
- by using global ordinals of the field and allocating one bucket per global ordinal (`global_ordinals`)
2014-07-21 05:55:22 -04:00
Elasticsearch tries to have sensible defaults so this is something that generally doesn't need to be configured.
2017-08-21 13:12:27 -04:00
`global_ordinals` is the default option for `keyword` field, it uses global ordinals to allocates buckets dynamically
so memory usage is linear to the number of values of the documents that are part of the aggregation scope.
2014-07-21 05:55:22 -04:00
2017-08-21 13:12:27 -04:00
`map` should only be considered when very few documents match a query. Otherwise the ordinals-based execution mode
is significantly faster. By default, `map` is only used when running an aggregation on scripts, since they don't have
ordinals.
2013-11-27 08:55:54 -05:00
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2013-11-27 08:55:54 -05:00
{
"aggs" : {
"tags" : {
"terms" : {
"field" : "tags",
"execution_hint": "map" <1>
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2013-11-27 08:55:54 -05:00
2017-08-21 13:12:27 -04:00
<1> The possible values are `map`, `global_ordinals`
2013-11-27 08:55:54 -05:00
2014-07-21 05:55:22 -04:00
Please note that Elasticsearch will ignore this execution hint if it is not applicable and that there is no backward compatibility guarantee on these hints.
2015-05-07 10:46:40 -04:00
==== Missing value
The `missing` parameter defines how documents that are missing a value should be treated.
By default they will be ignored but it is also possible to treat them as if they
had a value.
[source,js]
--------------------------------------------------
2017-08-30 06:11:10 -04:00
GET /_search
2015-05-07 10:46:40 -04:00
{
"aggs" : {
"tags" : {
"terms" : {
"field" : "tags",
"missing": "N/A" <1>
}
}
}
}
--------------------------------------------------
2017-08-30 06:11:10 -04:00
// CONSOLE
2015-05-07 10:46:40 -04:00
2016-06-14 06:37:10 -04:00
<1> Documents without a value in the `tags` field will fall into the same bucket as documents that have the value `N/A`.
2017-01-10 05:50:56 -05:00
==== Mixing field types
WARNING: When aggregating on multiple indices the type of the aggregated field may not be the same in all indices.
Some types are compatible with each other (`integer` and `long` or `float` and `double`) but when the types are a mix
of decimal and non-decimal number the terms aggregation will promote the non-decimal numbers to decimal numbers.
This can result in a loss of precision in the bucket values.