[[search-aggregations-bucket-significantterms-aggregation]] === Significant Terms An aggregation that returns interesting or unusual occurrences of terms in a set. .Experimental! [IMPORTANT] ===== This feature is marked as experimental, and may be subject to change in the future. If you use this feature, please let us know your experience with it! ===== coming[1.1.0] .Example use cases: * Suggesting "H5N1" when users search for "bird flu" in text * Identifying the merchant that is the "common point of compromise" from the transaction history of credit card owners reporting loss * Suggesting keywords relating to stock symbol $ATI for an automated news classifier * Spotting the fraudulent doctor who is diagnosing more than his fair share of whiplash injuries * Spotting the tire manufacturer who has a disproportionate number of blow-outs In all these cases the terms being selected are not simply the most popular terms in a set. They are the terms that have undergone a significant change in popularity measured between a _foreground_ and _background_ set. If the term "H5N1" only exists in 5 documents in a 10 million document index and yet is found in 4 of the 100 documents that make up a user's search results that is significant and probably very relevant to their search. 5/10,000,000 vs 4/100 is a big swing in frequency. ==== Single-set analysis In the simplest case, the _foreground_ set of interest is the search results matched by a query and the _background_ set used for statistical comparisons is the index or indices from which the results were gathered. Example: [source,js] -------------------------------------------------- { "query" : { "terms" : {"force" : [ "British Transport Police" ]} }, "aggregations" : { "significantCrimeTypes" : { "significant_terms" : { "field" : "crime_type" } } } } -------------------------------------------------- Response: [source,js] -------------------------------------------------- { ... "aggregations" : { "significantCrimeTypes" : { "doc_count": 47347, "buckets" : [ { "key": "Bicycle theft", "doc_count": 3640, "score": 0.371235374214817, "bg_count": 66799 } ... ] } } } -------------------------------------------------- When querying an index of all crimes from all police forces, what these results show is that the British Transport Police force stand out as a force dealing with a disproportionately large number of bicycle thefts. Ordinarily, bicycle thefts represent only 1% of crimes (66799/5064554) but for the British Transport Police, who handle crime on railways and stations, 7% of crimes (3640/47347) is a bike theft. This is a significant seven-fold increase in frequency and so this anomaly was highlighted as the top crime type. The problem with using a query to spot anomalies is it only gives us one subset to use for comparisons. To discover all the other police forces' anomalies we would have to repeat the query for each of the different forces. This can be a tedious way to look for unusual patterns in an index ==== Multi-set analysis A simpler way to perform analysis across multiple categories is to use a parent-level aggregation to segment the data ready for analysis. Example using a parent aggregation for segmentation: [source,js] -------------------------------------------------- { "aggregations": { "forces": { "terms": {"field": "force"}, "aggregations": { "significantCrimeTypes": { "significant_terms": {"field": "crime_type"} } } } } } -------------------------------------------------- Response: [source,js] -------------------------------------------------- { ... "aggregations": { "forces": { "buckets": [ { "key": "Metropolitan Police Service", "doc_count": 894038, "significantCrimeTypes": { "doc_count": 894038, "buckets": [ { "key": "Robbery", "doc_count": 27617, "score": 0.0599, "bg_count": 53182 }, ... ] } }, { "key": "British Transport Police", "doc_count": 47347, "significantCrimeTypes": { "doc_count": 47347, "buckets": [ { "key": "Bicycle theft", "doc_count": 3640, "score": 0.371, "bg_count": 66799 }, ... ] } } ] } } -------------------------------------------------- Now we have anomaly detection for each of the police forces using a single request. We can use other forms of top-level aggregations to segment our data, for example segmenting by geographic area to identify unusual hot-spots of a particular crime type: [source,js] -------------------------------------------------- { "aggs": { "hotspots": { "geohash_grid" : { "field":"location", "precision":5, }, "aggs": { "significantCrimeTypes": { "significant_terms": {"field": "crime_type"} } } } } } -------------------------------------------------- This example uses the `geohash_grid` aggregation to create result buckets that represent geographic areas, and inside each bucket we can identify anomalous levels of a crime type in these tightly-focused areas e.g. * Airports exhibit unusual numbers of weapon confiscations * Universities show uplifts of bicycle thefts At a higher geohash_grid zoom-level with larger coverage areas we would start to see where an entire police-force may be tackling an unusual volume of a particular crime type. Obviously a time-based top-level segmentation would help identify current trends for each point in time where a simple `terms` aggregation would typically show the very popular "constants" that persist across all time slots. .How are the scores calculated? ********************************** The numbers returned for scores are primarily intended for ranking different suggestions sensibly rather than something easily understood by end users. The scores are derived from the doc frequencies in _foreground_ and _background_ sets. The _absolute_ change in popularity (foregroundPercent - backgroundPercent) would favour common terms whereas the _relative_ change in popularity (foregroundPercent/ backgroundPercent) would favour rare terms. Rare vs common is essentially a precision vs recall balance and so the absolute and relative changes are multiplied to provide a sweet spot between precision and recall. ********************************** === Use on free-text fields The significant_terms aggregation can be used effectively on tokenized free-text fields to suggest: * keywords for refining end-user searches * keywords for use in percolator queries WARNING: Picking a free-text field as the subject of a significant terms analysis can be expensive! It will attempt to load every unique word into RAM. It is recommended to only use this on smaller indices. .Use the _"like this but not this"_ pattern ********************************** You can spot mis-categorized content by first searching a structured field e.g. `category:adultMovie` and use significant_terms on the free-text "movie_description" field. Take the suggested words (I'll leave them to your imagination) and then search for all movies NOT marked as category:adultMovie but containing these keywords. You now have a ranked list of badly-categorized movies that you should reclassify or at least remove from the "familyFriendly" category. The significance score from each term can also provide a useful `boost` setting to sort matches. Using the `minimum_should_match` setting of the `terms` query with the keywords will help control the balance of precision/recall in the result set i.e a high setting would have a small number of relevant results packed full of keywords and a setting of "1" would produce a more exhaustive results set with all documents containing _any_ keyword. ********************************** [TIP] ============ .Show significant_terms in context Free-text significant_terms are much more easily understood when viewed in context. Take the results of `significant_terms` suggestions from a free-text field and use them in a `terms` query on the same field with a `highlight` clause to present users with example snippets of documents. When the terms are presented unstemmed, highlighted, with the right case, in the right order and with some context, their significance/meaning is more readily apparent. ============ === Limitations ===== Single _background_ comparison base The above examples show how to select the _foreground_ set for analysis using a query or parent aggregation to filter but currently there is no means of specifying a _background_ set other than the index from which all results are ultimately drawn. Sometimes it may prove useful to use a different background set as the basis for comparisons e.g. to first select the tweets for the TV show "XFactor" and then look for significant terms in a subset of that content which is from this week. ===== Significant terms must be indexed values Unlike the terms aggregation it is currently not possible to use script-generated terms for counting purposes. Because of the way the significant_terms aggregation must consider both _foreground_ and _background_ frequencies it would be prohibitively expensive to use a script on the entire index to obtain background frequencies for comparisons. Also DocValues are not supported as sources of term data for similar reasons. ===== No analysis of floating point fields Floating point fields are currently not supported as the subject of significant_terms analysis. While integer or long fields can be used to represent concepts like bank account numbers or category numbers which can be interesting to track, floating point fields are usually used to represent quantities of something. As such, individual floating point terms are not useful for this form of frequency analysis. ===== Use as a parent aggregation If there is the equivalent of a `match_all` query or no query criteria providing a subset of the index the significant_terms aggregation should not be used as the top-most aggregation - in this scenario the _foreground_ set is exactly the same as the _background_ set and so there is no difference in document frequencies to observe and from which to make sensible suggestions. Another consideration is that the significant_terms aggregation produces many candidate results at shard level that are only later pruned on the reducing node once all statistics from all shards are merged. As a result, it can be inefficient and costly in terms of RAM to embed large child aggregations under a significant_terms aggregation that later discards many candidate terms. It is advisable in these cases to perform two searches - the first to provide a rationalized list of significant_terms and then add this shortlist of terms to a second query to go back and fetch the required child aggregations. ===== Approximate counts The counts of how many documents contain a term provided in results are based on summing the samples returned from each shard and as such may be: * low if certain shards did not provide figures for a given term in their top sample * high when considering the background frequency as it may count occurrences found in deleted documents Like most design decisions, this is the basis of a trade-off in which we have chosen to provide fast performance at the cost of some (typically small) inaccuracies. However, the `size` and `shard size` settings covered in the next section provide tools to help control the accuracy levels. === Parameters ==== Size & Shard Size The `size` parameter can be set to define how many term buckets should be returned out of the overall terms list. By default, the node coordinating the search process will request each shard to provide its own top term buckets and once all shards respond, it will reduce the results to the final list that will then be returned to the client. If the number of unique terms is greater than `size`, the returned list can be slightly off and not accurate (it could be that the term counts are slightly off and it could even be that a term that should have been in the top size buckets was not returned). To ensure better accuracy a multiple of the final `size` is used as the number of terms to request from each shard using a heuristic based on the number of shards. To take manual control of this setting the `shard_size` parameter can be used to control the volumes of candidate terms produced by each shard. Low-frequency terms can turn out to be the most interesting ones once all results are combined so the significant_terms aggregation can produce higher-quality results when the `shard_size` parameter is set to values significantly higher than the `size` setting. This ensures that a bigger volume of promising candidate terms are given a consolidated review by the reducing node before the final selection. Obviously large candidate term lists will cause extra network traffic and RAM usage so this is quality/cost trade off that needs to be balanced. NOTE: `shard_size` cannot be smaller than `size` (as it doesn't make much sense). When it is, elasticsearch will override it and reset it to be equal to `size`. ==== Minimum document count It is possible to only return terms that match more than a configured number of hits using the `min_doc_count` option: [source,js] -------------------------------------------------- { "aggs" : { "tags" : { "significant_terms" : { "field" : "tag", "min_doc_count": 10 } } } } -------------------------------------------------- The above aggregation would only return tags which have been found in 10 hits or more. Default value is `3`. WARNING: Setting `min_doc_count` to `1` is generally not advised as it tends to return terms that are typos or other bizarre curiosities. Finding more than one instance of a term helps reinforce that, while still rare, the term was not the result of a one-off accident. The default value of 3 is used to provide a minimum weight-of-evidence. ==== Filtering Values It is possible (although rarely required) to filter the values for which buckets will be created. This can be done using the `include` and `exclude` parameters which are based on regular expressions. This functionality mirrors the features offered by the `terms` aggregation. [source,js] -------------------------------------------------- { "aggs" : { "tags" : { "significant_terms" : { "field" : "tags", "include" : ".*sport.*", "exclude" : "water_.*" } } } } -------------------------------------------------- In the above example, buckets will be created for all the tags that has the word `sport` in them, except those starting with `water_` (so the tag `water_sports` will no be aggregated). The `include` regular expression will determine what values are "allowed" to be aggregated, while the `exclude` determines the values that should not be aggregated. When both are defined, the `exclude` has precedence, meaning, the `include` is evaluated first and only then the `exclude`. The regular expression are based on the Java(TM) http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html[Pattern], and as such, they it is also possible to pass in flags that will determine how the compiled regular expression will work: [source,js] -------------------------------------------------- { "aggs" : { "tags" : { "terms" : { "field" : "tags", "include" : { "pattern" : ".*sport.*", "flags" : "CANON_EQ|CASE_INSENSITIVE" <1> }, "exclude" : { "pattern" : "water_.*", "flags" : "CANON_EQ|CASE_INSENSITIVE" } } } } } -------------------------------------------------- <1> the flags are concatenated using the `|` character as a separator The possible flags that can be used are: http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#CANON_EQ[`CANON_EQ`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#CASE_INSENSITIVE[`CASE_INSENSITIVE`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#COMMENTS[`COMMENTS`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#DOTALL[`DOTALL`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#LITERAL[`LITERAL`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#MULTILINE[`MULTILINE`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#UNICODE_CASE[`UNICODE_CASE`], http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#UNICODE_CHARACTER_CLASS[`UNICODE_CHARACTER_CLASS`] and http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html#UNIX_LINES[`UNIX_LINES`] ==== Execution hint There are two mechanisms by which terms aggregations can be executed: either by using field values directly in order to aggregate data per-bucket (`map`), or by using ordinals of the field values instead of the values themselves (`ordinals`). Although the latter execution mode can be expected to be slightly faster, it is only available for use when the underlying data source exposes those terms ordinals. Moreover, it may actually be slower if most field values are unique. Elasticsearch tries to have sensible defaults when it comes to the execution mode that should be used, but in case you know that an execution mode may perform better than the other one, you have the ability to provide Elasticsearch with a hint: [source,js] -------------------------------------------------- { "aggs" : { "tags" : { "significant_terms" : { "field" : "tags", "execution_hint": "map" <1> } } } } -------------------------------------------------- <1> the possible values are `map` and `ordinals` Please note that Elasticsearch will ignore this execution hint if it is not applicable.