2015-05-05 17:07:32 -04:00
---
layout: doc_page
---
# Aggregations
2015-12-31 19:32:51 -05:00
Aggregations can be provided at ingestion time as part of the ingestion spec as a way of summarizing data before it enters Druid.
Aggregations can also be specified as part of many queries at query time.
2015-05-05 17:07:32 -04:00
Available aggregations are:
### Count aggregator
2015-12-31 19:32:51 -05:00
`count` computes the count of Druid rows that match the filters.
2015-05-05 17:07:32 -04:00
```json
{ "type" : "count", "name" : < output_name > }
```
2015-12-31 19:32:51 -05:00
Please note the count aggregator counts the number of Druid rows, which does not always reflect the number of raw events ingested.
This is because Druid rolls up data at ingestion time. To
count the number of ingested rows of data, include a count aggregator at ingestion time, and a longSum aggregator at
query time.
2015-05-05 17:07:32 -04:00
### Sum aggregators
#### `longSum` aggregator
computes the sum of values as a 64-bit, signed integer
```json
{ "type" : "longSum", "name" : < output_name > , "fieldName" : < metric_name > }
```
`name` – output name for the summed value
`fieldName` – name of the metric column to sum over
#### `doubleSum` aggregator
Computes the sum of values as 64-bit floating point value. Similar to `longSum`
```json
{ "type" : "doubleSum", "name" : < output_name > , "fieldName" : < metric_name > }
```
### Min / Max aggregators
#### `doubleMin` aggregator
`doubleMin` computes the minimum of all metric values and Double.POSITIVE_INFINITY
```json
{ "type" : "doubleMin", "name" : < output_name > , "fieldName" : < metric_name > }
```
#### `doubleMax` aggregator
`doubleMax` computes the maximum of all metric values and Double.NEGATIVE_INFINITY
```json
{ "type" : "doubleMax", "name" : < output_name > , "fieldName" : < metric_name > }
```
#### `longMin` aggregator
`longMin` computes the minimum of all metric values and Long.MAX_VALUE
```json
{ "type" : "longMin", "name" : < output_name > , "fieldName" : < metric_name > }
```
#### `longMax` aggregator
`longMax` computes the maximum of all metric values and Long.MIN_VALUE
```json
{ "type" : "longMax", "name" : < output_name > , "fieldName" : < metric_name > }
```
### JavaScript aggregator
Computes an arbitrary JavaScript function over a set of columns (both metrics and dimensions).
All JavaScript functions must return numerical values.
2016-01-06 00:27:52 -05:00
JavaScript aggregators are much slower than native Java aggregators and if performance is critical, you should implement
your functionality as a native Java aggregator.
2015-05-05 17:07:32 -04:00
```json
{ "type": "javascript",
"name": "< output_name > ",
"fieldNames" : [ < column1 > , < column2 > , ... ],
"fnAggregate" : "function(current, column1, column2, ...) {
< updates partial aggregate ( current ) based on the current row values >
return < updated partial aggregate >
}",
"fnCombine" : "function(partialA, partialB) { return < combined partial results > ; }",
"fnReset" : "function() { return < initial value > ; }"
}
```
**Example**
```json
{
"type": "javascript",
2015-06-24 09:24:07 -04:00
"name": "sum(log(x)*y) + 10",
2015-05-05 17:07:32 -04:00
"fieldNames": ["x", "y"],
"fnAggregate" : "function(current, a, b) { return current + (Math.log(a) * b); }",
"fnCombine" : "function(partialA, partialB) { return partialA + partialB; }",
"fnReset" : "function() { return 10; }"
}
```
2015-12-31 19:32:51 -05:00
The javascript aggregator is recommended for rapidly prototyping features. This aggregator will be much slower in production
use than a native Java aggregator.
2016-02-04 14:53:09 -05:00
## Approximate Aggregations
2015-05-05 17:07:32 -04:00
### Cardinality aggregator
2016-02-04 14:53:09 -05:00
Computes the cardinality of a set of Druid dimensions, using HyperLogLog to estimate the cardinality. Please note that this
aggregator will be much slower than indexing a column with the hyperUnique aggregator. This aggregator also runs over a dimension column, which
means the string dimension cannot be removed from the dataset to improve rollup. In general, we strongly recommend using the hyperUnique aggregator
instead of the cardinality aggregator if you do not care about the individual values of a dimension.
2015-05-05 17:07:32 -04:00
```json
{
"type": "cardinality",
"name": "< output_name > ",
"fieldNames": [ < dimension1 > , < dimension2 > , ... ],
"byRow": < false | true > # (optional, defaults to false)
}
```
#### Cardinality by value
When setting `byRow` to `false` (the default) it computes the cardinality of the set composed of the union of all dimension values for all the given dimensions.
* For a single dimension, this is equivalent to
```sql
SELECT COUNT(DISTINCT(dimension)) FROM < datasource >
```
* For multiple dimensions, this is equivalent to something akin to
```sql
SELECT COUNT(DISTINCT(value)) FROM (
SELECT dim_1 as value FROM < datasource >
UNION
SELECT dim_2 as value FROM < datasource >
UNION
SELECT dim_3 as value FROM < datasource >
)
```
#### Cardinality by row
2015-06-24 09:24:07 -04:00
When setting `byRow` to `true` it computes the cardinality by row, i.e. the cardinality of distinct dimension combinations.
2015-05-05 17:07:32 -04:00
This is equivalent to something akin to
```sql
2015-06-24 09:24:07 -04:00
SELECT COUNT(*) FROM ( SELECT DIM1, DIM2, DIM3 FROM < datasource > GROUP BY DIM1, DIM2, DIM3 )
2015-05-05 17:07:32 -04:00
```
**Example**
2015-08-04 14:08:44 -04:00
Determine the number of distinct countries people are living in or have come from.
2015-05-05 17:07:32 -04:00
```json
{
"type": "cardinality",
2015-08-04 14:08:44 -04:00
"name": "distinct_countries",
"fieldNames": [ "coutry_of_origin", "country_of_residence" ]
2015-05-05 17:07:32 -04:00
}
```
2015-08-04 14:08:44 -04:00
Determine the number of distinct people (i.e. combinations of first and last name).
2015-05-05 17:07:32 -04:00
```json
{
"type": "cardinality",
2015-08-04 14:08:44 -04:00
"name": "distinct_people",
"fieldNames": [ "first_name", "last_name" ],
2015-05-05 17:07:32 -04:00
"byRow" : true
}
```
### HyperUnique aggregator
Uses [HyperLogLog ](http://algo.inria.fr/flajolet/Publications/FlFuGaMe07.pdf ) to compute the estimated cardinality of a dimension that has been aggregated as a "hyperUnique" metric at indexing time.
```json
{ "type" : "hyperUnique", "name" : < output_name > , "fieldName" : < metric_name > }
```
2016-04-13 16:30:01 -04:00
For more approximate aggregators, please see [theta sketches ](../development/extensions-core/datasketches-aggregators.html ).
2016-02-04 14:53:09 -05:00
2015-05-05 17:07:32 -04:00
## Miscellaneous Aggregations
### Filtered Aggregator
A filtered aggregator wraps any given aggregator, but only aggregates the values for which the given dimension filter matches.
This makes it possible to compute the results of a filtered and an unfiltered aggregation simultaneously, without having to issue multiple queries, and use both results as part of post-aggregations.
*Note:* If only the filtered results are required, consider putting the filter on the query itself, which will be much faster since it does not require scanning all the data.
```json
{
"type" : "filtered",
"filter" : {
"type" : "selector",
"dimension" : < dimension > ,
"value" : < dimension value >
}
"aggregator" : < aggregation >
}
```