--- id: timeseriesquery title: "Timeseries queries" sidebar_label: "Timeseries" --- > Apache Druid supports two query languages: [Druid SQL](sql.md) and [native queries](querying.md). > This document describes a query > type in the native language. For information about when Druid SQL will use this query type, refer to the > [SQL documentation](sql-translation.md#query-types). These types of queries take a timeseries query object and return an array of JSON objects where each object represents a value asked for by the timeseries query. An example timeseries query object is shown below: ```json { "queryType": "timeseries", "dataSource": "sample_datasource", "granularity": "day", "descending": "true", "filter": { "type": "and", "fields": [ { "type": "selector", "dimension": "sample_dimension1", "value": "sample_value1" }, { "type": "or", "fields": [ { "type": "selector", "dimension": "sample_dimension2", "value": "sample_value2" }, { "type": "selector", "dimension": "sample_dimension3", "value": "sample_value3" } ] } ] }, "aggregations": [ { "type": "longSum", "name": "sample_name1", "fieldName": "sample_fieldName1" }, { "type": "doubleSum", "name": "sample_name2", "fieldName": "sample_fieldName2" } ], "postAggregations": [ { "type": "arithmetic", "name": "sample_divide", "fn": "/", "fields": [ { "type": "fieldAccess", "name": "postAgg__sample_name1", "fieldName": "sample_name1" }, { "type": "fieldAccess", "name": "postAgg__sample_name2", "fieldName": "sample_name2" } ] } ], "intervals": [ "2012-01-01T00:00:00.000/2012-01-03T00:00:00.000" ] } ``` There are 7 main parts to a timeseries query: |property|description|required?| |--------|-----------|---------| |queryType|This String should always be "timeseries"; this is the first thing Apache Druid looks at to figure out how to interpret the query|yes| |dataSource|A String or Object defining the data source to query, very similar to a table in a relational database. See [DataSource](../querying/datasource.md) for more information.|yes| |descending|Whether to make descending ordered result. Default is `false`(ascending).|no| |intervals|A JSON Object representing ISO-8601 Intervals. This defines the time ranges to run the query over.|yes| |granularity|Defines the granularity to bucket query results. See [Granularities](../querying/granularities.md)|yes| |filter|See [Filters](../querying/filters.md)|no| |virtualColumns|A JSON list of [virtual columns](./virtual-columns.md). You can reference the virtual columns in `aggregations` or `postAggregations`.| no (default none)| |aggregations|See [Aggregations](../querying/aggregations.md)|no| |postAggregations|See [Post Aggregations](../querying/post-aggregations.md)|no| |limit|An integer that limits the number of results. The default is unlimited.|no| |context|Can be used to modify query behavior, including [grand totals](#grand-totals) and [zero-filling](#zero-filling). See also [Context](../querying/query-context.md) for parameters that apply to all query types.|no| To pull it all together, the above query would return 2 data points, one for each day between 2012-01-01 and 2012-01-03, from the "sample\_datasource" table. Each data point would be the (long) sum of sample\_fieldName1, the (double) sum of sample\_fieldName2 and the (double) result of sample\_fieldName1 divided by sample\_fieldName2 for the filter set. The output looks like this: ```json [ { "timestamp": "2012-01-01T00:00:00.000Z", "result": { "sample_name1": , "sample_name2": , "sample_divide": } }, { "timestamp": "2012-01-02T00:00:00.000Z", "result": { "sample_name1": , "sample_name2": , "sample_divide": } } ] ``` ## Grand totals Druid can include an extra "grand totals" row as the last row of a timeseries result set. To enable this, add `"grandTotal" : true` to your query context. For example: ```json { "queryType": "timeseries", "dataSource": "sample_datasource", "intervals": [ "2012-01-01T00:00:00.000/2012-01-03T00:00:00.000" ], "granularity": "day", "aggregations": [ { "type": "longSum", "name": "sample_name1", "fieldName": "sample_fieldName1" }, { "type": "doubleSum", "name": "sample_name2", "fieldName": "sample_fieldName2" } ], "context": { "grandTotal": true } } ``` The grand totals row will appear as the last row in the result array, and will have no timestamp. It will be the last row even if the query is run in "descending" mode. Post-aggregations in the grand totals row will be computed based upon the grand total aggregations. ## Zero-filling Timeseries queries normally fill empty interior time buckets with zeroes. For example, if you issue a "day" granularity timeseries query for the interval 2012-01-01/2012-01-04, and no data exists for 2012-01-02, you will receive: ```json [ { "timestamp": "2012-01-01T00:00:00.000Z", "result": { "sample_name1": } }, { "timestamp": "2012-01-02T00:00:00.000Z", "result": { "sample_name1": 0 } }, { "timestamp": "2012-01-03T00:00:00.000Z", "result": { "sample_name1": } } ] ``` Time buckets that lie completely outside the data interval are not zero-filled. You can disable all zero-filling with the context flag "skipEmptyBuckets". In this mode, the data point for 2012-01-02 would be omitted from the results. A query with this context flag set would look like: ```json { "queryType": "timeseries", "dataSource": "sample_datasource", "granularity": "day", "aggregations": [ { "type": "longSum", "name": "sample_name1", "fieldName": "sample_fieldName1" } ], "intervals": [ "2012-01-01T00:00:00.000/2012-01-04T00:00:00.000" ], "context" : { "skipEmptyBuckets": "true" } } ```