CONSOLE-ify date histogram docs

This adds the `VIEW IN SENSE` and `COPY AS CURL` links and has
the build automatically execute the snippets and verify that they
work.

Relates to #18160
This commit is contained in:
Nik Everett 2017-01-20 16:23:28 -05:00
parent c2a580304b
commit f7524fbdef
2 changed files with 89 additions and 55 deletions

View File

@ -24,7 +24,6 @@ apply plugin: 'elasticsearch.docs-test'
* only remove entries from this list. When it is empty we'll remove it
* entirely and have a party! There will be cake and everything.... */
buildRestTests.expectedUnconvertedCandidates = [
'reference/aggregations/bucket/datehistogram-aggregation.asciidoc',
'reference/aggregations/bucket/daterange-aggregation.asciidoc',
'reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc',
'reference/aggregations/bucket/filter-aggregation.asciidoc',

View File

@ -12,9 +12,10 @@ Requesting bucket intervals of a month.
[source,js]
--------------------------------------------------
POST /sales/_search?size=0
{
"aggs" : {
"articles_over_time" : {
"sales_over_time" : {
"date_histogram" : {
"field" : "date",
"interval" : "month"
@ -23,6 +24,8 @@ Requesting bucket intervals of a month.
}
}
--------------------------------------------------
// CONSOLE
// TEST[setup:sales]
Available expressions for interval: `year`, `quarter`, `month`, `week`, `day`, `hour`, `minute`, `second`
@ -32,9 +35,10 @@ time unit (e.g., `1.5h` could instead be specified as `90m`).
[source,js]
--------------------------------------------------
POST /sales/_search?size=0
{
"aggs" : {
"articles_over_time" : {
"sales_over_time" : {
"date_histogram" : {
"field" : "date",
"interval" : "90m"
@ -43,6 +47,8 @@ time unit (e.g., `1.5h` could instead be specified as `90m`).
}
}
--------------------------------------------------
// CONSOLE
// TEST[setup:sales]
==== Keys
@ -56,9 +62,10 @@ TIP: If no `format` is specified, then it will use the first date
[source,js]
--------------------------------------------------
POST /sales/_search?size=0
{
"aggs" : {
"articles_over_time" : {
"sales_over_time" : {
"date_histogram" : {
"field" : "date",
"interval" : "1M",
@ -68,6 +75,8 @@ TIP: If no `format` is specified, then it will use the first date
}
}
--------------------------------------------------
// CONSOLE
// TEST[setup:sales]
<1> Supports expressive date <<date-format-pattern,format pattern>>
@ -76,25 +85,31 @@ Response:
[source,js]
--------------------------------------------------
{
...
"aggregations": {
"articles_over_time": {
"sales_over_time": {
"buckets": [
{
"key_as_string": "2013-02-02",
"key": 1328140800000,
"doc_count": 1
"key_as_string": "2015-01-01",
"key": 1420070400000,
"doc_count": 3
},
{
"key_as_string": "2013-03-02",
"key": 1330646400000,
"key_as_string": "2015-02-01",
"key": 1422748800000,
"doc_count": 2
},
...
{
"key_as_string": "2015-03-01",
"key": 1425168000000,
"doc_count": 2
}
]
}
}
}
--------------------------------------------------
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
==== Time Zone
@ -110,12 +125,12 @@ Consider the following example:
[source,js]
---------------------------------
PUT my_index/log/1
PUT my_index/log/1?refresh
{
"date": "2015-10-01T00:30:00Z"
}
PUT my_index/log/2
PUT my_index/log/2?refresh
{
"date": "2015-10-01T01:30:00Z"
}
@ -132,6 +147,7 @@ GET my_index/_search?size=0
}
}
---------------------------------
// CONSOLE
UTC is used if no time zone is specified, which would result in both of these
documents being placed into the same day bucket, which starts at midnight UTC
@ -139,7 +155,9 @@ on 1 October 2015:
[source,js]
---------------------------------
"aggregations": {
{
...
"aggregations": {
"by_day": {
"buckets": [
{
@ -149,8 +167,10 @@ on 1 October 2015:
}
]
}
}
}
---------------------------------
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
If a `time_zone` of `-01:00` is specified, then midnight starts at one hour before
midnight UTC:
@ -170,29 +190,36 @@ GET my_index/_search?size=0
}
}
---------------------------------
// CONSOLE
// TEST[continued]
Now the first document falls into the bucket for 30 September 2015, while the
second document falls into the bucket for 1 October 2015:
[source,js]
---------------------------------
"aggregations": {
{
...
"aggregations": {
"by_day": {
"buckets": [
{
"key_as_string": "2015-09-30T00:00:00.000-01:00", <1>
"key": 1443571200000,
"key": 1443574800000,
"doc_count": 1
},
{
"key_as_string": "2015-10-01T00:00:00.000-01:00", <1>
"key": 1443657600000,
"key": 1443661200000,
"doc_count": 1
}
]
}
}
}
---------------------------------
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
<1> The `key_as_string` value represents midnight on each day
in the specified time zone.
@ -220,12 +247,12 @@ to run from 6am to 6am:
[source,js]
-----------------------------
PUT my_index/log/1
PUT my_index/log/1?refresh
{
"date": "2015-10-01T05:30:00Z"
}
PUT my_index/log/2
PUT my_index/log/2?refresh
{
"date": "2015-10-01T06:30:00Z"
}
@ -243,13 +270,16 @@ GET my_index/_search?size=0
}
}
-----------------------------
// CONSOLE
Instead of a single bucket starting at midnight, the above request groups the
documents into buckets starting at 6am:
[source,js]
-----------------------------
"aggregations": {
{
...
"aggregations": {
"by_day": {
"buckets": [
{
@ -264,8 +294,10 @@ documents into buckets starting at 6am:
}
]
}
}
}
-----------------------------
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
NOTE: The start `offset` of each bucket is calculated after the `time_zone`
adjustments have been made.
@ -287,17 +319,20 @@ had a value.
[source,js]
--------------------------------------------------
POST /sales/_search?size=0
{
"aggs" : {
"publish_date" : {
"sale_date" : {
"date_histogram" : {
"field" : "publish_date",
"field" : "date",
"interval": "year",
"missing": "2000-01-01" <1>
"missing": "2000/01/01" <1>
}
}
}
}
--------------------------------------------------
// CONSOLE
// TEST[setup:sales]
<1> Documents without a value in the `publish_date` field will fall into the same bucket as documents that have the value `2000-01-01`.