CONSOLE-ify date histogram docs
This adds the `VIEW IN SENSE` and `COPY AS CURL` links and has the build automatically execute the snippets and verify that they work. Relates to #18160
This commit is contained in:
parent
c2a580304b
commit
f7524fbdef
|
@ -24,7 +24,6 @@ apply plugin: 'elasticsearch.docs-test'
|
||||||
* only remove entries from this list. When it is empty we'll remove it
|
* only remove entries from this list. When it is empty we'll remove it
|
||||||
* entirely and have a party! There will be cake and everything.... */
|
* entirely and have a party! There will be cake and everything.... */
|
||||||
buildRestTests.expectedUnconvertedCandidates = [
|
buildRestTests.expectedUnconvertedCandidates = [
|
||||||
'reference/aggregations/bucket/datehistogram-aggregation.asciidoc',
|
|
||||||
'reference/aggregations/bucket/daterange-aggregation.asciidoc',
|
'reference/aggregations/bucket/daterange-aggregation.asciidoc',
|
||||||
'reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc',
|
'reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc',
|
||||||
'reference/aggregations/bucket/filter-aggregation.asciidoc',
|
'reference/aggregations/bucket/filter-aggregation.asciidoc',
|
||||||
|
|
|
@ -12,9 +12,10 @@ Requesting bucket intervals of a month.
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
POST /sales/_search?size=0
|
||||||
{
|
{
|
||||||
"aggs" : {
|
"aggs" : {
|
||||||
"articles_over_time" : {
|
"sales_over_time" : {
|
||||||
"date_histogram" : {
|
"date_histogram" : {
|
||||||
"field" : "date",
|
"field" : "date",
|
||||||
"interval" : "month"
|
"interval" : "month"
|
||||||
|
@ -23,6 +24,8 @@ Requesting bucket intervals of a month.
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
// TEST[setup:sales]
|
||||||
|
|
||||||
Available expressions for interval: `year`, `quarter`, `month`, `week`, `day`, `hour`, `minute`, `second`
|
Available expressions for interval: `year`, `quarter`, `month`, `week`, `day`, `hour`, `minute`, `second`
|
||||||
|
|
||||||
|
@ -32,9 +35,10 @@ time unit (e.g., `1.5h` could instead be specified as `90m`).
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
POST /sales/_search?size=0
|
||||||
{
|
{
|
||||||
"aggs" : {
|
"aggs" : {
|
||||||
"articles_over_time" : {
|
"sales_over_time" : {
|
||||||
"date_histogram" : {
|
"date_histogram" : {
|
||||||
"field" : "date",
|
"field" : "date",
|
||||||
"interval" : "90m"
|
"interval" : "90m"
|
||||||
|
@ -43,6 +47,8 @@ time unit (e.g., `1.5h` could instead be specified as `90m`).
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
// TEST[setup:sales]
|
||||||
|
|
||||||
==== Keys
|
==== Keys
|
||||||
|
|
||||||
|
@ -56,9 +62,10 @@ TIP: If no `format` is specified, then it will use the first date
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
POST /sales/_search?size=0
|
||||||
{
|
{
|
||||||
"aggs" : {
|
"aggs" : {
|
||||||
"articles_over_time" : {
|
"sales_over_time" : {
|
||||||
"date_histogram" : {
|
"date_histogram" : {
|
||||||
"field" : "date",
|
"field" : "date",
|
||||||
"interval" : "1M",
|
"interval" : "1M",
|
||||||
|
@ -68,6 +75,8 @@ TIP: If no `format` is specified, then it will use the first date
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
// TEST[setup:sales]
|
||||||
|
|
||||||
<1> Supports expressive date <<date-format-pattern,format pattern>>
|
<1> Supports expressive date <<date-format-pattern,format pattern>>
|
||||||
|
|
||||||
|
@ -76,25 +85,31 @@ Response:
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
{
|
{
|
||||||
|
...
|
||||||
"aggregations": {
|
"aggregations": {
|
||||||
"articles_over_time": {
|
"sales_over_time": {
|
||||||
"buckets": [
|
"buckets": [
|
||||||
{
|
{
|
||||||
"key_as_string": "2013-02-02",
|
"key_as_string": "2015-01-01",
|
||||||
"key": 1328140800000,
|
"key": 1420070400000,
|
||||||
"doc_count": 1
|
"doc_count": 3
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"key_as_string": "2013-03-02",
|
"key_as_string": "2015-02-01",
|
||||||
"key": 1330646400000,
|
"key": 1422748800000,
|
||||||
"doc_count": 2
|
"doc_count": 2
|
||||||
},
|
},
|
||||||
...
|
{
|
||||||
|
"key_as_string": "2015-03-01",
|
||||||
|
"key": 1425168000000,
|
||||||
|
"doc_count": 2
|
||||||
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
|
||||||
|
|
||||||
==== Time Zone
|
==== Time Zone
|
||||||
|
|
||||||
|
@ -110,12 +125,12 @@ Consider the following example:
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
---------------------------------
|
---------------------------------
|
||||||
PUT my_index/log/1
|
PUT my_index/log/1?refresh
|
||||||
{
|
{
|
||||||
"date": "2015-10-01T00:30:00Z"
|
"date": "2015-10-01T00:30:00Z"
|
||||||
}
|
}
|
||||||
|
|
||||||
PUT my_index/log/2
|
PUT my_index/log/2?refresh
|
||||||
{
|
{
|
||||||
"date": "2015-10-01T01:30:00Z"
|
"date": "2015-10-01T01:30:00Z"
|
||||||
}
|
}
|
||||||
|
@ -132,6 +147,7 @@ GET my_index/_search?size=0
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
---------------------------------
|
---------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
|
||||||
UTC is used if no time zone is specified, which would result in both of these
|
UTC is used if no time zone is specified, which would result in both of these
|
||||||
documents being placed into the same day bucket, which starts at midnight UTC
|
documents being placed into the same day bucket, which starts at midnight UTC
|
||||||
|
@ -139,18 +155,22 @@ on 1 October 2015:
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
---------------------------------
|
---------------------------------
|
||||||
"aggregations": {
|
{
|
||||||
"by_day": {
|
...
|
||||||
"buckets": [
|
"aggregations": {
|
||||||
{
|
"by_day": {
|
||||||
"key_as_string": "2015-10-01T00:00:00.000Z",
|
"buckets": [
|
||||||
"key": 1443657600000,
|
{
|
||||||
"doc_count": 2
|
"key_as_string": "2015-10-01T00:00:00.000Z",
|
||||||
}
|
"key": 1443657600000,
|
||||||
]
|
"doc_count": 2
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
---------------------------------
|
---------------------------------
|
||||||
|
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
|
||||||
|
|
||||||
If a `time_zone` of `-01:00` is specified, then midnight starts at one hour before
|
If a `time_zone` of `-01:00` is specified, then midnight starts at one hour before
|
||||||
midnight UTC:
|
midnight UTC:
|
||||||
|
@ -170,29 +190,36 @@ GET my_index/_search?size=0
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
---------------------------------
|
---------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
// TEST[continued]
|
||||||
|
|
||||||
Now the first document falls into the bucket for 30 September 2015, while the
|
Now the first document falls into the bucket for 30 September 2015, while the
|
||||||
second document falls into the bucket for 1 October 2015:
|
second document falls into the bucket for 1 October 2015:
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
---------------------------------
|
---------------------------------
|
||||||
"aggregations": {
|
{
|
||||||
"by_day": {
|
...
|
||||||
"buckets": [
|
"aggregations": {
|
||||||
{
|
"by_day": {
|
||||||
"key_as_string": "2015-09-30T00:00:00.000-01:00", <1>
|
"buckets": [
|
||||||
"key": 1443571200000,
|
{
|
||||||
"doc_count": 1
|
"key_as_string": "2015-09-30T00:00:00.000-01:00", <1>
|
||||||
},
|
"key": 1443574800000,
|
||||||
{
|
"doc_count": 1
|
||||||
"key_as_string": "2015-10-01T00:00:00.000-01:00", <1>
|
},
|
||||||
"key": 1443657600000,
|
{
|
||||||
"doc_count": 1
|
"key_as_string": "2015-10-01T00:00:00.000-01:00", <1>
|
||||||
}
|
"key": 1443661200000,
|
||||||
]
|
"doc_count": 1
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
---------------------------------
|
---------------------------------
|
||||||
|
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
|
||||||
|
|
||||||
<1> The `key_as_string` value represents midnight on each day
|
<1> The `key_as_string` value represents midnight on each day
|
||||||
in the specified time zone.
|
in the specified time zone.
|
||||||
|
|
||||||
|
@ -220,12 +247,12 @@ to run from 6am to 6am:
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
-----------------------------
|
-----------------------------
|
||||||
PUT my_index/log/1
|
PUT my_index/log/1?refresh
|
||||||
{
|
{
|
||||||
"date": "2015-10-01T05:30:00Z"
|
"date": "2015-10-01T05:30:00Z"
|
||||||
}
|
}
|
||||||
|
|
||||||
PUT my_index/log/2
|
PUT my_index/log/2?refresh
|
||||||
{
|
{
|
||||||
"date": "2015-10-01T06:30:00Z"
|
"date": "2015-10-01T06:30:00Z"
|
||||||
}
|
}
|
||||||
|
@ -243,29 +270,34 @@ GET my_index/_search?size=0
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
-----------------------------
|
-----------------------------
|
||||||
|
// CONSOLE
|
||||||
|
|
||||||
Instead of a single bucket starting at midnight, the above request groups the
|
Instead of a single bucket starting at midnight, the above request groups the
|
||||||
documents into buckets starting at 6am:
|
documents into buckets starting at 6am:
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
-----------------------------
|
-----------------------------
|
||||||
"aggregations": {
|
{
|
||||||
"by_day": {
|
...
|
||||||
"buckets": [
|
"aggregations": {
|
||||||
{
|
"by_day": {
|
||||||
"key_as_string": "2015-09-30T06:00:00.000Z",
|
"buckets": [
|
||||||
"key": 1443592800000,
|
{
|
||||||
"doc_count": 1
|
"key_as_string": "2015-09-30T06:00:00.000Z",
|
||||||
},
|
"key": 1443592800000,
|
||||||
{
|
"doc_count": 1
|
||||||
"key_as_string": "2015-10-01T06:00:00.000Z",
|
},
|
||||||
"key": 1443679200000,
|
{
|
||||||
"doc_count": 1
|
"key_as_string": "2015-10-01T06:00:00.000Z",
|
||||||
}
|
"key": 1443679200000,
|
||||||
]
|
"doc_count": 1
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
-----------------------------
|
-----------------------------
|
||||||
|
// TESTRESPONSE[s/\.\.\./"took": $body.took,"timed_out": false,"_shards": $body._shards,"hits": $body.hits,/]
|
||||||
|
|
||||||
NOTE: The start `offset` of each bucket is calculated after the `time_zone`
|
NOTE: The start `offset` of each bucket is calculated after the `time_zone`
|
||||||
adjustments have been made.
|
adjustments have been made.
|
||||||
|
@ -287,17 +319,20 @@ had a value.
|
||||||
|
|
||||||
[source,js]
|
[source,js]
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
POST /sales/_search?size=0
|
||||||
{
|
{
|
||||||
"aggs" : {
|
"aggs" : {
|
||||||
"publish_date" : {
|
"sale_date" : {
|
||||||
"date_histogram" : {
|
"date_histogram" : {
|
||||||
"field" : "publish_date",
|
"field" : "date",
|
||||||
"interval": "year",
|
"interval": "year",
|
||||||
"missing": "2000-01-01" <1>
|
"missing": "2000/01/01" <1>
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
// CONSOLE
|
||||||
|
// TEST[setup:sales]
|
||||||
|
|
||||||
<1> Documents without a value in the `publish_date` field will fall into the same bucket as documents that have the value `2000-01-01`.
|
<1> Documents without a value in the `publish_date` field will fall into the same bucket as documents that have the value `2000-01-01`.
|
||||||
|
|
Loading…
Reference in New Issue