2016-04-29 10:42:03 -04:00
|
|
|
/*
|
|
|
|
* Licensed to Elasticsearch under one or more contributor
|
|
|
|
* license agreements. See the NOTICE file distributed with
|
|
|
|
* this work for additional information regarding copyright
|
|
|
|
* ownership. Elasticsearch licenses this file to you under
|
|
|
|
* the Apache License, Version 2.0 (the "License"); you may
|
|
|
|
* not use this file except in compliance with the License.
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing,
|
|
|
|
* software distributed under the License is distributed on an
|
|
|
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
|
|
* KIND, either express or implied. See the License for the
|
|
|
|
* specific language governing permissions and limitations
|
|
|
|
* under the License.
|
|
|
|
*/
|
|
|
|
|
2016-05-05 16:46:40 -04:00
|
|
|
apply plugin: 'elasticsearch.docs-test'
|
2016-04-29 10:42:03 -04:00
|
|
|
|
2016-09-09 11:10:13 -04:00
|
|
|
/* List of files that have snippets that probably should be converted to
|
|
|
|
* `// CONSOLE` and `// TESTRESPONSE` but have yet to be converted. Try and
|
|
|
|
* only remove entries from this list. When it is empty we'll remove it
|
|
|
|
* entirely and have a party! There will be cake and everything.... */
|
|
|
|
buildRestTests.expectedUnconvertedCandidates = [
|
|
|
|
'reference/aggregations/bucket/iprange-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/bucket/nested-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/bucket/range-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/bucket/reverse-nested-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/bucket/significantterms-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/bucket/terms-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/matrix/stats-aggregation.asciidoc',
|
|
|
|
'reference/aggregations/metrics/tophits-aggregation.asciidoc',
|
|
|
|
'reference/cluster/allocation-explain.asciidoc',
|
|
|
|
'reference/cluster/nodes-info.asciidoc',
|
|
|
|
'reference/cluster/pending.asciidoc',
|
|
|
|
'reference/cluster/state.asciidoc',
|
|
|
|
'reference/cluster/stats.asciidoc',
|
|
|
|
'reference/cluster/tasks.asciidoc',
|
|
|
|
'reference/docs/delete-by-query.asciidoc',
|
|
|
|
'reference/docs/reindex.asciidoc',
|
|
|
|
'reference/docs/update-by-query.asciidoc',
|
|
|
|
'reference/index-modules/similarity.asciidoc',
|
|
|
|
'reference/index-modules/store.asciidoc',
|
|
|
|
'reference/index-modules/translog.asciidoc',
|
|
|
|
'reference/indices/recovery.asciidoc',
|
|
|
|
'reference/indices/segments.asciidoc',
|
|
|
|
'reference/indices/shard-stores.asciidoc',
|
2017-07-15 02:12:05 -04:00
|
|
|
'reference/migration/migrate_6_0/scripting.asciidoc',
|
2016-09-09 11:10:13 -04:00
|
|
|
'reference/search/profile.asciidoc',
|
|
|
|
]
|
|
|
|
|
Build: Rework integ test setup and shutdown to ensure stop runs when desired (#23304)
Gradle's finalizedBy on tasks only ensures one task runs after another,
but not immediately after. This is problematic for our integration tests
since it allows multiple project's integ test clusters to be
simultaneously. While this has not been a problem thus far (gradle 2.13
happened to keep the finalizedBy tasks close enough that no clusters
were running in parallel), with gradle 3.3 the task graph generation has
changed, and numerous clusters may be running simultaneously, causing
memory pressure, and thus generally slower tests, or even failure if the
system has a limited amount of memory (eg in a vagrant host).
This commit reworks how integ tests are configured. It adds an
`integTestCluster` extension to gradle which is equivalent to the current
`integTest.cluster` and moves the rest test runner task to
`integTestRunner`. The `integTest` task is then just a dummy task,
which depends on the cluster runner task, as well as the cluster stop
task. This means running `integTest` in one project will both run the
rest tests, and shut down the cluster, before running `integTest` in
another project.
2017-02-22 15:43:15 -05:00
|
|
|
integTestCluster {
|
|
|
|
setting 'script.max_compilations_per_minute', '1000'
|
|
|
|
/* Enable regexes in painless so our tests don't complain about example
|
|
|
|
* snippets that use them. */
|
|
|
|
setting 'script.painless.regex.enabled', 'true'
|
|
|
|
Closure configFile = {
|
|
|
|
extraConfigFile it, "src/test/cluster/config/$it"
|
2016-05-13 16:15:51 -04:00
|
|
|
}
|
2017-04-02 11:15:26 -04:00
|
|
|
configFile 'analysis/example_word_list.txt'
|
|
|
|
configFile 'analysis/hyphenation_patterns.xml'
|
2017-03-22 16:30:52 -04:00
|
|
|
configFile 'analysis/synonym.txt'
|
2017-03-22 17:56:53 -04:00
|
|
|
configFile 'analysis/stemmer_override.txt'
|
Build: Rework integ test setup and shutdown to ensure stop runs when desired (#23304)
Gradle's finalizedBy on tasks only ensures one task runs after another,
but not immediately after. This is problematic for our integration tests
since it allows multiple project's integ test clusters to be
simultaneously. While this has not been a problem thus far (gradle 2.13
happened to keep the finalizedBy tasks close enough that no clusters
were running in parallel), with gradle 3.3 the task graph generation has
changed, and numerous clusters may be running simultaneously, causing
memory pressure, and thus generally slower tests, or even failure if the
system has a limited amount of memory (eg in a vagrant host).
This commit reworks how integ tests are configured. It adds an
`integTestCluster` extension to gradle which is equivalent to the current
`integTest.cluster` and moves the rest test runner task to
`integTestRunner`. The `integTest` task is then just a dummy task,
which depends on the cluster runner task, as well as the cluster stop
task. This means running `integTest` in one project will both run the
rest tests, and shut down the cluster, before running `integTest` in
another project.
2017-02-22 15:43:15 -05:00
|
|
|
configFile 'userdict_ja.txt'
|
|
|
|
configFile 'KeywordTokenizer.rbbi'
|
2017-04-02 11:15:26 -04:00
|
|
|
extraConfigFile 'hunspell/en_US/en_US.aff', '../core/src/test/resources/indices/analyze/conf_dir/hunspell/en_US/en_US.aff'
|
|
|
|
extraConfigFile 'hunspell/en_US/en_US.dic', '../core/src/test/resources/indices/analyze/conf_dir/hunspell/en_US/en_US.dic'
|
Build: Rework integ test setup and shutdown to ensure stop runs when desired (#23304)
Gradle's finalizedBy on tasks only ensures one task runs after another,
but not immediately after. This is problematic for our integration tests
since it allows multiple project's integ test clusters to be
simultaneously. While this has not been a problem thus far (gradle 2.13
happened to keep the finalizedBy tasks close enough that no clusters
were running in parallel), with gradle 3.3 the task graph generation has
changed, and numerous clusters may be running simultaneously, causing
memory pressure, and thus generally slower tests, or even failure if the
system has a limited amount of memory (eg in a vagrant host).
This commit reworks how integ tests are configured. It adds an
`integTestCluster` extension to gradle which is equivalent to the current
`integTest.cluster` and moves the rest test runner task to
`integTestRunner`. The `integTest` task is then just a dummy task,
which depends on the cluster runner task, as well as the cluster stop
task. This means running `integTest` in one project will both run the
rest tests, and shut down the cluster, before running `integTest` in
another project.
2017-02-22 15:43:15 -05:00
|
|
|
// Whitelist reindexing from the local node so we can test it.
|
|
|
|
setting 'reindex.remote.whitelist', '127.0.0.1:*'
|
2016-05-13 16:15:51 -04:00
|
|
|
}
|
|
|
|
|
2017-02-15 18:13:27 -05:00
|
|
|
// Build the cluster with all plugins
|
2016-07-15 17:34:21 -04:00
|
|
|
|
2016-05-13 16:15:51 -04:00
|
|
|
project.rootProject.subprojects.findAll { it.parent.path == ':plugins' }.each { subproj ->
|
|
|
|
/* Skip repositories. We just aren't going to be able to test them so it
|
|
|
|
* doesn't make sense to waste time installing them. */
|
|
|
|
if (subproj.path.startsWith(':plugins:repository-')) {
|
|
|
|
return
|
|
|
|
}
|
2016-07-15 17:34:21 -04:00
|
|
|
subproj.afterEvaluate { // need to wait until the project has been configured
|
Build: Rework integ test setup and shutdown to ensure stop runs when desired (#23304)
Gradle's finalizedBy on tasks only ensures one task runs after another,
but not immediately after. This is problematic for our integration tests
since it allows multiple project's integ test clusters to be
simultaneously. While this has not been a problem thus far (gradle 2.13
happened to keep the finalizedBy tasks close enough that no clusters
were running in parallel), with gradle 3.3 the task graph generation has
changed, and numerous clusters may be running simultaneously, causing
memory pressure, and thus generally slower tests, or even failure if the
system has a limited amount of memory (eg in a vagrant host).
This commit reworks how integ tests are configured. It adds an
`integTestCluster` extension to gradle which is equivalent to the current
`integTest.cluster` and moves the rest test runner task to
`integTestRunner`. The `integTest` task is then just a dummy task,
which depends on the cluster runner task, as well as the cluster stop
task. This means running `integTest` in one project will both run the
rest tests, and shut down the cluster, before running `integTest` in
another project.
2017-02-22 15:43:15 -05:00
|
|
|
integTestCluster {
|
|
|
|
plugin subproj.path
|
2016-05-13 16:15:51 -04:00
|
|
|
}
|
2016-05-05 16:46:40 -04:00
|
|
|
}
|
2016-04-29 10:42:03 -04:00
|
|
|
}
|
|
|
|
|
2016-05-05 16:46:40 -04:00
|
|
|
buildRestTests.docs = fileTree(projectDir) {
|
|
|
|
// No snippets in here!
|
|
|
|
exclude 'build.gradle'
|
|
|
|
// That is where the snippets go, not where they come from!
|
|
|
|
exclude 'build'
|
2016-04-29 10:42:03 -04:00
|
|
|
}
|
|
|
|
|
2016-05-05 16:46:40 -04:00
|
|
|
Closure setupTwitter = { String name, int count ->
|
|
|
|
buildRestTests.setups[name] = '''
|
2016-07-22 18:51:36 -04:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: twitter
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
tweet:
|
|
|
|
properties:
|
|
|
|
user:
|
|
|
|
type: keyword
|
|
|
|
doc_values: true
|
|
|
|
date:
|
|
|
|
type: date
|
|
|
|
likes:
|
|
|
|
type: long
|
2016-04-29 10:42:03 -04:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: twitter
|
|
|
|
type: tweet
|
|
|
|
refresh: true
|
|
|
|
body: |'''
|
2016-04-29 10:42:03 -04:00
|
|
|
for (int i = 0; i < count; i++) {
|
2016-05-10 18:29:56 -04:00
|
|
|
String user, text
|
|
|
|
if (i == 0) {
|
|
|
|
user = 'kimchy'
|
|
|
|
text = 'trying out Elasticsearch'
|
|
|
|
} else {
|
|
|
|
user = 'test'
|
|
|
|
text = "some message with the number $i"
|
|
|
|
}
|
2016-05-05 16:46:40 -04:00
|
|
|
buildRestTests.setups[name] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{"_id": "$i"}}
|
|
|
|
{"user": "$user", "message": "$text", "date": "2009-11-15T14:12:12", "likes": $i}"""
|
2016-04-29 10:42:03 -04:00
|
|
|
}
|
|
|
|
}
|
2016-05-05 16:46:40 -04:00
|
|
|
setupTwitter('twitter', 5)
|
|
|
|
setupTwitter('big_twitter', 120)
|
2016-09-01 17:08:18 -04:00
|
|
|
setupTwitter('huge_twitter', 1200)
|
2016-05-19 12:40:22 -04:00
|
|
|
|
|
|
|
buildRestTests.setups['host'] = '''
|
|
|
|
# Fetch the http host. We use the host of the master because we know there will always be a master.
|
|
|
|
- do:
|
|
|
|
cluster.state: {}
|
|
|
|
- set: { master_node: master }
|
|
|
|
- do:
|
|
|
|
nodes.info:
|
|
|
|
metric: [ http ]
|
|
|
|
- is_true: nodes.$master.http.publish_address
|
|
|
|
- set: {nodes.$master.http.publish_address: host}
|
|
|
|
'''
|
2016-08-12 18:42:19 -04:00
|
|
|
|
2017-06-02 03:46:38 -04:00
|
|
|
buildRestTests.setups['node'] = '''
|
|
|
|
# Fetch the node name. We use the host of the master because we know there will always be a master.
|
|
|
|
- do:
|
|
|
|
cluster.state: {}
|
|
|
|
- is_true: master_node
|
|
|
|
- set: { master_node: node_name }
|
|
|
|
'''
|
|
|
|
|
2016-11-15 11:45:54 -05:00
|
|
|
// Used by scripted metric docs
|
|
|
|
buildRestTests.setups['ledger'] = '''
|
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: ledger
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 2
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
sale:
|
|
|
|
properties:
|
|
|
|
type:
|
|
|
|
type: keyword
|
|
|
|
amount:
|
|
|
|
type: double
|
2016-11-15 11:45:54 -05:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: ledger
|
2017-04-27 02:43:20 -04:00
|
|
|
type: sale
|
2017-03-02 06:43:20 -05:00
|
|
|
refresh: true
|
|
|
|
body: |
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "amount": 200, "type": "sale", "description": "something"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "amount": 10, "type": "expense", "decription": "another thing"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "amount": 150, "type": "sale", "description": "blah"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "amount": 50, "type": "expense", "description": "cost of blah"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "amount": 50, "type": "expense", "description": "advertisement"}'''
|
2016-11-15 11:45:54 -05:00
|
|
|
|
2017-02-07 14:17:54 -05:00
|
|
|
// Used by aggregation docs
|
2016-08-12 18:42:19 -04:00
|
|
|
buildRestTests.setups['sales'] = '''
|
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: sales
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 2
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
sale:
|
|
|
|
properties:
|
|
|
|
type:
|
|
|
|
type: keyword
|
2016-08-12 18:42:19 -04:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: sales
|
|
|
|
type: sale
|
|
|
|
refresh: true
|
|
|
|
body: |
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "price": 200, "promoted": true, "rating": 1, "type": "hat"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "price": 200, "promoted": true, "rating": 1, "type": "t-shirt"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/01/01 00:00:00", "price": 150, "promoted": true, "rating": 5, "type": "bag"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/02/01 00:00:00", "price": 50, "promoted": false, "rating": 1, "type": "hat"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/02/01 00:00:00", "price": 10, "promoted": true, "rating": 4, "type": "t-shirt"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/03/01 00:00:00", "price": 200, "promoted": true, "rating": 1, "type": "hat"}
|
|
|
|
{"index":{}}
|
|
|
|
{"date": "2015/03/01 00:00:00", "price": 175, "promoted": false, "rating": 2, "type": "t-shirt"}'''
|
2016-10-02 23:16:21 -04:00
|
|
|
|
|
|
|
// Dummy bank account data used by getting-started.asciidoc
|
|
|
|
buildRestTests.setups['bank'] = '''
|
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: bank
|
|
|
|
type: account
|
|
|
|
refresh: true
|
|
|
|
body: |
|
2016-10-02 23:16:21 -04:00
|
|
|
#bank_data#
|
|
|
|
'''
|
|
|
|
/* Load the actual accounts only if we're going to use them. This complicates
|
|
|
|
* dependency checking but that is a small price to pay for not building a
|
|
|
|
* 400kb string every time we start the build. */
|
|
|
|
File accountsFile = new File("$projectDir/src/test/resources/accounts.json")
|
|
|
|
buildRestTests.inputs.file(accountsFile)
|
|
|
|
buildRestTests.doFirst {
|
|
|
|
String accounts = accountsFile.getText('UTF-8')
|
|
|
|
// Indent like a yaml test needs
|
2017-03-02 06:43:20 -05:00
|
|
|
accounts = accounts.replaceAll('(?m)^', ' ')
|
2016-10-02 23:16:21 -04:00
|
|
|
buildRestTests.setups['bank'] =
|
|
|
|
buildRestTests.setups['bank'].replace('#bank_data#', accounts)
|
|
|
|
}
|
Add RangeFieldMapper for numeric and date range types
Lucene 6.2 added index and query support for numeric ranges. This commit adds a new RangeFieldMapper for indexing numeric (int, long, float, double) and date ranges and creating appropriate range and term queries. The design is similar to NumericFieldMapper in that it uses a RangeType enumerator for implementing the logic specific to each type. The following range types are supported by this field mapper: int_range, float_range, long_range, double_range, date_range.
Lucene does not provide a DocValue field specific to RangeField types so the RangeFieldMapper implements a CustomRangeDocValuesField for handling doc value support.
When executing a Range query over a Range field, the RangeQueryBuilder has been enhanced to accept a new relation parameter for defining the type of query as one of: WITHIN, CONTAINS, INTERSECTS. This provides support for finding all ranges that are related to a specific range in a desired way. As with other spatial queries, DISJOINT can be achieved as a MUST_NOT of an INTERSECTS query.
2016-09-16 09:50:56 -04:00
|
|
|
|
|
|
|
buildRestTests.setups['range_index'] = '''
|
|
|
|
- do :
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: range_index
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 2
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
my_type:
|
|
|
|
properties:
|
|
|
|
expected_attendees:
|
|
|
|
type: integer_range
|
|
|
|
time_frame:
|
|
|
|
type: date_range
|
|
|
|
format: yyyy-MM-dd HH:mm:ss||yyyy-MM-dd||epoch_millis
|
Add RangeFieldMapper for numeric and date range types
Lucene 6.2 added index and query support for numeric ranges. This commit adds a new RangeFieldMapper for indexing numeric (int, long, float, double) and date ranges and creating appropriate range and term queries. The design is similar to NumericFieldMapper in that it uses a RangeType enumerator for implementing the logic specific to each type. The following range types are supported by this field mapper: int_range, float_range, long_range, double_range, date_range.
Lucene does not provide a DocValue field specific to RangeField types so the RangeFieldMapper implements a CustomRangeDocValuesField for handling doc value support.
When executing a Range query over a Range field, the RangeQueryBuilder has been enhanced to accept a new relation parameter for defining the type of query as one of: WITHIN, CONTAINS, INTERSECTS. This provides support for finding all ranges that are related to a specific range in a desired way. As with other spatial queries, DISJOINT can be achieved as a MUST_NOT of an INTERSECTS query.
2016-09-16 09:50:56 -04:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: range_index
|
|
|
|
type: my_type
|
|
|
|
refresh: true
|
|
|
|
body: |
|
|
|
|
{"index":{"_id": 1}}
|
|
|
|
{"expected_attendees": {"gte": 10, "lte": 20}, "time_frame": {"gte": "2015-10-31 12:00:00", "lte": "2015-11-01"}}'''
|
2016-11-07 03:20:06 -05:00
|
|
|
|
|
|
|
// Used by index boost doc
|
|
|
|
buildRestTests.setups['index_boost'] = '''
|
|
|
|
- do:
|
|
|
|
indices.create:
|
|
|
|
index: index1
|
|
|
|
- do:
|
|
|
|
indices.create:
|
|
|
|
index: index2
|
|
|
|
|
|
|
|
- do:
|
|
|
|
indices.put_alias:
|
|
|
|
index: index1
|
|
|
|
name: alias1
|
|
|
|
'''
|
2017-01-31 04:45:25 -05:00
|
|
|
// Used by sampler and diversified-sampler aggregation docs
|
|
|
|
buildRestTests.setups['stackoverflow'] = '''
|
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: stackoverflow
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
question:
|
|
|
|
properties:
|
|
|
|
author:
|
|
|
|
type: keyword
|
|
|
|
tags:
|
|
|
|
type: keyword
|
2017-01-31 04:45:25 -05:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: stackoverflow
|
|
|
|
type: question
|
|
|
|
refresh: true
|
|
|
|
body: |'''
|
2017-02-07 13:33:00 -05:00
|
|
|
|
2017-01-31 04:45:25 -05:00
|
|
|
// Make Kibana strongly connected to elasticsearch and logstash
|
|
|
|
// Make Kibana rarer (and therefore higher-ranking) than Javascript
|
|
|
|
// Make Javascript strongly connected to jquery and angular
|
|
|
|
// Make Cabana strongly connected to elasticsearch but only as a result of a single author
|
|
|
|
|
|
|
|
for (int i = 0; i < 150; i++) {
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{}}
|
|
|
|
{"author": "very_relevant_$i", "tags": ["elasticsearch", "kibana"]}"""
|
2017-01-31 04:45:25 -05:00
|
|
|
}
|
|
|
|
for (int i = 0; i < 50; i++) {
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{}}
|
|
|
|
{"author": "very_relevant_$i", "tags": ["logstash", "kibana"]}"""
|
2017-01-31 04:45:25 -05:00
|
|
|
}
|
|
|
|
for (int i = 0; i < 200; i++) {
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{}}
|
|
|
|
{"author": "partially_relevant_$i", "tags": ["javascript", "jquery"]}"""
|
2017-01-31 04:45:25 -05:00
|
|
|
}
|
|
|
|
for (int i = 0; i < 200; i++) {
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{}}
|
|
|
|
{"author": "partially_relevant_$i", "tags": ["javascript", "angular"]}"""
|
2017-01-31 04:45:25 -05:00
|
|
|
}
|
|
|
|
for (int i = 0; i < 50; i++) {
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
2017-03-02 06:43:20 -05:00
|
|
|
{"index":{}}
|
|
|
|
{"author": "noisy author", "tags": ["elasticsearch", "cabana"]}"""
|
2017-01-31 04:45:25 -05:00
|
|
|
}
|
|
|
|
buildRestTests.setups['stackoverflow'] += """
|
|
|
|
"""
|
2017-05-24 08:46:43 -04:00
|
|
|
// Used by significant_text aggregation docs
|
|
|
|
buildRestTests.setups['news'] = '''
|
|
|
|
- do:
|
|
|
|
indices.create:
|
|
|
|
index: news
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
question:
|
|
|
|
properties:
|
|
|
|
source:
|
|
|
|
type: keyword
|
|
|
|
content:
|
|
|
|
type: text
|
|
|
|
- do:
|
|
|
|
bulk:
|
|
|
|
index: news
|
|
|
|
type: article
|
|
|
|
refresh: true
|
|
|
|
body: |'''
|
|
|
|
|
|
|
|
// Make h5n1 strongly connected to bird flu
|
|
|
|
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"source": "very_relevant_$i", "content": "bird flu h5n1"}"""
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"source": "filler_$i", "content": "bird dupFiller "}"""
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"source": "filler_$i", "content": "flu dupFiller "}"""
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 20; i++) {
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"source": "partially_relevant_$i", "content": "elasticsearch dupFiller dupFiller dupFiller dupFiller pozmantier"}"""
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"source": "partially_relevant_$i", "content": "elasticsearch logstash kibana"}"""
|
|
|
|
}
|
|
|
|
buildRestTests.setups['news'] += """
|
|
|
|
"""
|
2017-02-07 15:59:40 -05:00
|
|
|
|
2017-02-07 13:33:00 -05:00
|
|
|
// Used by some aggregations
|
|
|
|
buildRestTests.setups['exams'] = '''
|
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
indices.create:
|
|
|
|
index: exams
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
exam:
|
|
|
|
properties:
|
|
|
|
grade:
|
|
|
|
type: byte
|
2017-02-07 13:33:00 -05:00
|
|
|
- do:
|
2017-03-02 06:43:20 -05:00
|
|
|
bulk:
|
|
|
|
index: exams
|
|
|
|
type: exam
|
|
|
|
refresh: true
|
|
|
|
body: |
|
|
|
|
{"index":{}}
|
|
|
|
{"grade": 100}
|
|
|
|
{"index":{}}
|
|
|
|
{"grade": 50}'''
|
2017-05-17 17:42:25 -04:00
|
|
|
|
|
|
|
buildRestTests.setups['stored_example_script'] = '''
|
|
|
|
# Simple script to load a field. Not really a good example, but a simple one.
|
|
|
|
- do:
|
|
|
|
put_script:
|
|
|
|
id: "my_script"
|
2017-06-09 11:29:25 -04:00
|
|
|
body: { "script": { "lang": "painless", "source": "doc[params.field].value" } }
|
2017-05-17 17:42:25 -04:00
|
|
|
- match: { acknowledged: true }
|
|
|
|
'''
|
|
|
|
|
|
|
|
buildRestTests.setups['stored_scripted_metric_script'] = '''
|
|
|
|
- do:
|
|
|
|
put_script:
|
|
|
|
id: "my_init_script"
|
2017-06-09 11:29:25 -04:00
|
|
|
body: { "script": { "lang": "painless", "source": "params._agg.transactions = []" } }
|
2017-05-17 17:42:25 -04:00
|
|
|
- match: { acknowledged: true }
|
2017-07-05 06:30:19 -04:00
|
|
|
|
2017-05-17 17:42:25 -04:00
|
|
|
- do:
|
|
|
|
put_script:
|
|
|
|
id: "my_map_script"
|
2017-06-09 11:29:25 -04:00
|
|
|
body: { "script": { "lang": "painless", "source": "params._agg.transactions.add(doc.type.value == 'sale' ? doc.amount.value : -1 * doc.amount.value)" } }
|
2017-05-17 17:42:25 -04:00
|
|
|
- match: { acknowledged: true }
|
|
|
|
|
|
|
|
- do:
|
|
|
|
put_script:
|
|
|
|
id: "my_combine_script"
|
2017-06-09 11:29:25 -04:00
|
|
|
body: { "script": { "lang": "painless", "source": "double profit = 0;for (t in params._agg.transactions) { profit += t; } return profit" } }
|
2017-05-17 17:42:25 -04:00
|
|
|
- match: { acknowledged: true }
|
|
|
|
|
|
|
|
- do:
|
|
|
|
put_script:
|
|
|
|
id: "my_reduce_script"
|
2017-06-09 11:29:25 -04:00
|
|
|
body: { "script": { "lang": "painless", "source": "double profit = 0;for (a in params._aggs) { profit += a; } return profit" } }
|
2017-05-17 17:42:25 -04:00
|
|
|
- match: { acknowledged: true }
|
|
|
|
'''
|
2017-07-04 06:16:56 -04:00
|
|
|
|
|
|
|
// Used by analyze api
|
|
|
|
buildRestTests.setups['analyze_sample'] = '''
|
|
|
|
- do:
|
|
|
|
indices.create:
|
|
|
|
index: analyze_sample
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 0
|
|
|
|
analysis:
|
|
|
|
normalizer:
|
|
|
|
my_normalizer:
|
|
|
|
type: custom
|
|
|
|
filter: [lowercase]
|
|
|
|
mappings:
|
|
|
|
tweet:
|
|
|
|
properties:
|
|
|
|
obj1.field1:
|
|
|
|
type: text'''
|
2017-08-02 17:47:27 -04:00
|
|
|
|
|
|
|
// Used by percentile/percentile-rank aggregations
|
|
|
|
buildRestTests.setups['latency'] = '''
|
|
|
|
- do:
|
|
|
|
indices.create:
|
|
|
|
index: latency
|
|
|
|
body:
|
|
|
|
settings:
|
|
|
|
number_of_shards: 1
|
|
|
|
number_of_replicas: 1
|
|
|
|
mappings:
|
|
|
|
data:
|
|
|
|
properties:
|
|
|
|
load_time:
|
|
|
|
type: long
|
|
|
|
- do:
|
|
|
|
bulk:
|
|
|
|
index: latency
|
|
|
|
type: data
|
|
|
|
refresh: true
|
|
|
|
body: |'''
|
|
|
|
|
|
|
|
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
|
|
def value = i
|
|
|
|
if (i % 10) {
|
|
|
|
value = i*10
|
|
|
|
}
|
|
|
|
buildRestTests.setups['latency'] += """
|
|
|
|
{"index":{}}
|
|
|
|
{"load_time": "$value"}"""
|
|
|
|
}
|