2016-06-21 10:31:44 -04:00
|
|
|
[[tune-for-search-speed]]
|
|
|
|
== Tune for search speed
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Give memory to the filesystem cache
|
|
|
|
|
|
|
|
Elasticsearch heavily relies on the filesystem cache in order to make search
|
|
|
|
fast. In general, you should make sure that at least half the available memory
|
2017-11-29 03:44:25 -05:00
|
|
|
goes to the filesystem cache so that Elasticsearch can keep hot regions of the
|
2016-06-21 10:31:44 -04:00
|
|
|
index in physical memory.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Use faster hardware
|
|
|
|
|
|
|
|
If your search is I/O bound, you should investigate giving more memory to the
|
|
|
|
filesystem cache (see above) or buying faster drives. In particular SSD drives
|
|
|
|
are known to perform better than spinning disks. Always use local storage,
|
|
|
|
remote filesystems such as `NFS` or `SMB` should be avoided. Also beware of
|
|
|
|
virtualized storage such as Amazon's `Elastic Block Storage`. Virtualized
|
|
|
|
storage works very well with Elasticsearch, and it is appealing since it is so
|
|
|
|
fast and simple to set up, but it is also unfortunately inherently slower on an
|
|
|
|
ongoing basis when compared to dedicated local storage. If you put an index on
|
|
|
|
`EBS`, be sure to use provisioned IOPS otherwise operations could be quickly
|
|
|
|
throttled.
|
|
|
|
|
|
|
|
If your search is CPU-bound, you should investigate buying faster CPUs.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Document modeling
|
|
|
|
|
|
|
|
Documents should be modeled so that search-time operations are as cheap as possible.
|
|
|
|
|
|
|
|
In particular, joins should be avoided. <<nested,`nested`>> can make queries
|
|
|
|
several times slower and <<mapping-parent-field,parent-child>> relations can make
|
|
|
|
queries hundreds of times slower. So if the same questions can be answered without
|
|
|
|
joins by denormalizing documents, significant speedups can be expected.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Pre-index data
|
|
|
|
|
|
|
|
You should leverage patterns in your queries to optimize the way data is indexed.
|
|
|
|
For instance, if all your documents have a `price` field and most queries run
|
|
|
|
<<search-aggregations-bucket-range-aggregation,`range`>> aggregations on a fixed
|
|
|
|
list of ranges, you could make this aggregation faster by pre-indexing the ranges
|
|
|
|
into the index and using a <<search-aggregations-bucket-terms-aggregation,`terms`>>
|
|
|
|
aggregations.
|
|
|
|
|
|
|
|
For instance, if documents look like:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-12-14 11:47:53 -05:00
|
|
|
PUT index/_doc/1
|
2016-06-21 10:31:44 -04:00
|
|
|
{
|
|
|
|
"designation": "spoon",
|
|
|
|
"price": 13
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
|
|
|
|
and search requests look like:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET index/_search
|
|
|
|
{
|
|
|
|
"aggs": {
|
|
|
|
"price_ranges": {
|
|
|
|
"range": {
|
|
|
|
"field": "price",
|
|
|
|
"ranges": [
|
|
|
|
{ "to": 10 },
|
|
|
|
{ "from": 10, "to": 100 },
|
|
|
|
{ "from": 100 }
|
|
|
|
]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
// TEST[continued]
|
|
|
|
|
|
|
|
Then documents could be enriched by a `price_range` field at index time, which
|
|
|
|
should be mapped as a <<keyword,`keyword`>>:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
PUT index
|
|
|
|
{
|
|
|
|
"mappings": {
|
2017-12-14 11:47:53 -05:00
|
|
|
"_doc": {
|
2016-06-21 10:31:44 -04:00
|
|
|
"properties": {
|
|
|
|
"price_range": {
|
|
|
|
"type": "keyword"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-12-14 11:47:53 -05:00
|
|
|
PUT index/_doc/1
|
2016-06-21 10:31:44 -04:00
|
|
|
{
|
|
|
|
"designation": "spoon",
|
|
|
|
"price": 13,
|
|
|
|
"price_range": "10-100"
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
|
|
|
|
And then search requests could aggregate this new field rather than running a
|
|
|
|
`range` aggregation on the `price` field.
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET index/_search
|
|
|
|
{
|
|
|
|
"aggs": {
|
|
|
|
"price_ranges": {
|
|
|
|
"terms": {
|
|
|
|
"field": "price_range"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
// TEST[continued]
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Mappings
|
|
|
|
|
|
|
|
The fact that some data is numeric does not mean it should always be mapped as a
|
|
|
|
<<number,numeric field>>. Typically, fields storing identifiers such as an `ISBN`
|
|
|
|
or any number identifying a record from another database, might benefit from
|
|
|
|
being mapped as <<keyword,`keyword`>> rather than `integer` or `long`.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Avoid scripts
|
|
|
|
|
|
|
|
In general, scripts should be avoided. If they are absolutely needed, you
|
|
|
|
should prefer the `painless` and `expressions` engines.
|
|
|
|
|
2016-09-02 07:30:36 -04:00
|
|
|
[float]
|
|
|
|
=== Search rounded dates
|
|
|
|
|
|
|
|
Queries on date fields that use `now` are typically not cacheable since the
|
|
|
|
range that is being matched changes all the time. However switching to a
|
|
|
|
rounded date is often acceptable in terms of user experience, and has the
|
|
|
|
benefit of making better use of the query cache.
|
|
|
|
|
|
|
|
For instance the below query:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-12-14 11:47:53 -05:00
|
|
|
PUT index/_doc/1
|
2016-09-02 07:30:36 -04:00
|
|
|
{
|
|
|
|
"my_date": "2016-05-11T16:30:55.328Z"
|
|
|
|
}
|
|
|
|
|
|
|
|
GET index/_search
|
|
|
|
{
|
|
|
|
"query": {
|
|
|
|
"constant_score": {
|
|
|
|
"filter": {
|
|
|
|
"range": {
|
|
|
|
"my_date": {
|
|
|
|
"gte": "now-1h",
|
|
|
|
"lte": "now"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
|
|
|
|
could be replaced with the following query:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET index/_search
|
|
|
|
{
|
|
|
|
"query": {
|
|
|
|
"constant_score": {
|
|
|
|
"filter": {
|
|
|
|
"range": {
|
|
|
|
"my_date": {
|
|
|
|
"gte": "now-1h/m",
|
|
|
|
"lte": "now/m"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
// TEST[continued]
|
|
|
|
|
|
|
|
In that case we rounded to the minute, so if the current time is `16:31:29`,
|
|
|
|
the range query will match everything whose value of the `my_date` field is
|
|
|
|
between `15:31:00` and `16:31:59`. And if several users run a query that
|
|
|
|
contains this range in the same minute, the query cache could help speed things
|
|
|
|
up a bit. The longer the interval that is used for rounding, the more the query
|
|
|
|
cache can help, but beware that too aggressive rounding might also hurt user
|
|
|
|
experience.
|
|
|
|
|
|
|
|
|
|
|
|
NOTE: It might be tempting to split ranges into a large cacheable part and
|
|
|
|
smaller not cacheable parts in order to be able to leverage the query cache,
|
|
|
|
as shown below:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET index/_search
|
|
|
|
{
|
|
|
|
"query": {
|
|
|
|
"constant_score": {
|
|
|
|
"filter": {
|
|
|
|
"bool": {
|
|
|
|
"should": [
|
|
|
|
{
|
|
|
|
"range": {
|
|
|
|
"my_date": {
|
|
|
|
"gte": "now-1h",
|
|
|
|
"lte": "now-1h/m"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
{
|
|
|
|
"range": {
|
|
|
|
"my_date": {
|
|
|
|
"gt": "now-1h/m",
|
|
|
|
"lt": "now/m"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
{
|
|
|
|
"range": {
|
|
|
|
"my_date": {
|
|
|
|
"gte": "now/m",
|
|
|
|
"lte": "now"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
// TEST[continued]
|
|
|
|
|
|
|
|
However such practice might make the query run slower in some cases since the
|
|
|
|
overhead introduced by the `bool` query may defeat the savings from better
|
|
|
|
leveraging the query cache.
|
|
|
|
|
2016-06-21 10:31:44 -04:00
|
|
|
[float]
|
|
|
|
=== Force-merge read-only indices
|
|
|
|
|
|
|
|
Indices that are read-only would benefit from being
|
|
|
|
<<indices-forcemerge,merged down to a single segment>>. This is typically the
|
|
|
|
case with time-based indices: only the index for the current time frame is
|
|
|
|
getting new documents while older indices are read-only.
|
|
|
|
|
|
|
|
IMPORTANT: Don't force-merge indices that are still being written to -- leave
|
|
|
|
merging to the background merge process.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Warm up global ordinals
|
|
|
|
|
|
|
|
Global ordinals are a data-structure that is used in order to run
|
|
|
|
<<search-aggregations-bucket-terms-aggregation,`terms`>> aggregations on
|
|
|
|
<<keyword,`keyword`>> fields. They are loaded lazily in memory because
|
2017-11-29 03:44:25 -05:00
|
|
|
Elasticsearch does not know which fields will be used in `terms` aggregations
|
|
|
|
and which fields won't. You can tell Elasticsearch to load global ordinals
|
2016-06-21 10:31:44 -04:00
|
|
|
eagerly at refresh-time by configuring mappings as described below:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
PUT index
|
|
|
|
{
|
|
|
|
"mappings": {
|
2017-12-14 11:47:53 -05:00
|
|
|
"_doc": {
|
2016-06-21 10:31:44 -04:00
|
|
|
"properties": {
|
|
|
|
"foo": {
|
|
|
|
"type": "keyword",
|
|
|
|
"eager_global_ordinals": true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Warm up the filesystem cache
|
|
|
|
|
2017-11-29 03:44:25 -05:00
|
|
|
If the machine running Elasticsearch is restarted, the filesystem cache will be
|
2016-06-21 10:31:44 -04:00
|
|
|
empty, so it will take some time before the operating system loads hot regions
|
|
|
|
of the index into memory so that search operations are fast. You can explicitly
|
|
|
|
tell the operating system which files should be loaded into memory eagerly
|
|
|
|
depending on the file extension using the <<file-system,`index.store.preload`>>
|
|
|
|
setting.
|
|
|
|
|
|
|
|
WARNING: Loading data into the filesystem cache eagerly on too many indices or
|
2016-06-27 08:49:37 -04:00
|
|
|
too many files will make search _slower_ if the filesystem cache is not large
|
2016-06-21 10:31:44 -04:00
|
|
|
enough to hold all the data. Use with caution.
|
2017-06-01 11:23:22 -04:00
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Map identifiers as `keyword`
|
|
|
|
|
|
|
|
When you have numeric identifiers in your documents, it is tempting to map them
|
|
|
|
as numbers, which is consistent with their json type. However, the way that
|
|
|
|
Elasticsearch indexes numbers optimizes for `range` queries while `keyword`
|
|
|
|
fields are better at `term` queries. Since identifiers are never used in `range`
|
|
|
|
queries, they should be mapped as a `keyword`.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Use index sorting to speed up conjunctions
|
|
|
|
|
|
|
|
<<index-modules-index-sorting,Index sorting>> can be useful in order to make
|
|
|
|
conjunctions faster at the cost of slightly slower indexing. Read more about it
|
|
|
|
in the <<index-modules-index-sorting-conjunctions,index sorting documentation>>.
|
2017-06-16 05:23:40 -04:00
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Use `preference` to optimize cache utilization
|
|
|
|
|
|
|
|
There are multiple caches that can help with search performance, such as the
|
|
|
|
https://en.wikipedia.org/wiki/Page_cache[filesystem cache], the
|
|
|
|
<<shard-request-cache,request cache>> or the <<query-cache,query cache>>. Yet
|
|
|
|
all these caches are maintained at the node level, meaning that if you run the
|
|
|
|
same request twice in a row, have 1 <<glossary-replica-shard,replica>> or more
|
|
|
|
and use https://en.wikipedia.org/wiki/Round-robin_DNS[round-robin], the default
|
|
|
|
routing algorithm, then those two requests will go to different shard copies,
|
|
|
|
preventing node-level caches from helping.
|
|
|
|
|
|
|
|
Since it is common for users of a search application to run similar requests
|
|
|
|
one after another, for instance in order to analyze a narrower subset of the
|
|
|
|
index, using a preference value that identifies the current user or session
|
|
|
|
could help optimize usage of the caches.
|
|
|
|
|
|
|
|
[float]
|
|
|
|
=== Replicas might help with throughput, but not always
|
|
|
|
|
|
|
|
In addition to improving resiliency, replicas can help improve throughput. For
|
|
|
|
instance if you have a single-shard index and three nodes, you will need to
|
|
|
|
set the number of replicas to 2 in order to have 3 copies of your shard in
|
|
|
|
total so that all nodes are utilized.
|
|
|
|
|
|
|
|
Now imagine that you have a 2-shards index and two nodes. In one case, the
|
|
|
|
number of replicas is 0, meaning that each node holds a single shard. In the
|
|
|
|
second case the number of replicas is 1, meaning that each node has two shards.
|
|
|
|
Which setup is going to perform best in terms of search performance? Usually,
|
|
|
|
the setup that has fewer shards per node in total will perform better. The
|
|
|
|
reason for that is that it gives a greater share of the available filesystem
|
|
|
|
cache to each shard, and the filesystem cache is probably Elasticsearch's
|
|
|
|
number 1 performance factor. At the same time, beware that a setup that does
|
|
|
|
not have replicas is subject to failure in case of a single node failure, so
|
|
|
|
there is a trade-off between throughput and availability.
|
|
|
|
|
|
|
|
So what is the right number of replicas? If you have a cluster that has
|
|
|
|
`num_nodes` nodes, `num_primaries` primary shards _in total_ and if you want to
|
|
|
|
be able to cope with `max_failures` node failures at once at most, then the
|
|
|
|
right number of replicas for you is
|
|
|
|
`max(max_failures, ceil(num_nodes / num_primaries) - 1)`.
|