🔎 Open source distributed and RESTful search engine.
Go to file
Jim Ferenczi 9e4105ec37 Validate PIT on _msearch (#63167)
This change ensures that we validate point in times provided by individual search
requests in _msearch.

Relates #63132
2020-11-05 15:38:28 +01:00
.ci [7.10] Bump version after 7.9.3 release (#63818) 2020-10-22 17:49:21 +02:00
.github
.idea [7.x] Mirror privileges over data streams to their backing indices (#58991) 2020-07-03 06:33:38 -05:00
benchmarks Determine shard size before allocating shards recovering from snapshots (#61906) (#63337) 2020-10-06 18:37:05 +02:00
buildSrc Upgrade to Lucene-8.7.0 (#64532) (#64537) 2020-11-03 16:57:04 +01:00
client Add REST Test for Snapshot Clone API (#63863) (#63881) 2020-10-20 09:48:03 +02:00
dev-tools Remove the last Perl scripts (#57767) 2020-06-09 10:12:34 +01:00
distribution Windows service installation should allow multiple values in ES_JAVA_OPTS (#64254) 2020-10-30 15:13:02 -04:00
docs [DOCS] Fix `remove_index` callout 2020-11-05 09:10:01 -05:00
gradle Switch bundled jdk back to Oracle JDK (#63288) (#63290) 2020-10-05 16:31:10 -07:00
libs Fix #invariant Assertion in CacheFile (#64180) (#64264) 2020-10-28 10:22:47 +01:00
licenses
modules Disable using unsigned_long in scripts (#64552) (#64557) 2020-11-03 16:38:50 -05:00
plugins Upgrade to Lucene-8.7.0 (#64532) (#64537) 2020-11-03 16:57:04 +01:00
qa Windows service installation should allow multiple values in ES_JAVA_OPTS (#64254) 2020-10-30 15:13:02 -04:00
rest-api-spec Make sure non-collecting aggs include sub-aggs (backport of #64214) (#64247) 2020-10-28 08:38:05 -04:00
server Validate PIT on _msearch (#63167) 2020-11-05 15:38:28 +01:00
test Fix merging of terms aggregation with compound order (#64469) 2020-11-05 12:05:19 +01:00
x-pack Validate PIT on _msearch (#63167) 2020-11-05 15:38:28 +01:00
.dir-locals.el
.editorconfig Remove default indent from .editorconfig (#49183) 2019-11-18 08:05:53 +00:00
.gitattributes
.gitignore Fix nasty errors when importing into IntelliJ 2020-03-23 21:32:37 -07:00
CONTRIBUTING.md [DOCS] Add Elastic Contributor Program (#64440) (#64527) 2020-11-03 09:34:02 -05:00
LICENSE.txt Clarify mixed license text (#45637) 2019-08-16 13:39:12 -04:00
NOTICE.txt
README.asciidoc [DOCS] Minor fixes to README.asciidoc 2020-11-02 18:01:12 -05:00
TESTING.asciidoc [DOCS] Fix "the the" typos (#64344) (#64354) 2020-10-29 10:47:42 -04:00
Vagrantfile Remove opensuse 42 from vagrant tests (#63759) 2020-10-15 13:05:46 -07:00
build.gradle Upgrade spotless dependency (#62857) 2020-09-24 12:52:45 +01:00
gradle.properties Explicitly use TLS 1.2 in Gradle (#63880) 2020-10-19 08:57:26 -07:00
gradlew Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
gradlew.bat Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
settings.gradle Upgrade to version 3.5 of Gradle Enterprise plugin (#64451) 2020-11-02 09:22:17 -08:00

README.asciidoc

= Elasticsearch

== A Distributed RESTful Search Engine

=== https://www.elastic.co/products/elasticsearch[https://www.elastic.co/products/elasticsearch]

Elasticsearch is a distributed RESTful search engine built for the cloud. Features include:

* Distributed and Highly Available Search Engine.
** Each index is fully sharded with a configurable number of shards.
** Each shard can have one or more replicas.
** Read / Search operations performed on any of the replica shards.
* Multi Tenant.
** Support for more than one index.
** Index level configuration (number of shards, index storage, ...).
* Various set of APIs
** HTTP RESTful API
** All APIs perform automatic node operation rerouting.
* Document oriented
** No need for upfront schema definition.
** Schema can be defined for customization of the indexing process.
* Reliable, Asynchronous Write Behind for long term persistency.
* (Near) Real Time Search.
* Built on top of Apache Lucene
** Each shard is a fully functional Lucene index
** All the power of Lucene easily exposed through simple configuration / plugins.
* Per operation consistency
** Single document level operations are atomic, consistent, isolated and durable.

== Getting Started

First of all, DON'T PANIC. It will take 5 minutes to get the gist of what Elasticsearch is all about.

=== Installation

* https://www.elastic.co/downloads/elasticsearch[Download] and unpack the Elasticsearch official distribution.
* Run `bin/elasticsearch` on Linux or macOS. Run `bin\elasticsearch.bat` on Windows.
* Run `curl -X GET http://localhost:9200/` to verify Elasticsearch is running.

=== Indexing

First, index some sample JSON documents. The first request automatically creates
the `my-index-000001` index.

----
curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T14:12:12",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T01:46:38",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'
----

=== Search

Next, use a search request to find any documents with a `user.id` of `kimchy`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?q=user.id:kimchy&pretty=true'
----

Instead of a query string, you can use Elasticsearch's
https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl.html[Query
DSL] in the request body.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match" : { "user.id": "kimchy" }
  }
}'
----

You can also retrieve all documents in `my-index-000001`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

During indexing, Elasticsearch automatically mapped the `@timestamp` field as a
date. This lets you run a range search.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "range" : {
      "@timestamp": {
        "from": "2099-11-15T13:00:00",
        "to": "2099-11-15T14:00:00"
      }
    }
  }
}'
----

=== Multiple indices

Elasticsearch supports multiple indices. The previous examples used an index
called `my-index-000001`. You can create another index, `my-index-000002`, to
store additional data when `my-index-000001` reaches a certain age or size. You
can also use separate indices to store different types of data.

You can configure each index differently. The following request
creates `my-index-000002` with two primary shards rather than the default of
one. This may be helpful for larger indices.

----
curl -X PUT 'http://localhost:9200/my-index-000002?pretty' -H 'Content-Type: application/json' -d '
{
  "settings" : {
    "index.number_of_shards" : 2
  }
}'
----

You can then add a document to `my-index-000002`.

----
curl -X POST 'http://localhost:9200/my-index-000002/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-16T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'
----

You can search and perform other operations on multiple indices with a single
request. The following request searches `my-index-000001` and `my-index-000002`.

----
curl -X GET 'http://localhost:9200/my-index-000001,my-index-000002/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

You can omit the index from the request path to search all indices.

----
curl -X GET 'http://localhost:9200/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

=== Distributed, highly available

Let's face it, things will fail....

Elasticsearch is a highly available and distributed search engine. Each index is broken down into shards, and each shard can have one or more replicas. By default, an index is created with 1 shard and 1 replica per shard (1/1). There are many topologies that can be used, including 1/10 (improve search performance), or 20/1 (improve indexing performance, with search executed in a map reduce fashion across shards).

In order to play with the distributed nature of Elasticsearch, simply bring more nodes up and shut down nodes. The system will continue to serve requests (make sure you use the correct http port) with the latest data indexed.

=== Where to go from here?

We have just covered a very small portion of what Elasticsearch is all about. For more information, please refer to the https://www.elastic.co/products/elasticsearch[elastic.co] website. General questions can be asked on the https://discuss.elastic.co[Elastic Forum] or https://ela.st/slack[on Slack]. The Elasticsearch GitHub repository is reserved for bug reports and feature requests only.

=== Building from source

Elasticsearch uses https://gradle.org[Gradle] for its build system.

To build a distribution for your local OS and print its output location upon 
completion, run:
----
./gradlew localDistro 
----

To build a distribution for another platform, run the related command:
----
./gradlew :distribution:archives:linux-tar:assemble
./gradlew :distribution:archives:darwin-tar:assemble
./gradlew :distribution:archives:windows-zip:assemble
----

To build distributions for all supported platforms, run:
----
./gradlew assemble
----

Finished distributions are output to `distributions/archives`.

See the xref:TESTING.asciidoc[TESTING] for more information about running the Elasticsearch test suite.

=== Upgrading from older Elasticsearch versions

In order to ensure a smooth upgrade process from earlier versions of Elasticsearch, please see our https://www.elastic.co/guide/en/elasticsearch/reference/current/setup-upgrade.html[upgrade documentation] for more details on the upgrade process.