🔎 Open source distributed and RESTful search engine.
Go to file
Dan Hermann 85886e71c2
Handle error conditions when simulating ingest pipelines with verbosity enabled (#63327) (#63484)
2020-10-08 09:21:05 -05:00
.ci Bump version after 7.9.2 release 2020-09-24 10:48:57 -05:00
.github
.idea [7.x] Mirror privileges over data streams to their backing indices (#58991) 2020-07-03 06:33:38 -05:00
benchmarks Determine shard size before allocating shards recovering from snapshots (#61906) (#63337) 2020-10-06 18:37:05 +02:00
buildSrc Upgrade to lucene-8.7.0-snapshot-5c4168d (#63466) 2020-10-08 08:20:43 -04:00
client [ML] Validate that AucRoc has the data necessary to be calculated (#63302) (#63454) 2020-10-08 09:52:15 +02:00
dev-tools Remove the last Perl scripts (#57767) 2020-06-09 10:12:34 +01:00
distribution Cleanup on integtest distribution setup (7.x backport) (#63189) 2020-10-06 15:58:42 +02:00
docs [DOCS] allow_duplicates option for append processor (#62336) (#63486) 2020-10-08 09:01:43 -05:00
gradle Switch bundled jdk back to Oracle JDK (#63288) (#63290) 2020-10-05 16:31:10 -07:00
libs [7.x] Add assert that raw and readable xcontent field names are different (#63332) (#63343) 2020-10-06 11:32:41 -06:00
licenses
modules Handle error conditions when simulating ingest pipelines with verbosity enabled (#63327) (#63484) 2020-10-08 09:21:05 -05:00
plugins Upgrade to lucene-8.7.0-snapshot-5c4168d (#63466) 2020-10-08 08:20:43 -04:00
qa Muting FullClusterRestartIT.testSystemIndexMetadataIsUpgraded 2020-10-08 12:02:23 +02:00
rest-api-spec Convert TextFieldMapper to parametrized form (#63269) (#63392) 2020-10-07 13:26:25 +01:00
server Handle error conditions when simulating ingest pipelines with verbosity enabled (#63327) (#63484) 2020-10-08 09:21:05 -05:00
test Deprecate week_year in favour of weekyear date format backport(63307) (#63308) 2020-10-07 09:16:27 +02:00
x-pack [ML] fix grabbing the doc value limit setting in _explain (#63402) (#63471) 2020-10-08 08:53:26 -04:00
.dir-locals.el
.editorconfig Remove default indent from .editorconfig (#49183) 2019-11-18 08:05:53 +00:00
.gitattributes
.gitignore Fix nasty errors when importing into IntelliJ 2020-03-23 21:32:37 -07:00
CONTRIBUTING.md [Docs] Fixed a couple of typos in CONTRIBUTING (#63205) 2020-10-05 16:25:57 +03:00
LICENSE.txt
NOTICE.txt
README.asciidoc [DOCS] Replace `twitter` dataset in API conventions + README (#60408) (#60410) 2020-07-29 14:14:01 -04:00
TESTING.asciidoc [7.xUpdate TESTING.asciidoc for recent REST test changes (#62841) (#62895) 2020-10-01 11:02:29 -05:00
Vagrantfile Cleanup vagrant setup (#60697) 2020-08-06 15:37:56 -07:00
build.gradle Upgrade spotless dependency (#62857) 2020-09-24 12:52:45 +01:00
gradle.properties Enforce fail on deprecated gradle usage (7.x backport) (#59758) 2020-07-20 08:52:30 +02:00
gradlew Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
gradlew.bat Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
settings.gradle Update gradle enterprise plugin to 3.4.1 (#62165) (#62178) 2020-09-09 18:30:34 +02:00

README.asciidoc

= Elasticsearch

== A Distributed RESTful Search Engine

=== https://www.elastic.co/products/elasticsearch[https://www.elastic.co/products/elasticsearch]

Elasticsearch is a distributed RESTful search engine built for the cloud. Features include:

* Distributed and Highly Available Search Engine.
** Each index is fully sharded with a configurable number of shards.
** Each shard can have one or more replicas.
** Read / Search operations performed on any of the replica shards.
* Multi Tenant.
** Support for more than one index.
** Index level configuration (number of shards, index storage, ...).
* Various set of APIs
** HTTP RESTful API
** All APIs perform automatic node operation rerouting.
* Document oriented
** No need for upfront schema definition.
** Schema can be defined for customization of the indexing process.
* Reliable, Asynchronous Write Behind for long term persistency.
* (Near) Real Time Search.
* Built on top of Apache Lucene
** Each shard is a fully functional Lucene index
** All the power of Lucene easily exposed through simple configuration / plugins.
* Per operation consistency
** Single document level operations are atomic, consistent, isolated and durable.

== Getting Started

First of all, DON'T PANIC. It will take 5 minutes to get the gist of what Elasticsearch is all about.

=== Installation

* https://www.elastic.co/downloads/elasticsearch[Download] and unpack the Elasticsearch official distribution.
* Run `bin/elasticsearch` on Linux or macOS. Run `bin\elasticsearch.bat` on Windows.
* Run `curl -X GET http://localhost:9200/` to verify Elasticsearch is running.

=== Indexing

First, index some sample JSON documents. The first request automatically creates
the `my-index-000001` index.

----
curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T14:12:12",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T01:46:38",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'
----

=== Search

Next, use a search request to find any documents with a `user.id` of `kimchy`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?q=user.id:kimchy&pretty=true'
----

Instead of a query string, you can use Elasticsearch's
https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl.html[Query
DSL] in the request body.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match" : { "user.id": "kimchy" }
  }
}'
----

You can also retrieve all documents in `my-index-000001`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

During indexing, Elasticsearch automatically mapped the `@timestamp` field as a
date. This lets you run a range search.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "range" : {
      "@timestamp": {
        "from": "2099-11-15T13:00:00",
        "to": "2099-11-15T14:00:00"
      }
    }
  }
}'
----

=== Multiple indices

Elasticsearch supports multiple indices. The previous examples used an index
called `my-index-000001`. You can create another index, `my-index-000002`, to
store additional data when `my-index-000001` reaches a certain age or size. You
can also use separate indices to store different types of data.

You can configure each index differently. The following request
creates `my-index-000002` with two primary shards rather than the default of
one. This may be helpful for larger indices.

----
curl -X PUT 'http://localhost:9200/my-index-000002?pretty' -H 'Content-Type: application/json' -d '
{
  "settings" : {
    "index.number_of_shards" : 2
  }
}'
----

You can then add a document to `my-index-000002`.

----
curl -X POST 'http://localhost:9200/my-index-000002/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-16T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'
----

You can search and perform other operations on multiple indices with a single
request. The following request searches `my-index-000001` and `my-index-000002`.

----
curl -X GET 'http://localhost:9200/my-index-000001,my-index-000002/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

You can omit the index from the request path to search all indices.

----
curl -X GET 'http://localhost:9200/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

=== Distributed, highly available

Let's face it, things will fail....

Elasticsearch is a highly available and distributed search engine. Each index is broken down into shards, and each shard can have one or more replicas. By default, an index is created with 1 shard and 1 replica per shard (1/1). There are many topologies that can be used, including 1/10 (improve search performance), or 20/1 (improve indexing performance, with search executed in a map reduce fashion across shards).

In order to play with the distributed nature of Elasticsearch, simply bring more nodes up and shut down nodes. The system will continue to serve requests (make sure you use the correct http port) with the latest data indexed.

=== Where to go from here?

We have just covered a very small portion of what Elasticsearch is all about. For more information, please refer to the https://www.elastic.co/products/elasticsearch[elastic.co] website. General questions can be asked on the https://discuss.elastic.co[Elastic Forum] or https://ela.st/slack[on Slack]. The Elasticsearch GitHub repository is reserved for bug reports and feature requests only.

=== Building from source

Elasticsearch uses https://gradle.org[Gradle] for its build system.

In order to create a distribution, simply run the `./gradlew assemble` command in the cloned directory.

The distribution for each project will be created under the `build/distributions` directory in that project.

See the xref:TESTING.asciidoc[TESTING] for more information about running the Elasticsearch test suite.

=== Upgrading from older Elasticsearch versions

In order to ensure a smooth upgrade process from earlier versions of Elasticsearch, please see our https://www.elastic.co/guide/en/elasticsearch/reference/current/setup-upgrade.html[upgrade documentation] for more details on the upgrade process.