🔎 Open source distributed and RESTful search engine.
Go to file
Nick Knize 3769b2c6a4 [PURIFY] remove all trace of x-pack rollups (#17)
This commit removes all trace of Elastic licensed rollups

Signed-off-by: Peter Nied <petern@amazon.com>
2021-03-13 10:36:08 -06:00
.ci Cleanup build-scan, remove publish scan to elastic server (#1) (#4) 2021-03-13 10:36:06 -06:00
.idea [7.x] Mirror privileges over data streams to their backing indices (#58991) 2020-07-03 06:33:38 -05:00
benchmarks Determine shard size before allocating shards recovering from snapshots (#61906) (#63337) 2020-10-06 18:37:05 +02:00
buildSrc [PURIFY] remove all trace of x-pack rollups (#17) 2021-03-13 10:36:08 -06:00
client [PURIFY] remove all trace of x-pack rollups (#17) 2021-03-13 10:36:08 -06:00
dev-tools Remove the last Perl scripts (#57767) 2020-06-09 10:12:34 +01:00
distribution [PURIFY] remove all trace of x-pack security (#16) 2021-03-13 10:36:08 -06:00
gradle [PURIFY] remove all trace of x-pack rollups (#17) 2021-03-13 10:36:08 -06:00
libs Fix #invariant Assertion in CacheFile (#64180) (#64264) 2020-10-28 10:22:47 +01:00
licenses Remove the Elastic license file, all checks for this license and the license REST APIs. (#12) 2021-03-13 10:36:07 -06:00
modules [PURIFY] remove all trace of x-pack eql (#5) 2021-03-13 10:36:06 -06:00
plugins Cleanup build script to exclude security-authorization-engine (#8) (#8) 2021-03-13 10:36:06 -06:00
qa [PURIFY] remove all trace of x-pack security (#16) 2021-03-13 10:36:08 -06:00
rest-api-spec Fix the skip version for inner hits REST test. 2021-01-06 11:46:17 -08:00
server [PURIFY] remove all trace of x-pack rollups (#17) 2021-03-13 10:36:08 -06:00
test [PURIFY] remove all trace of x-pack rollups (#17) 2021-03-13 10:36:08 -06:00
.dir-locals.el
.editorconfig Remove default indent from .editorconfig (#49183) 2019-11-18 08:05:53 +00:00
.gitattributes
.gitignore Fix nasty errors when importing into IntelliJ 2020-03-23 21:32:37 -07:00
CONTRIBUTING.md [DOCS] Add Elastic Contributor Program (#64440) (#64527) 2020-11-03 09:34:02 -05:00
LICENSE.txt Clarify mixed license text (#45637) 2019-08-16 13:39:12 -04:00
NOTICE.txt Restore date aggregation performance in UTC case (#38221) (#38700) 2019-02-11 16:30:48 +03:00
README.asciidoc [DOCS] Fix grammar and style in README (#67317) (#67365) 2021-01-12 13:51:52 -05:00
TESTING.asciidoc Add option to preserve data in test clusters (#65400) 2020-11-24 11:56:56 -08:00
Vagrantfile Remove opensuse 42 from vagrant tests (#63759) 2020-10-15 13:05:46 -07:00
build.gradle Remove the Elastic license file, all checks for this license and the license REST APIs. (#12) 2021-03-13 10:36:07 -06:00
gradle.properties Explicitly use TLS 1.2 in Gradle (#63880) 2020-10-19 08:57:26 -07:00
gradlew Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
gradlew.bat Update gradle wrapper to 6.6 (#59909) (#60949) 2020-08-11 11:03:19 +02:00
settings.gradle Always use "elasticsearch" as Gradle project name (#65709) 2020-12-02 09:29:36 -08:00

README.asciidoc

= Elasticsearch

== A Distributed RESTful Search Engine

=== https://www.elastic.co/products/elasticsearch[https://www.elastic.co/products/elasticsearch]

Elasticsearch is a distributed RESTful search engine built for the cloud. Features include:

* Distributed and Highly Available Search Engine.
** Each index is fully sharded with a configurable number of shards.
** Each shard can have one or more replicas.
** Read / Search operations performed on any of the replica shards.
* Multi-tenant.
** Support for more than one index.
** Index level configuration (number of shards, index storage, etc.).
* Various set of APIs
** HTTP RESTful API
** All APIs perform automatic node operation rerouting.
* Document oriented
** No need for upfront schema definition.
** Schema can be defined for customization of the indexing process.
* Reliable, Asynchronous Write Behind for long term persistency.
* Near real-time search.
* Built on top of Apache Lucene
** Each shard is a fully functional Lucene index
** All the power of Lucene easily exposed through simple configuration and plugins.
* Per operation consistency
** Single document-level operations are atomic, consistent, isolated, and durable.

== Getting Started

First of all, DON'T PANIC. It will take 5 minutes to get the gist of what Elasticsearch is all about.

=== Installation

* https://www.elastic.co/downloads/elasticsearch[Download] and unpack the Elasticsearch official distribution.
* Run `bin/elasticsearch` on Linux or macOS. Run `bin\elasticsearch.bat` on Windows.
* Run `curl -X GET http://localhost:9200/` to verify Elasticsearch is running.

=== Indexing

First, index some sample JSON documents. The first request automatically creates
the `my-index-000001` index.

----
curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T14:12:12",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'

curl -X POST 'http://localhost:9200/my-index-000001/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-15T01:46:38",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "elkbee"
  }
}'
----

=== Search

Next, use a search request to find any documents with a `user.id` of `kimchy`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?q=user.id:kimchy&pretty=true'
----

Instead of a query string, you can use Elasticsearch's
https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl.html[Query
DSL] in the request body.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match" : { "user.id": "kimchy" }
  }
}'
----

You can also retrieve all documents in `my-index-000001`.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

During indexing, Elasticsearch automatically mapped the `@timestamp` field as a
date. This lets you run a range search.

----
curl -X GET 'http://localhost:9200/my-index-000001/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "range" : {
      "@timestamp": {
        "from": "2099-11-15T13:00:00",
        "to": "2099-11-15T14:00:00"
      }
    }
  }
}'
----

=== Multiple indices

Elasticsearch supports multiple indices. The previous examples used an index
called `my-index-000001`. You can create another index, `my-index-000002`, to
store additional data when `my-index-000001` reaches a certain age or size. You
can also use separate indices to store different types of data.

You can configure each index differently. The following request
creates `my-index-000002` with two primary shards rather than the default of
one. This may be helpful for larger indices.

----
curl -X PUT 'http://localhost:9200/my-index-000002?pretty' -H 'Content-Type: application/json' -d '
{
  "settings" : {
    "index.number_of_shards" : 2
  }
}'
----

You can then add a document to `my-index-000002`.

----
curl -X POST 'http://localhost:9200/my-index-000002/_doc?pretty' -H 'Content-Type: application/json' -d '
{
  "@timestamp": "2099-11-16T13:12:00",
  "message": "GET /search HTTP/1.1 200 1070000",
  "user": {
    "id": "kimchy"
  }
}'
----

You can search and perform other operations on multiple indices with a single
request. The following request searches `my-index-000001` and `my-index-000002`.

----
curl -X GET 'http://localhost:9200/my-index-000001,my-index-000002/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

You can omit the index from the request path to search all indices.

----
curl -X GET 'http://localhost:9200/_search?pretty=true' -H 'Content-Type: application/json' -d '
{
  "query" : {
    "match_all" : {}
  }
}'
----

=== Distributed, highly available

Let's face it; things will fail...

Elasticsearch is a highly available and distributed search engine. Each index is broken down into shards, and each shard can have one or more replicas. By default, an index is created with 1 shard and 1 replica per shard (1/1). Many topologies can be used, including 1/10 (improve search performance) or 20/1 (improve indexing performance, with search executed in a MapReduce fashion across shards).

To play with the distributed nature of Elasticsearch, bring more nodes up and shut down nodes. The system will continue to serve requests (ensure you use the correct HTTP port) with the latest data indexed.

=== Where to go from here?

We have just covered a tiny portion of what Elasticsearch is all about. For more information, please refer to the https://www.elastic.co/products/elasticsearch[elastic.co] website. General questions can be asked on the https://discuss.elastic.co[Elastic Forum] or https://ela.st/slack[on Slack]. The Elasticsearch GitHub repository is reserved for bug reports and feature requests only.

=== Building from source

Elasticsearch uses https://gradle.org[Gradle] for its build system.

To build a distribution for your local OS and print its output location upon 
completion, run:
----
./gradlew localDistro 
----

To build a distribution for another platform, run the related command:
----
./gradlew :distribution:archives:linux-tar:assemble
./gradlew :distribution:archives:darwin-tar:assemble
./gradlew :distribution:archives:windows-zip:assemble
----

To build distributions for all supported platforms, run:
----
./gradlew assemble
----

Finished distributions are output to `distributions/archives`.

See the xref:TESTING.asciidoc[TESTING] for more information about running the Elasticsearch test suite.

=== Upgrading from older Elasticsearch versions

To ensure a smooth upgrade process from earlier versions of Elasticsearch, please see our https://www.elastic.co/guide/en/elasticsearch/reference/current/setup-upgrade.html[upgrade documentation] for more details on the upgrade process.