kolchfa-aws 91caf0cc09
Adds search backpressure documentation (#1790)
* Adds search backpressure documentation

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Updated response

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Removed last cancelled task

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Fixed typo

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Incorporated tech review feedback

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Incorporated doc review feedback

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Update _opensearch/search-backpressure.md

Co-authored-by: Nate Bower <nbower@amazon.com>

* Apply suggestions from code review

Co-authored-by: Nate Bower <nbower@amazon.com>

* Rewording

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

* Minor rewording for clarity

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>

Signed-off-by: Fanit Kolchina <kolchfa@amazon.com>
Co-authored-by: Nate Bower <nbower@amazon.com>
2022-11-10 11:17:53 -05:00

1.3 KiB

layout title nav_order has_children redirect_from
default Segment replication 64 true
/opensearch/segment-replication/

Segment replication

Segment replication is an experimental feature with OpenSearch 2.3. Therefore, we do not recommend the use of segment replication in a production environment. For updates on the progress of segment replication or if you want leave feedback that could help improve the feature, see the Segment replication git issue. {: .warning}

With segment replication, segment files are copied across shards instead of documents being indexed on each shard copy. This improves indexing throughput and lowers resource utilization at the expense of increased network utilization.

As an experimental feature, segment replication will be behind a feature flag and must be enabled on each node of a cluster and pass a new setting during index creation. {: .note }

Potential use cases

  • Users who have high write loads but do not have high search requirements and are comfortable with longer refresh times.
  • Users with very high loads who want to add new nodes, as you do not need to index all nodes when adding a new node to the cluster.

This is the first step in a series of features designed to decouple reads and writes in order to lower compute costs.