2013-11-08 19:20:43 -05:00
[[modules-snapshots]]
== Snapshot And Restore
2017-09-27 21:32:40 -04:00
You can store snapshots of individual indices or an entire cluster in
a remote repository like a shared file system, S3, or HDFS. These snapshots
are great for backups because they can be restored relatively quickly. However,
snapshots can only be restored to versions of Elasticsearch that can read the
indices:
2016-10-13 10:49:32 -04:00
2017-09-27 21:32:40 -04:00
* A snapshot of an index created in 5.x can be restored to 6.x.
2016-10-13 10:49:32 -04:00
* A snapshot of an index created in 2.x can be restored to 5.x.
* A snapshot of an index created in 1.x can be restored to 2.x.
2017-03-30 13:08:41 -04:00
2017-09-27 21:32:40 -04:00
Conversely, snapshots of indices created in 1.x **cannot** be restored to
5.x or 6.x, and snapshots of indices created in 2.x **cannot** be restored
to 6.x.
Snapshots are incremental and can contain indices created in various
versions of Elasticsearch. If any indices in a snapshot were created in an
incompatible version, you will not be able restore the snapshot.
IMPORTANT: When backing up your data prior to an upgrade, keep in mind that you
won't be able to restore snapshots after you upgrade if they contain indices
created in a version that's incompatible with the upgrade version.
If you end up in a situation where you need to restore a snapshot of an index
that is incompatible with the version of the cluster you are currently running,
you can restore it on the latest compatible version and use
<<reindex-from-remote,reindex-from-remote>> to rebuild the index on the current
version. Reindexing from remote is only possible if the original index has
source enabled. Retrieving and reindexing the data can take significantly longer
than simply restoring a snapshot. If you have a large amount of data, we
recommend testing the reindex from remote process with a subset of your data to
understand the time requirements before proceeding.
2017-03-30 13:08:41 -04:00
2013-11-08 19:20:43 -05:00
[float]
=== Repositories
2017-09-27 21:32:40 -04:00
You must register a snapshot repository before you can perform snapshot and
restore operations. We recommend creating a new snapshot repository for each
major version. The valid repository settings depend on the repository type.
If you register same snapshot repository with multiple clusters, only
one cluster should have write access to the repository. All other clusters
connected to that repository should set the repository to `readonly` mode.
NOTE: The snapshot format can change across major versions, so if you have
clusters on different major versions trying to write the same repository,
new snapshots written by one version will not be visible to the other. While
setting the repository to `readonly` on all but one of the clusters should work
with multiple clusters differing by one major version, it is not a supported
configuration.
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
PUT /_snapshot/my_backup
{
"type": "fs",
"settings": {
2017-02-15 18:13:27 -05:00
"location": "my_backup_location"
2015-06-19 12:04:52 -04:00
}
}
2013-11-08 19:20:43 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
// TESTSETUP
2013-11-08 19:20:43 -05:00
2017-09-27 21:32:40 -04:00
To retrieve information about a registered repository, use a GET request:
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2015-06-19 12:04:52 -04:00
which returns:
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
{
2015-06-19 12:04:52 -04:00
"my_backup": {
"type": "fs",
"settings": {
2017-02-15 18:13:27 -05:00
"location": "my_backup_location"
2013-11-08 19:20:43 -05:00
}
}
}
-----------------------------------
2017-02-15 18:13:27 -05:00
// TESTRESPONSE
2013-11-08 19:20:43 -05:00
2017-09-27 21:32:40 -04:00
To retrieve information about multiple repositories, specify a
a comma-delimited list of repositories. You can also use the * wildcard when
specifying repository names. For example, the following request retrieves
information about all of the snapshot repositories that start with `repo` or
contain `backup`:
2015-12-01 09:41:30 -05:00
[source,js]
-----------------------------------
GET /_snapshot/repo*,*backup*
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
2015-12-01 09:41:30 -05:00
2017-09-27 21:32:40 -04:00
To retrieve information about all registered snapshot repositories, omit the
repository name or specify `_all`:
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot
2013-11-08 19:20:43 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
2013-11-08 19:20:43 -05:00
or
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/_all
2013-11-08 19:20:43 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
2013-11-08 19:20:43 -05:00
[float]
===== Shared File System Repository
2015-05-13 20:03:50 -04:00
The shared file system repository (`"type": "fs"`) uses the shared file system to store snapshots. In order to register
the shared file system repository it is necessary to mount the same shared filesystem to the same location on all
2016-01-26 11:54:41 -05:00
master and data nodes. This location (or one of its parent directories) must be registered in the `path.repo`
2015-05-13 20:03:50 -04:00
setting on all master and data nodes.
Assuming that the shared filesystem is mounted to `/mount/backups/my_backup`, the following setting should be added to
`elasticsearch.yml` file:
[source,yaml]
--------------
path.repo: ["/mount/backups", "/mount/longterm_backups"]
--------------
2015-08-12 12:33:07 -04:00
The `path.repo` setting supports Microsoft Windows UNC paths as long as at least server name and share are specified as
a prefix and back slashes are properly escaped:
[source,yaml]
--------------
2015-08-19 21:55:32 -04:00
path.repo: ["\\\\MY_SERVER\\Snapshots"]
2015-08-12 12:33:07 -04:00
--------------
2015-05-13 20:03:50 -04:00
After all nodes are restarted, the following command can be used to register the shared file system repository with
the name `my_backup`:
[source,js]
-----------------------------------
2017-02-15 18:13:27 -05:00
PUT /_snapshot/my_fs_backup
{
2015-05-13 20:03:50 -04:00
"type": "fs",
"settings": {
2017-02-15 18:13:27 -05:00
"location": "/mount/backups/my_fs_backup_location",
2015-05-13 20:03:50 -04:00
"compress": true
}
2017-02-15 18:13:27 -05:00
}
2015-05-13 20:03:50 -04:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
// TEST[skip:no access to absolute path]
2015-05-13 20:03:50 -04:00
If the repository location is specified as a relative path this path will be resolved against the first path specified
in `path.repo`:
[source,js]
-----------------------------------
2017-02-15 18:13:27 -05:00
PUT /_snapshot/my_fs_backup
{
2015-05-13 20:03:50 -04:00
"type": "fs",
"settings": {
2017-02-15 18:13:27 -05:00
"location": "my_fs_backup_location",
2015-05-13 20:03:50 -04:00
"compress": true
}
2017-02-15 18:13:27 -05:00
}
2015-05-13 20:03:50 -04:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
// TEST[continued]
2015-05-13 20:03:50 -04:00
The following settings are supported:
2013-11-08 19:20:43 -05:00
[horizontal]
`location`:: Location of the snapshots. Mandatory.
2014-09-09 00:17:38 -04:00
`compress`:: Turns on compression of the snapshot files. Compression is applied only to metadata files (index mapping and settings). Data files are not compressed. Defaults to `true`.
2013-11-08 19:20:43 -05:00
`chunk_size`:: Big files can be broken down into chunks during snapshotting if needed. The chunk size can be specified in bytes or by
using size value notation, i.e. 1g, 10m, 5k. Defaults to `null` (unlimited chunk size).
2015-03-20 16:09:42 -04:00
`max_restore_bytes_per_sec`:: Throttles per node restore rate. Defaults to `40mb` per second.
`max_snapshot_bytes_per_sec`:: Throttles per node snapshot rate. Defaults to `40mb` per second.
2015-10-07 07:27:36 -04:00
`readonly`:: Makes repository read-only. Defaults to `false`.
2013-11-08 19:20:43 -05:00
2014-01-15 12:57:05 -05:00
[float]
===== Read-only URL Repository
2014-12-17 06:22:17 -05:00
The URL repository (`"type": "url"`) can be used as an alternative read-only way to access data created by the shared file
2015-07-14 18:37:52 -04:00
system repository. The URL specified in the `url` parameter should point to the root of the shared filesystem repository.
The following settings are supported:
2014-01-15 12:57:05 -05:00
[horizontal]
`url`:: Location of the snapshots. Mandatory.
2015-07-14 18:37:52 -04:00
URL Repository supports the following protocols: "http", "https", "ftp", "file" and "jar". URL repositories with `http:`,
`https:`, and `ftp:` URLs has to be whitelisted by specifying allowed URLs in the `repositories.url.allowed_urls` setting.
This setting supports wildcards in the place of host, path, query, and fragment. For example:
[source,yaml]
-----------------------------------
repositories.url.allowed_urls: ["http://www.example.org/root/*", "https://*.mydomain.com/*?*#*"]
-----------------------------------
2015-10-26 16:49:05 -04:00
URL repositories with `file:` URLs can only point to locations registered in the `path.repo` setting similar to
2015-07-14 18:37:52 -04:00
shared file system repository.
2014-02-05 11:43:14 -05:00
[float]
===== Repository plugins
Other repository backends are available in these official plugins:
2015-09-03 06:11:58 -04:00
* {plugins}/repository-s3.html[repository-s3] for S3 repository support
2016-01-11 03:56:17 -05:00
* {plugins}/repository-hdfs.html[repository-hdfs] for HDFS repository support in Hadoop environments
2015-09-03 13:12:52 -04:00
* {plugins}/repository-azure.html[repository-azure] for Azure storage repositories
2015-09-15 11:35:10 -04:00
* {plugins}/repository-gcs.html[repository-gcs] for Google Cloud Storage repositories
2014-02-05 11:43:14 -05:00
2014-10-07 12:33:21 -04:00
[float]
2015-01-29 14:57:12 -05:00
===== Repository Verification
2015-01-29 06:15:20 -05:00
When a repository is registered, it's immediately verified on all master and data nodes to make sure that it is functional
2015-04-01 04:18:29 -04:00
on all nodes currently present in the cluster. The `verify` parameter can be used to explicitly disable the repository
verification when registering or updating a repository:
[source,js]
-----------------------------------
2017-02-15 18:13:27 -05:00
PUT /_snapshot/my_unverified_backup?verify=false
2015-06-19 12:04:52 -04:00
{
2017-02-15 18:13:27 -05:00
"type": "fs",
2015-06-19 12:04:52 -04:00
"settings": {
2017-02-15 18:13:27 -05:00
"location": "my_unverified_backup_location"
2015-06-19 12:04:52 -04:00
}
}
2015-04-01 04:18:29 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2015-04-01 04:18:29 -04:00
The verification process can also be executed manually by running the following command:
2014-09-10 21:49:15 -04:00
[source,js]
-----------------------------------
2017-02-15 18:13:27 -05:00
POST /_snapshot/my_unverified_backup/_verify
2014-09-10 21:49:15 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-09-10 21:49:15 -04:00
2014-10-07 12:33:21 -04:00
It returns a list of nodes where repository was successfully verified or an error message if verification process failed.
2014-09-10 21:49:15 -04:00
2013-11-08 19:20:43 -05:00
[float]
=== Snapshot
2015-10-16 11:53:45 -04:00
A repository can contain multiple snapshots of the same cluster. Snapshots are identified by unique names within the
2013-11-08 19:20:43 -05:00
cluster. A snapshot with the name `snapshot_1` in the repository `my_backup` can be created by executing the following
command:
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
PUT /_snapshot/my_backup/snapshot_1?wait_for_completion=true
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
2014-10-08 08:43:06 -04:00
The `wait_for_completion` parameter specifies whether or not the request should return immediately after snapshot
initialization (default) or wait for snapshot completion. During snapshot initialization, information about all
previous snapshots is loaded into the memory, which means that in large repositories it may take several seconds (or
even minutes) for this command to return even if the `wait_for_completion` parameter is set to `false`.
2015-02-12 05:00:21 -05:00
By default a snapshot of all open and started indices in the cluster is created. This behavior can be changed by
2014-10-08 08:43:06 -04:00
specifying the list of indices in the body of the snapshot request.
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
2017-02-15 18:13:27 -05:00
PUT /_snapshot/my_backup/snapshot_2?wait_for_completion=true
2015-06-19 12:04:52 -04:00
{
"indices": "index_1,index_2",
2016-08-12 09:34:43 -04:00
"ignore_unavailable": true,
2015-06-19 12:04:52 -04:00
"include_global_state": false
}
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
The list of indices that should be included into the snapshot can be specified using the `indices` parameter that
supports <<search-multi-index-type,multi index syntax>>. The snapshot request also supports the
2014-04-03 19:54:37 -04:00
`ignore_unavailable` option. Setting it to `true` will cause indices that do not exist to be ignored during snapshot
2013-12-11 18:30:12 -05:00
creation. By default, when `ignore_unavailable` option is not set and an index is missing the snapshot request will fail.
2013-11-08 19:20:43 -05:00
By setting `include_global_state` to false it's possible to prevent the cluster global state to be stored as part of
2015-02-12 05:00:21 -05:00
the snapshot. By default, the entire snapshot will fail if one or more indices participating in the snapshot don't have
2014-01-13 11:22:13 -05:00
all primary shards available. This behaviour can be changed by setting `partial` to `true`.
2013-11-08 19:20:43 -05:00
The index snapshot process is incremental. In the process of making the index snapshot Elasticsearch analyses
the list of the index files that are already stored in the repository and copies only files that were created or
changed since the last snapshot. That allows multiple snapshots to be preserved in the repository in a compact form.
Snapshotting process is executed in non-blocking fashion. All indexing and searching operation can continue to be
executed against the index that is being snapshotted. However, a snapshot represents the point-in-time view of the index
2015-02-12 05:00:21 -05:00
at the moment when snapshot was created, so no records that were added to the index after the snapshot process was started
2014-05-12 17:19:10 -04:00
will be present in the snapshot. The snapshot process starts immediately for the primary shards that has been started
2014-07-17 15:25:34 -04:00
and are not relocating at the moment. Before version 1.2.0, the snapshot operation fails if the cluster has any relocating or
2014-05-12 17:19:10 -04:00
initializing primaries of indices participating in the snapshot. Starting with version 1.2.0, Elasticsearch waits for
2014-07-17 15:25:34 -04:00
relocation or initialization of shards to complete before snapshotting them.
2013-11-08 19:20:43 -05:00
Besides creating a copy of each index the snapshot process can also store global cluster metadata, which includes persistent
cluster settings and templates. The transient settings and registered snapshot repositories are not stored as part of
the snapshot.
Only one snapshot process can be executed in the cluster at any time. While snapshot of a particular shard is being
created this shard cannot be moved to another node, which can interfere with rebalancing process and allocation
2015-02-12 05:00:21 -05:00
filtering. Elasticsearch will only be able to move a shard to another node (according to the current allocation
filtering settings and rebalancing algorithm) once the snapshot is finished.
2013-11-08 19:20:43 -05:00
Once a snapshot is created information about this snapshot can be obtained using the following command:
2015-07-14 12:14:09 -04:00
[source,sh]
2013-11-08 19:20:43 -05:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/snapshot_1
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
2017-03-23 15:20:38 -04:00
This command returns basic information about the snapshot including start and end time, version of
elasticsearch that created the snapshot, the list of included indices, the current state of the
snapshot and the list of failures that occurred during the snapshot. The snapshot `state` can be
[horizontal]
`IN_PROGRESS`::
The snapshot is currently running.
`SUCCESS`::
The snapshot finished and all shards were stored successfully.
`FAILED`::
The snapshot finished with an error and failed to store any data.
`PARTIAL`::
The global cluster state was stored, but data of at least one shard wasn't stored successfully.
The `failure` section in this case should contain more detailed information about shards
that were not processed correctly.
`INCOMPATIBLE`::
The snapshot was created with an old version of elasticsearch and therefore is incompatible with
the current version of the cluster.
2015-12-01 09:41:30 -05:00
Similar as for repositories, information about multiple snapshots can be queried in one go, supporting wildcards as well:
[source,sh]
-----------------------------------
GET /_snapshot/my_backup/snapshot_*,some_other_snapshot
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2015-12-01 09:41:30 -05:00
2013-11-08 19:20:43 -05:00
All snapshots currently stored in the repository can be listed using the following command:
2015-07-14 12:14:09 -04:00
[source,sh]
2013-11-08 19:20:43 -05:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/_all
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
2016-10-10 16:51:47 -04:00
The command fails if some of the snapshots are unavailable. The boolean parameter `ignore_unavailable` can be used to
2015-11-18 01:36:30 -05:00
return all snapshots that are currently available.
Enhances get snapshots API to allow retrieving repository index only (#24477)
Currently, the get snapshots API (e.g. /_snapshot/{repositoryName}/_all)
provides information about snapshots in the repository, including the
snapshot state, number of shards snapshotted, failures, etc. In order
to provide information about each snapshot in the repository, the call
must read the snapshot metadata blob (`snap-{snapshot_uuid}.dat`) for
every snapshot. In cloud-based repositories, this can be expensive,
both from a cost and performance perspective. Sometimes, all the user
wants is to retrieve all the names/uuids of each snapshot, and the
indices that went into each snapshot, without any of the other status
information about the snapshot. This minimal information can be
retrieved from the repository index blob (`index-N`) without needing to
read each snapshot metadata blob.
This commit enhances the get snapshots API with an optional `verbose`
parameter. If `verbose` is set to false on the request, then the get
snapshots API will only retrieve the minimal information about each
snapshot (the name, uuid, and indices in the snapshot), and only read
this information from the repository index blob, thereby giving users
the option to retrieve the snapshots in a repository in a more
cost-effective and efficient manner.
Closes #24288
2017-05-10 15:48:40 -04:00
Getting all snapshots in the repository can be costly on cloud-based repositories,
both from a cost and performance perspective. If the only information required is
the snapshot names/uuids in the repository and the indices in each snapshot, then
the optional boolean parameter `verbose` can be set to `false` to execute a more
performant and cost-effective retrieval of the snapshots in the repository. Note
that setting `verbose` to `false` will omit all other information about the snapshot
such as status information, the number of snapshotted shards, etc. The default
value of the `verbose` parameter is `true`.
2015-10-07 07:27:36 -04:00
A currently running snapshot can be retrieved using the following command:
2015-01-23 15:31:29 -05:00
2015-07-14 12:14:09 -04:00
[source,sh]
2015-01-23 15:31:29 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
GET /_snapshot/my_backup/_current
2015-01-23 15:31:29 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
// CONSOLE
// TEST[continued]
2015-01-23 15:31:29 -05:00
2013-11-08 19:20:43 -05:00
A snapshot can be deleted from the repository using the following command:
2015-07-14 12:14:09 -04:00
[source,sh]
2013-11-08 19:20:43 -05:00
-----------------------------------
2017-02-15 18:13:27 -05:00
DELETE /_snapshot/my_backup/snapshot_2
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
When a snapshot is deleted from a repository, Elasticsearch deletes all files that are associated with the deleted
snapshot and not used by any other snapshots. If the deleted snapshot operation is executed while the snapshot is being
created the snapshotting process will be aborted and all files created as part of the snapshotting process will be
cleaned. Therefore, the delete snapshot operation can be used to cancel long running snapshot operations that were
started by mistake.
2014-10-20 21:04:34 -04:00
A repository can be deleted using the following command:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-10-20 21:04:34 -04:00
-----------------------------------
2017-02-15 18:13:27 -05:00
DELETE /_snapshot/my_fs_backup
2014-10-20 21:04:34 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-10-20 21:04:34 -04:00
When a repository is deleted, Elasticsearch only removes the reference to the location where the repository is storing
the snapshots. The snapshots themselves are left untouched and in place.
2013-11-08 19:20:43 -05:00
[float]
=== Restore
2014-03-15 18:58:18 -04:00
A snapshot can be restored using the following command:
2013-11-08 19:20:43 -05:00
2015-07-14 12:14:09 -04:00
[source,sh]
2013-11-08 19:20:43 -05:00
-----------------------------------
2015-06-19 12:04:52 -04:00
POST /_snapshot/my_backup/snapshot_1/_restore
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
2016-06-07 17:06:20 -04:00
By default, all indices in the snapshot are restored, and the cluster state is
*not* restored. It's possible to select indices that should be restored as well
as to allow the global cluster state from being restored by using `indices` and
`include_global_state` options in the restore request body. The list of indices
supports <<search-multi-index-type,multi index syntax>>. The `rename_pattern`
2016-06-13 14:04:20 -04:00
and `rename_replacement` options can be also used to rename indices on restore
2016-06-07 17:06:20 -04:00
using regular expression that supports referencing the original text as
explained
2013-11-08 19:20:43 -05:00
http://docs.oracle.com/javase/6/docs/api/java/util/regex/Matcher.html#appendReplacement(java.lang.StringBuffer,%20java.lang.String)[here].
2016-06-07 17:06:20 -04:00
Set `include_aliases` to `false` to prevent aliases from being restored together
with associated indices
2013-11-08 19:20:43 -05:00
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
POST /_snapshot/my_backup/snapshot_1/_restore
{
"indices": "index_1,index_2",
2016-08-12 09:34:43 -04:00
"ignore_unavailable": true,
2016-06-07 17:06:20 -04:00
"include_global_state": true,
2015-06-19 12:04:52 -04:00
"rename_pattern": "index_(.+)",
"rename_replacement": "restored_index_$1"
}
2013-11-08 19:20:43 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2013-11-08 19:20:43 -05:00
2016-06-07 17:06:20 -04:00
The restore operation can be performed on a functioning cluster. However, an
existing index can be only restored if it's <<indices-open-close,closed>> and
has the same number of shards as the index in the snapshot. The restore
operation automatically opens restored indices if they were closed and creates
new indices if they didn't exist in the cluster. If cluster state is restored
2016-12-07 08:12:10 -05:00
with `include_global_state` (defaults to `false`), the restored templates that
2016-06-07 17:06:20 -04:00
don't currently exist in the cluster are added and existing templates with the
same name are replaced by the restored templates. The restored persistent
settings are added to the existing persistent settings.
2014-03-15 18:58:18 -04:00
2014-06-01 15:13:13 -04:00
[float]
2015-05-12 20:59:24 -04:00
==== Partial restore
2014-06-01 15:13:13 -04:00
2015-02-12 05:00:21 -05:00
By default, the entire restore operation will fail if one or more indices participating in the operation don't have
2014-06-01 15:13:13 -04:00
snapshots of all shards available. It can occur if some shards failed to snapshot for example. It is still possible to
restore such indices by setting `partial` to `true`. Please note, that only successfully snapshotted shards will be
restored in this case and all missing shards will be recreated empty.
2014-03-15 18:58:18 -04:00
2015-01-13 19:16:37 -05:00
[float]
2015-05-12 20:59:24 -04:00
==== Changing index settings during restore
2015-01-13 19:16:37 -05:00
Most of index settings can be overridden during the restore process. For example, the following command will restore
the index `index_1` without creating any replicas while switching back to default refresh interval:
[source,js]
-----------------------------------
2015-06-19 12:04:52 -04:00
POST /_snapshot/my_backup/snapshot_1/_restore
{
"indices": "index_1",
"index_settings": {
"index.number_of_replicas": 0
},
"ignore_index_settings": [
"index.refresh_interval"
]
}
2015-01-13 19:16:37 -05:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2015-01-13 19:16:37 -05:00
2015-05-12 20:59:24 -04:00
Please note, that some settings such as `index.number_of_shards` cannot be changed during restore operation.
[float]
==== Restoring to a different cluster
The information stored in a snapshot is not tied to a particular cluster or a cluster name. Therefore it's possible to
restore a snapshot made from one cluster into another cluster. All that is required is registering the repository
containing the snapshot in the new cluster and starting the restore process. The new cluster doesn't have to have the
2016-08-09 21:22:29 -04:00
same size or topology. However, the version of the new cluster should be the same or newer (only 1 major version newer) than the cluster that was used to create the snapshot. For example, you can restore a 1.x snapshot to a 2.x cluster, but not a 1.x snapshot to a 5.x cluster.
2015-05-12 20:59:24 -04:00
If the new cluster has a smaller size additional considerations should be made. First of all it's necessary to make sure
that new cluster have enough capacity to store all indices in the snapshot. It's possible to change indices settings
during restore to reduce the number of replicas, which can help with restoring snapshots into smaller cluster. It's also
2017-10-31 20:04:00 -04:00
possible to select only subset of the indices using the `indices` parameter.
2015-05-12 20:59:24 -04:00
If indices in the original cluster were assigned to particular nodes using
<<shard-allocation-filtering,shard allocation filtering>>, the same rules will be enforced in the new cluster. Therefore
if the new cluster doesn't contain nodes with appropriate attributes that a restored index can be allocated on, such
index will not be successfully restored unless these index allocation settings are changed during restore operation.
2015-01-13 19:16:37 -05:00
2017-10-31 20:04:00 -04:00
The restore operation also checks that restored persistent settings are compatible with the current cluster to avoid accidentally
restoring an incompatible settings such as `discovery.zen.minimum_master_nodes` and as a result disable a smaller cluster until the
required number of master eligible nodes is added. If you need to restore a snapshot with incompatible persistent settings, try
restoring it without the global cluster state.
2014-03-15 18:58:18 -04:00
[float]
=== Snapshot status
A list of currently running snapshots with their detailed status information can be obtained using the following command:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-03-15 18:58:18 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/_status
2014-03-15 18:58:18 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-03-15 18:58:18 -04:00
In this format, the command will return information about all currently running snapshots. By specifying a repository name, it's possible
to limit the results to a particular repository:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-03-15 18:58:18 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/_status
2014-03-15 18:58:18 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-03-15 18:58:18 -04:00
If both repository name and snapshot id are specified, this command will return detailed status information for the given snapshot even
if it's not currently running:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-03-15 18:58:18 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/snapshot_1/_status
2014-03-15 18:58:18 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-03-15 18:58:18 -04:00
Multiple ids are also supported:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-03-15 18:58:18 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/snapshot_1,snapshot_2/_status
2014-03-15 18:58:18 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-06-11 20:52:45 -04:00
[float]
=== Monitoring snapshot/restore progress
There are several ways to monitor the progress of the snapshot and restores processes while they are running. Both
operations support `wait_for_completion` parameter that would block client until the operation is completed. This is
the simplest method that can be used to get notified about operation completion.
The snapshot operation can be also monitored by periodic calls to the snapshot info:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-06-11 20:52:45 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/snapshot_1
2014-06-11 20:52:45 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-06-11 20:52:45 -04:00
2014-12-17 06:22:17 -05:00
Please note that snapshot info operation uses the same resources and thread pool as the snapshot operation. So,
executing a snapshot info operation while large shards are being snapshotted can cause the snapshot info operation to wait
2014-06-11 20:52:45 -04:00
for available resources before returning the result. On very large shards the wait time can be significant.
To get more immediate and complete information about snapshots the snapshot status command can be used instead:
2015-07-14 12:14:09 -04:00
[source,sh]
2014-06-11 20:52:45 -04:00
-----------------------------------
2015-06-19 12:04:52 -04:00
GET /_snapshot/my_backup/snapshot_1/_status
2014-06-11 20:52:45 -04:00
-----------------------------------
2016-05-09 09:42:23 -04:00
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2014-06-11 20:52:45 -04:00
While snapshot info method returns only basic information about the snapshot in progress, the snapshot status returns
complete breakdown of the current state for each shard participating in the snapshot.
The restore process piggybacks on the standard recovery mechanism of the Elasticsearch. As a result, standard recovery
monitoring services can be used to monitor the state of restore. When restore operation is executed the cluster
typically goes into `red` state. It happens because the restore operation starts with "recovering" primary shards of the
restored indices. During this operation the primary shards become unavailable which manifests itself in the `red` cluster
state. Once recovery of primary shards is completed Elasticsearch is switching to standard replication process that
creates the required number of replicas at this moment cluster switches to the `yellow` state. Once all required replicas
are created, the cluster switches to the `green` states.
2017-10-31 20:04:00 -04:00
The cluster health operation provides only a high level status of the restore process. It's possible to get more
2014-06-11 20:52:45 -04:00
detailed insight into the current state of the recovery process by using <<indices-recovery, indices recovery>> and
<<cat-recovery, cat recovery>> APIs.
[float]
=== Stopping currently running snapshot and restore operations
2015-02-12 05:00:21 -05:00
The snapshot and restore framework allows running only one snapshot or one restore operation at a time. If a currently
running snapshot was executed by mistake, or takes unusually long, it can be terminated using the snapshot delete operation.
The snapshot delete operation checks if the deleted snapshot is currently running and if it does, the delete operation stops
that snapshot before deleting the snapshot data from the repository.
2014-06-11 20:52:45 -04:00
2016-08-25 06:59:56 -04:00
[source,sh]
-----------------------------------
DELETE /_snapshot/my_backup/snapshot_1
-----------------------------------
// CONSOLE
2017-02-15 18:13:27 -05:00
// TEST[continued]
2016-08-25 06:59:56 -04:00
2014-12-17 06:22:17 -05:00
The restore operation uses the standard shard recovery mechanism. Therefore, any currently running restore operation can
2014-06-11 20:52:45 -04:00
be canceled by deleting indices that are being restored. Please note that data for all deleted indices will be removed
from the cluster as a result of this operation.
2015-06-10 12:57:45 -04:00
[float]
=== Effect of cluster blocks on snapshot and restore operations
Many snapshot and restore operations are affected by cluster and index blocks. For example, registering and unregistering
repositories require write global metadata access. The snapshot operation requires that all indices and their metadata as
well as the global metadata were readable. The restore operation requires the global metadata to be writable, however
the index level blocks are ignored during restore because indices are essentially recreated during restore.
Please note that a repository content is not part of the cluster and therefore cluster blocks don't affect internal
repository operations such as listing or deleting snapshots from an already registered repository.