diff --git a/docs/plugins/repository-azure.asciidoc b/docs/plugins/repository-azure.asciidoc index e69c9fa8f1b..3743abb92ce 100644 --- a/docs/plugins/repository-azure.asciidoc +++ b/docs/plugins/repository-azure.asciidoc @@ -121,8 +121,8 @@ The Azure repository supports following settings: `chunk_size`:: Big files can be broken down into chunks during snapshotting if needed. - Specify the chunk size as a value and unit, for example: - `10MB`, `5KB`, `500B`. Defaults to `64MB` (64MB max). + The chunk size can be specified in bytes or by using size value notation, + i.e. `1g`, `10m`, `5k`. Defaults to `64m` (64m max) `compress`:: @@ -154,7 +154,7 @@ PUT _snapshot/my_backup2 "settings": { "container": "backup-container", "base_path": "backups", - "chunk_size": "32MB", + "chunk_size": "32m", "compress": true } } diff --git a/docs/plugins/repository-gcs.asciidoc b/docs/plugins/repository-gcs.asciidoc index 695daaffea7..2974fc2ded2 100644 --- a/docs/plugins/repository-gcs.asciidoc +++ b/docs/plugins/repository-gcs.asciidoc @@ -11,8 +11,8 @@ include::install_remove.asciidoc[] ==== Getting started The plugin uses the https://github.com/GoogleCloudPlatform/google-cloud-java/tree/master/google-cloud-clients/google-cloud-storage[Google Cloud Java Client for Storage] -to connect to the Storage service. If you are using -https://cloud.google.com/storage/[Google Cloud Storage] for the first time, you +to connect to the Storage service. If you are using +https://cloud.google.com/storage/[Google Cloud Storage] for the first time, you must connect to the https://console.cloud.google.com/[Google Cloud Platform Console] and create a new project. After your project is created, you must enable the Cloud Storage Service for your project. @@ -20,10 +20,10 @@ Cloud Storage Service for your project. [[repository-gcs-creating-bucket]] ===== Creating a Bucket -The Google Cloud Storage service uses the concept of a -https://cloud.google.com/storage/docs/key-terms[bucket] as a container for all -the data. Buckets are usually created using the -https://console.cloud.google.com/[Google Cloud Platform Console]. The plugin +The Google Cloud Storage service uses the concept of a +https://cloud.google.com/storage/docs/key-terms[bucket] as a container for all +the data. Buckets are usually created using the +https://console.cloud.google.com/[Google Cloud Platform Console]. The plugin does not automatically create buckets. To create a new bucket: @@ -43,12 +43,12 @@ https://cloud.google.com/storage/docs/quickstart-console#create_a_bucket[Google [[repository-gcs-service-authentication]] ===== Service Authentication -The plugin must authenticate the requests it makes to the Google Cloud Storage +The plugin must authenticate the requests it makes to the Google Cloud Storage service. It is common for Google client libraries to employ a strategy named https://cloud.google.com/docs/authentication/production#providing_credentials_to_your_application[application default credentials]. -However, that strategy is **not** supported for use with Elasticsearch. The -plugin operates under the Elasticsearch process, which runs with the security +However, that strategy is **not** supported for use with Elasticsearch. The +plugin operates under the Elasticsearch process, which runs with the security manager enabled. The security manager obstructs the "automatic" credential discovery. -Therefore, you must configure <> +Therefore, you must configure <> credentials even if you are using an environment that does not normally require this configuration (such as Compute Engine, Kubernetes Engine or App Engine). @@ -67,7 +67,7 @@ Here is a summary of the steps: 3. Got to the https://console.cloud.google.com/permissions[Permission] tab. 4. Select the https://console.cloud.google.com/permissions/serviceaccounts[Service Accounts] tab. 5. Click *Create service account*. -6. After the account is created, select it and download a JSON key file. +6. After the account is created, select it and download a JSON key file. A JSON service account file looks like this: @@ -92,13 +92,13 @@ To provide this file to the plugin, it must be stored in the {ref}/secure-settin add a `file` setting with the name `gcs.client.NAME.credentials_file` using the `add-file` subcommand. `NAME` is the name of the client configuration for the repository. The implicit client name is `default`, but a different client name can be specified in the -repository settings with the `client` key. +repository settings with the `client` key. -NOTE: Passing the file path via the GOOGLE_APPLICATION_CREDENTIALS environment +NOTE: Passing the file path via the GOOGLE_APPLICATION_CREDENTIALS environment variable is **not** supported. -For example, if you added a `gcs.client.my_alternate_client.credentials_file` -setting in the keystore, you can configure a repository to use those credentials +For example, if you added a `gcs.client.my_alternate_client.credentials_file` +setting in the keystore, you can configure a repository to use those credentials like this: [source,js] @@ -116,11 +116,11 @@ PUT _snapshot/my_gcs_repository // TEST[skip:we don't have gcs setup while testing this] The `credentials_file` settings are {ref}/secure-settings.html#reloadable-secure-settings[reloadable]. -After you reload the settings, the internal `gcs` clients, which are used to +After you reload the settings, the internal `gcs` clients, which are used to transfer the snapshot contents, utilize the latest settings from the keystore. NOTE: Snapshot or restore jobs that are in progress are not preempted by a *reload* -of the client's `credentials_file` settings. They complete using the client as +of the client's `credentials_file` settings. They complete using the client as it was built when the operation started. [[repository-gcs-client]] @@ -232,8 +232,8 @@ The following settings are supported: `chunk_size`:: Big files can be broken down into chunks during snapshotting if needed. - Specify the chunk size as a value and unit, for example: - `10MB` or `5KB`. Defaults to `100MB`, which is the maximum permitted. + The chunk size can be specified in bytes or by using size value notation, + e.g. , `10m` or `5k`. Defaults to `100m`, which is the maximum permitted. `compress`:: diff --git a/docs/plugins/repository-s3.asciidoc b/docs/plugins/repository-s3.asciidoc index 3949517a174..f1a00742bc8 100644 --- a/docs/plugins/repository-s3.asciidoc +++ b/docs/plugins/repository-s3.asciidoc @@ -32,7 +32,8 @@ PUT _snapshot/my_s3_repository } ---- // CONSOLE -// TEST[skip:we don't have s3 setup while testing this +// TEST[skip:we don't have s3 setup while testing this] + [[repository-s3-client]] ==== Client Settings @@ -243,9 +244,9 @@ The following settings are supported: `chunk_size`:: - Big files can be broken down into chunks during snapshotting if needed. - Specify the chunk size as a value and unit, for example: - `1GB`, `10MB`, `5KB`, `500B`. Defaults to `1GB`. + Big files can be broken down into chunks during snapshotting if needed. The + chunk size can be specified in bytes or by using size value notation, i.e. + `1gb`, `10mb`, `5kb`. Defaults to `1gb`. `compress`:: diff --git a/docs/reference/modules/snapshots.asciidoc b/docs/reference/modules/snapshots.asciidoc index 4da16eb9e78..d0b552f4365 100644 --- a/docs/reference/modules/snapshots.asciidoc +++ b/docs/reference/modules/snapshots.asciidoc @@ -218,9 +218,9 @@ The following settings are supported: [horizontal] `location`:: Location of the snapshots. Mandatory. -`compress`:: Turns on compression of the snapshot files. Compression is applied only to metadata files (index mapping and settings). Data files are not compressed. Defaults to `true`. -`chunk_size`:: Big files can be broken down into chunks during snapshotting if needed. Specify the chunk size as a value and -unit, for example: `1GB`, `10MB`, `5KB`, `500B`. Defaults to `null` (unlimited chunk size). +`compress`:: Turns on compression of the snapshot files. Compression is applied only to metadata files (index mapping and settings). Data files are not compressed. Defaults to `false`. +`chunk_size`:: Big files can be broken down into chunks during snapshotting if needed. The chunk size can be specified in bytes or by + using size value notation, i.e. 1g, 10m, 5k. Defaults to `null` (unlimited chunk size). `max_restore_bytes_per_sec`:: Throttles per node restore rate. Defaults to `40mb` per second. `max_snapshot_bytes_per_sec`:: Throttles per node snapshot rate. Defaults to `40mb` per second. `readonly`:: Makes repository read-only. Defaults to `false`.