HBASE-16412 Warnings from asciidoc
@ -710,66 +710,54 @@ image::backup-app-components.png[]
|
|||||||
The following is an outline of the steps and examples of commands that are used to backup the data for the _green_ application and
|
The following is an outline of the steps and examples of commands that are used to backup the data for the _green_ application and
|
||||||
to recover the data later. All commands are run when logged in as HBase superuser.
|
to recover the data later. All commands are run when logged in as HBase superuser.
|
||||||
|
|
||||||
1. A backup set called _green_set_ is created as an alias for both the transactions table and the customer table. The backup set can
|
* A backup set called _green_set_ is created as an alias for both the transactions table and the customer table. The backup set can
|
||||||
be used for all operations to avoid typing each table name. The backup set name is case-sensitive and should be formed with only
|
be used for all operations to avoid typing each table name. The backup set name is case-sensitive and should be formed with only
|
||||||
printable characters and without spaces.
|
printable characters and without spaces.
|
||||||
|
|
||||||
[source]
|
$ hbase backup set add green_set transactions
|
||||||
----
|
$ hbase backup set add green_set customer
|
||||||
$ hbase backup set add green_set transactions
|
|
||||||
$ hbase backup set add green_set customer
|
|
||||||
----
|
|
||||||
|
|
||||||
2. The first backup of green_set data must be a full backup. The following command example shows how credentials are passed to Amazon
|
* The first backup of green_set data must be a full backup. The following command example shows how credentials are passed to Amazon
|
||||||
S3 and specifies the file system with the s3a: prefix.
|
S3 and specifies the file system with the s3a: prefix.
|
||||||
|
|
||||||
[source]
|
$ ACCESS_KEY=ABCDEFGHIJKLMNOPQRST
|
||||||
----
|
$ SECRET_KEY=123456789abcdefghijklmnopqrstuvwxyzABCD
|
||||||
$ ACCESS_KEY=ABCDEFGHIJKLMNOPQRST
|
$ sudo -u hbase hbase backup create full\
|
||||||
$ SECRET_KEY=123456789abcdefghijklmnopqrstuvwxyzABCD
|
s3a://$ACCESS_KEY:SECRET_KEY@prodhbasebackups/backups -s green_set
|
||||||
$ sudo -u hbase hbase backup create full\
|
|
||||||
s3a://$ACCESS_KEY:SECRET_KEY@prodhbasebackups/backups -s green_set
|
|
||||||
----
|
|
||||||
|
|
||||||
3. Incremental backups should be run according to a schedule that ensures essential data recovery in the event of a catastrophe. At
|
* Incremental backups should be run according to a schedule that ensures essential data recovery in the event of a catastrophe. At
|
||||||
this retail company, the HBase admin team decides that automated daily backups secures the data sufficiently. The team decides that
|
this retail company, the HBase admin team decides that automated daily backups secures the data sufficiently. The team decides that
|
||||||
they can implement this by modifying an existing Cron job that is defined in `/etc/crontab`. Consequently, IT modifies the Cron job
|
they can implement this by modifying an existing Cron job that is defined in `/etc/crontab`. Consequently, IT modifies the Cron job
|
||||||
by adding the following line:
|
by adding the following line:
|
||||||
|
|
||||||
[source]
|
@daily hbase hbase backup create incremental s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups -s green_set
|
||||||
----
|
|
||||||
@daily hbase hbase backup create incremental s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups -s green_set
|
|
||||||
----
|
|
||||||
|
|
||||||
4. A catastrophic IT incident disables the production cluster that the green application uses. An HBase system administrator of the
|
* A catastrophic IT incident disables the production cluster that the green application uses. An HBase system administrator of the
|
||||||
backup cluster must restore the _green_set_ dataset to the point in time closest to the recovery objective.
|
backup cluster must restore the _green_set_ dataset to the point in time closest to the recovery objective.
|
||||||
|
+
|
||||||
NOTE: If the administrator of the backup HBase cluster has the backup ID with relevant details in accessible records, the following
|
NOTE: If the administrator of the backup HBase cluster has the backup ID with relevant details in accessible records, the following
|
||||||
search with the `hdfs dfs -ls` command and manually scanning the backup ID list can be bypassed. Consider continuously maintaining
|
search with the `hdfs dfs -ls` command and manually scanning the backup ID list can be bypassed. Consider continuously maintaining
|
||||||
and protecting a detailed log of backup IDs outside the production cluster in your environment.
|
and protecting a detailed log of backup IDs outside the production cluster in your environment.
|
||||||
|
+
|
||||||
The HBase administrator runs the following command on the directory where backups are stored to print the list of successful backup
|
The HBase administrator runs the following command on the directory where backups are stored to print the list of successful backup
|
||||||
IDs on the console:
|
IDs on the console:
|
||||||
|
|
||||||
`hdfs dfs -ls -t /prodhbasebackups/backups`
|
`hdfs dfs -ls -t /prodhbasebackups/backups`
|
||||||
|
|
||||||
5. The admin scans the list to see which backup was created at a date and time closest to the recovery objective. To do this, the
|
* The admin scans the list to see which backup was created at a date and time closest to the recovery objective. To do this, the
|
||||||
admin converts the calendar timestamp of the recovery point in time to Unix time because backup IDs are uniquely identified with
|
admin converts the calendar timestamp of the recovery point in time to Unix time because backup IDs are uniquely identified with
|
||||||
Unix time. The backup IDs are listed in reverse chronological order, meaning the most recent successful backup appears first.
|
Unix time. The backup IDs are listed in reverse chronological order, meaning the most recent successful backup appears first.
|
||||||
|
+
|
||||||
The admin notices that the following line in the command output corresponds with the _green_set_ backup that needs to be restored:
|
The admin notices that the following line in the command output corresponds with the _green_set_ backup that needs to be restored:
|
||||||
|
|
||||||
`/prodhbasebackups/backups/backup_1467823988425`
|
/prodhbasebackups/backups/backup_1467823988425`
|
||||||
|
|
||||||
6. The admin restores green_set invoking the backup ID and the -overwrite option. The -overwrite option truncates all existing data
|
* The admin restores green_set invoking the backup ID and the -overwrite option. The -overwrite option truncates all existing data
|
||||||
in the destination and populates the tables with data from the backup dataset. Without this flag, the backup data is appended to the
|
in the destination and populates the tables with data from the backup dataset. Without this flag, the backup data is appended to the
|
||||||
existing data in the destination. In this case, the admin decides to overwrite the data because it is corrupted.
|
existing data in the destination. In this case, the admin decides to overwrite the data because it is corrupted.
|
||||||
|
|
||||||
[source]
|
$ sudo -u hbase hbase restore -s green_set \
|
||||||
----
|
s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite
|
||||||
$ sudo -u hbase hbase restore -s green_set \
|
|
||||||
s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite
|
|
||||||
----
|
|
||||||
|
|
||||||
[[br.data.security]]
|
[[br.data.security]]
|
||||||
## Security of Backup Data
|
## Security of Backup Data
|
||||||
|
@ -786,14 +786,18 @@ Make sure you are pulling from the repository when tests run and that you are no
|
|||||||
====
|
====
|
||||||
|
|
||||||
See link:https://www.apache.org/dev/publishing-maven-artifacts.html[Publishing Maven Artifacts] for some pointers on this maven staging process.
|
See link:https://www.apache.org/dev/publishing-maven-artifacts.html[Publishing Maven Artifacts] for some pointers on this maven staging process.
|
||||||
|
+
|
||||||
If the HBase version ends in `-SNAPSHOT`, the artifacts go elsewhere.
|
If the HBase version ends in `-SNAPSHOT`, the artifacts go elsewhere.
|
||||||
They are put into the Apache snapshots repository directly and are immediately available.
|
They are put into the Apache snapshots repository directly and are immediately available.
|
||||||
Making a SNAPSHOT release, this is what you want to happen.
|
Making a SNAPSHOT release, this is what you want to happen.
|
||||||
|
+
|
||||||
At this stage, you have two tarballs in your 'build output directory' and a set of artifacts in a staging area of the maven repository, in the 'closed' state.
|
At this stage, you have two tarballs in your 'build output directory' and a set of artifacts
|
||||||
|
in a staging area of the maven repository, in the 'closed' state.
|
||||||
Next sign, fingerprint and then 'stage' your release candiate build output directory via svnpubsub by committing
|
Next sign, fingerprint and then 'stage' your release candiate build output directory via svnpubsub by committing
|
||||||
your directory to link:https://dist.apache.org/repos/dist/dev/hbase/[The 'dev' distribution directory] (See comments on link:https://issues.apache.org/jira/browse/HBASE-10554[HBASE-10554 Please delete old releases from mirroring system] but in essence it is an svn checkout of https://dist.apache.org/repos/dist/dev/hbase -- releases are at https://dist.apache.org/repos/dist/release/hbase). In the _version directory_ run the following commands:
|
your directory to link:https://dist.apache.org/repos/dist/dev/hbase/[The dev distribution directory]
|
||||||
|
(See comments on link:https://issues.apache.org/jira/browse/HBASE-10554[HBASE-10554 Please delete old releases from mirroring system]
|
||||||
|
but in essence it is an svn checkout of link:https://dist.apache.org/repos/dist/dev/hbase[dev/hbase] -- releases are at
|
||||||
|
link:https://dist.apache.org/repos/dist/release/hbase[release/hbase]). In the _version directory_ run the following commands:
|
||||||
|
|
||||||
[source,bourne]
|
[source,bourne]
|
||||||
----
|
----
|
||||||
|
Before Width: | Height: | Size: 24 KiB After Width: | Height: | Size: 24 KiB |
Before Width: | Height: | Size: 29 KiB After Width: | Height: | Size: 29 KiB |
Before Width: | Height: | Size: 24 KiB After Width: | Height: | Size: 24 KiB |
Before Width: | Height: | Size: 19 KiB After Width: | Height: | Size: 19 KiB |