HBASE-20344 Fix asciidoc warnings
Signed-off-by: Sean Busbey <busbey@apache.org>
This commit is contained in:
parent
d59a6c8166
commit
826909a59c
|
@ -175,7 +175,7 @@ and its options. The below information is captured in this help message for each
|
|||
// hbase backup create
|
||||
|
||||
[[br.creating.complete.backup]]
|
||||
### Creating a Backup Image
|
||||
=== Creating a Backup Image
|
||||
|
||||
[NOTE]
|
||||
====
|
||||
|
@ -204,7 +204,7 @@ dataset with a restore operation, having the backup ID readily available can sav
|
|||
====
|
||||
|
||||
[[br.create.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_type_::
|
||||
The type of backup to execute: _full_ or _incremental_. As a reminder, an _incremental_ backup requires a _full_ backup to
|
||||
|
@ -215,7 +215,7 @@ _backup_path_::
|
|||
are _hdfs:_, _webhdfs:_, _gpfs:_, and _s3fs:_.
|
||||
|
||||
[[br.create.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
_-t <table_name[,table_name]>_::
|
||||
A comma-separated list of tables to back up. If no tables are specified, all tables are backed up. No regular-expression or
|
||||
|
@ -242,7 +242,7 @@ _-q <name>_::
|
|||
is useful to prevent backup tasks from stealing resources away from other MapReduce jobs of high importance.
|
||||
|
||||
[[br.usage.examples]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -255,7 +255,7 @@ in the path _/data/backup_. The _-w_ option specifies that no more than three pa
|
|||
// hbase backup restore
|
||||
|
||||
[[br.restoring.backup]]
|
||||
### Restoring a Backup Image
|
||||
=== Restoring a Backup Image
|
||||
|
||||
Run the following command as an HBase superuser. You can only restore a backup on a running HBase cluster because the data must be
|
||||
redistributed the RegionServers for the operation to complete successfully.
|
||||
|
@ -266,7 +266,7 @@ hbase restore <backup_path> <backup_id>
|
|||
----
|
||||
|
||||
[[br.restore.positional.args]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_path_::
|
||||
The _backup_path_ argument specifies the full filesystem URI of where to store the backup image. Valid prefixes are
|
||||
|
@ -277,7 +277,7 @@ _backup_id_::
|
|||
|
||||
|
||||
[[br.restore.named.args]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
_-t <table_name[,table_name]>_::
|
||||
A comma-separated list of tables to restore. See <<br.using.backup.sets,Backup Sets>> for more
|
||||
|
@ -304,7 +304,7 @@ _-o_::
|
|||
|
||||
|
||||
[[br.restore.usage]]
|
||||
#### Example of Usage
|
||||
==== Example of Usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -319,7 +319,7 @@ This command restores two tables of an incremental backup image. In this example
|
|||
// hbase backup merge
|
||||
|
||||
[[br.merge.backup]]
|
||||
### Merging Incremental Backup Images
|
||||
=== Merging Incremental Backup Images
|
||||
|
||||
This command can be used to merge two or more incremental backup images into a single incremental
|
||||
backup image. This can be used to consolidate multiple, small incremental backup images into a single
|
||||
|
@ -332,18 +332,18 @@ $ hbase backup merge <backup_ids>
|
|||
----
|
||||
|
||||
[[br.merge.backup.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_ids_::
|
||||
A comma-separated list of incremental backup image IDs that are to be combined into a single image.
|
||||
|
||||
[[br.merge.backup.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.merge.backup.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -353,7 +353,7 @@ $ hbase backup merge backupId_1467823988425,backupId_1467827588425
|
|||
// hbase backup set
|
||||
|
||||
[[br.using.backup.sets]]
|
||||
### Using Backup Sets
|
||||
=== Using Backup Sets
|
||||
|
||||
Backup sets can ease the administration of HBase data backups and restores by reducing the amount of repetitive input
|
||||
of table names. You can group tables into a named backup set with the `hbase backup set add` command. You can then use
|
||||
|
@ -381,7 +381,7 @@ $ hbase backup set <subcommand> <backup_set_name> <tables>
|
|||
----
|
||||
|
||||
[[br.set.subcommands]]
|
||||
#### Backup Set Subcommands
|
||||
==== Backup Set Subcommands
|
||||
|
||||
The following list details subcommands of the hbase backup set command.
|
||||
|
||||
|
@ -406,7 +406,7 @@ _delete_::
|
|||
Deletes a backup set. Enter the value for the _backup_set_name_ option directly after the `hbase backup set delete` command.
|
||||
|
||||
[[br.set.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_set_name_::
|
||||
Use to assign or invoke a backup set name. The backup set name must contain only printable characters and cannot have any spaces.
|
||||
|
@ -419,7 +419,7 @@ TIP: Maintain a log or other record of the case-sensitive backup set names and t
|
|||
or remote cluster, backup strategy. This information can help you in case of failure on the primary cluster.
|
||||
|
||||
[[br.set.usage]]
|
||||
#### Example of Usage
|
||||
==== Example of Usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -432,7 +432,7 @@ Depending on the environment, this command results in _one_ of the following act
|
|||
* If the `Q1Data` backup set exists already, the tables `TEAM_3` and `TEAM_4` are added to the `Q1Data` backup set.
|
||||
|
||||
[[br.administration]]
|
||||
## Administration of Backup Images
|
||||
== Administration of Backup Images
|
||||
|
||||
The `hbase backup` command has several subcommands that help with administering backup images as they accumulate. Most production
|
||||
environments require recurring backups, so it is necessary to have utilities to help manage the data of the backup repository.
|
||||
|
@ -445,7 +445,7 @@ the HBase superuser.
|
|||
// hbase backup progress
|
||||
|
||||
[[br.managing.backup.progress]]
|
||||
### Managing Backup Progress
|
||||
=== Managing Backup Progress
|
||||
|
||||
You can monitor a running backup in another terminal session by running the _hbase backup progress_ command and specifying the backup ID as an argument.
|
||||
|
||||
|
@ -457,18 +457,18 @@ $ hbase backup progress <backup_id>
|
|||
----
|
||||
|
||||
[[br.progress.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_id_::
|
||||
Specifies the backup that you want to monitor by seeing the progress information. The backupId is case-sensitive.
|
||||
|
||||
[[br.progress.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.progress.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -478,7 +478,7 @@ hbase backup progress backupId_1467823988425
|
|||
// hbase backup history
|
||||
|
||||
[[br.managing.backup.history]]
|
||||
### Managing Backup History
|
||||
=== Managing Backup History
|
||||
|
||||
This command displays a log of backup sessions. The information for each session includes backup ID, type (full or incremental), the tables
|
||||
in the backup, status, and start and end time. Specify the number of backup sessions to display with the optional -n argument.
|
||||
|
@ -489,13 +489,13 @@ $ hbase backup history <backup_id>
|
|||
----
|
||||
|
||||
[[br.history.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_id_::
|
||||
Specifies the backup that you want to monitor by seeing the progress information. The backupId is case-sensitive.
|
||||
|
||||
[[br.history.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
_-n <num_records>_::
|
||||
(Optional) The maximum number of backup records (Default: 10).
|
||||
|
@ -510,7 +510,7 @@ _-t_ <table_name>::
|
|||
The name of table to obtain history for. Mutually exclusive with the _-s_ option.
|
||||
|
||||
[[br.history.backup.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -522,7 +522,7 @@ $ hbase backup history -t WebIndexRecords
|
|||
// hbase backup describe
|
||||
|
||||
[[br.describe.backup]]
|
||||
### Describing a Backup Image
|
||||
=== Describing a Backup Image
|
||||
|
||||
This command can be used to obtain information about a specific backup image.
|
||||
|
||||
|
@ -532,18 +532,18 @@ $ hbase backup describe <backup_id>
|
|||
----
|
||||
|
||||
[[br.describe.backup.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_id_::
|
||||
The ID of the backup image to describe.
|
||||
|
||||
[[br.describe.backup.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.describe.backup.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -553,7 +553,7 @@ $ hbase backup describe backupId_1467823988425
|
|||
// hbase backup delete
|
||||
|
||||
[[br.delete.backup]]
|
||||
### Deleting a Backup Image
|
||||
=== Deleting a Backup Image
|
||||
|
||||
This command can be used to delete a backup image which is no longer needed.
|
||||
|
||||
|
@ -563,18 +563,18 @@ $ hbase backup delete <backup_id>
|
|||
----
|
||||
|
||||
[[br.delete.backup.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
_backup_id_::
|
||||
The ID to the backup image which should be deleted.
|
||||
|
||||
[[br.delete.backup.named.cli.arguments]]
|
||||
#### Named Command-Line Arguments
|
||||
==== Named Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.delete.backup.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -584,7 +584,7 @@ $ hbase backup delete backupId_1467823988425
|
|||
// hbase backup repair
|
||||
|
||||
[[br.repair.backup]]
|
||||
### Backup Repair Command
|
||||
=== Backup Repair Command
|
||||
|
||||
This command attempts to correct any inconsistencies in persisted backup metadata which exists as
|
||||
the result of software errors or unhandled failure scenarios. While the backup implementation tries
|
||||
|
@ -597,17 +597,17 @@ $ hbase backup repair
|
|||
----
|
||||
|
||||
[[br.repair.backup.positional.cli.arguments]]
|
||||
#### Positional Command-Line Arguments
|
||||
==== Positional Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.repair.backup.named.cli.arguments]]
|
||||
### Named Command-Line Arguments
|
||||
=== Named Command-Line Arguments
|
||||
|
||||
None.
|
||||
|
||||
[[br.repair.backup.example]]
|
||||
#### Example usage
|
||||
==== Example usage
|
||||
|
||||
[source]
|
||||
----
|
||||
|
@ -615,11 +615,11 @@ $ hbase backup repair
|
|||
----
|
||||
|
||||
[[br.backup.configuration]]
|
||||
## Configuration keys
|
||||
== Configuration keys
|
||||
|
||||
The backup and restore feature includes both required and optional configuration keys.
|
||||
|
||||
### Required properties
|
||||
=== Required properties
|
||||
|
||||
_hbase.backup.enable_: Controls whether or not the feature is enabled (Default: `false`). Set this value to `true`.
|
||||
|
||||
|
@ -638,7 +638,7 @@ _hbase.coprocessor.region.classes_: A comma-separated list of RegionObservers de
|
|||
_hbase.master.hfilecleaner.plugins_: A comma-separated list of HFileCleaners deployed on the Master. Set this value
|
||||
to `org.apache.hadoop.hbase.backup.BackupHFileCleaner` or append it to the current value.
|
||||
|
||||
### Optional properties
|
||||
=== Optional properties
|
||||
|
||||
_hbase.backup.system.ttl_: The time-to-live in seconds of data in the `hbase:backup` tables (default: forever). This property
|
||||
is only relevant prior to the creation of the `hbase:backup` table. Use the `alter` command in the HBase shell to modify the TTL
|
||||
|
@ -653,9 +653,9 @@ _hbase.backup.logroll.timeout.millis_: The amount of time (in milliseconds) to w
|
|||
in the Master's procedure framework (default: 30000).
|
||||
|
||||
[[br.best.practices]]
|
||||
## Best Practices
|
||||
== Best Practices
|
||||
|
||||
### Formulate a restore strategy and test it.
|
||||
=== Formulate a restore strategy and test it.
|
||||
|
||||
Before you rely on a backup and restore strategy for your production environment, identify how backups must be performed,
|
||||
and more importantly, how restores must be performed. Test the plan to ensure that it is workable.
|
||||
|
@ -668,14 +668,14 @@ site renders locally stored backups useless. Consider storing the backup data an
|
|||
and operator expertise) to restore the data at a site sufficiently remote from the production site. In the case of a catastrophe
|
||||
at the whole primary site (fire, earthquake, etc.), the remote backup site can be very valuable.
|
||||
|
||||
### Secure a full backup image first.
|
||||
=== Secure a full backup image first.
|
||||
|
||||
As a baseline, you must complete a full backup of HBase data at least once before you can rely on incremental backups. The full
|
||||
backup should be stored outside of the source cluster. To ensure complete dataset recovery, you must run the restore utility
|
||||
with the option to restore baseline full backup. The full backup is the foundation of your dataset. Incremental backup data
|
||||
is applied on top of the full backup during the restore operation to return you to the point in time when backup was last taken.
|
||||
|
||||
### Define and use backup sets for groups of tables that are logical subsets of the entire dataset.
|
||||
=== Define and use backup sets for groups of tables that are logical subsets of the entire dataset.
|
||||
|
||||
You can group tables into an object called a backup set. A backup set can save time when you have a particular group of tables
|
||||
that you expect to repeatedly back up or restore.
|
||||
|
@ -684,7 +684,7 @@ When you create a backup set, you type table names to include in the group. The
|
|||
tables, but also retains the HBase backup metadata. Afterwards, you can invoke the backup set name to indicate what tables apply
|
||||
to the command execution instead of entering all the table names individually.
|
||||
|
||||
### Document the backup and restore strategy, and ideally log information about each backup.
|
||||
=== Document the backup and restore strategy, and ideally log information about each backup.
|
||||
|
||||
Document the whole process so that the knowledge base can transfer to new administrators after employee turnover. As an extra
|
||||
safety precaution, also log the calendar date, time, and other relevant details about the data of each backup. This metadata
|
||||
|
@ -693,7 +693,7 @@ copies of all documentation: one copy at the production cluster site and another
|
|||
accessed by an administrator remotely from the production cluster.
|
||||
|
||||
[[br.s3.backup.scenario]]
|
||||
## Scenario: Safeguarding Application Datasets on Amazon S3
|
||||
== Scenario: Safeguarding Application Datasets on Amazon S3
|
||||
|
||||
This scenario describes how a hypothetical retail business uses backups to safeguard application data and then restore the dataset
|
||||
after failure.
|
||||
|
@ -760,7 +760,7 @@ existing data in the destination. In this case, the admin decides to overwrite t
|
|||
s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite
|
||||
|
||||
[[br.data.security]]
|
||||
## Security of Backup Data
|
||||
== Security of Backup Data
|
||||
|
||||
With this feature which makes copying data to remote locations, it's important to take a moment to clearly state the procedural
|
||||
concerns that exist around data security. Like the HBase replication feature, backup and restore provides the constructs to automatically
|
||||
|
@ -774,7 +774,7 @@ being accessed via HBase, and its authentication and authorization controls, we
|
|||
providing a comparable level of security. This is a manual step which users *must* implement on their own.
|
||||
|
||||
[[br.technical.details]]
|
||||
## Technical Details of Incremental Backup and Restore
|
||||
== Technical Details of Incremental Backup and Restore
|
||||
|
||||
HBase incremental backups enable more efficient capture of HBase table images than previous attempts at serial backup and restore
|
||||
solutions, such as those that only used HBase Export and Import APIs. Incremental backups use Write Ahead Logs (WALs) to capture
|
||||
|
@ -790,7 +790,7 @@ Bulk Load utility automatically imports as restored data in the table.
|
|||
You can only restore on a live HBase cluster because the data must be redistributed to complete the restore operation successfully.
|
||||
|
||||
[[br.filesystem.growth.warning]]
|
||||
## A Warning on File System Growth
|
||||
== A Warning on File System Growth
|
||||
|
||||
As a reminder, incremental backups are implemented via retaining the write-ahead logs which HBase primarily uses for data durability.
|
||||
Thus, to ensure that all data needing to be included in a backup is still available in the system, the HBase backup and restore feature
|
||||
|
@ -806,14 +806,14 @@ more aggressive backup merges and deletions). As a reminder, the TTL can be alte
|
|||
in the HBase shell. Modifying the configuration property `hbase.backup.system.ttl` in hbase-site.xml after the system table exists has no effect.
|
||||
|
||||
[[br.backup.capacity.planning]]
|
||||
## Capacity Planning
|
||||
== Capacity Planning
|
||||
|
||||
When designing a distributed system deployment, it is critical that some basic mathmatical rigor is executed to ensure sufficient computational
|
||||
capacity is available given the data and software requirements of the system. For this feature, the availability of network capacity is the largest
|
||||
bottleneck when estimating the performance of some implementation of backup and restore. The second most costly function is the speed at which
|
||||
data can be read/written.
|
||||
|
||||
### Full Backups
|
||||
=== Full Backups
|
||||
|
||||
To estimate the duration of a full backup, we have to understand the general actions which are invoked:
|
||||
|
||||
|
@ -840,7 +840,7 @@ queue which can limit the specific nodes where the workers will be spawned -- th
|
|||
a set of non-critical nodes. Relating the `-b` and `-w` options to our earlier equations: `-b` would be used to restrict each node from reading
|
||||
data at the full 80MB/s and `-w` is used to limit the job from spawning 16 worker tasks.
|
||||
|
||||
### Incremental Backup
|
||||
=== Incremental Backup
|
||||
|
||||
Like we did for full backups, we have to understand the incremental backup process to approximate its runtime and cost.
|
||||
|
||||
|
@ -854,7 +854,7 @@ this would require approximately 15 minutes to perform this step for 50GB of dat
|
|||
DistCp MapReduce job would likely dominate the actual time taken to copy the data (50 / 1.25 = 40 seconds) and can be ignored.
|
||||
|
||||
[[br.limitations]]
|
||||
## Limitations of the Backup and Restore Utility
|
||||
== Limitations of the Backup and Restore Utility
|
||||
|
||||
*Serial backup operations*
|
||||
|
||||
|
|
|
@ -335,25 +335,18 @@ You do not need to re-create the table or copy data.
|
|||
If you are changing codecs, be sure the old codec is still available until all the old StoreFiles have been compacted.
|
||||
|
||||
.Enabling Compression on a ColumnFamily of an Existing Table using HBaseShell
|
||||
====
|
||||
----
|
||||
|
||||
hbase> disable 'test'
|
||||
hbase> alter 'test', {NAME => 'cf', COMPRESSION => 'GZ'}
|
||||
hbase> enable 'test'
|
||||
----
|
||||
====
|
||||
|
||||
.Creating a New Table with Compression On a ColumnFamily
|
||||
====
|
||||
----
|
||||
|
||||
hbase> create 'test2', { NAME => 'cf2', COMPRESSION => 'SNAPPY' }
|
||||
----
|
||||
====
|
||||
|
||||
.Verifying a ColumnFamily's Compression Settings
|
||||
====
|
||||
----
|
||||
|
||||
hbase> describe 'test'
|
||||
|
@ -366,7 +359,6 @@ DESCRIPTION ENABLED
|
|||
LOCKCACHE => 'true'}
|
||||
1 row(s) in 0.1070 seconds
|
||||
----
|
||||
====
|
||||
|
||||
==== Testing Compression Performance
|
||||
|
||||
|
@ -374,9 +366,7 @@ HBase includes a tool called LoadTestTool which provides mechanisms to test your
|
|||
You must specify either `-write` or `-update-read` as your first parameter, and if you do not specify another parameter, usage advice is printed for each option.
|
||||
|
||||
.+LoadTestTool+ Usage
|
||||
====
|
||||
----
|
||||
|
||||
$ bin/hbase org.apache.hadoop.hbase.util.LoadTestTool -h
|
||||
usage: bin/hbase org.apache.hadoop.hbase.util.LoadTestTool <options>
|
||||
Options:
|
||||
|
@ -429,16 +419,12 @@ Options:
|
|||
port numbers
|
||||
-zk_root <arg> name of parent znode in zookeeper
|
||||
----
|
||||
====
|
||||
|
||||
.Example Usage of LoadTestTool
|
||||
====
|
||||
----
|
||||
|
||||
$ hbase org.apache.hadoop.hbase.util.LoadTestTool -write 1:10:100 -num_keys 1000000
|
||||
-read 100:30 -num_tables 1 -data_block_encoding NONE -tn load_test_tool_NONE
|
||||
----
|
||||
====
|
||||
|
||||
[[data.block.encoding.enable]]
|
||||
=== Enable Data Block Encoding
|
||||
|
@ -449,9 +435,7 @@ Disable the table before altering its DATA_BLOCK_ENCODING setting.
|
|||
Following is an example using HBase Shell:
|
||||
|
||||
.Enable Data Block Encoding On a Table
|
||||
====
|
||||
----
|
||||
|
||||
hbase> disable 'test'
|
||||
hbase> alter 'test', { NAME => 'cf', DATA_BLOCK_ENCODING => 'FAST_DIFF' }
|
||||
Updating all regions with the new schema...
|
||||
|
@ -462,12 +446,9 @@ Done.
|
|||
hbase> enable 'test'
|
||||
0 row(s) in 0.1580 seconds
|
||||
----
|
||||
====
|
||||
|
||||
.Verifying a ColumnFamily's Data Block Encoding
|
||||
====
|
||||
----
|
||||
|
||||
hbase> describe 'test'
|
||||
DESCRIPTION ENABLED
|
||||
'test', {NAME => 'cf', DATA_BLOCK_ENCODING => 'FAST true
|
||||
|
@ -478,7 +459,6 @@ DESCRIPTION ENABLED
|
|||
e', BLOCKCACHE => 'true'}
|
||||
1 row(s) in 0.0650 seconds
|
||||
----
|
||||
====
|
||||
|
||||
:numbered:
|
||||
|
||||
|
|
|
@ -604,18 +604,14 @@ On each node of the cluster, run the `jps` command and verify that the correct p
|
|||
You may see additional Java processes running on your servers as well, if they are used for other purposes.
|
||||
+
|
||||
.`node-a` `jps` Output
|
||||
====
|
||||
----
|
||||
|
||||
$ jps
|
||||
20355 Jps
|
||||
20071 HQuorumPeer
|
||||
20137 HMaster
|
||||
----
|
||||
====
|
||||
+
|
||||
.`node-b` `jps` Output
|
||||
====
|
||||
----
|
||||
$ jps
|
||||
15930 HRegionServer
|
||||
|
@ -623,17 +619,14 @@ $ jps
|
|||
15838 HQuorumPeer
|
||||
16010 HMaster
|
||||
----
|
||||
====
|
||||
+
|
||||
.`node-c` `jps` Output
|
||||
====
|
||||
----
|
||||
$ jps
|
||||
13901 Jps
|
||||
13639 HQuorumPeer
|
||||
13737 HRegionServer
|
||||
----
|
||||
====
|
||||
+
|
||||
.ZooKeeper Process Name
|
||||
[NOTE]
|
||||
|
|
|
@ -61,12 +61,10 @@ an object is considered to be a MOB. Only `IS_MOB` is required. If you do not
|
|||
specify the `MOB_THRESHOLD`, the default threshold value of 100 KB is used.
|
||||
|
||||
.Configure a Column for MOB Using HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400}
|
||||
hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400}
|
||||
----
|
||||
====
|
||||
|
||||
.Configure a Column for MOB Using the Java API
|
||||
====
|
||||
|
@ -91,7 +89,6 @@ weekly policy - compact MOB Files for one week into one large MOB file
|
|||
montly policy - compact MOB Files for one month into one large MOB File
|
||||
|
||||
.Configure MOB compaction policy Using HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY => 'daily'}
|
||||
hbase> create 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY => 'weekly'}
|
||||
|
@ -101,7 +98,6 @@ hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_C
|
|||
hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY => 'weekly'}
|
||||
hbase> alter 't1', {NAME => 'f1', IS_MOB => true, MOB_THRESHOLD => 102400, MOB_COMPACT_PARTITION_POLICY => 'monthly'}
|
||||
----
|
||||
====
|
||||
|
||||
=== Configure MOB Compaction mergeable threshold
|
||||
|
||||
|
|
|
@ -1023,13 +1023,10 @@ The script requires you to set some environment variables before running it.
|
|||
Examine the script and modify it to suit your needs.
|
||||
|
||||
._rolling-restart.sh_ General Usage
|
||||
====
|
||||
----
|
||||
|
||||
$ ./bin/rolling-restart.sh --help
|
||||
Usage: rolling-restart.sh [--config <hbase-confdir>] [--rs-only] [--master-only] [--graceful] [--maxthreads xx]
|
||||
----
|
||||
====
|
||||
|
||||
Rolling Restart on RegionServers Only::
|
||||
To perform a rolling restart on the RegionServers only, use the `--rs-only` option.
|
||||
|
|
|
@ -188,11 +188,9 @@ It is useful for tuning the IO impact of prefetching versus the time before all
|
|||
To enable prefetching on a given column family, you can use HBase Shell or use the API.
|
||||
|
||||
.Enable Prefetch Using HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> create 'MyTable', { NAME => 'myCF', PREFETCH_BLOCKS_ON_OPEN => 'true' }
|
||||
----
|
||||
====
|
||||
|
||||
.Enable Prefetch Using the API
|
||||
====
|
||||
|
|
|
@ -504,11 +504,9 @@ Deleted cells are still subject to TTL and there will never be more than "maximu
|
|||
A new "raw" scan options returns all deleted rows and the delete markers.
|
||||
|
||||
.Change the Value of `KEEP_DELETED_CELLS` Using HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> hbase> alter ‘t1′, NAME => ‘f1′, KEEP_DELETED_CELLS => true
|
||||
----
|
||||
====
|
||||
|
||||
.Change the Value of `KEEP_DELETED_CELLS` Using the API
|
||||
====
|
||||
|
|
|
@ -1086,7 +1086,6 @@ public static void revokeFromTable(final HBaseTestingUtility util, final String
|
|||
. Showing a User's Effective Permissions
|
||||
+
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> user_permission 'user'
|
||||
|
||||
|
@ -1094,7 +1093,6 @@ hbase> user_permission '.*'
|
|||
|
||||
hbase> user_permission JAVA_REGEX
|
||||
----
|
||||
====
|
||||
|
||||
.API
|
||||
====
|
||||
|
@ -1234,11 +1232,9 @@ Refer to the official API for usage instructions.
|
|||
. Define the List of Visibility Labels
|
||||
+
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> add_labels [ 'admin', 'service', 'developer', 'test' ]
|
||||
----
|
||||
====
|
||||
+
|
||||
.Java API
|
||||
====
|
||||
|
@ -1265,7 +1261,6 @@ public static void addLabels() throws Exception {
|
|||
. Associate Labels with Users
|
||||
+
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> set_auths 'service', [ 'service' ]
|
||||
----
|
||||
|
@ -1281,7 +1276,6 @@ hbase> set_auths 'qa', [ 'test', 'developer' ]
|
|||
----
|
||||
hbase> set_auths '@qagroup', [ 'test' ]
|
||||
----
|
||||
====
|
||||
+
|
||||
.Java API
|
||||
====
|
||||
|
@ -1305,7 +1299,6 @@ public void testSetAndGetUserAuths() throws Throwable {
|
|||
. Clear Labels From Users
|
||||
+
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> clear_auths 'service', [ 'service' ]
|
||||
----
|
||||
|
@ -1321,7 +1314,6 @@ hbase> clear_auths 'qa', [ 'test', 'developer' ]
|
|||
----
|
||||
hbase> clear_auths '@qagroup', [ 'test', 'developer' ]
|
||||
----
|
||||
====
|
||||
+
|
||||
.Java API
|
||||
====
|
||||
|
@ -1345,7 +1337,6 @@ The label is only applied when data is written.
|
|||
The label is associated with a given version of the cell.
|
||||
+
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> set_visibility 'user', 'admin|service|developer', { COLUMNS => 'i' }
|
||||
----
|
||||
|
@ -1357,7 +1348,6 @@ hbase> set_visibility 'user', 'admin|service', { COLUMNS => 'pii' }
|
|||
----
|
||||
hbase> set_visibility 'user', 'test', { COLUMNS => [ 'i', 'pii' ], FILTER => "(PrefixFilter ('test'))" }
|
||||
----
|
||||
====
|
||||
+
|
||||
NOTE: HBase Shell support for applying labels or permissions to cells is for testing and verification support, and should not be employed for production use because it won't apply the labels to cells that don't exist yet.
|
||||
The correct way to apply cell level labels is to do so in the application code when storing the values.
|
||||
|
@ -1408,12 +1398,10 @@ set as an additional filter. It will further filter your results, rather than
|
|||
giving you additional authorization.
|
||||
|
||||
.HBase Shell
|
||||
====
|
||||
----
|
||||
hbase> get_auths 'myUser'
|
||||
hbase> scan 'table1', AUTHORIZATIONS => ['private']
|
||||
----
|
||||
====
|
||||
|
||||
.Java API
|
||||
====
|
||||
|
|
|
@ -145,7 +145,6 @@ For instance, if your script creates a table, but returns a non-zero exit value,
|
|||
You can enter HBase Shell commands into a text file, one command per line, and pass that file to the HBase Shell.
|
||||
|
||||
.Example Command File
|
||||
====
|
||||
----
|
||||
create 'test', 'cf'
|
||||
list 'test'
|
||||
|
@ -158,7 +157,6 @@ get 'test', 'row1'
|
|||
disable 'test'
|
||||
enable 'test'
|
||||
----
|
||||
====
|
||||
|
||||
.Directing HBase Shell to Execute the Commands
|
||||
====
|
||||
|
|
Loading…
Reference in New Issue