7.2 KiB
layout |
---|
doc_page |
Metadata Storage
The Metadata Storage is an external dependency of Druid. Druid uses it to store various metadata about the system, but not to store the actual data. There are a number of tables used for various purposes described below.
Supported Metadata Storages
The following metadata storage engines are supported:
- Derby (default, but not suitable for production)
- MySQL
- PostgreSQL
Even though Derby is the default, it works only if you have all Druid processes running on the same host, and should be used only for experimentation. For production, MySQL or PostgreSQL should be used.
To choose the metadata storage type, set druid.metadata.storage.type
to
mysql
, postgres
or derby
.
Set other druid.metadata.storage
configuration
keywords as shown below to give Druid information about how to connect to
the database.
As discussed in Including Extensions,
there are two ways for giving Druid the extension files it needs for the
database you are using.
The first is to put the extension files in the classpath. The second is to
put the extension files in a subdirectory of
druid.extensions.directory
(by default extensions
under the Druid working directory) and list the subdirectory name in
druid.extensions.loadList
. The example properties below show the second
way.
Setting up MySQL
- Install MySQL
Use your favorite package manager to install mysql, e.g.:
- on Ubuntu/Debian using apt
apt-get install mysql-server
- on OS X, using Homebrew
brew install mysql
Alternatively, download and follow installation instructions for MySQL Community Server here: http://dev.mysql.com/downloads/mysql/
- Create a druid database and user
Connect to MySQL from the machine where it is installed.
> mysql -u root
Paste the following snippet into the mysql prompt:
-- create a druid database, make sure to use utf8 as encoding
CREATE DATABASE druid DEFAULT CHARACTER SET utf8;
-- create a druid user, and grant it all permission on the database we just created
GRANT ALL ON druid.* TO 'druid'@'localhost' IDENTIFIED BY 'diurd';
- Configure your Druid metadata storage extension:
Add the following parameters to your Druid configuration, replacing <host>
with the location (host name and port) of the database.
druid.extensions.loadList=["mysql-metadata-storage"]
druid.metadata.storage.type=mysql
druid.metadata.storage.connector.connectURI=jdbc:mysql://<host>/druid
druid.metadata.storage.connector.user=druid
druid.metadata.storage.connector.password=diurd
Note: the metadata storage extension is not packaged within the main Druid tarball; it is packaged in a separate tarball that can be downloaded from here. You can also get it using pull-deps, or you can build it from source code; see Build from Source.
Setting up PostgreSQL
- Install PostgreSQL
Use your favorite package manager to install PostgreSQL, e.g.:
- on Ubuntu/Debian using apt
apt-get install postgresql
- on OS X, using Homebrew
brew install postgresql
- Create a druid database and user
On the machine where PostgreSQL is installed, using an account with proper postgresql permissions:
Create a druid user, enter diurd
when prompted for the password.
createuser druid -P
Create a druid database owned by the user we just created
createdb druid -O druid
Note: On Ubuntu / Debian you may have to prefix the createuser
and
createdb
commands with sudo -u postgres
in order to gain proper
permissions.
- Configure your Druid metadata storage extension:
Add the following parameters to your Druid configuration, replacing <host>
with the location (host name and port) of the database.
druid.extensions.loadList=["postgresql-metadata-storage"]
druid.metadata.storage.type=postgresql
druid.metadata.storage.connector.connectURI=jdbc:postgresql://<host>/druid
druid.metadata.storage.connector.user=druid
druid.metadata.storage.connector.password=diurd
Using derby
Add the following to your Druid configuration.
druid.metadata.storage.type=derby
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//home/y/var/druid_state/derby;create=true
Metadata Storage Tables
Segments Table
This is dictated by the druid.metadata.storage.tables.segments
property.
This table stores metadata about the segments that are available in the system. The table is polled by the Coordinator to determine the set of segments that should be available for querying in the system. The table has two main functional columns, the other columns are for indexing purposes.
The used
column is a boolean "tombstone". A 1 means that the segment should
be "used" by the cluster (i.e. it should be loaded and available for requests).
A 0 means that the segment should not be actively loaded into the cluster. We
do this as a means of removing segments from the cluster without actually
removing their metadata (which allows for simpler rolling back if that is ever
an issue).
The payload
column stores a JSON blob that has all of the metadata for the segment (some of the data stored in this payload is redundant with some of the columns in the table, that is intentional). This looks something like
{
"dataSource":"wikipedia",
"interval":"2012-05-23T00:00:00.000Z/2012-05-24T00:00:00.000Z",
"version":"2012-05-24T00:10:00.046Z",
"loadSpec":{"type":"s3_zip",
"bucket":"bucket_for_segment",
"key":"path/to/segment/on/s3"},
"dimensions":"comma-delimited-list-of-dimension-names",
"metrics":"comma-delimited-list-of-metric-names",
"shardSpec":{"type":"none"},
"binaryVersion":9,
"size":size_of_segment,
"identifier":"wikipedia_2012-05-23T00:00:00.000Z_2012-05-24T00:00:00.000Z_2012-05-23T00:10:00.046Z"
}
Note that the format of this blob can and will change from time-to-time.
Rule Table
The rule table is used to store the various rules about where segments should land. These rules are used by the Coordinator when making segment (re-)allocation decisions about the cluster.
Config Table
The config table is used to store runtime configuration objects. We do not have many of these yet and we are not sure if we will keep this mechanism going forward, but it is the beginnings of a method of changing some configuration parameters across the cluster at runtime.
Task-related Tables
There are also a number of tables created and used by the Indexing Service in the course of its work.
Audit Table
The Audit table is used to store the audit history for configuration changes e.g rule changes done by Coordinator and other config changes.
+##Accessed By: ## + +The Metadata Storage is accessed only by: + +1. Realtime Nodes 2. Indexing Service Nodes (if any) +3. Coordinator Nodes + +Thus you need to give permissions (eg in AWS Security Groups) only for these machines to access the Metadata storage.