druid/docs/dependencies/metadata-storage.md

142 lines
5.5 KiB
Markdown
Raw Normal View History

---
id: metadata-storage
title: "Metadata storage"
---
<!--
~ Licensed to the Apache Software Foundation (ASF) under one
~ or more contributor license agreements. See the NOTICE file
~ distributed with this work for additional information
~ regarding copyright ownership. The ASF licenses this file
~ to you under the Apache License, Version 2.0 (the
~ "License"); you may not use this file except in compliance
~ with the License. You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing,
~ software distributed under the License is distributed on an
~ "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
~ KIND, either express or implied. See the License for the
~ specific language governing permissions and limitations
~ under the License.
-->
The Metadata Storage is an external dependency of Apache Druid (incubating). Druid uses it to store
2015-08-21 16:17:01 -04:00
various metadata about the system, but not to store the actual data. There are
a number of tables used for various purposes described below.
Derby is the default metadata store for Druid, however, it is not suitable for production.
[MySQL](../development/extensions-core/mysql.md) and [PostgreSQL](../development/extensions-core/postgresql.md) are more production suitable metadata stores.
2015-08-21 16:17:01 -04:00
> The Metadata Storage stores the entire metadata which is essential for a Druid cluster to work.
> For production clusters, consider using MySQL or PostgreSQL instead of Derby.
> Also, it's highly recommended to set up a high availability environment
> because there is no way to restore if you lose any metadata.
2015-08-21 16:17:01 -04:00
## Using Derby
Add the following to your Druid configuration.
```properties
druid.metadata.storage.type=derby
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//opt/var/druid_state/derby;create=true
```
## MySQL
See [mysql-metadata-storage extension documentation](../development/extensions-core/mysql.md).
## PostgreSQL
See [postgresql-metadata-storage](../development/extensions-core/postgresql.md).
## Adding custom dbcp properties
NOTE: These properties are not settable through the `druid.metadata.storage.connector.dbcp properties`: `username`, `password`, `connectURI`, `validationQuery`, `testOnBorrow`. These must be set through `druid.metadata.storage.connector` properties.
Example supported properties:
```properties
druid.metadata.storage.connector.dbcp.maxConnLifetimeMillis=1200000
druid.metadata.storage.connector.dbcp.defaultQueryTimeout=30000
```
See [BasicDataSource Configuration](https://commons.apache.org/proper/commons-dbcp/configuration.html) for full list.
## Metadata storage tables
2015-07-28 14:36:48 -04:00
### Segments table
This is dictated by the `druid.metadata.storage.tables.segments` property.
2015-08-21 16:17:01 -04:00
This table stores metadata about the segments that are available in the system.
The table is polled by the [Coordinator](../design/coordinator.md) to
2015-08-21 16:17:01 -04:00
determine the set of segments that should be available for querying in the
system. The table has two main functional columns, the other columns are for
indexing purposes.
2015-08-21 16:17:01 -04:00
The `used` column is a boolean "tombstone". A 1 means that the segment should
be "used" by the cluster (i.e., it should be loaded and available for requests).
2015-08-21 16:17:01 -04:00
A 0 means that the segment should not be actively loaded into the cluster. We
do this as a means of removing segments from the cluster without actually
removing their metadata (which allows for simpler rolling back if that is ever
an issue).
The `payload` column stores a JSON blob that has all of the metadata for the segment (some of the data stored in this payload is redundant with some of the columns in the table, that is intentional). This looks something like
2016-02-04 14:53:09 -05:00
```json
{
"dataSource":"wikipedia",
"interval":"2012-05-23T00:00:00.000Z/2012-05-24T00:00:00.000Z",
"version":"2012-05-24T00:10:00.046Z",
2016-02-04 14:53:09 -05:00
"loadSpec":{
"type":"s3_zip",
"bucket":"bucket_for_segment",
"key":"path/to/segment/on/s3"
},
"dimensions":"comma-delimited-list-of-dimension-names",
"metrics":"comma-delimited-list-of-metric-names",
"shardSpec":{"type":"none"},
"binaryVersion":9,
"size":size_of_segment,
"identifier":"wikipedia_2012-05-23T00:00:00.000Z_2012-05-24T00:00:00.000Z_2012-05-23T00:10:00.046Z"
}
```
Note that the format of this blob can and will change from time-to-time.
### Rule table
2015-08-21 16:17:01 -04:00
The rule table is used to store the various rules about where segments should
land. These rules are used by the [Coordinator](../design/coordinator.md)
2015-08-21 16:17:01 -04:00
when making segment (re-)allocation decisions about the cluster.
### Config table
2015-08-21 16:17:01 -04:00
The config table is used to store runtime configuration objects. We do not have
many of these yet and we are not sure if we will keep this mechanism going
forward, but it is the beginnings of a method of changing some configuration
parameters across the cluster at runtime.
### Task-related tables
There are also a number of tables created and used by the [Overlord](../design/overlord.md) and [MiddleManager](../design/middlemanager.md) when managing tasks.
### Audit table
2015-08-21 16:17:01 -04:00
The Audit table is used to store the audit history for configuration changes
e.g rule changes done by [Coordinator](../design/coordinator.md) and other
2015-08-21 16:17:01 -04:00
config changes.
##Accessed by: ##
2016-02-04 14:53:09 -05:00
The Metadata Storage is accessed only by:
2019-02-28 21:10:39 -05:00
1. Indexing Service Processes (if any)
2. Realtime Processes (if any)
3. Coordinator Processes
2016-02-04 14:53:09 -05:00
Thus you need to give permissions (e.g., in AWS Security Groups) only for these machines to access the Metadata storage.