mirror of https://github.com/apache/druid.git
163 lines
8.3 KiB
Markdown
163 lines
8.3 KiB
Markdown
---
|
|
layout: doc_page
|
|
title: "Updating Existing Data"
|
|
---
|
|
|
|
<!--
|
|
~ Licensed to the Apache Software Foundation (ASF) under one
|
|
~ or more contributor license agreements. See the NOTICE file
|
|
~ distributed with this work for additional information
|
|
~ regarding copyright ownership. The ASF licenses this file
|
|
~ to you under the Apache License, Version 2.0 (the
|
|
~ "License"); you may not use this file except in compliance
|
|
~ with the License. You may obtain a copy of the License at
|
|
~
|
|
~ http://www.apache.org/licenses/LICENSE-2.0
|
|
~
|
|
~ Unless required by applicable law or agreed to in writing,
|
|
~ software distributed under the License is distributed on an
|
|
~ "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
~ KIND, either express or implied. See the License for the
|
|
~ specific language governing permissions and limitations
|
|
~ under the License.
|
|
-->
|
|
|
|
# Updating Existing Data
|
|
|
|
Once you ingest some data in a dataSource for an interval and create Druid segments, you might want to make changes to
|
|
the ingested data. There are several ways this can be done.
|
|
|
|
##### Updating Dimension Values
|
|
|
|
If you have a dimension where values need to be updated frequently, try first using [lookups](../querying/lookups.html). A
|
|
classic use case of lookups is when you have an ID dimension stored in a Druid segment, and want to map the ID dimension to a
|
|
human-readable String value that may need to be updated periodically.
|
|
|
|
##### Rebuilding Segments (Reindexing)
|
|
|
|
If lookups are not sufficient, you can entirely rebuild Druid segments for specific intervals of time. Rebuilding a segment
|
|
is known as reindexing the data. For example, if you want to add or remove columns from your existing segments, or you want to
|
|
change the rollup granularity of your segments, you will have to reindex your data.
|
|
|
|
We recommend keeping a copy of your raw data around in case you ever need to reindex your data.
|
|
|
|
##### Dealing with Delayed Events (Delta Ingestion)
|
|
|
|
If you have a batch ingestion pipeline and have delayed events come in and want to append these events to existing
|
|
segments and avoid the overhead of rebuilding new segments with reindexing, you can use delta ingestion.
|
|
|
|
### Reindexing and Delta Ingestion with Hadoop Batch Ingestion
|
|
|
|
This section assumes the reader understands how to do batch ingestion using Hadoop. See
|
|
[Hadoop batch ingestion](./hadoop.html) for more information. Hadoop batch-ingestion can be used for reindexing and delta ingestion.
|
|
|
|
Druid uses an `inputSpec` in the `ioConfig` to know where the data to be ingested is located and how to read it.
|
|
For simple Hadoop batch ingestion, `static` or `granularity` spec types allow you to read data stored in deep storage.
|
|
|
|
There are other types of `inputSpec` to enable reindexing and delta ingestion.
|
|
|
|
#### `dataSource`
|
|
|
|
This is a type of `inputSpec` that reads data already stored inside Druid. This is used to allow "re-indexing" data and for "delta-ingestion" described later in `multi` type inputSpec.
|
|
|
|
|Field|Type|Description|Required|
|
|
|-----|----|-----------|--------|
|
|
|type|String.|This should always be 'dataSource'.|yes|
|
|
|ingestionSpec|JSON object.|Specification of Druid segments to be loaded. See below.|yes|
|
|
|maxSplitSize|Number|Enables combining multiple segments into single Hadoop InputSplit according to size of segments. With -1, druid calculates max split size based on user specified number of map task(mapred.map.tasks or mapreduce.job.maps). By default, one split is made for one segment. maxSplitSize is specified in bytes.|no|
|
|
|useNewAggs|Boolean|If "false", then list of aggregators in "metricsSpec" of hadoop indexing task must be same as that used in original indexing task while ingesting raw data. Default value is "false". This field can be set to "true" when "inputSpec" type is "dataSource" and not "multi" to enable arbitrary aggregators while reindexing. See below for "multi" type support for delta-ingestion.|no|
|
|
|
|
Here is what goes inside `ingestionSpec`:
|
|
|
|
|Field|Type|Description|Required|
|
|
|-----|----|-----------|--------|
|
|
|dataSource|String|Druid dataSource name from which you are loading the data.|yes|
|
|
|intervals|List|A list of strings representing ISO-8601 Intervals.|yes|
|
|
|segments|List|List of segments from which to read data from, by default it is obtained automatically. You can obtain list of segments to put here by making a POST query to coordinator at url /druid/coordinator/v1/metadata/datasources/segments?full with list of intervals specified in the request paylod e.g. ["2012-01-01T00:00:00.000/2012-01-03T00:00:00.000", "2012-01-05T00:00:00.000/2012-01-07T00:00:00.000"]. You may want to provide this list manually in order to ensure that segments read are exactly same as they were at the time of task submission, task would fail if the list provided by the user does not match with state of database when the task actually runs.|no|
|
|
|filter|JSON|See [Filters](../querying/filters.html)|no|
|
|
|dimensions|Array of String|Name of dimension columns to load. By default, the list will be constructed from parseSpec. If parseSpec does not have an explicit list of dimensions then all the dimension columns present in stored data will be read.|no|
|
|
|metrics|Array of String|Name of metric columns to load. By default, the list will be constructed from the "name" of all the configured aggregators.|no|
|
|
|ignoreWhenNoSegments|boolean|Whether to ignore this ingestionSpec if no segments were found. Default behavior is to throw error when no segments were found.|no|
|
|
|
|
For example
|
|
|
|
```json
|
|
"ioConfig" : {
|
|
"type" : "hadoop",
|
|
"inputSpec" : {
|
|
"type" : "dataSource",
|
|
"ingestionSpec" : {
|
|
"dataSource": "wikipedia",
|
|
"intervals": ["2014-10-20T00:00:00Z/P2W"]
|
|
}
|
|
},
|
|
...
|
|
}
|
|
```
|
|
|
|
#### `multi`
|
|
|
|
This is a composing inputSpec to combine other inputSpecs. This inputSpec is used for delta ingestion. You can also use a `multi` inputSpec to combine data from multiple dataSources. However, each particular dataSource can only be specified one time.
|
|
Note that, "useNewAggs" must be set to default value false to support delta-ingestion.
|
|
|
|
|Field|Type|Description|Required|
|
|
|-----|----|-----------|--------|
|
|
|children|Array of JSON objects|List of JSON objects containing other inputSpecs.|yes|
|
|
|
|
For example:
|
|
|
|
```json
|
|
"ioConfig" : {
|
|
"type" : "hadoop",
|
|
"inputSpec" : {
|
|
"type" : "multi",
|
|
"children": [
|
|
{
|
|
"type" : "dataSource",
|
|
"ingestionSpec" : {
|
|
"dataSource": "wikipedia",
|
|
"intervals": ["2012-01-01T00:00:00.000/2012-01-03T00:00:00.000", "2012-01-05T00:00:00.000/2012-01-07T00:00:00.000"],
|
|
"segments": [
|
|
{
|
|
"dataSource": "test1",
|
|
"interval": "2012-01-01T00:00:00.000/2012-01-03T00:00:00.000",
|
|
"version": "v2",
|
|
"loadSpec": {
|
|
"type": "local",
|
|
"path": "/tmp/index1.zip"
|
|
},
|
|
"dimensions": "host",
|
|
"metrics": "visited_sum,unique_hosts",
|
|
"shardSpec": {
|
|
"type": "none"
|
|
},
|
|
"binaryVersion": 9,
|
|
"size": 2,
|
|
"identifier": "test1_2000-01-01T00:00:00.000Z_3000-01-01T00:00:00.000Z_v2"
|
|
}
|
|
]
|
|
}
|
|
},
|
|
{
|
|
"type" : "static",
|
|
"paths": "/path/to/more/wikipedia/data/"
|
|
}
|
|
]
|
|
},
|
|
...
|
|
}
|
|
```
|
|
|
|
It is STRONGLY RECOMMENDED to provide list of segments in `dataSource` inputSpec explicitly so that your delta ingestion task is idempotent. You can obtain that list of segments by making following call to the coordinator.
|
|
POST `/druid/coordinator/v1/metadata/datasources/{dataSourceName}/segments?full`
|
|
Request Body: [interval1, interval2,...] for example ["2012-01-01T00:00:00.000/2012-01-03T00:00:00.000", "2012-01-05T00:00:00.000/2012-01-07T00:00:00.000"]
|
|
|
|
|
|
### Reindexing with Native Batch Ingestion
|
|
|
|
This section assumes the reader understands how to do batch ingestion without Hadoop using [Native Batch Indexing](../ingestion/native_tasks.html),
|
|
which uses a "firehose" to know where and how to read the input data. [IngestSegmentFirehose](firehose.html#ingestsegmentfirehose)
|
|
can be used to read data from segments inside Druid. Note that IndexTask is to be used for prototyping purposes only as
|
|
it has to do all processing inside a single process and can't scale. Please use Hadoop batch ingestion for production
|
|
scenarios dealing with more than 1GB of data.
|