20 KiB
layout |
---|
doc_page |
Tutorial: A First Look at Druid
Greetings! This tutorial will help clarify some core Druid concepts. We will use a real-time dataset and issue some basic Druid queries. If you are ready to explore Druid, and learn a thing or two, read on!
About the data
The data source we'll be working with is Wikipedia edits. Each time an edit is made in Wikipedia, an event gets pushed to an IRC channel associated with the language of the Wikipedia page. We scrape IRC channels for several different languages and load this data into Druid.
Each event has a timestamp indicating the time of the edit (in UTC time), a list of dimensions indicating various metadata about the event (such as information about the user editing the page and where the user is a bot), and a list of metrics associated with the event (such as the number of characters added and deleted).
Specifically. the data schema looks like so:
Dimensions (things to filter on):
"page"
"language"
"user"
"unpatrolled"
"newPage"
"robot"
"anonymous"
"namespace"
"continent"
"country"
"region"
"city"
Metrics (things to aggregate over):
"count"
"added"
"delta"
"deleted"
Setting Up
To start, we need to get our hands on a Druid build. There are two ways to get Druid: download a tarball, or Build From Source. You only need to do one of these.
Download a Tarball
We've built a tarball that contains everything you'll need. You'll find it here. Download this file to a directory of your choosing.
Build From Source
Follow the Build From Source guide to build from source. Then grab the tarball from services/target/druid--bin.tar.gz.
Unpack the Tarball
You can extract the content within by issuing:
tar -zxvf druid-<version>-bin.tar.gz
If you cd into the directory:
cd druid-<version>
You should see a bunch of files:
- run_example_server.sh
- run_example_client.sh
- LICENSE, config, examples, lib directories
External Dependencies
Druid requires 3 external dependencies.
- A "deep storage" that acts as a data repository. This is generally distributed storage like HDFS or S3. For prototyping or experimentation on a single machine, Druid can use the local filesystem.
- A "metadata storage" to hold configuration and metadata information. This is generally a small, shared database like MySQL or Postgres. For prototyping or experimentation on a single machine, Druid can use a local instance of Apache Derby.
- Apache Zookeeper for coordination among different pieces of the cluster.
Set up Zookeeper
- Download zookeeper from http://www.apache.org/dyn/closer.cgi/zookeeper/
- Install zookeeper.
curl http://www.gtlib.gatech.edu/pub/apache/zookeeper/zookeeper-3.4.6/zookeeper-3.4.6.tar.gz -o zookeeper-3.4.6.tar.gz
tar xzf zookeeper-3.4.6.tar.gz
cd zookeeper-3.4.6
cp conf/zoo_sample.cfg conf/zoo.cfg
./bin/zkServer.sh start
cd ..
Running Example Scripts
Let's start doing stuff. You can start an example Druid Realtime node by issuing:
./run_example_server.sh
Select the "wikipedia" example.
Note that the first time you start the example, it may take some extra time due to its fetching various dependencies. Once the node starts up you will see a bunch of logs about setting up properties and connecting to the data source. If everything was successful, you should see messages of the form shown below.
2015-02-17T21:46:36,804 INFO [main] org.eclipse.jetty.server.ServerConnector - Started ServerConnector@79b6cf95{HTTP/1.1}{0.0.0.0:8084}
2015-02-17T21:46:36,804 INFO [main] org.eclipse.jetty.server.Server - Started @9580ms
2015-02-17T21:46:36,862 INFO [ApiDaemon] io.druid.segment.realtime.firehose.IrcFirehoseFactory - irc connection to server [irc.wikimedia.org] established
2015-02-17T21:46:36,862 INFO [ApiDaemon] io.druid.segment.realtime.firehose.IrcFirehoseFactory - Joining channel #en.wikipedia
2015-02-17T21:46:36,863 INFO [ApiDaemon] io.druid.segment.realtime.firehose.IrcFirehoseFactory - Joining channel #fr.wikipedia
2015-02-17T21:46:36,863 INFO [ApiDaemon] io.druid.segment.realtime.firehose.IrcFirehoseFactory - Joining channel #de.wikipedia
2015-02-17T21:46:36,863 INFO [ApiDaemon] io.druid.segment.realtime.firehose.IrcFirehoseFactory - Joining channel #ja.wikipedia
2015-02-17T21:46:37,009 INFO [ServerInventoryView-0] io.druid.client.BatchServerInventoryView - Inventory Initialized
The Druid real time-node ingests events in an in-memory buffer. Periodically, these events will be persisted to disk. If you are interested in the details of our real-time architecture and why we persist indexes to disk, we suggest you read our White Paper.
To query the real-time node you've spun up, you can issue:
./run_example_client.sh
Select "wikipedia" once again. This script issues TimeBoundary to the data we've been ingesting. The query looks like this:
{
"queryType":"timeBoundary",
"dataSource":"wikipedia"
}
The timeBoundary query is one of the simplest queries you can make in Druid. It gives you the boundaries of the ingested data.
The result looks something like this (when it's prettified):
[ {
"timestamp" : "2013-09-04T21:44:00.000Z",
"result" : {
"minTime" : "2013-09-04T21:44:00.000Z",
"maxTime" : "2013-09-04T21:47:00.000Z"
}
} ]
If you are having problems with getting results back, make sure you have curl installed. Control+C to break out of the client script.
Querying Druid
In your favorite editor, create the file:
timeseries.json
We are going to make a slightly more complicated query, the TimeseriesQuery. Copy and paste the following into the file:
{
"queryType": "timeseries",
"dataSource": "wikipedia",
"intervals": [ "2010-01-01/2020-01-01" ],
"granularity": "all",
"aggregations": [
{"type": "longSum", "fieldName": "count", "name": "edit_count"},
{"type": "doubleSum", "fieldName": "added", "name": "chars_added"}
]
}
Our query has now expanded to include a time interval, Granularities, and Aggregations. What the query is doing is aggregating a set of metrics over a span of time, and the results are grouped into a single time bucket. To issue the query and get some results, run the following in your command line:
curl -X POST 'http://localhost:8084/druid/v2/?pretty' -H 'content-type: application/json' -d @timeseries.json
Once again, you should get a JSON blob of text back with your results, that looks something like this:
[ {
"timestamp" : "2013-09-04T21:44:00.000Z",
"result" : { "chars_added" : 312670.0, "edit_count" : 733 }
} ]
If you issue the query again, you should notice your results updating.
Right now all the results you are getting back are being aggregated into a single timestamp bucket. What if we wanted to see our aggregations on a per minute basis?
We can change granularity for the results to "minute". To specify different granularities to bucket our results, we change our query like so:
{
"queryType": "timeseries",
"dataSource": "wikipedia",
"intervals": [ "2010-01-01/2020-01-01" ],
"granularity": "minute",
"aggregations": [
{"type": "longSum", "fieldName": "count", "name": "edit_count"},
{"type": "doubleSum", "fieldName": "added", "name": "chars_added"}
]
}
This gives us results like the following:
[
{
"timestamp" : "2013-09-04T21:44:00.000Z",
"result" : { "chars_added" : 30665.0, "edit_count" : 128 }
},
{
"timestamp" : "2013-09-04T21:45:00.000Z",
"result" : { "chars_added" : 122637.0, "edit_count" : 167 }
},
{
"timestamp" : "2013-09-04T21:46:00.000Z",
"result" : { "chars_added" : 78938.0, "edit_count" : 159 }
},
...
]
Solving a Problem
One of Druid's main powers is to provide answers to problems, so let's pose a problem. What if we wanted to know what the top pages in the US are, ordered by the number of edits over the last few minutes you've been going through this tutorial? To solve this problem, we can use the TopN.
Let's create the file:
topn.json
and put the following in there:
{
"queryType": "topN",
"dataSource": "wikipedia",
"granularity": "all",
"dimension": "page",
"metric": "edit_count",
"threshold" : 10,
"aggregations": [
{"type": "longSum", "fieldName": "count", "name": "edit_count"}
],
"filter": { "type": "selector", "dimension": "country", "value": "United States" },
"intervals": ["2012-10-01T00:00/2020-01-01T00"]
}
Note that our query now includes Filters. Filters are like WHERE
clauses in SQL and help narrow down the data that needs to be scanned.
If you issue the query:
curl -X POST 'http://localhost:8084/druid/v2/?pretty' -H 'content-type: application/json' -d @topn.json
You should see an answer to our question. As an example, some results are shown below:
[
{
"timestamp" : "2013-09-04T21:00:00.000Z",
"result" : [
{ "page" : "RTC_Transit", "edit_count" : 6 },
{ "page" : "List_of_Deadly_Women_episodes", "edit_count" : 4 },
{ "page" : "User_talk:David_Biddulph", "edit_count" : 4 },
...
]
}
]
Feel free to tweak other query parameters to answer other questions you may have about the data. Druid also includes more complex query types such as groupBy queries. For more information on querying, see this link.
Next Steps
This tutorial only covered the basic operations of a single Druid node. For production, you'll likely need a full Druid cluster. Check out our next tutorial The Druid Cluster to learn more.
To learn more about loading streaming data, see Loading Streaming Data.
To learn more about loading batch data, see Loading Batch Data.
What to Do When You Have a Firewall
When you are behind a firewall, the Maven Druid dependencies will not be accessible, as well as the IRC wikipedia channels that feed realtime data into Druid. To workaround those two challenges, you will need to:
- make the Maven Druid dependencies available offline
- make the Wikipedia example GeoLite DB dependency available offline
- use an alternative approach to load data into Druid
Making Maven Druid Dependencies Available Offline
-
Extract Druid to a machine that has internet access; e.g.
C:\druid-<version>
-
Create a repository directory to download the dependencies to; e.g.
C:\druid-<version>\repo
-
Create property
druid.extensions.localRepository=
path to repo directory
in theDruid Directory
\config\_common/common.runtime.properties
file; e.g.druid.extensions.localRepository=C\:/druid-<version>/repo
-
From within Druid directory, run the
pull-deps
command to download all Druid dependencies to the repository specified in thecommon.runtime.properties
file:java -classpath "config\_common;lib\*" io.druid.cli.Main tools pull-deps
-
Once all dependencies have been downloaded successfully, replicate the
repo
directory to the machine behind the firewall; e.g./opt/druid-<version>/repo
-
Create property
druid.extensions.localRepository=
path to repo directory
in theDruid Directory
/config/_common/common.runtime.properties
file; e.g.druid.extensions.localRepository=/opt/druid-<version>/repo
Making the Wikipedia Example GeoLite DB Dependency Available Offline
-
Download GeoLite2 City DB from http://dev.maxmind.com/geoip/geoip2/geolite2/
-
Copy and extract the DB to
java.io.tmpdir
/io.druid.segment.realtime.firehose.WikipediaIrcDecoder.GeoLite2-City.mmdb
; e.g./tmp/io.druid.segment.realtime.firehose.WikipediaIrcDecoder.GeoLite2-City.mmdb
Note: depending on the machine's reboot policy, if the
java.io.tmpdir
resolves to the/tmp
directory, you may have to create this file again in thetmp
directory after a machine reboot
Loading the Data into Druid
As an alternative to reading the data from the IRC channels, which is a challenge to try to do it from behind a firewall, we will use Kafka to stream the data to Druid. To do so, we will need to:
- configure the Wikipedia example to read streaming data from Kafka
- set up and configure Kafka
Wikipedia Example Configuration
-
In your favorite editor, open the file
druid-<version>/examples/wikipedia/wikipedia_realtime.spec
-
Backup the file, if necessary, then replace the file content with the following:
[ { "dataSchema": { "dataSource": "wikipedia", "parser": { "type": "string", "parseSpec": { "format": "json", "timestampSpec": { "column": "timestamp", "format": "auto" }, "dimensionsSpec": { "dimensions": [ "page", "language", "user", "unpatrolled", "newPage", "robot", "anonymous", "namespace", "continent", "country", "region", "city" ], "dimensionExclusions": [], "spatialDimensions": [] } } }, "metricsSpec": [ { "type": "count", "name": "count" }, { "type": "doubleSum", "name": "added", "fieldName": "added" }, { "type": "doubleSum", "name": "deleted", "fieldName": "deleted" }, { "type": "doubleSum", "name": "delta", "fieldName": "delta" } ], "granularitySpec": { "type": "uniform", "segmentGranularity": "DAY", "queryGranularity": "NONE" } }, "ioConfig": { "type": "realtime", "firehose": { "type": "kafka-0.8", "consumerProps": { "zookeeper.connect": "localhost:2181", "zookeeper.connection.timeout.ms": "15000", "zookeeper.session.timeout.ms": "15000", "zookeeper.sync.time.ms": "5000", "group.id": "druid-example", "fetch.message.max.bytes": "1048586", "auto.offset.reset": "largest", "auto.commit.enable": "false" }, "feed": "wikipedia" }, "plumber": { "type": "realtime" } }, "tuningConfig": { "type": "realtime", "maxRowsInMemory": 500000, "intermediatePersistPeriod": "PT10m", "windowPeriod": "PT10m", "basePersistDirectory": "/tmp/realtime/basePersist", "rejectionPolicy": { "type": "messageTime" } } } ]
-
Refer to the Running Example Scripts section to start the example Druid Realtime node by issuing the following from within your Druid directory:
./run_example_server.sh
Kafka Setup and Configuration
-
Download Kafka
For this tutorial we will [download Kafka 0.8.2.1] (https://www.apache.org/dyn/closer.cgi?path=/kafka/0.8.2.1/kafka_2.10-0.8.2.1.tgz)
tar -xzf kafka_2.10-0.8.2.1.tgz cd kafka_2.10-0.8.2.1
-
Start Kafka
First, launch ZooKeeper (refer to the Set up Zookeeper section for details), then start the Kafka server (in a separate console):
./bin/kafka-server-start.sh config/server.properties
-
Create a topic named
wikipedia
./bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic wikipedia
-
Launch a console producer for the topic
wikipedia
./bin/kafka-console-producer.sh --broker-list localhost:9092 --topic wikipedia
-
Copy and paste the following data into the terminal where we launched the Kafka console producer in the previous step:
{"timestamp": "2013-08-31T01:02:33Z", "page": "Gypsy Danger", "language" : "en", "user" : "nuclear", "unpatrolled" : "true", "newPage" : "true", "robot": "false", "anonymous": "false", "namespace":"article", "continent":"North America", "country":"United States", "region":"Bay Area", "city":"San Francisco", "added": 57, "deleted": 200, "delta": -143} {"timestamp": "2013-08-31T03:32:45Z", "page": "Striker Eureka", "language" : "en", "user" : "speed", "unpatrolled" : "false", "newPage" : "true", "robot": "true", "anonymous": "false", "namespace":"wikipedia", "continent":"Australia", "country":"Australia", "region":"Cantebury", "city":"Syndey", "added": 459, "deleted": 129, "delta": 330} {"timestamp": "2013-08-31T07:11:21Z", "page": "Cherno Alpha", "language" : "ru", "user" : "masterYi", "unpatrolled" : "false", "newPage" : "true", "robot": "true", "anonymous": "false", "namespace":"article", "continent":"Asia", "country":"Russia", "region":"Oblast", "city":"Moscow", "added": 123, "deleted": 12, "delta": 111} {"timestamp": "2013-08-31T11:58:39Z", "page": "Crimson Typhoon", "language" : "zh", "user" : "triplets", "unpatrolled" : "true", "newPage" : "false", "robot": "true", "anonymous": "false", "namespace":"wikipedia", "continent":"Asia", "country":"China", "region":"Shanxi", "city":"Taiyuan", "added": 905, "deleted": 5, "delta": 900} {"timestamp": "2013-08-31T12:41:27Z", "page": "Coyote Tango", "language" : "ja", "user" : "stringer", "unpatrolled" : "true", "newPage" : "false", "robot": "true", "anonymous": "false", "namespace":"wikipedia", "continent":"Asia", "country":"Japan", "region":"Kanto", "city":"Tokyo", "added": 1, "deleted": 10, "delta": -9}
Finally
Now, that data has been fed into Druid, refer to the Running Example Scripts section to query the real-time node by issuing the following from within the Druid directory:
./run_example_client.sh
The Querying Druid section also has further querying examples.
Additional Information
This tutorial is merely showcasing a small fraction of what Druid can do. If you are interested in more information about Druid, including setting up a more sophisticated Druid cluster, read more of the Druid documentation and blogs found on druid.io.
Hopefully you learned a thing or two about Druid real-time ingestion, querying Druid, and how Druid can be used to solve problems. If you have additional questions, feel free to post in our google groups page.