Merge pull request #2763 from druid-io/b-docs

clean up for extensions docs
This commit is contained in:
Fangjin Yang 2016-03-30 21:53:03 -07:00
commit 2fc5918e69
9 changed files with 23 additions and 1 deletions

View File

@ -4,6 +4,8 @@ layout: doc_page
# Microsoft Azure
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-azure-extensions` extension.
## Deep Storage
[Microsoft Azure Storage](http://azure.microsoft.com/en-us/services/storage/) is another option for deep storage. This requires some additional druid configuration.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Apache Cassandra
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-cassandra-storage` extension.
[Apache Cassandra](http://www.datastax.com/what-we-offer/products-services/datastax-enterprise/apache-cassandra) can also
be leveraged for deep storage. This requires some additional druid configuration as well as setting up the necessary
schema within a Cassandra keystore.

View File

@ -4,6 +4,10 @@ layout: doc_page
# DistinctCount aggregator
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-distinctcount` extension.
Additionally, follow these steps:
(1) First use single dimension hash-based partitioning to partition data by a dimension for example visitor_id, this to make sure all rows with a particular value for that dimension will go into the same segment or this might over count.
(2) Second use distinctCount to calculate exact distinct count, make sure queryGranularity is divide exactly by segmentGranularity or else the result will be wrong.
There is some limitations, when use with groupBy, the groupBy keys' numbers should not exceed maxIntermediateRows in every segment, if exceed the result will wrong. And when use with topN, numValuesPerPass should not too big, if too big the distinctCount will use many memory and cause the JVM out of service.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Graphite Emitter
To use this extension, make sure to [include](../../operations/including-extensions.html) `graphite-emitter` extension.
## Introduction
This extension emits druid metrics to a graphite carbon server.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Kafka Simple Consumer
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-kafka-eight-simpleConsumer` extension.
## Firehose
This is an experimental firehose to ingest data from kafka using kafka simple consumer api. Currently, this firehose would only work inside standalone realtime nodes.

View File

@ -1,5 +1,11 @@
---
layout: doc_page
---
# Parquet
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-avro-extensions` and `druid-parquet-extensions`.
This extension enables Druid to ingest and understand the Apache Parquet data format offline.
## Parquet Hadoop Parser

View File

@ -4,6 +4,8 @@ layout: doc_page
# RabbitMQ
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rabbitmq` extension.
## Firehose
#### RabbitMQFirehose

View File

@ -4,4 +4,6 @@ layout: doc_page
# RocketMQ
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rocketmq` extension.
Original author: [https://github.com/lizhanhui](https://github.com/lizhanhui).

View File

@ -45,7 +45,7 @@ If you'd like to take on maintenance for a community extension, please post on [
|druid-cloudfiles-extensions|Rackspace Cloudfiles deep storage and firehose.|[link](../development/extensions-contrib/cloudfiles.html)|
|druid-distinctcount|DistinctCount aggregator|[link](../development/extensions-contrib/distinctcount.html)|
|druid-kafka-eight-simpleConsumer|Kafka ingest firehose (low level consumer).|[link](../development/extensions-contrib/kafka-simple.html)|
|druid-parquet-extensions|Support for data in Apache Parquet data format.|[link](../development/extensions-contrib/parquet.html)|
|druid-parquet-extensions|Support for data in Apache Parquet data format. Requires druid-avro-extensions to be loaded.|[link](../development/extensions-contrib/parquet.html)|
|druid-rabbitmq|RabbitMQ firehose.|[link](../development/extensions-contrib/rabbitmq.html)|
|druid-rocketmq|RocketMQ firehose.|[link](../development/extensions-contrib/rocketmq.html)|
|graphite-emitter|Graphite metrics emitter|[link](../development/extensions-contrib/graphite.html)|