clean up for extensions docs

This commit is contained in:
fjy 2016-03-30 17:14:58 -07:00
parent 595d359c3b
commit 14dbc431ef
9 changed files with 23 additions and 1 deletions

View File

@ -4,6 +4,8 @@ layout: doc_page
# Microsoft Azure # Microsoft Azure
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-azure-extensions` extension.
## Deep Storage ## Deep Storage
[Microsoft Azure Storage](http://azure.microsoft.com/en-us/services/storage/) is another option for deep storage. This requires some additional druid configuration. [Microsoft Azure Storage](http://azure.microsoft.com/en-us/services/storage/) is another option for deep storage. This requires some additional druid configuration.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Apache Cassandra # Apache Cassandra
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-cassandra-storage` extension.
[Apache Cassandra](http://www.datastax.com/what-we-offer/products-services/datastax-enterprise/apache-cassandra) can also [Apache Cassandra](http://www.datastax.com/what-we-offer/products-services/datastax-enterprise/apache-cassandra) can also
be leveraged for deep storage. This requires some additional druid configuration as well as setting up the necessary be leveraged for deep storage. This requires some additional druid configuration as well as setting up the necessary
schema within a Cassandra keystore. schema within a Cassandra keystore.

View File

@ -4,6 +4,10 @@ layout: doc_page
# DistinctCount aggregator # DistinctCount aggregator
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-distinctcount` extension.
Additionally, follow these steps:
(1) First use single dimension hash-based partitioning to partition data by a dimension for example visitor_id, this to make sure all rows with a particular value for that dimension will go into the same segment or this might over count. (1) First use single dimension hash-based partitioning to partition data by a dimension for example visitor_id, this to make sure all rows with a particular value for that dimension will go into the same segment or this might over count.
(2) Second use distinctCount to calculate exact distinct count, make sure queryGranularity is divide exactly by segmentGranularity or else the result will be wrong. (2) Second use distinctCount to calculate exact distinct count, make sure queryGranularity is divide exactly by segmentGranularity or else the result will be wrong.
There is some limitations, when use with groupBy, the groupBy keys' numbers should not exceed maxIntermediateRows in every segment, if exceed the result will wrong. And when use with topN, numValuesPerPass should not too big, if too big the distinctCount will use many memory and cause the JVM out of service. There is some limitations, when use with groupBy, the groupBy keys' numbers should not exceed maxIntermediateRows in every segment, if exceed the result will wrong. And when use with topN, numValuesPerPass should not too big, if too big the distinctCount will use many memory and cause the JVM out of service.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Graphite Emitter # Graphite Emitter
To use this extension, make sure to [include](../../operations/including-extensions.html) `graphite-emitter` extension.
## Introduction ## Introduction
This extension emits druid metrics to a graphite carbon server. This extension emits druid metrics to a graphite carbon server.

View File

@ -4,6 +4,8 @@ layout: doc_page
# Kafka Simple Consumer # Kafka Simple Consumer
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-kafka-eight-simpleConsumer` extension.
## Firehose ## Firehose
This is an experimental firehose to ingest data from kafka using kafka simple consumer api. Currently, this firehose would only work inside standalone realtime nodes. This is an experimental firehose to ingest data from kafka using kafka simple consumer api. Currently, this firehose would only work inside standalone realtime nodes.

View File

@ -1,5 +1,11 @@
---
layout: doc_page
---
# Parquet # Parquet
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-avro-extensions` and `druid-parquet-extensions`.
This extension enables Druid to ingest and understand the Apache Parquet data format offline. This extension enables Druid to ingest and understand the Apache Parquet data format offline.
## Parquet Hadoop Parser ## Parquet Hadoop Parser

View File

@ -4,6 +4,8 @@ layout: doc_page
# RabbitMQ # RabbitMQ
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rabbitmq` extension.
## Firehose ## Firehose
#### RabbitMQFirehose #### RabbitMQFirehose

View File

@ -4,4 +4,6 @@ layout: doc_page
# RocketMQ # RocketMQ
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rocketmq` extension.
Original author: [https://github.com/lizhanhui](https://github.com/lizhanhui). Original author: [https://github.com/lizhanhui](https://github.com/lizhanhui).

View File

@ -45,7 +45,7 @@ If you'd like to take on maintenance for a community extension, please post on [
|druid-cloudfiles-extensions|Rackspace Cloudfiles deep storage and firehose.|[link](../development/extensions-contrib/cloudfiles.html)| |druid-cloudfiles-extensions|Rackspace Cloudfiles deep storage and firehose.|[link](../development/extensions-contrib/cloudfiles.html)|
|druid-distinctcount|DistinctCount aggregator|[link](../development/extensions-contrib/distinctcount.html)| |druid-distinctcount|DistinctCount aggregator|[link](../development/extensions-contrib/distinctcount.html)|
|druid-kafka-eight-simpleConsumer|Kafka ingest firehose (low level consumer).|[link](../development/extensions-contrib/kafka-simple.html)| |druid-kafka-eight-simpleConsumer|Kafka ingest firehose (low level consumer).|[link](../development/extensions-contrib/kafka-simple.html)|
|druid-parquet-extensions|Support for data in Apache Parquet data format.|[link](../development/extensions-contrib/parquet.html)| |druid-parquet-extensions|Support for data in Apache Parquet data format. Requires druid-avro-extensions to be loaded.|[link](../development/extensions-contrib/parquet.html)|
|druid-rabbitmq|RabbitMQ firehose.|[link](../development/extensions-contrib/rabbitmq.html)| |druid-rabbitmq|RabbitMQ firehose.|[link](../development/extensions-contrib/rabbitmq.html)|
|druid-rocketmq|RocketMQ firehose.|[link](../development/extensions-contrib/rocketmq.html)| |druid-rocketmq|RocketMQ firehose.|[link](../development/extensions-contrib/rocketmq.html)|
|graphite-emitter|Graphite metrics emitter|[link](../development/extensions-contrib/graphite.html)| |graphite-emitter|Graphite metrics emitter|[link](../development/extensions-contrib/graphite.html)|