mirror of https://github.com/apache/druid.git
clean up for extensions docs
This commit is contained in:
parent
595d359c3b
commit
14dbc431ef
|
@ -4,6 +4,8 @@ layout: doc_page
|
||||||
|
|
||||||
# Microsoft Azure
|
# Microsoft Azure
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-azure-extensions` extension.
|
||||||
|
|
||||||
## Deep Storage
|
## Deep Storage
|
||||||
|
|
||||||
[Microsoft Azure Storage](http://azure.microsoft.com/en-us/services/storage/) is another option for deep storage. This requires some additional druid configuration.
|
[Microsoft Azure Storage](http://azure.microsoft.com/en-us/services/storage/) is another option for deep storage. This requires some additional druid configuration.
|
||||||
|
|
|
@ -4,6 +4,8 @@ layout: doc_page
|
||||||
|
|
||||||
# Apache Cassandra
|
# Apache Cassandra
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-cassandra-storage` extension.
|
||||||
|
|
||||||
[Apache Cassandra](http://www.datastax.com/what-we-offer/products-services/datastax-enterprise/apache-cassandra) can also
|
[Apache Cassandra](http://www.datastax.com/what-we-offer/products-services/datastax-enterprise/apache-cassandra) can also
|
||||||
be leveraged for deep storage. This requires some additional druid configuration as well as setting up the necessary
|
be leveraged for deep storage. This requires some additional druid configuration as well as setting up the necessary
|
||||||
schema within a Cassandra keystore.
|
schema within a Cassandra keystore.
|
||||||
|
|
|
@ -4,6 +4,10 @@ layout: doc_page
|
||||||
|
|
||||||
# DistinctCount aggregator
|
# DistinctCount aggregator
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-distinctcount` extension.
|
||||||
|
|
||||||
|
Additionally, follow these steps:
|
||||||
|
|
||||||
(1) First use single dimension hash-based partitioning to partition data by a dimension for example visitor_id, this to make sure all rows with a particular value for that dimension will go into the same segment or this might over count.
|
(1) First use single dimension hash-based partitioning to partition data by a dimension for example visitor_id, this to make sure all rows with a particular value for that dimension will go into the same segment or this might over count.
|
||||||
(2) Second use distinctCount to calculate exact distinct count, make sure queryGranularity is divide exactly by segmentGranularity or else the result will be wrong.
|
(2) Second use distinctCount to calculate exact distinct count, make sure queryGranularity is divide exactly by segmentGranularity or else the result will be wrong.
|
||||||
There is some limitations, when use with groupBy, the groupBy keys' numbers should not exceed maxIntermediateRows in every segment, if exceed the result will wrong. And when use with topN, numValuesPerPass should not too big, if too big the distinctCount will use many memory and cause the JVM out of service.
|
There is some limitations, when use with groupBy, the groupBy keys' numbers should not exceed maxIntermediateRows in every segment, if exceed the result will wrong. And when use with topN, numValuesPerPass should not too big, if too big the distinctCount will use many memory and cause the JVM out of service.
|
||||||
|
|
|
@ -4,6 +4,8 @@ layout: doc_page
|
||||||
|
|
||||||
# Graphite Emitter
|
# Graphite Emitter
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `graphite-emitter` extension.
|
||||||
|
|
||||||
## Introduction
|
## Introduction
|
||||||
|
|
||||||
This extension emits druid metrics to a graphite carbon server.
|
This extension emits druid metrics to a graphite carbon server.
|
||||||
|
|
|
@ -4,6 +4,8 @@ layout: doc_page
|
||||||
|
|
||||||
# Kafka Simple Consumer
|
# Kafka Simple Consumer
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-kafka-eight-simpleConsumer` extension.
|
||||||
|
|
||||||
## Firehose
|
## Firehose
|
||||||
|
|
||||||
This is an experimental firehose to ingest data from kafka using kafka simple consumer api. Currently, this firehose would only work inside standalone realtime nodes.
|
This is an experimental firehose to ingest data from kafka using kafka simple consumer api. Currently, this firehose would only work inside standalone realtime nodes.
|
||||||
|
|
|
@ -1,5 +1,11 @@
|
||||||
|
---
|
||||||
|
layout: doc_page
|
||||||
|
---
|
||||||
|
|
||||||
# Parquet
|
# Parquet
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-avro-extensions` and `druid-parquet-extensions`.
|
||||||
|
|
||||||
This extension enables Druid to ingest and understand the Apache Parquet data format offline.
|
This extension enables Druid to ingest and understand the Apache Parquet data format offline.
|
||||||
|
|
||||||
## Parquet Hadoop Parser
|
## Parquet Hadoop Parser
|
||||||
|
|
|
@ -4,6 +4,8 @@ layout: doc_page
|
||||||
|
|
||||||
# RabbitMQ
|
# RabbitMQ
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rabbitmq` extension.
|
||||||
|
|
||||||
## Firehose
|
## Firehose
|
||||||
|
|
||||||
#### RabbitMQFirehose
|
#### RabbitMQFirehose
|
||||||
|
|
|
@ -4,4 +4,6 @@ layout: doc_page
|
||||||
|
|
||||||
# RocketMQ
|
# RocketMQ
|
||||||
|
|
||||||
|
To use this extension, make sure to [include](../../operations/including-extensions.html) `druid-rocketmq` extension.
|
||||||
|
|
||||||
Original author: [https://github.com/lizhanhui](https://github.com/lizhanhui).
|
Original author: [https://github.com/lizhanhui](https://github.com/lizhanhui).
|
||||||
|
|
|
@ -45,7 +45,7 @@ If you'd like to take on maintenance for a community extension, please post on [
|
||||||
|druid-cloudfiles-extensions|Rackspace Cloudfiles deep storage and firehose.|[link](../development/extensions-contrib/cloudfiles.html)|
|
|druid-cloudfiles-extensions|Rackspace Cloudfiles deep storage and firehose.|[link](../development/extensions-contrib/cloudfiles.html)|
|
||||||
|druid-distinctcount|DistinctCount aggregator|[link](../development/extensions-contrib/distinctcount.html)|
|
|druid-distinctcount|DistinctCount aggregator|[link](../development/extensions-contrib/distinctcount.html)|
|
||||||
|druid-kafka-eight-simpleConsumer|Kafka ingest firehose (low level consumer).|[link](../development/extensions-contrib/kafka-simple.html)|
|
|druid-kafka-eight-simpleConsumer|Kafka ingest firehose (low level consumer).|[link](../development/extensions-contrib/kafka-simple.html)|
|
||||||
|druid-parquet-extensions|Support for data in Apache Parquet data format.|[link](../development/extensions-contrib/parquet.html)|
|
|druid-parquet-extensions|Support for data in Apache Parquet data format. Requires druid-avro-extensions to be loaded.|[link](../development/extensions-contrib/parquet.html)|
|
||||||
|druid-rabbitmq|RabbitMQ firehose.|[link](../development/extensions-contrib/rabbitmq.html)|
|
|druid-rabbitmq|RabbitMQ firehose.|[link](../development/extensions-contrib/rabbitmq.html)|
|
||||||
|druid-rocketmq|RocketMQ firehose.|[link](../development/extensions-contrib/rocketmq.html)|
|
|druid-rocketmq|RocketMQ firehose.|[link](../development/extensions-contrib/rocketmq.html)|
|
||||||
|graphite-emitter|Graphite metrics emitter|[link](../development/extensions-contrib/graphite.html)|
|
|graphite-emitter|Graphite metrics emitter|[link](../development/extensions-contrib/graphite.html)|
|
||||||
|
|
Loading…
Reference in New Issue