Apache Druid: a high performance real-time analytics database.
Go to file
Fangjin Yang edb0eca3a9 fix docs (#3370) 2016-08-16 16:25:50 -07:00
api Add timestampSpec to metadata.drd and SegmentMetadataQuery (#3227) 2016-07-25 15:45:30 -07:00
aws-common Update master version to 0.9.2-SNAPSHOT. (#3133) 2016-06-13 13:10:38 -07:00
benchmarks Restore optimizations in BoundFilter. (#3343) 2016-08-10 08:53:17 -07:00
common Rename fields in OrderedMergeIterator (#3149) 2016-08-11 09:42:12 -07:00
distribution Support variance and standard deviation (#2525) 2016-08-04 17:32:58 -07:00
docs fix docs (#3370) 2016-08-16 16:25:50 -07:00
examples Quickstart: Use hadoopyString for batch indexing instead of string. (#3263) 2016-07-19 10:18:10 -07:00
extensions-contrib fix can't get latest offset in KafkaEightSimpleConsumerFirehoseFactory (#3355) 2016-08-11 18:00:24 -07:00
extensions-core allow registrants to opt out of announcing themselves when registering as a chat handler (#3360) 2016-08-16 10:51:28 +05:30
indexing-hadoop Fix issue #2707 (#2708) 2016-08-16 12:19:44 -05:00
indexing-service FileTaskLogsTest: Throw unthrown exception. (#3352) 2016-08-11 09:40:28 -07:00
integration-tests fixing expected result for segmentMetadata query in integration tests (#3318) 2016-08-03 12:13:27 -07:00
processing Adding filters for TimeBoundary on backend (#3168) 2016-08-15 10:25:24 -07:00
publications Support min/max values for metadata query (#2208) 2016-02-12 09:35:58 +09:00
server fix old usage of dimension as string instead of dimensionSchema in DataSchema (#3365) 2016-08-16 09:58:04 -07:00
services Mask properties from logging (#3332) 2016-08-08 21:36:10 +05:30
.gitignore move distribution artifacts to distribution/target 2015-10-30 12:40:05 -05:00
.travis.yml Disable cobertura travis portion (#3122) 2016-06-13 12:27:35 +05:30
CONTRIBUTING.md Add doc link to eclipse formatting settings as well (#3131) 2016-06-24 15:27:50 -07:00
DruidCorporateCLA.pdf fix CLA email / mailing address 2014-04-17 15:26:28 -07:00
DruidIndividualCLA.pdf fix CLA email / mailing address 2014-04-17 15:26:28 -07:00
LICENSE Clean up README and license 2015-02-18 23:09:28 -08:00
NOTICE Add variance aggregator from hive to NOTICE (#3327) 2016-08-04 17:43:55 -07:00
README.md update readme (#2830) 2016-04-13 11:33:31 -07:00
druid_intellij_formatting.xml Make formatting IntelliJ 2016 friendly (#2978) 2016-05-18 12:42:21 -07:00
eclipse.importorder Merge pull request #2905 from javasoze/eclipse_formatting 2016-04-29 18:42:03 -07:00
eclipse_formatting.xml Merge pull request #2905 from javasoze/eclipse_formatting 2016-04-29 18:42:03 -07:00
pom.xml Update java-util to 0.27.10. (#3337) 2016-08-09 13:37:30 +05:30
upload.sh upload.sh: Use awscli if s3cmd is not available. (#3114) 2016-06-08 17:01:46 -07:00

README.md

Build Status Coverage Status

Druid

Druid is a distributed, column-oriented, real-time analytics data store that is commonly used to power exploratory dashboards in multi-tenant environments.

Druid excels as a data warehousing solution for fast aggregate queries on petabyte sized data sets. Druid supports a variety of flexible filters, exact calculations, approximate algorithms, and other useful calculations.

Druid can load both streaming and batch data and integrates with Samza, Kafka, Storm, Spark, and Hadoop.

License

Apache License, Version 2.0

More Information

More information about Druid can be found on http://www.druid.io.

Documentation

You can find the documentation for the latest Druid release on the project website.

If you would like to contribute documentation, please do so under /docs/content in this repository and submit a pull request.

Getting Started

You can get started with Druid with our quickstart.

Reporting Issues

If you find any bugs, please file a GitHub issue.

Community

Community support is available on the druid-user mailing list(druid-user@googlegroups.com).

Development discussions occur on the druid-development list(druid-development@googlegroups.com).

We also have a couple people hanging out on IRC in #druid-dev on irc.freenode.net.

Contributing

Please follow the guidelines listed here.