Apache Druid: a high performance real-time analytics database.
Go to file
Jihoon Son 0cc9eb4903
Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288)
* Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided

* query context

* fix tests; add more test

* javadoc

* docs and more tests

* remove default and hadoop tests

* consistent name and fix javadoc

* spelling and field name

* default function for partitionsSpec

* other comments

* address comments

* fix tests and spelling

* test

* doc
2020-09-24 16:32:56 -07:00
.github Make stale bot less aggressive (#10261) 2020-08-10 20:59:02 -07:00
.idea IntelliJ inspection and checkstyle rule for "Collection.EMPTY_* field accesses replaceable with Collections.empty*()" (#9690) 2020-06-18 09:47:07 -07:00
benchmarks vectorized expressions and expression virtual columns (#10401) 2020-09-23 13:56:38 -07:00
cloud bump version to 0.20.0-SNAPSHOT (#10124) 2020-07-06 15:08:32 -07:00
codestyle Add "offset" parameter to the Scan query. (#10233) 2020-08-13 14:56:24 -07:00
core Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
dev Set default server.maxsize to the sum of segment cache (#10255) 2020-08-10 09:21:22 -07:00
distribution Upgrade ORC to 1.5.10 version (#10291) 2020-09-18 13:38:45 -07:00
docs Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
examples Support combining inputsource for parallel ingestion (#10387) 2020-09-15 16:25:35 -07:00
extendedset bump version to 0.20.0-SNAPSHOT (#10124) 2020-07-06 15:08:32 -07:00
extensions-contrib Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
extensions-core vectorized expressions and expression virtual columns (#10401) 2020-09-23 13:56:38 -07:00
hll bump version to 0.20.0-SNAPSHOT (#10124) 2020-07-06 15:08:32 -07:00
hooks Add git pre-commit hook to source control (#9554) 2020-06-05 11:19:42 -10:00
indexing-hadoop Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
indexing-service Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
integration-tests Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
licenses Fix for [CVE-2020-1958]: Apache Druid LDAP injection vulnerability (#9600) 2020-04-01 14:52:01 -07:00
processing Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
publications De-incubation cleanup in code, docs, packaging (#9108) 2020-01-03 12:33:19 -05:00
server Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
services Add last_compaction_state to sys.segments table (#10413) 2020-09-23 15:29:36 -07:00
sql Add last_compaction_state to sys.segments table (#10413) 2020-09-23 15:29:36 -07:00
web-console better query view initial state (#10431) 2020-09-24 09:49:58 -07:00
website Store hash partition function in dataSegment and allow segment pruning only when hash partition function is provided (#10288) 2020-09-24 16:32:56 -07:00
.asf.yaml Add .asf.yaml. (#9083) 2019-12-20 16:45:38 -08:00
.backportrc.json Add 0.18.0 to .backportrc.json to facilitate backport. (#9661) 2020-04-11 13:49:04 -07:00
.codecov.yml Use Codecov (#8388) 2019-08-28 08:49:30 -07:00
.dockerignore Add docker container for druid (#6896) 2019-02-08 12:12:28 +00:00
.gitignore Web console basic end-to-end-test (#9595) 2020-04-09 12:38:09 -07:00
.lgtm.yml Suppress LGTM warnings about stack trace exposure (#9631) 2020-04-09 17:31:03 -07:00
.travis.yml Integration tests and docs for auto compaction with different partitioning (#10354) 2020-09-15 11:28:09 -07:00
CONTRIBUTING.md Fix numbered list formatting in markdown. (#9664) 2020-04-21 20:18:12 -07:00
LABELS Add plain text README.txt, use relative link from README.md to build.md (#7611) 2019-05-09 21:29:26 -07:00
LICENSE support Aliyun OSS service as deep storage (#9898) 2020-07-01 22:20:53 -07:00
NOTICE Fixed the Copyright year of Druid (#9859) 2020-05-20 21:13:34 -07:00
README.md add link to Docker quickstart in github README (#10299) 2020-09-02 01:17:34 -07:00
README.template De-incubation cleanup in code, docs, packaging (#9108) 2020-01-03 12:33:19 -05:00
licenses.yaml Web console: fix lookup edit dialog, allow column renaming (#10406) 2020-09-20 14:10:05 -07:00
owasp-dependency-check-suppressions.xml Ignore CVEs from htrace and ambari transitive deps (#10353) 2020-09-04 15:22:26 -07:00
pom.xml Upgrade ORC to 1.5.10 version (#10291) 2020-09-18 13:38:45 -07:00
setup-hooks.sh Add git pre-commit hook to source control (#9554) 2020-06-05 11:19:42 -10:00
upload.sh Adding licenses and enable apache-rat-plugin. (#6215) 2018-09-18 08:39:26 -07:00

README.md

Slack Build Status Language grade: Java Coverage Status Docker


Website | Documentation | Developer Mailing List | User Mailing List | Slack | Twitter | Download


Apache Druid

Druid is a high performance real-time analytics database. Druid's main value add is to reduce time to insight and action.

Druid is designed for workflows where fast queries and ingest really matter. Druid excels at powering UIs, running operational (ad-hoc) queries, or handling high concurrency. Consider Druid as an open source alternative to data warehouses for a variety of use cases.

Getting started

You can get started with Druid with our local or Docker quickstart.

Druid provides a rich set of APIs (via HTTP and JDBC) for loading, managing, and querying your data. You can also interact with Druid via the built-in console (shown below).

Load data

data loader Kafka

Load streaming and batch data using a point-and-click wizard to guide you through ingestion setup. Monitor one off tasks and ingestion supervisors.

Manage the cluster

management

Manage your cluster with ease. Get a view of your datasources, segments, ingestion tasks, and services from one convenient location. All powered by SQL systems tables, allowing you to see the underlying query for each view.

Issue queries

query view combo

Use the built-in query workbench to prototype DruidSQL and native queries or connect one of the many tools that help you make the most out of Druid.

Documentation

You can find the documentation for the latest Druid release on the project website.

If you would like to contribute documentation, please do so under /docs in this repository and submit a pull request.

Community

Community support is available on the druid-user mailing list, which is hosted at Google Groups.

Development discussions occur on dev@druid.apache.org, which you can subscribe to by emailing dev-subscribe@druid.apache.org.

Chat with Druid committers and users in real-time on the #druid channel in the Apache Slack team. Please use this invitation link to join the ASF Slack, and once joined, go into the #druid channel.

Building from source

Please note that JDK 8 is required to build Druid.

For instructions on building Druid from source, see docs/development/build.md

Contributing

Please follow the community guidelines for contributing.

For instructions on setting up IntelliJ dev/intellij-setup.md

License

Apache License, Version 2.0