Updated docs (#12978)

This commit is contained in:
Jill Osborne 2022-08-26 08:08:40 +01:00 committed by GitHub
parent eb3a12ac7d
commit ca2b540693
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
8 changed files with 2 additions and 26 deletions

View File

@ -22,10 +22,6 @@ title: "Router Process"
~ under the License.
-->
> The Router is an optional and [experimental](../development/experimental.md) feature due to the fact that its recommended place in the Druid cluster architecture is still evolving.
> However, it has been battle-tested in production, and it hosts the powerful [Druid console](../operations/druid-console.md), so you should feel safe deploying it.
The Apache Druid Router process can be used to route queries to different Broker processes. By default, the broker routes queries based on how [Rules](../operations/rule-configuration.md) are set up. For example, if 1 month of recent data is loaded into a `hot` cluster, queries that fall within the recent month can be routed to a dedicated set of brokers. Queries outside this range are routed to another set of brokers. This set up provides query isolation such that queries for more important data are not impacted by queries for less important data.
For query routing purposes, you should only ever need the Router process if you have a Druid cluster well into the terabyte range.

View File

@ -43,13 +43,7 @@ to compute approximate histograms, with the following modifications:
increasing accuracy when there are few data points, or when dealing with
discrete data points. You can find some of the details in [this post](https://metamarkets.com/2013/histograms/).
Approximate histogram sketches are still experimental for a reason, and you
should understand the limitations of the current implementation before using
them. The approximation is heavily data-dependent, which makes it difficult to
give good general guidelines, so you should experiment and see what parameters
work well for your data.
Here are a few things to note before using them:
Here are a few things to note before using approximate histograms:
- As indicated in the original paper, there are no formal error bounds on the
approximation. In practice, the approximation gets worse if the distribution

View File

@ -22,8 +22,6 @@ title: "Apache Kafka Lookups"
~ under the License.
-->
> Lookups are an [experimental](../experimental.md) feature.
To use this Apache Druid extension, [include](../../development/extensions.md#loading-extensions) `druid-lookups-cached-global` and `druid-kafka-extraction-namespace` in the extensions load list.
If you need updates to populate as promptly as possible, it is possible to plug into a Kafka topic whose key is the old value and message is the desired new value (both in UTF-8) as a LookupExtractorFactory.

View File

@ -134,8 +134,6 @@ If you want to ingest data from other fields in addition to the Kafka message co
- the Kafka event timestamp
- the Kafka event value that stores the payload.
> The Kafka inputFormat is currently designated as experimental.
For example, consider the following structure for a message that represents a fictitious wiki edit in a development environment:
- **Event headers**: {"environment": "development"}
- **Event key**: {"key: "wiki-edit"}

View File

@ -22,9 +22,6 @@ title: "Globally Cached Lookups"
~ under the License.
-->
> Lookups are an [experimental](../experimental.md) feature.
To use this Apache Druid extension, [include](../extensions.md#loading-extensions) `druid-lookups-cached-global` in the extensions load list.
## Configuration

View File

@ -155,8 +155,6 @@ Be sure to change the `delimiter` to the appropriate delimiter for your data. Li
Configure the Kafka `inputFormat` to load complete kafka records including header, key, and value.
> That Kafka `inputFormat` is currently designated as experimental.
| Field | Type | Description | Required |
|-------|------|-------------|----------|
| type | String | Set value to `kafka`. | yes |

View File

@ -393,9 +393,7 @@ them to create the final segments. Finally, they push the final segments to the
#### Multi-dimension range partitioning
> Multiple dimension (multi-dimension) range partitioning is an experimental feature.
> Multi-dimension range partitioning is not supported in the sequential mode of the
> `index_parallel` task type.
> Multi-dimension range partitioning is not supported in the sequential mode of the `index_parallel` task type.
Range partitioning has [several benefits](#benefits-of-range-partitioning) related to storage footprint and query
performance. Multi-dimension range partitioning improves over single-dimension range partitioning by allowing

View File

@ -22,9 +22,6 @@ title: "Lookups"
~ under the License.
-->
> Lookups are an [experimental](../development/experimental.md) feature.
Lookups are a concept in Apache Druid where dimension values are (optionally) replaced with new values, allowing join-like
functionality. Applying lookups in Druid is similar to joining a dimension table in a data warehouse. See
[dimension specs](../querying/dimensionspecs.md) for more information. For the purpose of these documents, a "key"