mirror of https://github.com/apache/druid.git
Merge pull request #2699 from gianm/docs-fix-broken-links
Fix a bunch of broken links in the docs.
This commit is contained in:
commit
a6af1974f7
|
@ -21,7 +21,7 @@ Derby is not suitable for production use as a metadata store. Use MySQL or Postg
|
||||||
|
|
||||||
```properties
|
```properties
|
||||||
druid.metadata.storage.type=derby
|
druid.metadata.storage.type=derby
|
||||||
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//home/y/var/druid_state/derby;create=true
|
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//opt/var/druid_state/derby;create=true
|
||||||
```
|
```
|
||||||
|
|
||||||
## MySQL
|
## MySQL
|
||||||
|
|
|
@ -239,7 +239,7 @@ classification=yarn-site,properties=[mapreduce.reduce.memory.mb=6144,mapreduce.r
|
||||||
```
|
```
|
||||||
|
|
||||||
- Follow the instructions under "[Configure Hadoop for data
|
- Follow the instructions under "[Configure Hadoop for data
|
||||||
loads](cluster.html#configure-cluster-for-hadoop-data-loads)" using the XML files from
|
loads](../tutorials/cluster.html#configure-cluster-for-hadoop-data-loads)" using the XML files from
|
||||||
`/etc/hadoop/conf` on your EMR master.
|
`/etc/hadoop/conf` on your EMR master.
|
||||||
|
|
||||||
#### Loading from S3 with EMR
|
#### Loading from S3 with EMR
|
||||||
|
@ -269,7 +269,7 @@ Druid works out of the box with many Hadoop distributions.
|
||||||
|
|
||||||
If you are having dependency conflicts between Druid and your version of Hadoop, you can try
|
If you are having dependency conflicts between Druid and your version of Hadoop, you can try
|
||||||
searching for a solution in the [Druid user groups](https://groups.google.com/forum/#!forum/druid-
|
searching for a solution in the [Druid user groups](https://groups.google.com/forum/#!forum/druid-
|
||||||
user), or reading the Druid [Different Hadoop Versions](..//operations/other-hadoop.html) documentation.
|
user), or reading the Druid [Different Hadoop Versions](../operations/other-hadoop.html) documentation.
|
||||||
|
|
||||||
## Command Line Hadoop Indexer
|
## Command Line Hadoop Indexer
|
||||||
|
|
||||||
|
|
|
@ -293,9 +293,6 @@ results.
|
||||||
Is this always a problem? No. If your data is small enough to fit on a single Kafka partition, you can replicate without issues.
|
Is this always a problem? No. If your data is small enough to fit on a single Kafka partition, you can replicate without issues.
|
||||||
Otherwise, you can run real-time nodes without replication.
|
Otherwise, you can run real-time nodes without replication.
|
||||||
|
|
||||||
There is now also an [experimental low level Kafka firehose](../development/kafka-simple-consumer-firehose.html) which
|
|
||||||
solves the issues described above with using the high level Kafka consumer.
|
|
||||||
|
|
||||||
Please note that druid will skip over event that failed its checksum and it is corrupt.
|
Please note that druid will skip over event that failed its checksum and it is corrupt.
|
||||||
|
|
||||||
### Locking
|
### Locking
|
||||||
|
|
|
@ -28,7 +28,7 @@ segments and avoid the overhead of rebuilding new segments with reindexing, you
|
||||||
### Reindexing and Delta Ingestion with Hadoop Batch Ingestion
|
### Reindexing and Delta Ingestion with Hadoop Batch Ingestion
|
||||||
|
|
||||||
This section assumes the reader understands how to do batch ingestion using Hadoop. See
|
This section assumes the reader understands how to do batch ingestion using Hadoop. See
|
||||||
[batch-ingestion](batch-ingestion.md) for more information. Hadoop batch-ingestion can be used for reindexing and delta ingestion.
|
[batch-ingestion](batch-ingestion.html) for more information. Hadoop batch-ingestion can be used for reindexing and delta ingestion.
|
||||||
|
|
||||||
Druid uses an `inputSpec` in the `ioConfig` to know where the data to be ingested is located and how to read it.
|
Druid uses an `inputSpec` in the `ioConfig` to know where the data to be ingested is located and how to read it.
|
||||||
For simple Hadoop batch ingestion, `static` or `granularity` spec types allow you to read data stored in deep storage.
|
For simple Hadoop batch ingestion, `static` or `granularity` spec types allow you to read data stored in deep storage.
|
||||||
|
|
|
@ -353,7 +353,7 @@ For example if you want to concat "[" and "]" before and after the actual dimens
|
||||||
|
|
||||||
### Filtered DimensionSpecs
|
### Filtered DimensionSpecs
|
||||||
|
|
||||||
These are only valid for multi-value dimensions. If you have a row in druid that has a multi-value dimension with values ["v1", "v2", "v3"] and you send a groupBy/topN query grouping by that dimension with [query filter](filter.html) for value "v1". In the response you will get 3 rows containing "v1", "v2" and "v3". This behavior might be unintuitive for some use cases.
|
These are only valid for multi-value dimensions. If you have a row in druid that has a multi-value dimension with values ["v1", "v2", "v3"] and you send a groupBy/topN query grouping by that dimension with [query filter](filters.html) for value "v1". In the response you will get 3 rows containing "v1", "v2" and "v3". This behavior might be unintuitive for some use cases.
|
||||||
|
|
||||||
It happens because "query filter" is internally used on the bitmaps and only used to match the row to be included in the query result processing. With multi-value dimensions, "query filter" behaves like a contains check, which will match the row with dimension value ["v1", "v2", "v3"]. Please see the section on "Multi-value columns" in [segment](../design/segments.html) for more details.
|
It happens because "query filter" is internally used on the bitmaps and only used to match the row to be included in the query result processing. With multi-value dimensions, "query filter" behaves like a contains check, which will match the row with dimension value ["v1", "v2", "v3"]. Please see the section on "Multi-value columns" in [segment](../design/segments.html) for more details.
|
||||||
Then groupBy/topN processing pipeline "explodes" all multi-value dimensions resulting 3 rows for "v1", "v2" and "v3" each.
|
Then groupBy/topN processing pipeline "explodes" all multi-value dimensions resulting 3 rows for "v1", "v2" and "v3" each.
|
||||||
|
|
|
@ -68,13 +68,13 @@ In this package, you'll find:
|
||||||
|
|
||||||
|
|
||||||
* `LICENSE` - the license files.
|
* `LICENSE` - the license files.
|
||||||
* `bin/` - scripts related to the [single-machine quickstart](quickstart.md).
|
* `bin/` - scripts related to the [single-machine quickstart](quickstart.html).
|
||||||
* `conf/*` - template configurations for a clustered setup.
|
* `conf/*` - template configurations for a clustered setup.
|
||||||
* `conf-quickstart/*` - configurations for the [single-machine quickstart](quickstart.md).
|
* `conf-quickstart/*` - configurations for the [single-machine quickstart](quickstart.html).
|
||||||
* `extensions/*` - all Druid extensions.
|
* `extensions/*` - all Druid extensions.
|
||||||
* `hadoop-dependencies/*` - Druid Hadoop dependencies.
|
* `hadoop-dependencies/*` - Druid Hadoop dependencies.
|
||||||
* `lib/*` - all included software packages for core Druid.
|
* `lib/*` - all included software packages for core Druid.
|
||||||
* `quickstart/*` - files related to the [single-machine quickstart](quickstart.md).
|
* `quickstart/*` - files related to the [single-machine quickstart](quickstart.html).
|
||||||
|
|
||||||
We'll be editing the files in `conf/` in order to get things running.
|
We'll be editing the files in `conf/` in order to get things running.
|
||||||
|
|
||||||
|
|
|
@ -174,7 +174,7 @@ bin/tranquility server -configFile <path_to_druid_distro>/conf-quickstart/tranqu
|
||||||
|
|
||||||
<div class="note info">
|
<div class="note info">
|
||||||
This section shows you how to load data using Tranquility Server, but Druid also supports a wide
|
This section shows you how to load data using Tranquility Server, but Druid also supports a wide
|
||||||
variety of <a href="ingestion-streams.html#stream-push">other streaming ingestion options</a>, including from
|
variety of <a href="../ingestion/stream-ingestion.html#stream-push">other streaming ingestion options</a>, including from
|
||||||
popular streaming systems like Kafka, Storm, Samza, and Spark Streaming.
|
popular streaming systems like Kafka, Storm, Samza, and Spark Streaming.
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
@ -229,7 +229,7 @@ visualize and explore data in Druid. We recommend trying [Pivot](https://github.
|
||||||
[Panoramix](https://github.com/mistercrunch/panoramix), or [Metabase](https://github.com/metabase/metabase) to start
|
[Panoramix](https://github.com/mistercrunch/panoramix), or [Metabase](https://github.com/metabase/metabase) to start
|
||||||
visualizing the data you just ingested.
|
visualizing the data you just ingested.
|
||||||
|
|
||||||
If you installed Pivot for example, you should be able to view your data in your browser at [localhost:9090](localhost:9090).
|
If you installed Pivot for example, you should be able to view your data in your browser at [localhost:9090](http://localhost:9090/).
|
||||||
|
|
||||||
### SQL and other query libraries
|
### SQL and other query libraries
|
||||||
|
|
||||||
|
|
|
@ -16,7 +16,7 @@ Once that's complete, you can load your own dataset by writing a custom ingestio
|
||||||
|
|
||||||
## Writing an ingestion spec
|
## Writing an ingestion spec
|
||||||
|
|
||||||
When loading files into Druid, you will use Druid's [batch loading](ingestion-batch.html) process.
|
When loading files into Druid, you will use Druid's [batch loading](../ingestion/batch-ingestion.html) process.
|
||||||
There's an example batch ingestion spec in `quickstart/wikiticker-index.json` that you can modify
|
There's an example batch ingestion spec in `quickstart/wikiticker-index.json` that you can modify
|
||||||
for your own needs.
|
for your own needs.
|
||||||
|
|
||||||
|
|
|
@ -45,7 +45,7 @@ Run this command to create a Kafka topic called *metrics*, to which we'll send d
|
||||||
|
|
||||||
## Enable Druid Kafka ingestion
|
## Enable Druid Kafka ingestion
|
||||||
|
|
||||||
Druid includes configs for [Tranquility Kafka](ingestion-streams.md#kafka) to support loading data from Kafka.
|
Druid includes configs for [Tranquility Kafka](../ingestion/stream-pull.html#kafka) to support loading data from Kafka.
|
||||||
To enable this in the quickstart-based configuration:
|
To enable this in the quickstart-based configuration:
|
||||||
|
|
||||||
- Stop your Tranquility command (CTRL-C) and then start it up again.
|
- Stop your Tranquility command (CTRL-C) and then start it up again.
|
||||||
|
|
Loading…
Reference in New Issue