Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 6 additions & 6 deletions docs/content/dependencies/metadata-storage.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,13 +17,13 @@ Derby is not suitable for production use as a metadata store. Use MySQL or Postg

## Using derby

Add the following to your Druid configuration.
Add the following to your Druid configuration.

```properties
druid.metadata.storage.type=derby
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//opt/var/druid_state/derby;create=true
```

```properties
druid.metadata.storage.type=derby
druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527//home/y/var/druid_state/derby;create=true
```

## MySQL

See [mysql-metadata-storage extension documentation](../development/extensions-core/mysql.html).
Expand Down
4 changes: 2 additions & 2 deletions docs/content/ingestion/batch-ingestion.md
Original file line number Diff line number Diff line change
Expand Up @@ -239,7 +239,7 @@ classification=yarn-site,properties=[mapreduce.reduce.memory.mb=6144,mapreduce.r
```

- Follow the instructions under "[Configure Hadoop for data
loads](cluster.html#configure-cluster-for-hadoop-data-loads)" using the XML files from
loads](../tutorials/cluster.html#configure-cluster-for-hadoop-data-loads)" using the XML files from
`/etc/hadoop/conf` on your EMR master.

#### Loading from S3 with EMR
Expand Down Expand Up @@ -269,7 +269,7 @@ Druid works out of the box with many Hadoop distributions.

If you are having dependency conflicts between Druid and your version of Hadoop, you can try
searching for a solution in the [Druid user groups](https://groups.google.com/forum/#!forum/druid-
user), or reading the Druid [Different Hadoop Versions](..//operations/other-hadoop.html) documentation.
user), or reading the Druid [Different Hadoop Versions](../operations/other-hadoop.html) documentation.

## Command Line Hadoop Indexer

Expand Down
3 changes: 0 additions & 3 deletions docs/content/ingestion/stream-pull.md
Original file line number Diff line number Diff line change
Expand Up @@ -293,9 +293,6 @@ results.
Is this always a problem? No. If your data is small enough to fit on a single Kafka partition, you can replicate without issues.
Otherwise, you can run real-time nodes without replication.

There is now also an [experimental low level Kafka firehose](../development/kafka-simple-consumer-firehose.html) which
solves the issues described above with using the high level Kafka consumer.

Please note that druid will skip over event that failed its checksum and it is corrupt.

### Locking
Expand Down
2 changes: 1 addition & 1 deletion docs/content/ingestion/update-existing-data.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ segments and avoid the overhead of rebuilding new segments with reindexing, you
### Reindexing and Delta Ingestion with Hadoop Batch Ingestion

This section assumes the reader understands how to do batch ingestion using Hadoop. See
[batch-ingestion](batch-ingestion.md) for more information. Hadoop batch-ingestion can be used for reindexing and delta ingestion.
[batch-ingestion](batch-ingestion.html) for more information. Hadoop batch-ingestion can be used for reindexing and delta ingestion.

Druid uses an `inputSpec` in the `ioConfig` to know where the data to be ingested is located and how to read it.
For simple Hadoop batch ingestion, `static` or `granularity` spec types allow you to read data stored in deep storage.
Expand Down
2 changes: 1 addition & 1 deletion docs/content/querying/dimensionspecs.md
Original file line number Diff line number Diff line change
Expand Up @@ -353,7 +353,7 @@ For example if you want to concat "[" and "]" before and after the actual dimens

### Filtered DimensionSpecs

These are only valid for multi-value dimensions. If you have a row in druid that has a multi-value dimension with values ["v1", "v2", "v3"] and you send a groupBy/topN query grouping by that dimension with [query filter](filter.html) for value "v1". In the response you will get 3 rows containing "v1", "v2" and "v3". This behavior might be unintuitive for some use cases.
These are only valid for multi-value dimensions. If you have a row in druid that has a multi-value dimension with values ["v1", "v2", "v3"] and you send a groupBy/topN query grouping by that dimension with [query filter](filters.html) for value "v1". In the response you will get 3 rows containing "v1", "v2" and "v3". This behavior might be unintuitive for some use cases.

It happens because "query filter" is internally used on the bitmaps and only used to match the row to be included in the query result processing. With multi-value dimensions, "query filter" behaves like a contains check, which will match the row with dimension value ["v1", "v2", "v3"]. Please see the section on "Multi-value columns" in [segment](../design/segments.html) for more details.
Then groupBy/topN processing pipeline "explodes" all multi-value dimensions resulting 3 rows for "v1", "v2" and "v3" each.
Expand Down
6 changes: 3 additions & 3 deletions docs/content/tutorials/cluster.md
Original file line number Diff line number Diff line change
Expand Up @@ -68,13 +68,13 @@ In this package, you'll find:


* `LICENSE` - the license files.
* `bin/` - scripts related to the [single-machine quickstart](quickstart.md).
* `bin/` - scripts related to the [single-machine quickstart](quickstart.html).
* `conf/*` - template configurations for a clustered setup.
* `conf-quickstart/*` - configurations for the [single-machine quickstart](quickstart.md).
* `conf-quickstart/*` - configurations for the [single-machine quickstart](quickstart.html).
* `extensions/*` - all Druid extensions.
* `hadoop-dependencies/*` - Druid Hadoop dependencies.
* `lib/*` - all included software packages for core Druid.
* `quickstart/*` - files related to the [single-machine quickstart](quickstart.md).
* `quickstart/*` - files related to the [single-machine quickstart](quickstart.html).

We'll be editing the files in `conf/` in order to get things running.

Expand Down
4 changes: 2 additions & 2 deletions docs/content/tutorials/quickstart.md
Original file line number Diff line number Diff line change
Expand Up @@ -174,7 +174,7 @@ bin/tranquility server -configFile <path_to_druid_distro>/conf-quickstart/tranqu

<div class="note info">
This section shows you how to load data using Tranquility Server, but Druid also supports a wide
variety of <a href="ingestion-streams.html#stream-push">other streaming ingestion options</a>, including from
variety of <a href="../ingestion/stream-ingestion.html#stream-push">other streaming ingestion options</a>, including from
popular streaming systems like Kafka, Storm, Samza, and Spark Streaming.
</div>

Expand Down Expand Up @@ -229,7 +229,7 @@ visualize and explore data in Druid. We recommend trying [Pivot](https://github.
[Panoramix](https://github.com/mistercrunch/panoramix), or [Metabase](https://github.com/metabase/metabase) to start
visualizing the data you just ingested.

If you installed Pivot for example, you should be able to view your data in your browser at [localhost:9090](localhost:9090).
If you installed Pivot for example, you should be able to view your data in your browser at [localhost:9090](http://localhost:9090/).

### SQL and other query libraries

Expand Down
2 changes: 1 addition & 1 deletion docs/content/tutorials/tutorial-batch.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ Once that's complete, you can load your own dataset by writing a custom ingestio

## Writing an ingestion spec

When loading files into Druid, you will use Druid's [batch loading](ingestion-batch.html) process.
When loading files into Druid, you will use Druid's [batch loading](../ingestion/batch-ingestion.html) process.
There's an example batch ingestion spec in `quickstart/wikiticker-index.json` that you can modify
for your own needs.

Expand Down
2 changes: 1 addition & 1 deletion docs/content/tutorials/tutorial-kafka.md
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ Run this command to create a Kafka topic called *metrics*, to which we'll send d

## Enable Druid Kafka ingestion

Druid includes configs for [Tranquility Kafka](ingestion-streams.md#kafka) to support loading data from Kafka.
Druid includes configs for [Tranquility Kafka](../ingestion/stream-pull.html#kafka) to support loading data from Kafka.
To enable this in the quickstart-based configuration:

- Stop your Tranquility command (CTRL-C) and then start it up again.
Expand Down