From 71a9b43c83995f62e42170e50ec0c783088ad1ee Mon Sep 17 00:00:00 2001 From: fjy Date: Wed, 16 Oct 2013 11:37:56 -0700 Subject: [PATCH] fix docs for deep storage --- docs/content/Deep-Storage.md | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/docs/content/Deep-Storage.md b/docs/content/Deep-Storage.md index 529aeeb80af..5e8f668fe36 100644 --- a/docs/content/Deep-Storage.md +++ b/docs/content/Deep-Storage.md @@ -12,21 +12,21 @@ S3-compatible deep storage is basically either S3 or something like riak-cs whic S3 configuration parameters are ``` -com.metamx.aws.accessKey= -com.metamx.aws.secretKey= -druid.storage.s3.bucket= -druid.storage.s3.baseKey= +druid.s3.accessKey= +druid.s3.secretKey= +druid.storage.bucket= +druid.storage.baseKey= ``` ## HDFS As of 0.4.0, HDFS can be used for storage of segments as well. -In order to use hdfs for deep storage, you need to set the following configuration on your realtime nodes. +In order to use hdfs for deep storage, you need to set the following configuration on your real-time nodes. ``` -druid.storage.hdfs=true -druid.storage.hdfs.storageDirectory= +druid.storage.type=hdfs +druid.storage.storageDirectory= ``` If you are using the Hadoop indexer, set your output directory to be a location on Hadoop and it will work @@ -36,13 +36,13 @@ If you are using the Hadoop indexer, set your output directory to be a location A local mount can be used for storage of segments as well. This allows you to use just your local file system or anything else that can be mount locally like NFS, Ceph, etc. -In order to use a local mount for deep storage, you need to set the following configuration on your realtime nodes. +In order to use a local mount for deep storage, you need to set the following configuration on your real-time nodes. ``` -druid.storage.local=true -druid.storage.local.storageDirectory= +druid.storage.type=local +druid.storage.storageDirectory= ``` -Note that you should generally set `druid.storage.local.storageDirectory` to something different from `druid.paths.indexCache`. +Note that you should generally set `druid.storage.storageDirectory` to something different from `druid.segmentCache.locations` and `druid.segmentCache.infoDir`. If you are using the Hadoop indexer in local mode, then just give it a local file as your output directory and it will work.