--- id: input-sources title: "Input sources" sidebar_label: "Input sources" --- The input source defines where your index task reads data for Apache Druid native batch ingestion. Only the native parallel task and simple task support the input source. For general information on native batch indexing and parallel task indexing, see [Native batch ingestion](./native-batch.md). ## S3 input source :::info Required extension To use the S3 input source, load the extension [`druid-s3-extensions`](../development/extensions-core/s3.md) in your `common.runtime.properties` file. ::: The S3 input source reads objects directly from S3. You can specify either: * a list of S3 URI strings * a list of S3 location prefixes that attempts to list the contents and ingest all objects contained within the locations. The S3 input source is splittable. Therefore, you can use it with the [parallel task](./native-batch.md). Each worker task of `index_parallel` reads one or multiple objects. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "objectGlob": "**.json", "uris": ["s3://foo/bar/file.json", "s3://bar/foo/file2.json"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "objectGlob": "**.parquet", "prefixes": ["s3://foo/bar/", "s3://bar/foo/"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "objectGlob": "**.json", "objects": [ { "bucket": "foo", "path": "bar/file1.json"}, { "bucket": "bar", "path": "foo/file2.json"} ] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "objectGlob": "**.json", "uris": ["s3://foo/bar/file.json", "s3://bar/foo/file2.json"], "properties": { "accessKeyId": "KLJ78979SDFdS2", "secretAccessKey": "KLS89s98sKJHKJKJH8721lljkd" } }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "objectGlob": "**.json", "uris": ["s3://foo/bar/file.json", "s3://bar/foo/file2.json"], "properties": { "accessKeyId": "KLJ78979SDFdS2", "secretAccessKey": "KLS89s98sKJHKJKJH8721lljkd", "assumeRoleArn": "arn:aws:iam::2981002874992:role/role-s3" } }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "s3", "uris": ["s3://foo/bar/file.json", "s3://bar/foo/file2.json"], "endpointConfig": { "url" : "s3-store.aws.com", "signingRegion" : "us-west-2" }, "clientConfig": { "protocol" : "http", "disableChunkedEncoding" : true, "enablePathStyleAccess" : true, "forceGlobalBucketAccessEnabled" : false }, "proxyConfig": { "host" : "proxy-s3.aws.com", "port" : 8888, "username" : "admin", "password" : "admin" }, "properties": { "accessKeyId": "KLJ78979SDFdS2", "secretAccessKey": "KLS89s98sKJHKJKJH8721lljkd", "assumeRoleArn": "arn:aws:iam::2981002874992:role/role-s3" } }, "inputFormat": { "type": "json" }, ... }, ... ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `s3`.|None|yes| |uris|JSON array of URIs where S3 objects to be ingested are located.|None|`uris` or `prefixes` or `objects` must be set| |prefixes|JSON array of URI prefixes for the locations of S3 objects to be ingested. Empty objects starting with one of the given prefixes will be skipped.|None|`uris` or `prefixes` or `objects` must be set| |objects|JSON array of S3 Objects to be ingested.|None|`uris` or `prefixes` or `objects` must be set| |objectGlob|A glob for the object part of the S3 URI. In the URI `s3://foo/bar/file.json`, the glob is applied to `bar/file.json`.

The glob must match the entire object part, not just the filename. For example, the glob `*.json` does not match `s3://foo/bar/file.json`, because the object part is `bar/file.json`, and the`*` does not match the slash. To match all objects ending in `.json`, use `**.json` instead.

For more information, refer to the documentation for [`FileSystem#getPathMatcher`](https://docs.oracle.com/javase/8/docs/api/java/nio/file/FileSystem.html#getPathMatcher-java.lang.String-).|None|no| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (S3 URI starting with `s3://`), `__file_bucket` (S3 bucket), and `__file_path` (S3 object key).|None|no| | endpointConfig |Config for overriding the default S3 endpoint and signing region. This would allow ingesting data from a different S3 store. Please see [s3 config](../development/extensions-core/s3.md#connecting-to-s3-configuration) for more information.|None|No (defaults will be used if not given) | clientConfig |S3 client properties for the overridden s3 endpoint. This is used in conjunction with `endPointConfig`. Please see [s3 config](../development/extensions-core/s3.md#connecting-to-s3-configuration) for more information.|None|No (defaults will be used if not given) | proxyConfig |Properties for specifying proxy information for the overridden s3 endpoint. This is used in conjunction with `clientConfig`. Please see [s3 config](../development/extensions-core/s3.md#connecting-to-s3-configuration) for more information.|None|No (defaults will be used if not given) |properties|Properties Object for overriding the default S3 configuration. See below for more information.|None|No (defaults will be used if not given) Note that the S3 input source will skip all empty objects only when `prefixes` is specified. S3 Object: |Property|Description|Default|Required| |--------|-----------|-------|---------| |bucket|Name of the S3 bucket|None|yes| |path|The path where data is located.|None|yes| Properties Object: |Property|Description|Default|Required| |--------|-----------|-------|---------| |accessKeyId|The [Password Provider](../operations/password-provider.md) or plain text string of this S3 input source access key|None|yes if secretAccessKey is given| |secretAccessKey|The [Password Provider](../operations/password-provider.md) or plain text string of this S3 input source secret key|None|yes if accessKeyId is given| |assumeRoleArn|AWS ARN of the role to assume [see](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp_request.html). **assumeRoleArn** can be used either with the ingestion spec AWS credentials or with the default S3 credentials|None|no| |assumeRoleExternalId|A unique identifier that might be required when you assume a role in another account [see](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_credentials_temp_request.html)|None|no| :::info If `accessKeyId` and `secretAccessKey` are not given, the default [S3 credentials provider chain](../development/extensions-core/s3.md#s3-authentication-methods) is used. ::: ## Google Cloud Storage input source :::info Required extension To use the Google Cloud Storage input source, load the extension [`druid-google-extensions`](../development/extensions-core/google.md) in your `common.runtime.properties` file. ::: The Google Cloud Storage input source is to support reading objects directly from Google Cloud Storage. Objects can be specified as list of Google Cloud Storage URI strings. The Google Cloud Storage input source is splittable and can be used by the [parallel task](./native-batch.md), where each worker task of `index_parallel` will read one or multiple objects. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "google", "objectGlob": "**.json", "uris": ["gs://foo/bar/file.json", "gs://bar/foo/file2.json"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "google", "objectGlob": "**.parquet", "prefixes": ["gs://foo/bar/", "gs://bar/foo/"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "google", "objectGlob": "**.json", "objects": [ { "bucket": "foo", "path": "bar/file1.json"}, { "bucket": "bar", "path": "foo/file2.json"} ] }, "inputFormat": { "type": "json" }, ... }, ... ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `google`.|None|yes| |uris|JSON array of URIs where Google Cloud Storage objects to be ingested are located.|None|`uris` or `prefixes` or `objects` must be set| |prefixes|JSON array of URI prefixes for the locations of Google Cloud Storage objects to be ingested. Empty objects starting with one of the given prefixes will be skipped.|None|`uris` or `prefixes` or `objects` must be set| |objects|JSON array of Google Cloud Storage objects to be ingested.|None|`uris` or `prefixes` or `objects` must be set| |objectGlob|A glob for the object part of the S3 URI. In the URI `s3://foo/bar/file.json`, the glob is applied to `bar/file.json`.

The glob must match the entire object part, not just the filename. For example, the glob `*.json` does not match `s3://foo/bar/file.json`, because the object part is `bar/file.json`, and the`*` does not match the slash. To match all objects ending in `.json`, use `**.json` instead.

For more information, refer to the documentation for [`FileSystem#getPathMatcher`](https://docs.oracle.com/javase/8/docs/api/java/nio/file/FileSystem.html#getPathMatcher-java.lang.String-).|None|no| Note that the Google Cloud Storage input source will skip all empty objects only when `prefixes` is specified. Google Cloud Storage object: |Property|Description|Default|Required| |--------|-----------|-------|---------| |bucket|Name of the Google Cloud Storage bucket|None|yes| |path|The path where data is located.|None|yes| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (Google Cloud Storage URI starting with `gs://`), `__file_bucket` (GCS bucket), and `__file_path` (GCS key).|None|no| ## Azure input source :::info Required extension To use the Azure input source, load the extension [`druid-azure-extensions`](../development/extensions-core/azure.md) in your `common.runtime.properties` file. ::: The Azure input source (that uses the type `azureStorage`) reads objects directly from Azure Blob store or Azure Data Lake sources. You can specify objects as a list of file URI strings or prefixes. You can split the Azure input source for use with [parallel task](./native-batch.md) indexing and each worker task reads one chunk of the split data. The `azureStorage` input source is a new schema for Azure input sources that allows you to specify which storage account files should be ingested from. We recommend that you update any specs that use the old `azure` schema to use the new `azureStorage` schema. The new schema provides more functionality than the older `azure` schema. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azureStorage", "objectGlob": "**.json", "uris": ["azureStorage://storageAccount/container/prefix1/file.json", "azureStorage://storageAccount/container/prefix2/file2.json"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azureStorage", "objectGlob": "**.parquet", "prefixes": ["azureStorage://storageAccount/container/prefix1/", "azureStorage://storageAccount/container/prefix2/"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azureStorage", "objectGlob": "**.json", "objects": [ { "bucket": "storageAccount", "path": "container/prefix1/file1.json"}, { "bucket": "storageAccount", "path": "container/prefix2/file2.json"} ], "properties": { "sharedAccessStorageToken": "?sv=......", } }, "inputFormat": { "type": "json" }, ... }, ... ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `azureStorage`.|None|yes| |uris|JSON array of URIs where the Azure objects to be ingested are located. Use this format: `azureStorage://STORAGE_ACCOUNT/CONTAINER/PATH_TO_FILE`|None|One of the following must be set:`uris`, `prefixes`, or `objects`.| |prefixes|JSON array of URI prefixes for the locations of Azure objects to ingest. Use this format`azureStorage://STORAGE_ACCOUNT/CONTAINER/PREFIX`. Empty objects starting with any of the given prefixes are skipped.|None|One of the following must be set:`uris`, `prefixes`, or `objects`.| |objects|JSON array of Azure objects to ingest.|None|One of the following must be set:`uris`, `prefixes`, or `objects`.| |objectGlob|A glob for the object part of the Azure URI. In the URI `azureStorage://foo/bar/file.json`, the glob is applied to `bar/file.json`.

The glob must match the entire object part, not just the filename. For example, the glob `*.json` does not match `azureStorage://foo/bar/file.json` because the object part is `bar/file.json`, and the`*` does not match the slash. To match all objects ending in `.json`, use `**.json` instead.

For more information, refer to the documentation for [`FileSystem#getPathMatcher`](https://docs.oracle.com/javase/8/docs/api/java/nio/file/FileSystem.html#getPathMatcher-java.lang.String-).|None|no| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (Azure blob URI starting with `azureStorage://`), `__file_bucket` (Azure bucket), and `__file_path` (Azure object path).|None|no| |properties|Properties object for overriding the default Azure configuration. See below for more information.|None|No (defaults will be used if not given)| Note that the Azure input source skips all empty objects only when `prefixes` is specified. The `objects` property can one of the following: |Property|Description|Default|Required| |--------|-----------|-------|---------| |bucket|Name of the Azure Blob Storage or Azure Data Lake storage account|None|yes| |path|The container and path where data is located.|None|yes| The `properties` property can be one of the following: * `sharedAccessStorageToken` * `key` * `appRegistrationClientId`, `appRegistrationClientSecret`, and `tenantId` * empty |Property|Description|Default|Required| |--------|-----------|-------|---------| |sharedAccessStorageToken|The plain text string of this Azure Blob Storage Shared Access Token|None|No| |key|The root key of Azure Blob Storage Account|None|no| |appRegistrationClientId|The client ID of the Azure App registration to authenticate as|None|No| |appRegistrationClientSecret|The client secret of the Azure App registration to authenticate as|None|Yes if `appRegistrationClientId` is provided| |tenantId|The tenant ID of the Azure App registration to authenticate as|None|Yes if `appRegistrationClientId` is provided| ### Legacy `azure` input source The Azure input source that uses the type `azure` is an older version of the Azure input type and is not recommended. It doesn't support specifying which storage account to ingest from. We recommend using the [`azureStorage` input source schema](#azure-input-source) instead since it provides more functionality. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azure", "objectGlob": "**.json", "uris": ["azure://container/prefix1/file.json", "azure://container/prefix2/file2.json"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azure", "objectGlob": "**.parquet", "prefixes": ["azure://container/prefix1/", "azure://container/prefix2/"] }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "azure", "objectGlob": "**.json", "objects": [ { "bucket": "container", "path": "prefix1/file1.json"}, { "bucket": "container", "path": "prefix2/file2.json"} ] }, "inputFormat": { "type": "json" }, ... }, ... ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `azure`.|None|yes| |uris|JSON array of URIs where the Azure objects to be ingested are located, in the form `azure:///`|None|`uris` or `prefixes` or `objects` must be set| |prefixes|JSON array of URI prefixes for the locations of Azure objects to ingest, in the form `azure:///`. Empty objects starting with one of the given prefixes are skipped.|None|`uris` or `prefixes` or `objects` must be set| |objects|JSON array of Azure objects to ingest.|None|`uris` or `prefixes` or `objects` must be set| |objectGlob|A glob for the object part of the Azure URI. In the URI `azure://foo/bar/file.json`, the glob is applied to `bar/file.json`.

The glob must match the entire object part, not just the filename. For example, the glob `*.json` does not match `azure://foo/bar/file.json`, because the object part is `bar/file.json`, and the`*` does not match the slash. To match all objects ending in `.json`, use `**.json` instead.

For more information, refer to the documentation for [`FileSystem#getPathMatcher`](https://docs.oracle.com/javase/8/docs/api/java/nio/file/FileSystem.html#getPathMatcher-java.lang.String-).|None|no| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (Azure blob URI starting with `azure://`), `__file_bucket` (Azure bucket), and `__file_path` (Azure object path).|None|no| Note that the Azure input source skips all empty objects only when `prefixes` is specified. The `objects` property is: |Property|Description|Default|Required| |--------|-----------|-------|---------| |bucket|Name of the Azure Blob Storage or Azure Data Lake container|None|yes| |path|The path where data is located.|None|yes| ## HDFS input source :::info Required extension To use the HDFS input source, load the extension [`druid-hdfs-storage`](../development/extensions-core/hdfs.md) in your `common.runtime.properties` file. ::: The HDFS input source is to support reading files directly from HDFS storage. File paths can be specified as an HDFS URI string or a list of HDFS URI strings. The HDFS input source is splittable and can be used by the [parallel task](./native-batch.md), where each worker task of `index_parallel` will read one or multiple files. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "hdfs", "paths": "hdfs://namenode_host/foo/bar/", "hdfs://namenode_host/bar/foo" }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "hdfs", "paths": "hdfs://namenode_host/foo/bar/", "hdfs://namenode_host/bar/foo" }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "hdfs", "paths": "hdfs://namenode_host/foo/bar/file.json", "hdfs://namenode_host/bar/foo/file2.json" }, "inputFormat": { "type": "json" }, ... }, ... ``` ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "hdfs", "paths": ["hdfs://namenode_host/foo/bar/file.json", "hdfs://namenode_host/bar/foo/file2.json"] }, "inputFormat": { "type": "json" }, ... }, ... ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `hdfs`.|None|yes| |paths|HDFS paths. Can be either a JSON array or comma-separated string of paths. Wildcards like `*` are supported in these paths. Empty files located under one of the given paths will be skipped.|None|yes| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (URI) and `__file_path` (path component of URI).|None|no| You can also ingest from other storage using the HDFS input source if the HDFS client supports that storage. However, if you want to ingest from cloud storage, consider using the service-specific input source for your data storage. If you want to use a non-hdfs protocol with the HDFS input source, include the protocol in `druid.ingestion.hdfs.allowedProtocols`. See [HDFS input source security configuration](../configuration/index.md#hdfs-input-source) for more details. ## HTTP input source The HTTP input source is to support reading files directly from remote sites via HTTP. :::info Security notes Ingestion tasks run under the operating system account that runs the Druid processes, for example the Indexer, Middle Manager, and Peon. This means any user who can submit an ingestion task can specify an input source referring to any location that the Druid process can access. For example, using `http` input source, users may have access to internal network servers. The `http` input source is not limited to the HTTP or HTTPS protocols. It uses the Java URI class that supports HTTP, HTTPS, FTP, file, and jar protocols by default. ::: For more information about security best practices, see [Security overview](../operations/security-overview.md#best-practices). The HTTP input source is _splittable_ and can be used by the [parallel task](./native-batch.md), where each worker task of `index_parallel` will read only one file. This input source does not support Split Hint Spec. Sample specs: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "http", "uris": ["http://example.com/uri1", "http://example2.com/uri2"] }, "inputFormat": { "type": "json" }, ... }, ... ``` Example with authentication fields using the DefaultPassword provider (this requires the password to be in the ingestion spec): ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "http", "uris": ["http://example.com/uri1", "http://example2.com/uri2"], "httpAuthenticationUsername": "username", "httpAuthenticationPassword": "password123" }, "inputFormat": { "type": "json" }, ... }, ... ``` You can also use the other existing Druid PasswordProviders. Here is an example using the EnvironmentVariablePasswordProvider: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "http", "uris": ["http://example.com/uri1", "http://example2.com/uri2"], "httpAuthenticationUsername": "username", "httpAuthenticationPassword": { "type": "environment", "variable": "HTTP_INPUT_SOURCE_PW" } }, "inputFormat": { "type": "json" }, ... }, ... } ``` |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set the value to `http`.|None|yes| |uris|URIs of the input files. See below for the protocols allowed for URIs.|None|yes| |httpAuthenticationUsername|Username to use for authentication with specified URIs. Can be optionally used if the URIs specified in the spec require a Basic Authentication Header.|None|no| |httpAuthenticationPassword|PasswordProvider to use with specified URIs. Can be optionally used if the URIs specified in the spec require a Basic Authentication Header.|None|no| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (URI including scheme) and `__file_path` (path component of URI).|None|no| You can only use protocols listed in the `druid.ingestion.http.allowedProtocols` property as HTTP input sources. The `http` and `https` protocols are allowed by default. See [HTTP input source security configuration](../configuration/index.md#http-input-source) for more details. ## Inline input source The Inline input source can be used to read the data inlined in its own spec. It can be used for demos or for quickly testing out parsing and schema. Sample spec: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "inline", "data": "0,values,formatted\n1,as,CSV" }, "inputFormat": { "type": "csv" }, ... }, ... ``` |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `inline`.|yes| |data|Inlined data to ingest.|yes| ## Local input source The Local input source is to support reading files directly from local storage, and is mainly intended for proof-of-concept testing. The Local input source is _splittable_ and can be used by the [parallel task](./native-batch.md), where each worker task of `index_parallel` will read one or multiple files. Sample spec: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "local", "filter" : "*.csv", "baseDir": "/data/directory", "files": ["/bar/foo", "/foo/bar"] }, "inputFormat": { "type": "csv" }, ... }, ... ``` |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `local`.|yes| |filter|A wildcard filter for files. See [here](http://commons.apache.org/proper/commons-io/apidocs/org/apache/commons/io/filefilter/WildcardFileFilter) for more information. Files matching the filter criteria are considered for ingestion. Files not matching the filter criteria are ignored.|yes if `baseDir` is specified| |baseDir|Directory to search recursively for files to be ingested. Empty files under the `baseDir` will be skipped.|At least one of `baseDir` or `files` should be specified| |files|File paths to ingest. Some files can be ignored to avoid ingesting duplicate files if they are located under the specified `baseDir`. Empty files will be skipped.|At least one of `baseDir` or `files` should be specified| |systemFields|JSON array of system fields to return as part of input rows. Possible values: `__file_uri` (File URI starting with `file:`) and `__file_path` (file path).|no| ## Druid input source The Druid input source is to support reading data directly from existing Druid segments, potentially using a new schema and changing the name, dimensions, metrics, rollup, etc. of the segment. The Druid input source is _splittable_ and can be used by the [parallel task](./native-batch.md). This input source has a fixed input format for reading from Druid segments; no `inputFormat` field needs to be specified in the ingestion spec when using this input source. |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `druid`.|yes| |dataSource|A String defining the Druid datasource to fetch rows from|yes| |interval|A String representing an ISO-8601 interval, which defines the time range to fetch the data over.|yes| |filter| See [Filters](../querying/filters.md). Only rows that match the filter, if specified, will be returned.|no| The Druid input source can be used for a variety of purposes, including: * Creating new datasources that are rolled-up copies of existing datasources. * Changing the [partitioning or sorting](./partitioning.md) of a datasource to improve performance. * Updating or removing rows using a [`transformSpec`](./ingestion-spec.md#transformspec). When using the Druid input source, the timestamp column shows up as a numeric field named `__time` set to the number of milliseconds since the epoch (January 1, 1970 00:00:00 UTC). It is common to use this in the timestampSpec, if you want the output timestamp to be equivalent to the input timestamp. In this case, set the timestamp column to `__time` and the format to `auto` or `millis`. It is OK for the input and output datasources to be the same. In this case, newly generated data will overwrite the previous data for the intervals specified in the `granularitySpec`. Generally, if you are going to do this, it is a good idea to test out your reindexing by writing to a separate datasource before overwriting your main one. Alternatively, if your goals can be satisfied by [compaction](../data-management/compaction.md), consider that instead as a simpler approach. An example task spec is shown below. It reads from a hypothetical raw datasource `wikipedia_raw` and creates a new rolled-up datasource `wikipedia_rollup` by grouping on hour, "countryName", and "page". ```json { "type": "index_parallel", "spec": { "dataSchema": { "dataSource": "wikipedia_rollup", "timestampSpec": { "column": "__time", "format": "millis" }, "dimensionsSpec": { "dimensions": [ "countryName", "page" ] }, "metricsSpec": [ { "type": "count", "name": "cnt" } ], "granularitySpec": { "type": "uniform", "queryGranularity": "HOUR", "segmentGranularity": "DAY", "intervals": ["2016-06-27/P1D"], "rollup": true } }, "ioConfig": { "type": "index_parallel", "inputSource": { "type": "druid", "dataSource": "wikipedia_raw", "interval": "2016-06-27/P1D" } }, "tuningConfig": { "type": "index_parallel", "partitionsSpec": { "type": "hashed" }, "forceGuaranteedRollup": true, "maxNumConcurrentSubTasks": 1 } } } ``` :::info Older versions (0.19 and earlier) did not respect the timestampSpec when using the Druid input source. If you have ingestion specs that rely on this and cannot rewrite them, set [`druid.indexer.task.ignoreTimestampSpecForDruidInputSource`](../configuration/index.md#indexer-general-configuration) to `true` to enable a compatibility mode where the timestampSpec is ignored. ::: The [secondary partitioning method](native-batch.md#partitionsspec) determines the requisite number of concurrent worker tasks that run in parallel to complete ingestion with the Combining input source. Set this value in `maxNumConcurrentSubTasks` in `tuningConfig` based on the secondary partitioning method: * `range` or `single_dim` partitioning: greater than or equal to 1 * `hashed` or `dynamic` partitioning: greater than or equal to 2 For more information on the `maxNumConcurrentSubTasks` field, see [Implementation considerations](native-batch.md#implementation-considerations). ## SQL input source :::info Required extension To use the SQL input source, you must load the appropriate extension in your `common.runtime.properties` file. * To connect to MySQL, load the extension [`mysql-metadata-storage`](../development/extensions-core/mysql.md). * To connect to PostgreSQL, load the extension [`postgresql-metadata-storage`](../development/extensions-core/postgresql.md). The MySQL extension requires a JDBC driver. For more information, see the [Installing the MySQL connector library](../development/extensions-core/mysql.md). ::: The SQL input source is used to read data directly from RDBMS. You can _split_ the ingestion tasks for a SQL input source. When you use the [parallel task](./native-batch.md) type, each worker task reads from one SQL query from the list of queries. This input source does not support Split Hint Spec. The SQL input source has a fixed input format for reading events. Don't specify `inputFormat` when using this input source. Refer to the [recommended practices](#recommended-practices) before using this input source. |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `sql`.|Yes| |database|Specifies the database connection details. The database type corresponds to the extension that supplies the `connectorConfig` support.

You can selectively allow JDBC properties in `connectURI`. See [JDBC connections security config](../configuration/index.md#jdbc-connections-to-external-databases) for more details.|Yes| |foldCase|Boolean to toggle case folding of database column names. For example, to ingest a database column named `Entry_Date` as `entry_date`, set `foldCase` to true and include `entry_date` in the [`dimensionsSpec`](ingestion-spec.md#dimensionsspec).|No| |sqls|List of SQL queries where each SQL query would retrieve the data to be indexed.|Yes| The following is an example of an SQL input source spec: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "sql", "database": { "type": "mysql", "connectorConfig": { "connectURI": "jdbc:mysql://host:port/schema", "user": "user", "password": "password" } }, "sqls": ["SELECT * FROM table1 WHERE timestamp BETWEEN '2013-01-01 00:00:00' AND '2013-01-01 11:59:59'", "SELECT * FROM table2 WHERE timestamp BETWEEN '2013-01-01 00:00:00' AND '2013-01-01 11:59:59'"] } }, ... ``` The spec above will read all events from two separate SQLs for the interval `2013-01-01/2013-01-02`. Each of the SQL queries will be run in its own sub-task and thus for the above example, there would be two sub-tasks. ### Recommended practices Compared to the other native batch input sources, SQL input source behaves differently in terms of reading the input data. Therefore, consider the following points before using this input source in a production environment: * During indexing, each sub-task would execute one of the SQL queries and the results are stored locally on disk. The sub-tasks then proceed to read the data from these local input files and generate segments. Presently, there isn’t any restriction on the size of the generated files and this would require the Middle Managers or Indexers to have sufficient disk capacity based on the volume of data being indexed. * Filtering the SQL queries based on the intervals specified in the `granularitySpec` can avoid unwanted data being retrieved and stored locally by the indexing sub-tasks. For example, if the `intervals` specified in the `granularitySpec` is `["2013-01-01/2013-01-02"]` and the SQL query is `SELECT * FROM table1`, `SqlInputSource` will read all the data for `table1` based on the query, even though only data between the intervals specified will be indexed into Druid. * Pagination may be used on the SQL queries to ensure that each query pulls a similar amount of data, thereby improving the efficiency of the sub-tasks. * Similar to file-based input formats, any updates to existing data will replace the data in segments specific to the intervals specified in the `granularitySpec`. ## Combining input source The Combining input source lets you read data from multiple input sources. It identifies the splits from delegate input sources and uses a worker task to process each split. Each delegate input source must be splittable and compatible with the [parallel task type](./native-batch.md). Similar to other input sources, the Combining input source supports a single `inputFormat`. Delegate input sources that require an `inputFormat` must have the same format for input data. If you include the [Druid input source](#druid-input-source), the timestamp column is stored in the `__time` field. To correctly combine the data from the Druid input source with another source, ensure that other delegate input sources also store the timestamp column in `__time`. |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `combining`.|Yes| |delegates|List of splittable input sources to read data from.|Yes| The following is an example of a Combining input source spec: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "combining", "delegates" : [ { "type": "local", "filter" : "*.csv", "baseDir": "/data/directory", "files": ["/bar/foo", "/foo/bar"] }, { "type": "druid", "dataSource": "wikipedia", "interval": "2013-01-01/2013-01-02" } ] }, "inputFormat": { "type": "csv" }, ... }, ... ``` ## Iceberg input source :::info Required extension To use the Iceberg input source, load the extension [`druid-iceberg-extensions`](../development/extensions-contrib/iceberg.md) in your `common.runtime.properties` file. ::: You use the Iceberg input source to read data stored in the Iceberg table format. For a given table, the input source scans up to the latest Iceberg snapshot from the configured Hive catalog. Druid ingests the underlying live data files using the existing input source formats. The Iceberg input source cannot be independent as it relies on the existing input sources to read from the data files. For example, if the warehouse associated with an Iceberg catalog is on S3, you must also load the [`druid-s3-extensions`](../development/extensions-core/s3.md) extension. The following is a sample spec for a HDFS warehouse source: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "iceberg", "tableName": "iceberg_table", "namespace": "iceberg_namespace", "icebergCatalog": { "type": "hive", "warehousePath": "hdfs://warehouse/path", "catalogUri": "thrift://hive-metastore.x.com:8970", "catalogProperties": { "hive.metastore.connect.retries": "1", "hive.metastore.execute.setugi": "false", "hive.metastore.kerberos.principal": "KRB_PRINCIPAL", "hive.metastore.sasl.enabled": "true", "metastore.catalog.default": "catalog_test", "hadoop.security.authentication": "kerberos", "hadoop.security.authorization": "true" } }, "icebergFilter": { "type": "interval", "filterColumn": "event_time", "intervals": [ "2023-05-10T19:00:00.000Z/2023-05-10T20:00:00.000Z" ] }, "warehouseSource": { "type": "hdfs" }, "snapshotTime": "2023-06-01T00:00:00.000Z", }, "inputFormat": { "type": "parquet" } }, ... }, ... ``` The following is a sample spec for a S3 warehouse source: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "iceberg", "tableName": "iceberg_table", "namespace": "iceberg_namespace", "icebergCatalog": { "type": "hive", "warehousePath": "s3://warehouse/path", "catalogUri": "thrift://hive-metastore.x.com:8970", "catalogProperties": { "hive.metastore.connect.retries": "1", "hive.metastore.execute.setugi": "false", "hive.metastore.kerberos.principal": "KRB_PRINCIPAL", "hive.metastore.sasl.enabled": "true", "metastore.catalog.default": "default_catalog", "fs.s3a.access.key" : "S3_ACCESS_KEY", "fs.s3a.secret.key" : "S3_SECRET_KEY", "fs.s3a.endpoint" : "S3_API_ENDPOINT" } }, "icebergFilter": { "type": "interval", "filterColumn": "event_time", "intervals": [ "2023-05-10T19:00:00.000Z/2023-05-10T20:00:00.000Z" ] }, "warehouseSource": { "type": "s3", "endpointConfig": { "url": "teststore.aws.com", "signingRegion": "us-west-2a" }, "clientConfig": { "protocol": "http", "disableChunkedEncoding": true, "enablePathStyleAccess": true, "forceGlobalBucketAccessEnabled": false }, "properties": { "accessKeyId": { "type": "default", "password": "foo" }, "secretAccessKey": { "type": "default", "password": "bar" } }, } }, "inputFormat": { "type": "parquet" } }, ... }, ``` |Property|Description|Required| |--------|-----------|---------| |type|Set the value to `iceberg`.|yes| |tableName|The Iceberg table name configured in the catalog.|yes| |namespace|The Iceberg namespace associated with the table.|yes| |icebergFilter|The JSON Object that filters data files within a snapshot.|no| |icebergCatalog|The JSON Object used to define the catalog that manages the configured Iceberg table.|yes| |warehouseSource|The JSON Object that defines the native input source for reading the data files from the warehouse.|yes| |snapshotTime|Timestamp in ISO8601 DateTime format that will be used to fetch the most recent snapshot as of this time.|no| ### Catalog Object The catalog object supports `rest`, `hive` and `local` catalog types. The following table lists the properties of a `local` catalog: |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set this value to `local`.|None|yes| |warehousePath|The location of the warehouse associated with the catalog.|None|yes| |catalogProperties|Map of any additional properties that needs to be attached to the catalog.|None|no| |caseSensitive|Toggle case sensitivity for column names during Iceberg table reads.|true|no| The following table lists the properties of a `hive` catalog: |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set this value to `hive`.|None|yes| |warehousePath|The location of the warehouse associated with the catalog.|None|yes| |catalogUri|The URI associated with the hive catalog.|None|yes| |catalogProperties|Map of any additional properties that needs to be attached to the catalog.|None|no| |caseSensitive|Toggle case sensitivity for column names during Iceberg table reads.|true|no| The following table lists the properties of a `rest` catalog: |Property|Description|Default|Required| |--------|-----------|-------|---------| |type|Set this value to `rest`.|None|yes| |catalogUri|The URI associated with the catalog's HTTP endpoint.|None|yes| |catalogProperties|Map of any additional properties that needs to be attached to the catalog.|None|no| ### Iceberg filter object This input source provides the following filters: `and`, `equals`, `interval`, and `or`. You can use these filters to filter out data files from a snapshot, reducing the number of files Druid has to ingest. It is strongly recommended to apply filtering only on Iceberg partition columns. When filtering on non-partition columns, Iceberg filters may return rows that do not fully match the expression. To address this, it may help to define an additional filter in the [`transformSpec`](./ingestion-spec.md#transformspec) to remove residual rows. `equals` Filter: |Property|Description|Required| |--------|-----------|---------| |type|Set this value to `equals`.|yes| |filterColumn|The name of the column from the Iceberg table schema to use for filtering.|yes| |filterValue|The value to filter on.|yes| `interval` Filter: |Property|Description|Required| |--------|-----------|---------| |type|Set this value to `interval`.|yes| |filterColumn|The column name from the iceberg table schema based on which filtering needs to happen|yes| |intervals|A JSON array containing ISO 8601 interval strings. This defines the time ranges to filter on. The start interval is inclusive and the end interval is exclusive. |yes| `and` Filter: |Property|Description|Required| |--------|-----------|---------| |type|Set this value to `and`.|yes| |filters|List of iceberg filters that needs to be AND-ed|yes| `or` Filter: |Property|Description|Required| |--------|-----------|---------| |type|Set this value to `or`.|yes| |filters|List of iceberg filters that needs to be OR-ed|yes| `not` Filter: |Property|Description|Required| |--------|-----------|---------| |type|Set this value to `not`.|yes| |filter|The iceberg filter on which logical NOT is applied|yes| `range` Filter: |Property|Description|Default|Required| |--------|-----------|-------|--------| |type|Set this value to `range`.|None|yes| |filterColumn|The column name from the iceberg table schema based on which range filtering needs to happen.|None|yes| |lower|Lower bound value to match.|None|no. At least one of `lower` or `upper` must not be null.| |upper|Upper bound value to match. |None|no. At least one of `lower` or `upper` must not be null.| |lowerOpen|Boolean indicating if lower bound is open in the interval of values defined by the range (">" instead of ">="). |false|no| |upperOpen|Boolean indicating if upper bound is open on the interval of values defined by range ("<" instead of "<="). |false|no| ## Delta Lake input source :::info Required extension To use the Delta Lake input source, load the extension [`druid-deltalake-extensions`](../development/extensions-contrib/delta-lake.md) in your `common.runtime.properties` file. ::: You can use the Delta input source to read data stored in a Delta Lake table. For a given table, the input source scans the latest snapshot from the configured table. Druid ingests the underlying delta files from the table. | Property|Description| Default|Required | |---------|-----------|-----------------| |type|Set this value to `delta`.| None|yes| |tablePath|The location of the Delta table.|None|yes| |filter|The JSON Object that filters data files within a snapshot.|None|no| |snapshotVersion|The snapshot version to read from the Delta table. An integer value must be specified.|Latest|no| ### Delta filter object You can use these filters to filter out data files from a snapshot, reducing the number of files Druid has to ingest from a Delta table. This input source provides the following filters: `and`, `or`, `not`, `=`, `>`, `>=`, `<`, `<=`. When a filter is applied on non-partitioned columns, the filtering is best-effort as the Delta Kernel solely relies on statistics collected when the non-partitioned table is created. In this scenario, this Druid connector may ingest data that doesn't match the filter. To guarantee that the Delta Kernel prunes out unnecessary column values, only use filters on partitioned columns. `and` filter: | Property | Description | Required | |----------|---------------------------------------------------------------------------------------------------------------------------------------------------------------|----------| | type | Set this value to `and`. | yes | | filters | List of Delta filter predicates that get evaluated using logical AND where both conditions need to be true. `and` filter requires two filter predicates. | yes | `or` filter: | Property | Description | Required | |----------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------|----------| | type | Set this value to `or`. | yes | | filters | List of Delta filter predicates that get evaluated using logical OR where only one condition needs to be true. `or` filter requires two filter predicates. | yes | `not` filter: | Property | Description | Required | |----------|---------------------------------------------------------------------------------------------------------------|----------| | type | Set this value to `not`. | yes | | filter | The Delta filter predicate that gets evaluated using logical NOT. `not` filter requires one filter predicate. | yes | `=` filter: | Property | Description | Required | |----------|------------------------------------------|----------| | type | Set this value to `=`. | yes | | column | The table column to apply the filter on. | yes | | value | The value to use in the filter. | yes | `>` filter: | Property | Description | Required | |----------|------------------------------------------|----------| | type | Set this value to `>`. | yes | | column | The table column to apply the filter on. | yes | | value | The value to use in the filter. | yes | `>=` filter: | Property | Description | Required | |----------|------------------------------------------|----------| | type | Set this value to `>=`. | yes | | column | The table column to apply the filter on. | yes | | value | The value to use in the filter. | yes | `<` filter: | Property | Description | Required | |----------|------------------------------------------|----------| | type | Set this value to `<`. | Yes | | column | The table column to apply the filter on. | Yes | | value | The value to use in the filter. | Yes | `<=` filter: | Property | Description | Required | |----------|------------------------------------------|----------| | type | Set this value to `<=`. | yes | | column | The table column to apply the filter on. | yes | | value | The value to use in the filter. | yes | The following is a sample spec to read all records from the latest snapshot from Delta table `/delta-table/foo`: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "delta", "tablePath": "/delta-table/foo" }, } ``` The following is a sample spec to read records from the Delta table `/delta-table/foo` snapshot version `3` to select records where `name = 'Employee4' and age >= 30`: ```json ... "ioConfig": { "type": "index_parallel", "inputSource": { "type": "delta", "tablePath": "/delta-table/foo", "filter": { "type": "and", "filters": [ { "type": "=", "column": "name", "value": "Employee4" }, { "type": ">=", "column": "age", "value": "30" } ] }, "snapshotVersion": 3 }, } ```