4.4 KiB
layout |
---|
doc_page |
Avro
This extension enables Druid to ingest and understand the Apache Avro data format. Make sure to include druid-avro-extensions
as an extension.
Avro Stream Parser
This is for streaming/realtime ingestion.
Field | Type | Description | Required |
---|---|---|---|
type | String | This should say avro_stream . |
no |
avroBytesDecoder | JSON Object | Specifies how to decode bytes to Avro record. | yes |
parseSpec | JSON Object | Specifies the timestamp and dimensions of the data. Should be a timeAndDims parseSpec. | yes |
For example, using Avro stream parser with schema repo Avro bytes decoder:
"parser" : {
"type" : "avro_stream",
"avroBytesDecoder" : {
"type" : "schema_repo",
"subjectAndIdConverter" : {
"type" : "avro_1124",
"topic" : "${YOUR_TOPIC}"
},
"schemaRepository" : {
"type" : "avro_1124_rest_client",
"url" : "${YOUR_SCHEMA_REPO_END_POINT}",
}
},
"parseSpec" : {
"type": "timeAndDims",
"timestampSpec": <standard timestampSpec>,
"dimensionsSpec": <standard dimensionsSpec>
}
}
Avro Bytes Decoder
If type
is not included, the avroBytesDecoder defaults to schema_repo
.
SchemaRepo Based Avro Bytes Decoder
This Avro bytes decoder first extract subject
and id
from input message bytes, then use them to lookup the Avro schema with which to decode Avro record from bytes. Details can be found in schema repo and AVRO-1124. You will need an http service like schema repo to hold the avro schema. Towards schema registration on the message producer side, you can refer to io.druid.data.input.AvroStreamInputRowParserTest#testParse()
.
Field | Type | Description | Required |
---|---|---|---|
type | String | This should say schema_repo . |
no |
subjectAndIdConverter | JSON Object | Specifies the how to extract subject and id from message bytes. | yes |
schemaRepository | JSON Object | Specifies the how to lookup Avro schema from subject and id. | yes |
Avro-1124 Subject And Id Converter
Field | Type | Description | Required |
---|---|---|---|
type | String | This should say avro_1124 . |
no |
topic | String | Specifies the topic of your kafka stream. | yes |
Avro-1124 Schema Repository
Field | Type | Description | Required |
---|---|---|---|
type | String | This should say avro_1124_rest_client . |
no |
url | String | Specifies the endpoint url of your Avro-1124 schema repository. | yes |
Avro Hadoop Parser
This is for batch ingestion using the HadoopDruidIndexer. The inputFormat
of inputSpec
in ioConfig
must be set to "io.druid.data.input.avro.AvroValueInputFormat"
. You may want to set Avro reader's schema in jobProperties
in tuningConfig
, eg: "avro.schema.path.input.value": "/path/to/your/schema.avsc"
or "avro.schema.input.value": "your_schema_JSON_object"
, if reader's schema is not set, the schema in Avro object container file will be used, see Avro specification. Make sure to include "io.druid.extensions:druid-avro-extensions" as an extension.
Field | Type | Description | Required |
---|---|---|---|
type | String | This should say avro_hadoop . |
no |
parseSpec | JSON Object | Specifies the timestamp and dimensions of the data. Should be a timeAndDims parseSpec. | yes |
fromPigAvroStorage | Boolean | Specifies whether the data file is stored using AvroStorage. | no(default == false) |
For example, using Avro Hadoop parser with custom reader's schema file:
{
"type" : "index_hadoop",
"spec" : {
"dataSchema" : {
"dataSource" : "",
"parser" : {
"type" : "avro_hadoop",
"parseSpec" : {
"type": "timeAndDims",
"timestampSpec": <standard timestampSpec>,
"dimensionsSpec": <standard dimensionsSpec>
}
}
},
"ioConfig" : {
"type" : "hadoop",
"inputSpec" : {
"type" : "static",
"inputFormat": "io.druid.data.input.avro.AvroValueInputFormat",
"paths" : ""
}
},
"tuningConfig" : {
"jobProperties" : {
"avro.schema.path.input.value" : "/path/to/my/schema.avsc",
}
}
}
}