Is schema repo mandatory for Avro druid extension

Hi All,

I am trying to get binary avro data ingested in druid using tranquility. Looks like this can only be done using schema repo. I need to encode schema in the avro file for other hadoop ecosystem tools anyways (hive, spark etc). May be there is a way to make them use schema repo as well, however at this point its a luxury for us.

Looking at the code - https://github.com/druid-io/druid/blob/master/extensions-core/avro-extensions/src/main/java/io/druid/data/input/avro/SchemaRepoBasedAvroBytesDecoder.java schema repo seems to be a mandatory requirement and there is no way to specify the schema inline in the kafka.json. Is this correct? Or are there any alternatives?

regards

Sunita

Hi Sunita,

Just tried to look for it quickly, looks like there are no alternatives right now.
But you can easily add another impl for AvroBytesDecoder which can read the schema from inline json spec.

(https://github.com/druid-io/druid/blob/master/extensions-core/avro-extensions/src/main/java/io/druid/data/input/avro/AvroBytesDecoder.java)

Thanks for your time and suggestion Nishant.Yea, looks like the only option.

regards

Sunita

Just saw this https://groups.google.com/d/msg/druid-development/tg2q3VHN5rQ/sQVN6Hi6AAAJ

regards

Sunita

Great, Charles has marked the PR to be included in 0.9.2. Hopefully it will be there soon.