i am out of ideas.
I made druid work with hadoop 2.6. Then i made druid work with jackson 2.3.5. So finally i have no classpath issues anymore. OK.
Then i use an AvroParquetInputFormat, that returns me from the files on HDFS the object, that i have stored in those parquet files.
That object is not of type Writable, but Druid complains, that it must be a Writable. I get a ClasscastException.
Ok. I thought, i can work around that, by creating an extra InputFormat, that wraps this object into a Writable. Now my parser is able to convert it into InputRow.
Finally Druid tries to figure out the segments by doing sorting magic. But no, now i get another ClasscastException. It expects my Writeable to be of Text.class
And here i am lost.
I get the feeling, that the whole HadoopIndexTask is only for Text.
Is this true?
If so, one has to write its own HadoopIndexer. Well, this is probably not a big deal, if it wasnt about the whole packaging and uploading to hadoop.
What do other experience?