I have a similar problem, I have been using kafka indexing service in a fresh new cluster without hadoop job. But I have the same problem, all tasks failed because [com.metamx.common.ISE: Could not allocate segment for row with timestamp] when I have no segments at all. I have tried with or without lateMessageRejectionPeriod in the ingest json file. Here is some of my task logs.
2016-07-11T09:53:38,270 DEBUG [task-runner-0-priority-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] starting
2016-07-11T09:53:38,272 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,298 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,299 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,300 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,301 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,301 INFO [task-runner-0-priority-0] com.metamx.http.client.pool.ChannelResourceFactory - Generating: http://host
2016-07-11T09:53:38,398 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] messageReceived: DefaultHttpResponse(chunked: true)
HTTP/1.1 200 OK
Content-Type: application/json
Vary: Accept-Encoding, User-Agent
Transfer-Encoding: chunked
Server: Jetty(9.2.5.v20141112)
2016-07-11T09:53:38,398 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] Got response: 200 OK
2016-07-11T09:53:38,400 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] messageReceived: org.jboss.netty.handler.codec.http.DefaultHttpChunk@6e3e91e4
2016-07-11T09:53:38,401 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] Got chunk: 15B, last=false
2016-07-11T09:53:38,401 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] messageReceived: org.jboss.netty.handler.codec.http.HttpChunk$1@71a7c8e9
2016-07-11T09:53:38,401 DEBUG [HttpClient-Netty-Worker-0] com.metamx.http.client.NettyHttpClient - [POST http://host/druid/indexer/v1/action] Got chunk: 0B, last=true
2016-07-11T09:53:38,406 WARN [task-runner-0-priority-0] io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver - Cannot allocate segment for timestamp[2016-07-11T08:52:13.795Z], sequenceName[index_kafka_test_c634fdcd8e2d84d_3].
2016-07-11T09:53:38,406 INFO [task-runner-0-priority-0] io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver - Persisting data.
2016-07-11T09:53:38,409 INFO [task-runner-0-priority-0] io.druid.segment.realtime.appenderator.AppenderatorImpl - Submitting persist runnable for dataSource[test]
2016-07-11T09:53:38,412 INFO [test-incremental-persist] io.druid.segment.realtime.appenderator.AppenderatorImpl - Committing metadata[FiniteAppenderatorDriverMetadata{activeSegments={}, lastSegmentIds={}, callerMetadata={nextPartitions=KafkaPartitions{topic=‘druid_metrics_final’, partitionOffsetMap={0=305318, 1=624366, 2=512090, 3=2157602, 4=305308}}}}] for sinks.
2016-07-11T09:53:38,419 INFO [task-runner-0-priority-0] io.druid.segment.realtime.appenderator.FiniteAppenderatorDriver - Persisted pending data in 13ms.
2016-07-11T09:53:38,420 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-closed:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-created:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent-received:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-received:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name select-time:client-id-consumer-1
2016-07-11T09:53:38,421 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name io-time:client-id-consumer-1
2016-07-11T09:53:38,422 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node–5.bytes-sent
2016-07-11T09:53:38,422 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node–5.bytes-received
2016-07-11T09:53:38,422 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node–5.latency
2016-07-11T09:53:38,422 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-3.bytes-sent
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-3.bytes-received
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-3.latency
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-2147483647.bytes-sent
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-2147483647.bytes-received
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-2147483647.latency
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-4.bytes-sent
2016-07-11T09:53:38,423 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-4.bytes-received
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-4.latency
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-sent
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-received
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.latency
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-1.bytes-sent
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-1.bytes-received
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-1.latency
2016-07-11T09:53:38,424 DEBUG [task-runner-0-priority-0] org.apache.kafka.clients.consumer.KafkaConsumer - The Kafka consumer has closed.
2016-07-11T09:53:38,425 INFO [task-runner-0-priority-0] io.druid.segment.realtime.appenderator.AppenderatorImpl - Shutting down…
2016-07-11T09:53:38,428 ERROR [task-runner-0-priority-0] io.druid.indexing.overlord.ThreadPoolTaskRunner - Exception while running task[KafkaIndexTask{id=index_kafka_test_c634fdcd8e2d84d_gdooapdf, type=index_kafka, dataSource=test}]
com.metamx.common.ISE: Could not allocate segment for row with timestamp[2016-07-11T08:52:13.795Z]
at io.druid.indexing.kafka.KafkaIndexTask.run(KafkaIndexTask.java:427) ~[?:?]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:436) [druid-indexing-service-0.9.1.1.jar:0.9.1.1]
at io.druid.indexing.overlord.ThreadPoolTaskRunner$ThreadPoolTaskRunnerCallable.call(ThreadPoolTaskRunner.java:408) [druid-indexing-service-0.9.1.1.jar:0.9.1.1]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_73]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_73]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_73]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_73]
2016-07-11T09:53:38,428 INFO [task-runner-0-priority-0] io.druid.indexing.overlord.TaskRunnerUtils - Task [index_kafka_test_c634fdcd8e2d84d_gdooapdf] status changed to [FAILED].
2016-07-11T09:53:38,430 INFO [task-runner-0-priority-0] io.druid.indexing.worker.executor.ExecutorLifecycle - Task completed with status: {
“id” : “index_kafka_test_c634fdcd8e2d84d_gdooapdf”,
“status” : “FAILED”,
“duration” : 23633
}
And here is my spec.json
{
“type”: “kafka”,
“dataSchema”: {
“dataSource”: “internal_druid_ingest_metrics”,
“parser”: {
“type”: “string”,
“parseSpec”: {
“format”: “json”,
“timestampSpec”: {
“column”: “timestamp”,
“format”: “auto”
},
“dimensionsSpec”: {
“dimensions”: [
“host”,
“service”,
“dataSource”
]
}
}
},
“metricsSpec”: [
{
“type”: “doubleSum”,
“name”: “ingest_events_thrownAway”,
“fieldName”: “ingest/events/thrownAway”
},
{
“type”: “doubleSum”,
“name”: “ingest_events_unparseable”,
“fieldName”: “ingest/events/unparseable”
},
{
“type”: “doubleSum”,
“name”: “ingest_events_processed”,
“fieldName”: “ingest/events/processed”
},
{
“type”: “doubleSum”,
“name”: “ingest_rows_output”,
“fieldName”: “ingest/rows/output”
},
{
“type”: “doubleSum”,
“name”: “ingest_persists_count”,
“fieldName”: “ingest/persists/count”
},
{
“type”: “doubleSum”,
“name”: “ingest_persists_time”,
“fieldName”: “ingest/persists/time”
},
{
“type”: “doubleSum”,
“name”: “ingest_persists_cpu”,
“fieldName”: “ingest/persists/cpu”
},
{
“type”: “doubleSum”,
“name”: “ingest_persists_backPressure”,
“fieldName”: “ingest/persists/backPressure”
},
{
“type”: “doubleSum”,
“name”: “ingest_persists_failed”,
“fieldName”: “ingest/persists/failed”
},
{
“type”: “doubleSum”,
“name”: “ingest_handoff_failed”,
“fieldName”: “ingest/handoff/failed”
},
{
“type”: “doubleSum”,
“name”: “ingest_merge_time”,
“fieldName”: “ingest/merge/time”
},
{
“type”: “doubleSum”,
“name”: “ingest_merge_cpu”,
“fieldName”: “ingest/merge/cpu”
},
{
“type”: “doubleSum”,
“name”: “ingest_handoff_count”,
“fieldName”: “ingest/handoff/count”
}
],
“granularitySpec”: {
“segmentGranularity”: “hour”,
“queryGranularity”: “none”,
“type”: “uniform”
}
},
“tuningConfig”: {
“type”: “kafka”,
“maxRowsPerSegment”: 5000000
},
“ioConfig”: {
“topic”: “druid_metrics_final”,
“consumerProperties”: {
“bootstrap.servers”: “host1:9092,host2:9092”
},
“taskCount”: 1,
“replicas”: 1,
“taskDuration”: “PT1H”
}
}
在 2016年6月29日星期三 UTC+8下午11:39:39,Gian Merlino写道: