Hi MRB,
Thank you for the logs. I have been through the logs and observe below :
As indexing task log is not attached I have picked up index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z
from the MM log to see what happens.
[1] MM log :
2019-10-29T20:32:22,897 INFO [WorkerTaskManager-CompletedTasksCleaner] org.apache.druid.indexing.worker.WorkerTaskManager - Deleting completed task[index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z] information, overlord task status[SUCCESS].
[2] Coordinator Overlord lask:
2019-10-29T20:31:21,271 INFO [qtp1677115089-82] org.apache.druid.indexing.common.actions.LocalTaskActionClient - Performing action for task[index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z]: SegmentTransactionalInsertAction{segmentsToBeOverwritten=null, segments=[DataSegment{binaryVersion=9, id=wikiticker-2015-09-12-sampled-test_2015-09-12T00:00:00.000Z_2015-09-13T00:00:00.000Z_2019-10-29T20:31:04.031Z, loadSpec={type=>local, path=>/home/mrb/apache-druid-0.16.0-incubating/var/druid/segments/wikiticker-2015-09-12-sampled-test/2015-09-12T00:00:00.000Z_2015-09-13T00:00:00.000Z/2019-10-29T20:31:04.031Z/0/index.zip}, dimensions=[channel, cityName, comment, countryIsoCode, countryName, isAnonymous, isMinor, isNew, isRobot, isUnpatrolled, namespace, page, regionIsoCode, regionName, user], metrics=[count, sum_added, sum_deleted, sum_delta, sum_metroCode], shardSpec=NumberedShardSpec{partitionNum=0, partitions=0}, size=4322529}], startMetadata=null, endMetadata=null}
2019-10-29T20:31:21,385 INFO [qtp1677115089-82] org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator - Published segment [wikiticker-2015-09-12-sampled-test_2015-09-12T00:00:00.000Z_2015-09-13T00:00:00.000Z_2019-10-29T20:31:04.031Z] to DB with used flag [true], json[{“dataSource”:“wikiticker-2015-09-12-sampled-test”,“interval”:“2015-09-12T00:00:00.000Z/2015-09-13T00:00:00.000Z”,“version”:“2019-10-29T20:31:04.031Z”,“loadSpec”:{“type”:“local”,“path”:"/home/mrb/apache-druid-0.16.0-incubating/var/druid/segments/wikiticker-2015-09-12-sampled-test/2015-09-12T00:00:00.000Z_2015-09-13T00:00:00.000Z/2019-10-29T20:31:04.031Z/0/index.zip"},“dimensions”:“channel,cityName,comment,countryIsoCode,countryName,isAnonymous,isMinor,isNew,isRobot,isUnpatrolled,namespace,page,regionIsoCode,regionName,user”,“metrics”:“count,sum_added,sum_deleted,sum_delta,sum_metroCode”,“shardSpec”:{“type”:“numbered”,“partitionNum”:0,“partitions”:0},“binaryVersion”:9,“size”:4322529,“identifier”:“wikiticker-2015-09-12-sampled-test_2015-09-12T00:00:00.000Z_2015-09-13T00:00:00.000Z_2019-10-29T20:31:04.031Z”}]
2019-10-29T20:31:22,390 INFO [Curator-PathChildrenCache-1] org.apache.druid.indexing.overlord.RemoteTaskRunner - Worker[localhost:8091] wrote SUCCESS status for task [index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z] on [TaskLocation{host=‘localhost’, port=8100, tlsPort=-1}]
2019-10-29T20:31:22,390 INFO [Curator-PathChildrenCache-1] org.apache.druid.indexing.overlord.RemoteTaskRunner - Worker[localhost:8091] completed task[index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z] with status[SUCCESS]
2019-10-29T20:31:22,391 INFO [Curator-PathChildrenCache-1] org.apache.druid.indexing.overlord.TaskQueue - Received SUCCESS status for task: index_parallel_wikiticker-2015-09-12-sampled-test_2019-10-29T20:30:40.053Z
2019-10-29T20:31:24,814 INFO [Coordinator-Exec–0] org.apache.druid.server.coordinator.helper.DruidCoordinatorSegmentInfoLoader - Found [187] used segments.
2019-10-29T20:37:54,188 INFO [org.apache.druid.metadata.SQLMetadataSegmentManager-Exec–0] org.apache.druid.metadata.SQLMetadataSegmentManager - Polled and found 188 segments in the database
[3] Historical log:
2019-10-29T20:11:48,561 INFO [main] org.apache.druid.java.util.common.lifecycle.Lifecycle - Successfully started lifecycle [module]
All the highlighted parts are self-explanatory, which shows that the indexing job ran successfully / segments are created into deep storage/ MetaDb database was updated and the coordinator is able to see 188 segments from DB. Till here everything seems ok !!!
Now, As the Coordinator does not directly communicate with a historical process when assigning it a new segment; instead, the Coordinator creates some temporary information about the new segment under load queue path of the historical process. Once this request is seen, the historical process will load the segment and begin servicing it.
It seems the Issue to be somewhere at COORDINATOR/ZOOKEEPER/HISTORICAL while communicating to historical to load the segments. I also don’t see further log printed after 20:11:48 in the historical log.
[a] Could you connect to your zookeeper server and see if these paths exist:
ls /druid
[servedSegments, coordinator, listeners, loadQueue, discovery, internal-discovery, indexer, announcements, overlord, segments]