Currently I’m monitoring several stats of 10K host, each host reports its stats every 10s.
Segments are split by hour, each segment contains about 600 * 10k * 10 = 60M rows.
Our system always query a stats of several host or several stats of a host of a range of time.
If I aggregate 600 samples of a stats in hour to one row, the number of rows will reduced by 600 times.
I don’t understand how druid store the metrics, if rows are aggregated, will the query boosts significantly ? (reduce IOs and delay ?)