The recommended size of a production broker node is very high.
It seems to be almost the same as the size of a historical node.
Since we have been running druid with large historicals and small broker, I am curious what factors affect broker performance.
It seems most of the computation is run on historical nodes.
If all the broker nodes do is combine data, then does the broker size depend on:
number of historical nodes
number of data points in the final result
my assumption here is:
increasing the number of historicals increases the amount of data to be combined
more data points = more memory needed to hold them
however, the caveat here is:
given a query, the same data is returned no matter the number of historicals, so the amount of historicals shouldnt affect the broker size.
does time range play a factor too?
if i query for 90 days of data and then query for 1 day of data, and in my chart I am going to show only 60 points (aggregated of course), will that affect the broker requirements?