-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Lower the default number of primary shards #3431
Comments
+1 on changing the default number of shards. I would also suggest to have 2 or 3 by default. But we should make it possible that this default can be changed in the filebeat config. As soon as we have #3603 we can generate a new template as soon as a change is made. The challenge we then have is what happens, if 2 filebeat instances for example have a different number of shards configured? |
+1 On making it easily configurable to the user. I think conflicting shard values would be treated the same as other conflicting template values. AFAIK this means last one to write it's template to ES wins? Do we have any other strategy? |
If we don't overwrite templates, first one wins :-) I think that is fine and is an edge we do not have to worry about at the moment. |
This PR is good news for us as the "soft limit" of 500 shards will be soonish increased and searching across lots of shards should become more efficient: elastic/elasticsearch#23253 (comment) I still think the default should be a lower number then now. |
Closing in favour of #5095. |
We should consider lowering the default number of shards created by the Beats' daily indices. Each index uses the Elasticsearch default of 5 primaries with 1 replica (5p1r).
I think it would improve our out of the box experience for new users to reduce the level of sharding. In my experience with only a few hosts reporting data, you end up with lots of tiny shards (indices with hundreds of MBs).
Each shard can easily handle many GB of data. So we could lower the sharding values in our index templates, and more experienced users can tune these numbers as they scale up.
We have 5 Beats now (FB, PB, MB, WLB, HB). Here are some sample numbers:
* default from Elasticsearch
Keeping more than 1p will allow users to take advantage of a multi-node cluster. I would go with 2p1r as it the smallest number that can take advantage of distributing the ingest load across nodes.
The text was updated successfully, but these errors were encountered: