Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove max_num_segments from _forcemerge #32323

Closed
jpountz opened this issue Jul 24, 2018 · 8 comments
Closed

Remove max_num_segments from _forcemerge #32323

jpountz opened this issue Jul 24, 2018 · 8 comments
Labels
blocker :Distributed Indexing/Engine Anything around managing Lucene and the Translog in an open shard. >enhancement

Comments

@jpountz
Copy link
Contributor

jpountz commented Jul 24, 2018

The _forcemerge API has a max_num_segments whose only useful value is 1. It also has the drawback of potentially conflicting with the maximum merged segment size (index.merge.policy.max_merged_segment).

We could remove the max_num_segments setting and make _forcemerge merge down to the minimum number of segments that honors the maximum merged segment size.

Marking as a blocker because of interactions with https://issues.apache.org/jira/browse/LUCENE-7976.

@jpountz jpountz added >enhancement blocker v7.0.0 :Distributed Indexing/Engine Anything around managing Lucene and the Translog in an open shard. labels Jul 24, 2018
@elasticmachine
Copy link
Collaborator

Pinging @elastic/es-distributed

@danielkasen
Copy link

I'm not sure you'd want to move the ability to slowly move down to this. Currently, we use a stepping policy as it's more efficient (fast and less memory utilization). So going from 600 -> 20 -> 10 is quicker then going 600 -> 10

@jpountz
Copy link
Contributor Author

jpountz commented Feb 1, 2019

@danielkasen Are you really having shards that have 600 segments?

We made the decision that forcemerge should keep the max_num_segments parameter in #32291, closing.

@jpountz jpountz closed this as completed Feb 1, 2019
@danielkasen
Copy link

Sry that's the total primary index for a 750gb index. So was this backported to 6.5 then? Or do I have to upgrade to 6.6.0 at this point?

@danielkasen
Copy link

So currently on 6.5.4 when ever I try to do a FM it reduces the entire shard to 1 segment. I've been looking for a fix for this ever since this update and am curious about my options at this time for our production systems.

@jpountz
Copy link
Contributor Author

jpountz commented Feb 8, 2019

Thanks for reporting this, I looked into it and this seems to be an unexpected side-effect of https://issues.apache.org/jira/browse/LUCENE-7679. I just opened https://issues.apache.org/jira/browse/LUCENE-8688 to discuss it on the Lucene side.

@jpountz
Copy link
Contributor Author

jpountz commented Mar 19, 2019

@danielkasen LUCENE-8688 will be fixed in Lucene 8.1. Thanks again for reporting.

@danielkasen
Copy link

Any chance this will be back ported? As isn't 8.1 a long way out at this point? Just saw there was a vote for 8.0 the other day and trying to get a gauge of what version this would actually get released in and what month. Thank you for fixing this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
blocker :Distributed Indexing/Engine Anything around managing Lucene and the Translog in an open shard. >enhancement
Projects
None yet
Development

No branches or pull requests

5 participants