Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make Stored Compression Configurable #4160

Closed
kimchy opened this issue Nov 13, 2013 · 8 comments
Closed

Make Stored Compression Configurable #4160

kimchy opened this issue Nov 13, 2013 · 8 comments

Comments

@kimchy
Copy link
Member

kimchy commented Nov 13, 2013

Allow to make the stored compression configurable on the index level (index.store.compression.level), and use that when constructing the codec to use.

If possible (will be tricky), allow to set it on a live index, so for example, for time base indices, older indices can move to use higher compression mode and optimized (which should probably be done over a different issue once the setting is implemented).

@ghost ghost assigned jpountz Nov 13, 2013
@avleen
Copy link

avleen commented Nov 13, 2013

This would be a huge plus to us right now - we're getting ready to set up a very large cluster (~30Tb+ data), and the current compression (which is designed to be a somewhat compressive, but mostly fast) is painful.

Being able to set a high compression rate, and accept that it will make things slightly slower, is a really big win.

@kimchy
Copy link
Member Author

kimchy commented Nov 14, 2013

Indeed, note, this will require some changes in Lucene as well, which we are working on to make it happen. We obviously will keep this issue updated with progress.

@kevinkluge kevinkluge added v1.1.0 and removed v1.0.1 labels Feb 11, 2014
@s1monw s1monw added v1.2.0 and removed v1.1.0 labels Mar 12, 2014
@s1monw
Copy link
Contributor

s1monw commented Mar 12, 2014

pushing out to 1.2

@jpountz jpountz removed the v1.2.0 label Mar 13, 2014
@avleen
Copy link

avleen commented Jul 5, 2014

Hi folks, we're into 1.2 now. Is this still on the map? I would still gladly give up some Elasticsearch CPU time for better compression :-)

@kimchy
Copy link
Member Author

kimchy commented Jul 5, 2014

@avleen this is tricky to implement, effectively, it means that we would somehow change Lucene configuration (codec) to have high LZ4 compression on the fly, which is not possible today. We are investigating how to properly expose it in Lucene safely, which proves to be a bit tricky. The hope is to address it soon, but no concrete timeframe. /cc @jpountz

@avleen
Copy link

avleen commented Jul 7, 2014

Thanks Shay!

On Sat, Jul 5, 2014 at 1:15 PM, Shay Banon [email protected] wrote:

@avleen https://github.com/avleen this is tricky to implement,
effectively, it means that we would somehow change Lucene configuration
(codec) to have high LZ4 compression on the fly, which is not possible
today. We are investigating how to properly expose it in Lucene safely,
which proves to be a bit tricky. The hope is to address it soon, but no
concrete timeframe. /cc @jpountz https://github.com/jpountz

Reply to this email directly or view it on GitHub
#4160 (comment)
.

@jpountz
Copy link
Contributor

jpountz commented Sep 3, 2014

FYI there is now an open issue on the Lucene side: https://issues.apache.org/jira/browse/LUCENE-5914

@kimchy
Copy link
Member Author

kimchy commented Dec 11, 2014

closed in favor of #8863

@kimchy kimchy closed this as completed Dec 11, 2014
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

6 participants