0.0.7post2
Fixes & tweaks to the previous release:
- Automatically adjust batch size on longer contexts (32 for 512 tokens, 16 for 1024, 8 for 2048, decreasing like this until a minimum of 1)
- Apply dynamic max context length to reranking
Fixes & tweaks to the previous release: