-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Fix ExLlamaV2 context length setting (closes #5750)
- Loading branch information
Showing
1 changed file
with
4 additions
and
3 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
624faa1
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note that this isn't directly related to EXL2. EXL2 just retains the config.json from the original model (apart from adding the quantization_config key). So the Transformers loader would need this change as well for DRBX.
If
rope_theta
is used anywhere, also be aware that DBRX moved it into theattn_config
section for some reason.624faa1
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I hadn't noticed they had moved rope_theta somewhere else, thanks. That should be accounted for as well now 9ab7365