You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This causes problems with at least one model (Yi), see discussion here: 01-ai/Yi#5
The automatic BOS that gets prepended apparently confuses the model.
SpecialVocab in gguf.py already loads tokenizer_config.json (although only as a fallback currently). The main question is probably how to add it to the GGUF file - what key, etc.
The text was updated successfully, but these errors were encountered:
Have been testing Yi-6B over the server API, it becomes very repetitive and incoherent quickly. So instead of passing the prompt as a string (in Python) I tried passing it as a list like prompt = [2, "my prompt here"] and (2 being the override BOS Token ID I believe). This seems to completely fix the model, with no more repetition or incoherency issues. So yea something wrong with the automatic BOS.
These models are looking pretty interesting now that the 200K (!) context versions were released. If it works as described, a 34B with 200K context is pretty insane.
This causes problems with at least one model (Yi), see discussion here: 01-ai/Yi#5
The automatic BOS that gets prepended apparently confuses the model.
SpecialVocab
ingguf.py
already loadstokenizer_config.json
(although only as a fallback currently). The main question is probably how to add it to the GGUF file - what key, etc.The text was updated successfully, but these errors were encountered: