-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Give Leo information about when text is being truncated so it can better formulate an answer. #33006
Comments
Is it enough to make the user aware it's cut off? #31405 |
For MVP? I think so |
After some testing, here are my findings: List of links tested
MethodologyTested summary plus a few questions about the context inside and outside of scope. Prompt changes tested:(changes signified in bold)
Models checked
ResultsMode 1 - PrependWorks okay, does not make much of a difference, unless you explicitly ask if the whole content was consumed. 13B model does not answer that correctly (assumes the whole article was read). Mode 2 - AppendWorks okay. Stronger conditioning that the whole output was not consumed. Mode 3 - SeedThis did not work at all. It affects the task completion significantly. RecommendationMaking the prompt truncation aware does not monotonically change the behaviour towards better responses. However, under some queries, it may yield that the whole content was not consumed instead of e.g. the cutoff point of training or lack of access to real information. The downside is that I also saw the behaviour of refusing to respond due to "ethics, etc.". For Claude-Instant, I did not test thoroughly as the context limit is quite high anyways. If we are to integrated it, I would vote for Mode 2. I can push the changes made. QA can further test more wide into the impact of this prompt change. |
An option is to just close this with that investigation too. Do you recommend that or would you prefer Mode 2? |
Keeping in mind that we do warn the user about it now even if the model doesn't know. |
Given that Llama-13B does not particularly "care" about this prompt change anyways (which is our public offering), I propose that we close this for now. |
thanks for checking 👍 |
The text was updated successfully, but these errors were encountered: