Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improve gpt-3.5-turbo performance #1

Closed
cpacker opened this issue Oct 15, 2023 · 1 comment · Fixed by #58
Closed

Improve gpt-3.5-turbo performance #1

cpacker opened this issue Oct 15, 2023 · 1 comment · Fixed by #58
Assignees
Labels

Comments

@cpacker
Copy link
Collaborator

cpacker commented Oct 15, 2023

WIP

Current gpt-3.5-turbo performance is not very good due primarily due to bad understanding of function set (functions called at the wrong times, messages sent w/o call to send_message, etc.). Goal is to upload specific variations of the key prompts (tailored towards worse instruction following ability of 3.5) that provide reasonable performance.

@cpacker cpacker self-assigned this Oct 15, 2023
@cpacker cpacker added the enhancement New feature or request label Oct 15, 2023
@tomsib2001
Copy link

I'm not sure that this is possible for chatgpt-3.5 (or any version of chatGPT), but for open LLMs, there is this technique called grammar-based sampling which basically forces the LLM to answer according to some predetermined grammar, here is a Hacker News thread and link to a ggml PR describing it: https://news.ycombinator.com/item?id=36819906 . This could at least force the LLM, when applicable, to only answer using a preset list of commands.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants