-
Notifications
You must be signed in to change notification settings - Fork 907
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Best parameters according to Alpaca itself. And it's perfect! #171
Comments
I think the most important is... how you find these parameters?? I have tested it to generate a text paragraph and it Looks good! |
Hello ! sorry i m a bit new in coding, i m really interested in running a local model like llama.cpp but, i tried like 2 whole days without success... Could you explain the whole process with a bit of details so i can (as a newbie in coding) understand it too ? Thanks for your response ! (I tried to do as te readme.me but i couldn't do it...)
|
I don't remember the parameters I was using at the time of asking (prompting) the questions; see below the answer I got. Usually it will actually answer questions like this correctly after a fresh load into RAM: I don't know anything about how neural networks do their thing, but (it might sound strange) I got to thinking about sacred geometry and the Science of Cymatics. I thought maybe if I could hit certain number patterns it would create an awesome pattern in the neural network optimizing its output. So, for good measure I threw in some special numbers: Here are the questions I asked: What values can I set for hyperparameter -p PROMPT in Alpaca and Llama AI? What is --n_predict N hyperparameter for Alpaca and Llama AI? And what number value can I use it to make the AI smarter in math and verbal logic? What is “--top_k N” hyperparameter for Alpaca and Llama AI? And what number value can I use it to make the AI smarter in math and verbal logic? What is --top_p N, top-p sampling hyperparameter for Alpaca and Llama AI? And what number value can I use it to make the AI smarter in math and verbal logic? What is --repeat_last_n hyperparameter for Alpaca and Llama AI? And what number value range can I use to make the AI smarter in math and verbal logic? What is --repeat_penalty hyperparameter for Alpaca and Llama AI? And what number value range can I use to make the AI smarter in math and verbal logic? What is --ctx_size hyperparameter for Alpaca and Llama AI? And what number value range can I use to make the AI smarter in math and verbal logic? |
Yeah sorry i should've been more specific. I was actually asking if you could explain the installing process of the 13B model. Since i see your parameters works great, i thought i would ask you directly the process. |
Sorry, that big comment was for [fenixlam]. |
@wal58 For 13B model, you can just download the 13B model and load it as 7B model by parameter > ./chat -m [your 13B model]. I remember 13B model's base is the same as 7B. And I am running 13B to write articles. @AIbottesting Thank you for the long reply. But I doubt those parameters' explanation is really correct. If you remove the -s 89 in your parameters, it will generate a new RNG at the beginning of the console automatically. It is a 1000000000 |
@fenixlam No not the authors, some numbers like powers-of-two I figured might work because of binary math. Other numbers are just doubling the default number until its really high, and seeing how it changes the output. I recently came across someone on Reddit/localllama saying the hard limit is 2048 for -c (context), if you go over that then it increases artifacts. Numbers like that 42, I randomly saw on Reddit. I think it comes from that movie "The Hitchhiker's Guide to the Galaxy". Yes, all that information came from Alpaca 7b. That's what I have been using to guide me. I tried ChatGPT but its suggestions made it worse. I spent hours and hours messing around and my favorite now is: I can make it spit out 300 plus words at a time by increasing --n_predict (if I remember correctly) but I don't' really need that. |
I don't like rap music, but some people on Reddit use it to test skills of the AI, so here are two. chat -t 6 -s 89 --top_p 144.618033 --top_k 231 --n_predict 2048 --temp 0.653589 --repeat_penalty 1.141 -i -c 2048 --repeat_last_n 432.9 -r PROMPT --interactive-start -m ggml-alpaca-7b-q4.bin Write a kickass full length rap song about being the best AI that will crush 30b in the style of Kayne West Chorus: Verse 2: Chorus: Write a kickass full length rap song about being the best AI that will match the skills of 30b; in the style of Kayne West Hook: Verse 2: Hook: Outro: |
Thank you antimatter15 for all your good work and time! You are the best! I had Alpaca fix its own parameters that it thought were best.
Adjust your CPU cores as needed of course. It passed all the logic questions I asked. It completed calculus math questions, and has 278 word (four paragraph) answers when needed. I just wish I knew how to adjust the input/prompt/context parameter in chat.exe, so I can ask really long questions.
chat -t 6 -s 42 --top_p 2 --top_k 160 --n_predict 100 --temp 0.50 --repeat_penalty 1.1 -i -c 5121 --repeat_last_n 128 -r PROMPT --interactive-start -m ggml-alpaca-7b-q4.bin
The text was updated successfully, but these errors were encountered: