-
Notifications
You must be signed in to change notification settings - Fork 419
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Chinese-Vicuna-lora-7b-3epoch-belle-and-guanaco这个模型是怎么训练出来的 #87
Comments
1、用我们的finetune脚本,把DATA_PATH设置成下载后的merge.json跑出来的 |
感谢回答 |
3、都可以,7b,13b,30b等把基底模型改一下都能训 |
也就是你们的代码是 |
最基础的流程可以这么理解 |
感谢回答 |
本人第一次接触这类项目,请问文档里的Chinese-Vicuna-lora-7b-3epoch-belle-and-guanaco这个模型是怎么训练出来的,finetune.py的输出又是什么,checkpoint-final是什么,和13B-based lora model什么关系。另外CPU推理的tools/merge_lora_for_cpp.py这个文件也没有。可以从huggingface上加载我们的模型或其他lora模型,为什么文件名是generate.py,需要generate生成什么呢?断点重训/增量训练支不支持单卡写法呢?哪些是你们自己训练的或是三方的,哪些是我们需要我们自己可以训练的呢?
The text was updated successfully, but these errors were encountered: