Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Problem with finetuning model #459

Closed
GooDRomka opened this issue May 17, 2023 · 2 comments · May be fixed by #471
Closed

Problem with finetuning model #459

GooDRomka opened this issue May 17, 2023 · 2 comments · May be fixed by #471

Comments

@GooDRomka
Copy link

I tried to fine-tune the model, took one example, and trained it for many epochs. I noticed that the loss decreased significantly, indicating that the model was learning. However, when I utilized the saved LoRa, I found that model was identical to the base model. I loaded both an overfitted model (trained for 25 epochs with 100 identical examples each) and an underfitted model (trained for 1 epoch). Surprisingly, both models produced the same text and achieved identical scores.

What are the chances that the LoRa was not saved correctly? Or maybe the problem is in the training parameters or in the training process?

@DingShi0
Copy link

DingShi0 commented May 18, 2023

I had a similar problem, and I solved it by looking at this answer. #363
You can start by looking at the size of the model file you have trained.

@GooDRomka
Copy link
Author

Solved. Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants