Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Progressive training? #547

Open
immartian opened this issue Aug 22, 2024 · 4 comments
Open

Progressive training? #547

immartian opened this issue Aug 22, 2024 · 4 comments

Comments

@immartian
Copy link

Hi @karpathy , thanks for sharing this interesting series. They are not only educational but also inspiring.

While I enjoy the tutorial together with my own copycat of your code, I wonder if you can advise whether it's possible to incrementally train a model, say adding Leo Tolstoy onto Shakespere, rather than retraining a data set combined time by time?

I'm asking because I wonder any new methods different from GPTs that can learn progressively like the human brain. If so, we can probably realize some level of baby intelligence without directly going to LLM.

@WeileiZeng
Copy link

Surely, you can do it. You can change the data source in any epoch, and the model will learn from it.

@immartian
Copy link
Author

I guess what I'm asking is when you've pretrained a model, can you add incremental states upon it?

@chris-aeviator
Copy link

@immartian that's exactly what @WeileiZeng said - an epoch is the incremental next state that you are referring to and you can take a pre trained model and add more epochs to it, also with different data as the previous. You can find some useful info on what to train in which order in the Llama 3 paper (94 pages).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants