Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning code? #1

Open
StrangeTcy opened this issue Jul 7, 2023 · 7 comments
Open

Finetuning code? #1

StrangeTcy opened this issue Jul 7, 2023 · 7 comments

Comments

@StrangeTcy
Copy link

That sounds massively interesting, and while we try to run inference and read the paper, should we expect the release of the finetuning code?

@syzymon
Copy link
Collaborator

syzymon commented Jul 8, 2023

Hi, thanks for interest in our work! That's right, we are currently supporting only inference. We are considering releasing examples for finetuning of our models in pytorch/huggingface API.

@memray
Copy link

memray commented Jul 11, 2023

@syzymon Is there any plan of releasing the training pipeline (is it based on the EasyLM library)?
Thank you!

@SUSTechBruce
Copy link

Hope to see your finetune code ASAP, since your work is very interesting!!!!

@syzymon
Copy link
Collaborator

syzymon commented Jul 25, 2023

The continued pretraining pipeline (used to train long_llama_3b base model) is based on EasyLM.

We are planning to release instruction tuning code in pytorch & checkpoints & examples early next week. Stay tuned!

@puddleglum56
Copy link

Will you also be releasing pretraining code? Since the contrastive training seems to be a very important element of your great results, it would be nice if we could try recreating it

@syzymon
Copy link
Collaborator

syzymon commented Jul 30, 2023

We are working on LongLLaMA v2, which will be a bigger release. After that we will release the pretraining code which is in JAX, based on EasyLM codebase - same as used for openllama pretraining. You can expect the instruction finetuning code in pytorch to be out very soon (basically next week). There are no plans to implement FoT pretraining in PyTorch on our side, as our compute is based on TPUs. Stay tuned for LongLLaMA v2 which will definitely be out there in August!

@syzymon
Copy link
Collaborator

syzymon commented Aug 5, 2023

In case you haven't seen, the instruction code is already there! see https://twitter.com/s_tworkowski/status/1687620785379360768 and READMEs in this repo for more details

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants