GPT2-Fineweb10B #33
Replies: 3 comments 1 reply
-
Hi dustin, Thank you |
Beta Was this translation helpful? Give feedback.
-
@Zemulax yes no problem. I will post a link soon along with upload all the files to github and huggingface. I have 4 different models each with different parameters. I should be able to get this done tomorrow or the next day. |
Beta Was this translation helpful? Give feedback.
-
Here are the 2 repos I will add documentation soon. |
Beta Was this translation helpful? Give feedback.
-
model hosted on huggingface : https://huggingface.co/AIGym/gpt2-fineweb10
The model I trained is hosted on huggingface and is available to download and use as you want. I also included the log file with it from its training.
The model was trained on 1 4060ti over 5 days. It was trained on the 10b sample from the fineweb dataset.
We have now began tokienizing the 350B sample and will be training a much bigger version with a custom bpe tokenizer.
Just thought I would share what I have so far.
Beta Was this translation helpful? Give feedback.
All reactions