Skip to content

uk4b Transformers

Compare
Choose a tag to compare
@proger proger released this 07 May 15:08
· 235 commits to main since this release

This release introduces a REPL for models trained for my and @dchaplinsky's paper on GPT-2 Metadata Pretraining Towards Instruction Finetuning for Ukrainian.

The REPL is accessible via a new CLI program, hat.

To use hat, first install some additional dependencies and models:

pip install haloop --upgrade               # make sure you have at least 0.0.7
pip install bitsandbytes sentencepiece     # I opted for not installing these as dependencies for now

wget https://a.wilab.org.ua/gpt/wiki.model  # sentencepiece tokenizer
wget https://a.wilab.org.ua/gpt/ckpt10m.pt  # model checkpoint for GPT-2 Large

Now, start the REPL:

hat --spm wiki.model ckpt10m.pt