Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Phi2 multipack #1173

Merged
merged 11 commits into from
Jan 23, 2024
Merged

Phi2 multipack #1173

merged 11 commits into from
Jan 23, 2024

Conversation

winglian
Copy link
Collaborator

Description

enables multipack for phi and phi2 now that it's stable and merged into transformers.

@winglian winglian merged commit 814aee6 into main Jan 23, 2024
7 checks passed
@winglian winglian deleted the phi2-multipack branch January 23, 2024 17:54
djsaunde pushed a commit that referenced this pull request Dec 17, 2024
* phi2 multipack

* update validation and examples for phi

* more updates to phi examples

* make sure to use the correct collator for phi multipack

* phi needs attention mask now for multipack

* if the special token already exists in the tokenizer, don't require in lora modules to save

* fix qlora yml for phi, fix phi test validation

* test qlora too

* make sure flash attention is enabled for the test

* don't use remote code for phi anymore

* reduce sequence len for sample packing phi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant