Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deepspeed-chat: train v_head when only optimizing lora #758

Merged
merged 2 commits into from
Oct 16, 2023

Conversation

mosheisland
Copy link
Contributor

When using only optimize lora, we still need to train the v_head parameter.

Change-Id: I252c3ee69819997bf336482c6779b070f2e76df8

When using only optimize lora, we still need to train the v_head parameter.

Change-Id: I252c3ee69819997bf336482c6779b070f2e76df8
Signed-off-by: Moshe Island <[email protected]>
@tjruwase tjruwase merged commit 5161c0f into microsoft:master Oct 16, 2023
2 checks passed
@mosheisland mosheisland deleted the 8_train_v_head_lora branch October 17, 2023 06:40
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants