Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unusual Model Utility Gap: Gradient Difference vs Ascent (Llama2) #45

Open
jeongjin0 opened this issue Dec 2, 2024 · 0 comments
Open

Comments

@jeongjin0
Copy link

jeongjin0 commented Dec 2, 2024

Hi, I noticed that for Llama2(forget10), gradient difference shows much lower model utility (~0.27) than gradient ascent (~0.63) in the leaderboard. This seems unusual since gradient difference is designed to maintain performance on the retain set while unlearning.

Interestingly, in Phi model results, gradient difference shows higher utility than ascent as expected. Could you help explain this significant performance gap in Llama2 implementation?

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant