You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I noticed that for Llama2(forget10), gradient difference shows much lower model utility (~0.27) than gradient ascent (~0.63) in the leaderboard. This seems unusual since gradient difference is designed to maintain performance on the retain set while unlearning.
Interestingly, in Phi model results, gradient difference shows higher utility than ascent as expected. Could you help explain this significant performance gap in Llama2 implementation?
Thanks!
The text was updated successfully, but these errors were encountered:
Hi, I noticed that for Llama2(forget10), gradient difference shows much lower model utility (~0.27) than gradient ascent (~0.63) in the leaderboard. This seems unusual since gradient difference is designed to maintain performance on the retain set while unlearning.
Interestingly, in Phi model results, gradient difference shows higher utility than ascent as expected. Could you help explain this significant performance gap in Llama2 implementation?
Thanks!
The text was updated successfully, but these errors were encountered: