-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prodigy is not working well with Stable Diffusion 3.5 Medium training #27
Comments
Hey Bocchi-Chan2023, are you able to take note of the value of For example, here's the results of an SDXL LoRA training, batch 8, with a modified Prodigy that treats each parameter group independently (#20). From left to right, the graphs show the As you can see, both TEs hit a good LR quickly, but the Unet took until steps 200-300 to find a decent LR, and even then it continued to search. I've been experimenting with ways to combat this but haven't been successful so far. Also double check you're setting the regular LRs to 1 (as the LR is multiplied by Not sure if any of this will help, but sharing my experiences while playing around with the internals. |
Okay, I will try to record the value of d while adjusting the value of d0 |
I set d0 from 1e-6 to 1e-5 and prodigy seems to be working well! |
Thanks for sharing your experience and a especially thanks to @LoganBooker for giving a solution to the problem. Since the problem seems to be solved, I'm closing the issue, but feel free to reopen it if you have more questions. I'll also add a comment on changing |
I have been trying to train with the stock settings of this optimizer and have not been successful yet. Specifically, it seems that it is not learning nearly as well as it should.
adamW8bit seems to be working with 4e-4 lr
The text was updated successfully, but these errors were encountered: