-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to load layer weights on Comfyui node #13
Comments
There should be directories named after the epoch, e.g. "epoch7", that have a safetensors file in them which is the lora. The files in the screenshot are deepspeed internal checkpoint files. The save frequency is controlled by the "save_every_n_epochs" config field which is 2 in the example config file. |
@tdrussell
so The config I used was
Perhaps that's why it didn't save any directories named after the epoch as you said because |
The checkpointing settings are only for deepspeed checkpoints which is this per-layer thing in your screenshots. And yes, probably you didn't train long enough for it to save even once. The num_repeats works like it does in Kohya sd-scripts, the items in your dataset are logically being repeated that many times, so one epoch takes longer. You can still "salvage" the training run by setting save_every_n_epochs to 1, then resuming using --resume_from_checkpoint (since you have deepspeed checkpoints), and then waiting for 1 epoch. Or even go into the train.py code and put a "saver.save_model('some_name')" right before the training loop to save immediately. |
@tdrussell
|
It will copy the current config file, whatever it is, at the moment it saves. Did you change the config file after starting training? Probably it should be changed to read all the file bytes at training startup and just write that. |
Thanks for sharing your work.
I don't know how to go about loading LoRA model on "Hunyuanvideo lora select" node.
Following is the weights I have after training LoRA with your script.
Are "layer_xx-model_states.pt" all different weights? Am I not supposed to use all layer weights together?
The text was updated successfully, but these errors were encountered: