-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Resuming training resets the logged step number #12274
Comments
Any progress on this issue? |
Same bug here for PL 1.6.1. Any progress on this issue? |
Or how can we hack this to work around? |
just wondering if you could just point |
Don’t know how to achieve this. What do you mean by global_step / number of optimizers ? and why the _batches_that_stepped should be pointed to it? |
Hi, this workaround seems to work for my use case:
|
cc: @carmocca wdyt? |
🐛 Bug
The change introduced in #11805 causes a reset to the logged step number.
https://github.com/PyTorchLightning/pytorch-lightning/blob/49a4a36ad45b937dd0124ecfb08eb7400dbf3950/pytorch_lightning/trainer/connectors/logger_connector/logger_connector.py#L122
To Reproduce
The script will create two tensorboard logs:
Expected behavior
This was the behavior before #11805
Environment
cc @tchaton @rohitgr7 @akihironitta @awaelchli @ananthsub @ninginthecloud @carmocca
The text was updated successfully, but these errors were encountered: