Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

espaloma training with reweighting (train_sampler) is not well tested #10

Open
kntkb opened this issue Mar 16, 2024 · 1 comment
Open
Labels
bug Something isn't working

Comments

@kntkb
Copy link
Collaborator

kntkb commented Mar 16, 2024

CUDA out of memory error is raised for tag <=0.1.2

torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 80.00 MiB (GPU 0; 10.75 GiB total capacity; 9.76 GiB >already allocated; 7.62 MiB free; 10.40 GiB reserved in total by PyTorch)
If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation.
See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF

From tag 0.1.3, gradient accumulation is used to resolve this problem but have not been well tested.

@kntkb
Copy link
Collaborator Author

kntkb commented Mar 16, 2024

Unexpected error from espfit/tests/test_app_train_sampler.py

RuntimeError: Trying to backward through the graph a second time (or directly access saved tensors after they have >already been freed). Saved intermediate values of the graph are freed when you call .backward() or autograd.grad(). >Specify retain_graph=True if you need to backward through the graph a second time or if you need to access saved >tensors after calling backward.

@kntkb kntkb added the bug Something isn't working label Mar 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant