Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

KeyError in multihead.py example when not using num_workers on MPS #1656

Open
guilhermegog opened this issue Jul 9, 2024 · 0 comments
Open
Labels
bug Something isn't working

Comments

@guilhermegog
Copy link

guilhermegog commented Jul 9, 2024

Hey everyone, first off thank you for developing this amazing library.

🐛 Describe the bug
So apparently there is a bug, which I am not sure if it is a bug related to avalanche or mps, but if you attempt to run the multihead example without the num_workers parameter set (which I would assume would default to 1), the training loop stops working and a KeyError is raised, with a seemingly random key.

🐜 To Reproduce
In lines 71 and 72 of the multihead.py example just replace with the following lines:
strategy.train(train_task)
strategy.eval(test_stream)

🦋 Additional context
The issue only seems to occur when running the script on 'mps' devices, as when running the same piece of code on a server with cuda the issue does not persust

@guilhermegog guilhermegog added the bug Something isn't working label Jul 9, 2024
@guilhermegog guilhermegog changed the title KeyError in Multihead example when not using num_workers on MPS KeyError in multihead.py example when not using num_workers on MPS Jul 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant