Replies: 1 comment
-
i found an issue. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I got an error when only subgraph forwarded using DDP training mode.
Dataloader provides batches with annotation only for several heads for one step, and leads to gradient calculation only for subgraph.
Env settings
Error :
ModelWrapper used
Possible solution could be link. passing allow_grad to autograd or distributed.autograd. But i need a little guidance to find the right code module
And one more important thing - train code works fine with single gpu (without DDP).
Beta Was this translation helpful? Give feedback.
All reactions