Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

some question for dataparallel #33

Open
WFLiu0327 opened this issue May 6, 2023 · 2 comments
Open

some question for dataparallel #33

WFLiu0327 opened this issue May 6, 2023 · 2 comments

Comments

@WFLiu0327
Copy link

when i try to train this model on double GPUs, an error occurred, suggesting that in models.maniqa.py in line 117 x = torch.cat((x6, x7, x8, x9), dim=2) appeared some data in CUDA0, some data in cuda1, please ask how this problem should be solved? I don't see some suitable solution on the internet.

@TianheWu
Copy link
Collaborator

Hi, I know this question.
This reason is the problem of implementation. It just support one GPU to training.

@TianheWu
Copy link
Collaborator

One solution is to return the output of each layer directly, without using the SaveOut module

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

2 participants