-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support to run distributed training tests in CI #926
Labels
Comments
This seems straightforward. The CI job running inside the container can see all gpus. So we can either add one item in the bash script, or run it as a separate step in CI. I will try to include it. |
just a follow-up on this, the CI environment already provides 2 GPUs per job, the remaining tasks are to
|
wyli
added a commit
that referenced
this issue
Nov 27, 2020
* fixes #926 Signed-off-by: Wenqi Li <[email protected]>
wyli
added a commit
to wyli/MONAI
that referenced
this issue
Dec 15, 2020
Signed-off-by: Wenqi Li <[email protected]>
wyli
added a commit
to wyli/MONAI
that referenced
this issue
Dec 15, 2020
Signed-off-by: Wenqi Li <[email protected]>
IsaacYangSLA
added a commit
that referenced
this issue
Jan 4, 2021
Signed-off-by: Wenqi Li <[email protected]> Co-authored-by: Nic Ma <[email protected]> Co-authored-by: Isaac Yang <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Is your feature request related to a problem? Please describe.
Now we already have some distributed training test cases, like: https://github.com/Project-MONAI/MONAI/blob/master/tests/test_handler_rocauc_dist.py
Need to be executed in our CI system.
We can run with 2 GPUs in 1 node first.
The text was updated successfully, but these errors were encountered: