-
-
Notifications
You must be signed in to change notification settings - Fork 142
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Should we merge Dask HPC Runners in here? #638
Comments
Thanks @jacobtomlinson for sharing this and opening this discussion!
I totally agree with this statement. I would love to see a single |
This is fantastic, @jacobtomlinson! Very cool stuff. I agree with @guillaumeeb that it would be really nice to have a Dask HPC package with everything in one place. I would volunteer to help maintain that package, since Dask MPI would get rolled into it. This reminds me that I have had a major PR in Dask MPI waiting for quite a while now. I should finish that and merge it in before moving DaskMPI into any new package. |
Awesome thanks @kmpaul. Your new implementation looks interesting, feel free to ping me again for review once the conflicts are worked out. I think my plan will look like the following:
I've submitted an abstract to EuroSciPy to talk about Dask and HPC (and GPUs/UCX) quite broadly, but if it gets accepted that could make a nice deadline to work towards and announce these changes. |
For a while I've been playing around with this prototype repo which implements Dask Runners for HPC systems. I'm motivated to reduce the fragmentation and confusion around tooling in the Dask HPC community, so perhaps this new code should live here.
In dask/community#346 I wrote up the difference between Dask Clusters and Dask Runners. The TL;DR is that a Cluster creates the scheduler and worker tasks directly, for example
dask_jobqueue.SLURMCluster
submits jobs to SLURM for each worker. A Dask Runner is different because it is invoked from within an existing allocation and populates that job with Dask processes. This the same as how Dask MPI works.SlurmRunner
ExampleIf I write a Python script and call it with
srun -n 6 python myscript.py
the script will be invoked by Slurm 6 times in parallel on 6 different nodes/cores on the HPC. The Dask Runner class then uses the Slurm process ID environment variable to decide what role reach process should play and uses the shared filesystem to bootstrap communications with a scheduler file.Should this live in
dask-jobqueue
?I'm at the point of trying to decide where this code should live within the Dask ecosystem. So far I have implemented
MPIRunner
andSlurmRunner
as a proof-of-concept. It would be very straight forward to write runners for other batch systems provided it is possible to detect the process ID/rank from the environment.I can imagine users choosing between
SLURMCluster
andSlurmRunner
depending on their use case and how they want to deploy Dask. There are pros/cons to each deployment model, for example the cluster can adaptively scale, but the runner only requires a single job submission which will guarantee better node locality. So perhaps it makes sense forSlurmRunner
to live here indask-jobqueue
and we can use documentation to help users choose the right one for them? (We can make the name casing more consistent).The
MPIRunner
andSlurmRunner
share a common base class, so I'm not sure if that meansMPIRunner
should also live here, or whether we should accept some code duplication and put it indask-mpi
?Alternatively my prototype repo could just move to
dask-contrib
and become a separate project?Or we could roll all of
dask-jobqueue
,dask-mpi
and the newdask-hpc-runners
into a singledask-hpc
package? Or pull everything intodask-jobqueue
?The Dask HPC tooling is currently very fragmented and I'm keen to make things better, not worse. But I'm very keen to hear opinions from folks like @guillaumeeb @lesteve @jrbourbeau @kmpaul @mrocklin on what we should do here.
The text was updated successfully, but these errors were encountered: