We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi team,
From the code I can see the multiple gpu deployment is supported for persistent deployment with server method. Is it supported with pipeline method?
Thanks
The text was updated successfully, but these errors were encountered:
Yes we support multi-GPU support for the mii.pipeline: https://github.com/microsoft/DeepSpeed-MII#tensor-parallelism
mii.pipeline
You will need to launch script using the deepspeed launcher to enable model parallelism with the pipeline: deepspeed --num_gpus 2 pipeline.py
deepspeed
deepspeed --num_gpus 2 pipeline.py
Where pipeline.py could be:
pipeline.py
import mii pipe = mii.pipeline("mistralai/Mistral-7B-v0.1") response = pipe("DeepSpeed is", max_new_tokens=128) print(response)
Sorry, something went wrong.
mrwyattii
No branches or pull requests
Hi team,
From the code I can see the multiple gpu deployment is supported for persistent deployment with server method. Is it supported with pipeline method?
Thanks
The text was updated successfully, but these errors were encountered: