Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is multiple gpu supported with non-persistent pipeline #310

Closed
yaliqin opened this issue Nov 15, 2023 · 1 comment
Closed

Is multiple gpu supported with non-persistent pipeline #310

yaliqin opened this issue Nov 15, 2023 · 1 comment
Assignees

Comments

@yaliqin
Copy link

yaliqin commented Nov 15, 2023

Hi team,

From the code I can see the multiple gpu deployment is supported for persistent deployment with server method. Is it supported with pipeline method?

Thanks

@mrwyattii
Copy link
Contributor

Yes we support multi-GPU support for the mii.pipeline: https://github.com/microsoft/DeepSpeed-MII#tensor-parallelism

You will need to launch script using the deepspeed launcher to enable model parallelism with the pipeline:
deepspeed --num_gpus 2 pipeline.py

Where pipeline.py could be:

import mii
pipe = mii.pipeline("mistralai/Mistral-7B-v0.1")
response = pipe("DeepSpeed is", max_new_tokens=128)
print(response)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants