Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Specify GPU Selection (e.g., CUDA:0, CUDA:1) #1816

Open
RakshitAralimatti opened this issue Oct 30, 2024 · 4 comments
Open

Specify GPU Selection (e.g., CUDA:0, CUDA:1) #1816

RakshitAralimatti opened this issue Oct 30, 2024 · 4 comments

Comments

@RakshitAralimatti
Copy link

Hi,

Is there a way to specify which GPU to use for inference, such as restricting it to only cuda:0 or cuda:1 in the code? Or are there any workarounds for achieving this?

Thanks in advance.

@ExtReMLapin
Copy link
Contributor

You can use tensor split [1,0,0] to ignore cuda 1 and 2 and keep on 0.

Also use split mode none to increase perf if it stays on only one gpu

@RakshitAralimatti
Copy link
Author

Hi @ExtReMLapin Thanks for your reply!
I tried the way you said but got stuck so can you please elaborate in more detail way

@ExtReMLapin
Copy link
Contributor

It’s the Llama class arguments

@RakshitAralimatti
Copy link
Author

@ExtReMLapin Got it Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants