-
Notifications
You must be signed in to change notification settings - Fork 4.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] [master] Garbage GPT-Neo-X output when using multi-gpu inference #2293
Comments
Hey, any updates on this? |
@mrwyattii That would be great, thanks! |
@mrwyattii |
@ryanai3 i have try the same deepspeed inference code , i encounter the errors as follows, does any suggestion? thanks !
|
This was referenced Mar 7, 2023
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Describe the bug
Similar to #2233 and #2133 I'm seeing garbage output when using multi-gpu fp16 inference for gpt-neo-x. Running the script below, replacing Gpt-Neo-X with GPT-Neo-2.7B works fine.
Output from 2 3090s with Deepspeed inference:
"Deepspeed is BytePtrFromStringgranwasysym BytePtrFromString BytePtrFromString BytePtrFromString BytePtrFromString BytePtrFromString _ BytePtrFromStringHypergranTal 2011 BytePtrFromString BytePtrFromString **j BytePtrFromString BytePtrFromString BytePtrFromStringgran¶Enggrantwgran _ BytePtrFromStringgran ausgranENTRY¶`){#Delta¶sysEveramssymbitgran`Ever last`grangran ** deliberate ENTRY stag Eng` BytePtrFromStringwasysym _ BytePtrFromStringwasysymBOX Eng...](granModelupgreek BytePtrFromStringamssymb BytePtrFromStringwasysym BytePtrFromStringSegment BytePtrFromString BytePtrFromString _ BytePtrFromString BytePtrFromStringupgreekEverEng_( **gran mistENTRY BytePtrFromString BytePtrFromString _amssymbwasysym..." last BytePtrFromStringwasysym BytePtrFromString BytePtrFromStringgrangran ever"
Note that 'BytePtrFromString' has shown up as the beginning of the generated tokens for every prompt I've used.
Output from 2 3090s with huggingface accelerate (way slower than deepspeed):
"Deepspeed is \nan on-line digital media company created in January 2002. Over the past 10 \nyears, Deepspeed has provided a comprehensive digital entertainment network to\n businesses throughout the US"
To Reproduce
Steps to reproduce the behavior:
Expected behavior
I would expect output that makes sense, like when using accelerate.
ds_report output
System info (please complete the following information):
Launcher context
launching with deepspeed:
deepspeed --num_gpus 2 script.py
Docker context
Additional context
When in docker, run
eval $(pdm venv activate docker)
to activate the venv, then run the deepspeed commandThe text was updated successfully, but these errors were encountered: