-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Issues: microsoft/DeepSpeedExamples
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
How can I change the master_port when using deepspeed for multi-GPU on single node, i.e. localhost
#936
opened Nov 20, 2024 by
lovedoubledan
Does DeepSpeed's Pipeline-Parallelism optimizer supports skip connections?
#932
opened Oct 17, 2024 by
RoyMahlab
How to calculate training efficiency ,i.e tokens/sec of step 1 fine tuning of llama2 model ?
#923
opened Aug 29, 2024 by
sowmya04101998
an error with gradient checkpointing in DeepspeedChat step 3
#908
opened Jun 28, 2024 by
wangyuwen1999
[Error] AutoTune:
connect to host localhost port 22: Connection refused
#894
opened Apr 20, 2024 by
wqw547243068
How to use deepspeed for multi-node and multi-card task in slurm cluster
#893
opened Apr 18, 2024 by
dshwei
About multiple-thread attention computation on CPU using zero-inference example.
#886
opened Apr 4, 2024 by
luckyq
Suggested GPU to run the demo code of step2_reward_model_finetuning (DeepSpeed-Chat)
#885
opened Apr 4, 2024 by
wenbozhangjs
[REQUEST] More fine-grained distributed strategies for RLHF training
#884
opened Apr 3, 2024 by
krisx0101
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.