Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Disable custom all reduce by default #2808

Merged
merged 1 commit into from
Feb 8, 2024
Merged

Disable custom all reduce by default #2808

merged 1 commit into from
Feb 8, 2024

Conversation

WoosukKwon
Copy link
Collaborator

This PR temporarily disables the custom all-reduce kernels. We will enable them once the stability issues are resolved.

@simon-mo simon-mo merged commit 3711811 into main Feb 8, 2024
17 checks passed
hanzhi713 added a commit to hanzhi713/vllm that referenced this pull request Feb 10, 2024
yhu422 added a commit to yhu422/vllm that referenced this pull request Feb 13, 2024
[ROCm] Fix build problem resulted from previous commit related to FP8 kv-cache support  (vllm-project#2790)

Add documentation on how to do incremental builds (vllm-project#2796)

[Ray] Integration compiled DAG off by default (vllm-project#2471)

Disable custom all reduce by default (vllm-project#2808)

add usage context

removed usage_context from Engine_args

Move IO to another process

added http request

[ROCm] support Radeon™ 7900 series (gfx1100) without using flash-attention (vllm-project#2768)

Add documentation section about LoRA (vllm-project#2834)

Refactor 2 awq gemm kernels into m16nXk32 (vllm-project#2723)

Co-authored-by: Chunan Zeng <[email protected]>

Added additional arg for from_engine_args

comments
alexm-neuralmagic pushed a commit to neuralmagic/nm-vllm that referenced this pull request Feb 13, 2024
jvmncs pushed a commit to jvmncs/vllm that referenced this pull request Feb 14, 2024
@WoosukKwon WoosukKwon deleted the disable-custom-ar branch February 15, 2024 06:47
xjpang pushed a commit to xjpang/vllm that referenced this pull request Feb 20, 2024
xjpang pushed a commit to xjpang/vllm that referenced this pull request Feb 22, 2024
xjpang pushed a commit to xjpang/vllm that referenced this pull request Mar 4, 2024
Temirulan pushed a commit to Temirulan/vllm-whisper that referenced this pull request Sep 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants