Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature request:support ExLlama #296

Closed
alanxmay opened this issue Jun 28, 2023 · 15 comments
Closed

Feature request:support ExLlama #296

alanxmay opened this issue Jun 28, 2023 · 15 comments

Comments

@alanxmay
Copy link

ExLlama (https://github.com/turboderp/exllama)

It's currently the fastest and most memory-efficient executor of models that I'm aware of.

Is there an interest from the maintainers in adding this support?

@SinanAkkoyun
Copy link

How do you plan on adding batched support for Exllama? I am very interested in your approach as I am trying to work on that too

@iibw
Copy link

iibw commented Dec 21, 2023

ExLlamaV2 has taken over ExLlama in quantization performance for most cases. I hope we can get it implemented in vLLM because it is also an incredible quantization technique. Benchmarks between all the big quantization techniques indicate ExLlamaV2 is the best out of all of them. Have there been any new developments since it was added to the roadmap?

@SinanAkkoyun
Copy link

Please, having exllamav2 with paged attention and with continuous batching would be a big win for the LLM world

@DaBossCoda
Copy link

Also looking forward to exllamav2 support

@RuntimeRacer
Copy link

RuntimeRacer commented Jan 1, 2024

I was hoping this would be possible, too. I recently worked with the Mixtral-8x7b Model; AWQ 4-bit had significant OOM / Memory overhead compared to ExLlama2 in 4-Bit; also I ended up just running the model in 8-bit using ExLlama2, since that turned out to be the best compromise between model capabilities and VRAM footprint. I can run it in 8-bit on 3x3090 and use full 32k context with ExLlama2; but I need 4x3090 to be even able to load it in 16-bit within VLLM; and i reach OOM when I try to use full context.

So this would definitely be an amazing addition to have more flexibility in terms of VRAM-Resources.

@theobjectivedad
Copy link

+1

1 similar comment
@tolecy
Copy link

tolecy commented Jan 5, 2024

+1

@chricro
Copy link

chricro commented Feb 23, 2024

+1

2 similar comments
@agahEbrahimi
Copy link

+1

@a-creation
Copy link

+1

@rjmehta1993
Copy link

This will be the biggest release for vllm to support exllamav2. +1

@sapountzis
Copy link

+1

1 similar comment
@kulievvitaly
Copy link

+1

yukavio pushed a commit to yukavio/vllm that referenced this issue Jul 3, 2024
…ct#296)

SUMMARY:
* update `nightly` workflow to actually use the `skip-for-nightly.txt`
skip list
Copy link

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!

@github-actions github-actions bot added the stale label Oct 31, 2024
Copy link

This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant. Thank you!

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Nov 30, 2024
billishyahao pushed a commit to billishyahao/vllm that referenced this issue Dec 31, 2024
* Run clang-format on develop

* applying the missing formatting
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

15 participants