Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bump llama-cpp-python to 0.2.36 #5397

Merged
merged 1 commit into from
Jan 30, 2024
Merged

Bump llama-cpp-python to 0.2.36 #5397

merged 1 commit into from
Jan 30, 2024

Conversation

oobabooga
Copy link
Owner

@oobabooga oobabooga commented Jan 30, 2024

I had to remove older Python/CUDA versions due to GitHub rate limiting my Github Actions jobs.

@oobabooga oobabooga changed the base branch from main to dev January 30, 2024 02:30
@oobabooga oobabooga marked this pull request as draft January 30, 2024 02:30
@Ph0rk0z
Copy link
Contributor

Ph0rk0z commented Jan 30, 2024

It needs split by rows or layers setting as doing it by layers is slower, even with 3090s.

abetlen/llama-cpp-python#1085

@oobabooga oobabooga marked this pull request as ready for review January 30, 2024 16:15
@oobabooga
Copy link
Owner Author

Could you make a PR? I don't have 2 GPUs to properly test this.

@oobabooga oobabooga merged commit 89f6036 into dev Jan 30, 2024
@Ph0rk0z
Copy link
Contributor

Ph0rk0z commented Jan 30, 2024

Sure, I can PR it later on. Can just make a checkbox to uncheck.

@oobabooga oobabooga deleted the llamacpp_0.2.36 branch February 4, 2024 04:16
PoetOnTheRun pushed a commit to PoetOnTheRun/text-generation-webui that referenced this pull request Feb 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants