Skip to content

Bump llama-cpp-python to 0.2.36#5397

Merged
oobabooga merged 1 commit intodevfrom
llamacpp_0.2.36
Jan 30, 2024
Merged

Bump llama-cpp-python to 0.2.36#5397
oobabooga merged 1 commit intodevfrom
llamacpp_0.2.36

Conversation

@oobabooga
Copy link
Copy Markdown
Owner

@oobabooga oobabooga commented Jan 30, 2024

I had to remove older Python/CUDA versions due to GitHub rate limiting my Github Actions jobs.

@oobabooga oobabooga changed the base branch from main to dev January 30, 2024 02:30
@oobabooga oobabooga marked this pull request as draft January 30, 2024 02:30
@Ph0rk0z
Copy link
Copy Markdown
Contributor

Ph0rk0z commented Jan 30, 2024

It needs split by rows or layers setting as doing it by layers is slower, even with 3090s.

abetlen/llama-cpp-python#1085

@oobabooga oobabooga marked this pull request as ready for review January 30, 2024 16:15
@oobabooga
Copy link
Copy Markdown
Owner Author

Could you make a PR? I don't have 2 GPUs to properly test this.

@oobabooga oobabooga merged commit 89f6036 into dev Jan 30, 2024
@Ph0rk0z
Copy link
Copy Markdown
Contributor

Ph0rk0z commented Jan 30, 2024

Sure, I can PR it later on. Can just make a checkbox to uncheck.

@oobabooga oobabooga deleted the llamacpp_0.2.36 branch February 4, 2024 04:16
PoetOnTheRun pushed a commit to PoetOnTheRun/text-generation-webui that referenced this pull request Feb 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants