Skip to content

Conversation

@moejay
Copy link

@moejay moejay commented May 13, 2023

Add n_gpu_layers param to llama-cpp that allows some processing on the GPU

This is in preparation for added GPU support in llama-cpp
abetlen/llama-cpp-python#203 for when is approved and merged

This is the llama-cpp commit that this PR adds support for

Before submitting

  • No New integration

Who can review?

Thanks all for the wonderful project!

@m0sh1x2
Copy link

m0sh1x2 commented May 14, 2023

Is there an easy way to test all of the functionality in one go or there is quite a lot of chained merges below?

@moejay
Copy link
Author

moejay commented May 14, 2023

Is there an easy way to test all of the functionality in one go or there is quite a lot of chained merges below?

not that I can think of (This is dependent on one PR though) so maybe it's not too bad to test

you need the llama-cpp-python binding to be updated, and for that
you could pip install from https://github.com/moejay/llama-cpp-python (That's my branch with this updated)

I can add some more detailed testing instructions in the PR description when I come back later this evening if that helps.

I did notice that a bunch of the examples were broken (There are probably issues for those that exist at the moment, maybe I'll be able to contribute some fixes there later on)

@moejay
Copy link
Author

moejay commented May 15, 2023

Looks like the same change was merged a bit later here
So, this should now work with the latest llama-cpp-bindings

@dev2049
Copy link
Contributor

dev2049 commented May 17, 2023

duplicate of #4739

@dev2049 dev2049 closed this May 17, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants