Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama.cpp Vulkan support #126

Open
atirut-w opened this issue Jan 6, 2025 · 1 comment
Open

Llama.cpp Vulkan support #126

atirut-w opened this issue Jan 6, 2025 · 1 comment
Labels
enhancement New feature or request

Comments

@atirut-w
Copy link

atirut-w commented Jan 6, 2025

Using Llama.cpp with Vulkan support should provide GPU acceleration on a wide range of graphics cards. It is also easier to integrate than Ollama. For a good example, see LM Studio.

@FrancescoCaracciolo FrancescoCaracciolo added the enhancement New feature or request label Jan 6, 2025
@FrancescoCaracciolo
Copy link
Collaborator

LM studio supports OpenAI like API, so you can use the OpenAI Handler and running it externally, if it solves your use case.
Maybe we can add a specialized handler that can do a few other things like for ollama.

For direct llama.cpp support, getting hardware acceleration work under Flatpak always requires more effort. I will check what I can do

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants