Try it out! Benchmarks here: https://github.com/ggml-org/llama.cpp/discussions/10879
(ollama doesn’t support vulkan for some weird reason. I guess they never pulled the code from llama.cpp)
Try it out! Benchmarks here: https://github.com/ggml-org/llama.cpp/discussions/10879
(ollama doesn’t support vulkan for some weird reason. I guess they never pulled the code from llama.cpp)