llama-cpp: maybe use vulkan?
All checks were successful
Build and Deploy / deploy (push) Successful in 8m30s

This commit is contained in:
2026-04-06 02:12:46 -04:00
parent 3e46c5bfa5
commit 0a927ea893
3 changed files with 11 additions and 10 deletions

View File

@@ -29,7 +29,8 @@
};
llamacpp = {
url = "github:TheTom/llama-cpp-turboquant/feature/turboquant-kv-cache";
# url = "github:TheTom/llama-cpp-turboquant/feature/turboquant-kv-cache";
url = "github:apollosenvy/llama-cpp-turboquant/pr/vulkan-turbo3";
inputs.nixpkgs.follows = "nixpkgs";
};