Reapply "llama-cpp: maybe use vulkan?"
All checks were successful
Build and Deploy / deploy (push) Successful in 2m17s

This reverts commit 9addb1569a.
This commit is contained in:
2026-04-07 19:12:57 -04:00
parent 88fc219f2d
commit 778b04a80f
3 changed files with 11 additions and 10 deletions

12
flake.lock generated
View File

@@ -325,16 +325,16 @@
]
},
"locked": {
"lastModified": 1775603401,
"narHash": "sha256-kp+cnqLX+K4M6gBc5Iy4S+G0xkz78qVEcO1xmNTrtgM=",
"owner": "TheTom",
"lastModified": 1774922513,
"narHash": "sha256-TKk1i8AZzxy4/z0MkqKxoGf/CQDvoL+jo8JDtZeCRy8=",
"owner": "apollosenvy",
"repo": "llama-cpp-turboquant",
"rev": "a4e8af4455d34d4872f967e615c8212643c2123e",
"rev": "9e80e93ceb115bc5055997c373d8c09bfa47a565",
"type": "github"
},
"original": {
"owner": "TheTom",
"ref": "feature/turboquant-kv-cache",
"owner": "apollosenvy",
"ref": "pr/vulkan-turbo3",
"repo": "llama-cpp-turboquant",
"type": "github"
}

View File

@@ -29,7 +29,8 @@
};
llamacpp = {
url = "github:TheTom/llama-cpp-turboquant/feature/turboquant-kv-cache";
# url = "github:TheTom/llama-cpp-turboquant/feature/turboquant-kv-cache";
url = "github:apollosenvy/llama-cpp-turboquant/pr/vulkan-turbo3";
inputs.nixpkgs.follows = "nixpkgs";
};

View File

@@ -23,10 +23,10 @@ in
);
port = service_configs.ports.private.llama_cpp.port;
host = "0.0.0.0";
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.default);
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
extraFlags = [
# "-ngl"
# "12"
"-ngl"
"999"
"-c"
"65536"
"-ctk"