llama-cpp: maybe use vulkan?

This commit is contained in:
2026-04-06 02:12:46 -04:00
parent 3e46c5bfa5
commit 0a927ea893
3 changed files with 11 additions and 10 deletions

View File

@@ -23,10 +23,10 @@ in
);
port = service_configs.ports.private.llama_cpp.port;
host = "0.0.0.0";
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.default);
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
extraFlags = [
# "-ngl"
# "12"
"-ngl"
"999"
"-c"
"65536"
"-ctk"