llama-cpp: cpu only
All checks were successful
Build and Deploy / deploy (push) Successful in 20m0s
All checks were successful
Build and Deploy / deploy (push) Successful in 20m0s
This commit is contained in:
@@ -17,10 +17,10 @@
|
||||
);
|
||||
port = service_configs.ports.private.llama_cpp.port;
|
||||
host = "0.0.0.0";
|
||||
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
|
||||
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.default);
|
||||
extraFlags = [
|
||||
"-ngl"
|
||||
"12"
|
||||
# "-ngl"
|
||||
# "12"
|
||||
"-c"
|
||||
"16384"
|
||||
"-ctk"
|
||||
|
||||
Reference in New Issue
Block a user