llama-cpp: disable flash attn

This commit is contained in:
2025-05-28 21:00:08 -07:00
parent 0f46de5eb7
commit efb0bd38e8

View File

@@ -20,7 +20,6 @@
host = "0.0.0.0";
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
extraFlags = [
"--flash-attn"
"-ngl"
"9999"
];