llama-cpp: patch for vulkan support instead
All checks were successful
Build and Deploy / deploy (push) Successful in 7m23s

This commit is contained in:
2026-04-07 20:04:06 -04:00
parent fdc1596bce
commit 2884a39eb1
4 changed files with 404 additions and 9 deletions

12
flake.lock generated
View File

@@ -325,16 +325,16 @@
]
},
"locked": {
"lastModified": 1774922513,
"narHash": "sha256-TKk1i8AZzxy4/z0MkqKxoGf/CQDvoL+jo8JDtZeCRy8=",
"owner": "apollosenvy",
"lastModified": 1775603401,
"narHash": "sha256-kp+cnqLX+K4M6gBc5Iy4S+G0xkz78qVEcO1xmNTrtgM=",
"owner": "TheTom",
"repo": "llama-cpp-turboquant",
"rev": "9e80e93ceb115bc5055997c373d8c09bfa47a565",
"rev": "a4e8af4455d34d4872f967e615c8212643c2123e",
"type": "github"
},
"original": {
"owner": "apollosenvy",
"ref": "pr/vulkan-turbo3",
"owner": "TheTom",
"ref": "feature/turboquant-kv-cache",
"repo": "llama-cpp-turboquant",
"type": "github"
}