llama-cpp: patch for vulkan support instead
This commit is contained in:
12
flake.lock
generated
12
flake.lock
generated
@@ -325,16 +325,16 @@
|
||||
]
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1774922513,
|
||||
"narHash": "sha256-TKk1i8AZzxy4/z0MkqKxoGf/CQDvoL+jo8JDtZeCRy8=",
|
||||
"owner": "apollosenvy",
|
||||
"lastModified": 1775603401,
|
||||
"narHash": "sha256-kp+cnqLX+K4M6gBc5Iy4S+G0xkz78qVEcO1xmNTrtgM=",
|
||||
"owner": "TheTom",
|
||||
"repo": "llama-cpp-turboquant",
|
||||
"rev": "9e80e93ceb115bc5055997c373d8c09bfa47a565",
|
||||
"rev": "a4e8af4455d34d4872f967e615c8212643c2123e",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "apollosenvy",
|
||||
"ref": "pr/vulkan-turbo3",
|
||||
"owner": "TheTom",
|
||||
"ref": "feature/turboquant-kv-cache",
|
||||
"repo": "llama-cpp-turboquant",
|
||||
"type": "github"
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user