This repository has been archived on 2026-04-18. You can view files and clone it. You cannot open issues or pull requests or push a commit.
Simon Gardling 9baeaa5c23 llama-cpp: add grafana annotations for inference requests
Poll /slots endpoint, create annotations when slots start processing,
close with token count when complete. Includes NixOS VM test with
mock llama-cpp and grafana servers. Dashboard annotation entry added.
2026-04-02 17:43:49 -04:00
2026-03-03 14:26:42 -05:00
2026-04-02 13:13:38 -04:00
2026-03-21 10:26:28 -04:00
2026-03-03 14:30:43 -05:00
2026-03-31 12:47:09 -04:00
2026-03-03 14:29:00 -05:00
2026-03-03 14:30:47 -05:00
2026-04-02 13:42:39 -04:00
2026-04-02 13:42:39 -04:00
Description
No description provided
7.2 MiB
Languages
Nix 83.1%
Python 15.4%
Shell 1.5%