This repository has been archived on 2026-04-18. You can view files and clone it. You cannot open issues or pull requests or push a commit.
Simon Gardling df15be01ea llama-cpp: pause xmrig during active inference requests
Add sidecar service that polls llama-cpp /slots endpoint every 3s.
When any slot is processing, stops xmrig. Restarts xmrig after 10s
grace period when all slots are idle. Handles unreachable llama-cpp
gracefully (leaves xmrig untouched).
2026-04-02 17:43:07 -04:00
2026-03-03 14:26:42 -05:00
2026-04-02 13:13:38 -04:00
2026-03-21 10:26:28 -04:00
2026-03-03 14:30:43 -05:00
2026-03-31 12:47:09 -04:00
2026-03-03 14:29:00 -05:00
2026-03-03 14:30:47 -05:00
2026-04-02 13:42:39 -04:00
2026-04-02 13:42:39 -04:00
Description
No description provided
7.2 MiB
Languages
Nix 83.1%
Python 15.4%
Shell 1.5%