I had one Ollama instance and a question: can I queue inference tasks and route them with priorities? That question turned into two open source tools over about three days, from planning through to shipped repos. I don't fully need either of them yet.
I built them anyway. The starting point Ollama ships with no authentication. There's an OLLAMA_API_KEY option, but it's a single shared key — ever