Lead DevSecOps, Accenture Interactive FR
•
4 reviewsUsed as an LLM proxy, it allows the caching and load balancing between multiple AI services (Groq, OpenRouter, etc.) and even local with Ollama.
It uses an OpenAI-compatible API that allows (when we can set the base URL) to use it in many apps or services.
I use it configured with Langfuse which provides the performance analysis (monitoring) of each prompt/session.