GPTRouter is an open source LLM API Gateway that offers a universal API for 30+ LLMs, vision, and image models, with smart fallbacks based on uptime and latency, automatic retries, and streaming. Stay operational even when OpenAI is down.
Hi Product Hunt,
Today, we are excited to be open-sourcing GPTRouter, an LLMOps tool we have been using internally at Writesonic for handling millions of monthly requests for our users.
✅ Universal API for 30+ LLMs, Vision and Image Models
✅ Smart Fallbacks based on latency and uptime
✅ Automatic Retries
✅ Supports streaming
Since embracing OpenAI GPT-3 in production in 2020, we at Writesonic have been serving millions of users and faced the typical scaling pains with generative AI models:
1. Dependency on a single model risked total downtime.
2. Latency issues with models like GPT-4 affected user experience.
3. Integrating various models was tough due to different APIs and SDKs.
🌟 Early this year at Writesonic, we set out with a clear vision: to become model agnostic.
Faced with single-model limitations and diverse AI challenges, we began building GPTRouter - our bespoke solution to navigate and thrive in a multi-model AI world.
🔗 With GPTRouter's Universal API, you're the master of AI models.
Swap between OpenAI, Azure, Anthropic, Replicate, Cohere & more with just one line of code.
It simplifies model management to a great extent.
🛡️ Downtime isn't an option.
GPTRouter's Smart Fallbacks mean your service is always on.
You can define a hierarchy of models for each use case. GPTRouter will constantly check for uptime/downtime, latency and other factors, and automatically fallback to the next best model with zero interruption.
♻️ Say goodbye to manual retries.
GPTRouter does the heavy lifting with Automatic Retries for failed requests, keeping your AI services sharp and consistent.
🌟 GPTRouter's Edge:
✅ Universal API for seamless model switching.
✅ Smart, automatic fallbacks for continuous service.
✅ Reduced latencies for quick interactions.
This is just the starting point. We are also working on integrations with LangChain and LlamaIndex.
Additionally, we will also be open sourcing our frontend LLMOps layer that provides a playground to test multiple models in parallel, keep a tab on the latencies for each model, track tokens and costs for each model and user all in one place.
📈 With GPTRouter, we're not just solving our problems at Writesonic; we're offering a solution to the startups and companies looking to make use of Generative AI in production.
We are looking forward to seeing how developers leverage GPTRouter in their own use cases.
Thank you!
Does this handle anything with rate limiting, or does it allow for multiple instances of the same kind of model? we have azure openai deployments in multiple regions for instance.
Congratulations on the launch. I was looking for something like this. Just tried to find detailed documentation but can't find it. For example, I want to set 3 Open AI models as fallbacks if my primary model fails. Where should I make the changes?
Kudos to Writesonic for releasing GPTRouter! This Universal API for 30+ LLMs, with smart fallbacks, automatic retries, and streaming support, is a game-changer. Seamlessly switch models, ensure continuous service, and reduce latencies—GPTRouter is the future of Generative AI!
I think this is amazing since it's open source and customizable so people can use it as a base and build new products or modify it as per their needs instead of the modification living behind an enterprise paywall it's out in wild that makes the MOAT for this
The automatic retries and streaming add even more robustness to this already impressive gateway. The fact that it can stay operational even when OpenAI is down is a testament to its resilience and reliability.
An absolute must-have tool for anyone working with LLMs. Kudos to the team behind GPTRouter!
Great work Samanayou! You've made an amazing way to help us stay connected and avoid downtimes with one unique API. Kudos to you for creating this open source powerhouse!
Strengths:
Clear value proposition: Highlights the key benefits of avoiding OpenAI downtimes and accessing multiple LLMs with a single API.
Technical details: Mentions important features like smart fallbacks, retries, and streaming, appealing to developers.
Open source: Showcases transparency and community appeal.
Writesonic