Arch
p/arch-2
Build fast, specialized agents with intelligent infra
Kevin William David
Arch — Build fast, hyper-personalized agents with intelligent infra
Featured
44
Arch is an intelligent infrastructure primitive to help developers build fast, personalized agents in mins. Arch is a gateway engineered with LLMs to seamlessly integrate prompts with APIs, and to transparently add safety and tracing features outside app logic
Replies
Salman Paracha
Hello PH! My name is Salman and I work on Arch - an open source infrastructure primitive to help developers build fast, personalized agent in minus. Arch is an intelligent prompt gateway engineered with (fast) LLMs for the secure handling, robust observability, and seamless integration of prompts with your APIs - all outside business logic. Arch is built on (and by the contributors of) Envoy with the belief that: Prompts are nuanced and opaque user requests, which require the same capabilities as traditional HTTP requests including secure handling, intelligent routing, robust observability, and integration with backend (API) systems for personalization – all outside business logic. Arch handles the critical but undifferentiated tasks related to the handling and processing of prompts, including detecting and rejecting jailbreak attempts, intelligently calling "backend" APIs to fulfill the user's request represented in a prompt, routing to and offering disaster recovery between upstream LLMs, and managing the observability of prompts and LLM interactions in a centralized way. ⭐ Core Features: 🏗️ Built on Envoy: Arch runs alongside application servers, and builds on top of Envoy's proven HTTP management and scalability features to handle ingress and egress traffic related to prompts and LLMs. 🤖 Function Calling: For fast agentic and RAG apps. Engineered with SOTA.LLMs to handle fast, cost-effective, and accurate prompt-based tasks like function calling, and parameter extraction from prompts. Our models can run under <200 ms!! 🛡️ Prompt Guard: Arch centralizes prompt guards to prevent jailbreak attempts and ensure safe user interactions without writing a single line of code. 🚦 Traffic Management: Arch manages LLM calls, offering smart retries, automatic cut over, and resilient upstream connections for continuous availability between LLMs or a single LLM provider with multiple versions 👀 OpenTelemetry Tracing, Metrics and Logs : Arch uses the W3C Trace Context standard to enable complete request tracing across applications, ensuring compatibility with exiting observability tools, and provides metrics to monitor latency, token usage, and error rates, helping optimize AI application performance. - Visit our Github page to get started (and ⭐️ the project 🙏) : https://github.com/katanemo/arch - To learn more about Arch our docs: https://docs.archgw.com/ A big thanks 🙏 to my incredibly talented team who helped us to our first milestone as we re:invent infrastructure primitives for Generative AI.
Salman Paracha
@alex_tartach Thanks Alex - building this was a lot of fun, and its early days for us. Packing intelligence in infrastructure to help developers build fast agents (faster than before) is the ultimate goal
Naeem ul Haq
Congrats on the launch. Would love to try it, especially the prompt guard. Onward!
Salman Paracha
@_naeemulhaq Thanks Naeem - deeply appreciate the kind words. Would love to hack away with your team and see how we can help you move faster in building fast personalized agents...
Adil Hafeez
@_naeemulhaq thanks Naeem. Try it out here at https://github.com/katanemo/arch/. We would love to know your feedback.
Ishwar Jha
congratulations 👏 I am going to dive in
Salman Paracha
@ishwarjha than you. Would love the feedback as we build Arch as an open source project
Adil Hafeez
@ishwarjha thank you @ishwarjha 🙏
José Ulises Niño Rivera
Hi I am Jose, current owner of Pukka Built. I contracted with Katanemo to help the team get Arch off the ground. It was great to bootstrap Arch alongside Adil and Salman as they get to launch it for developers and platform teams. As someone who built Envoy at Lyft, I can attest to the durability of Envoy as a design choice. For those who need efficient, reliable handling of LLM requests, Arch is a strong addition to the stack. For the same reasons the out of process architecture for Envoy was a solid design choice, Arch benefits as well: giving teams more control over how prompts are managed without impacting existing services.
Salman Paracha
@junr03 Deeply appreciate your help and support. You are a critical part of our journey and couldn't have built this without you Jose!
Adil Hafeez
@junr03 thanks Jose
Tanmay Parekh
All the best for the launch @salman_paracha & team!
Salman Paracha
@parekh_tanmay thank you 🙏 🙏
Adil Hafeez
Kane
Launching soon!
Congrats, Salman! Sounds awesome! I’m following this project on GitHub. Keep it going! 🚀
Adil Hafeez
@blueeon thanks Kane. Do give it a shot at https://github.com/katanemo/arch/. And we will love to hear your feedback.
Simon Peter Damian
Congrats on your launch 🚀🚀🚀 --- Arch seems cool and promising. I haven't tried it out yet, but I do have a few questions. Langchain currently dominates this space, I see your pitch, but Arch doesn't seem to fit in the toolbox for users new to building GenAI apps. It does have its benefits for prompt-heavy applications, but I don't see the appeal for users new to building GenAI apps except that these folks aren't your target audience.
Salman Paracha
@theterminalguy Simon thanks for the question. Langchain and Arch are better together - for all the business logic around agents like chunking data, chaining calls to LLMs before returning a final response Langchain will continue to offer developers the fine-grained control over the business logic of agents. Arch handles all the crufty work that can be transparently managed outside business logic like guardrails, end-to-end tracing, and being able to route calls to APIs based on the true intent of the user prompt in under 300 ms - Arch is uniquely intelligent, distributed and out of process giving time back to developers to think about their differentiation.
Adil Hafeez
Hello! My name is Adil Hafeez, and I am the Co-Founder at Katanemo and the lead developer behind Arch. Previously I worked on Envoy at Lyft. Arch is engineered with purpose-built LLMs, it handles the critical but undifferentiated tasks related to the handling and processing of prompts, including detecting and rejecting jailbreak attempts, intelligently calling “backend” APIs to fulfill the user’s request represented in a prompt, routing to and offering disaster recovery between upstream LLMs, and managing the observability of prompts and LLM interactions in a centralized way - all outside business logic. Here are some additional key details of the project, * Built on top of Envoy and is written in rust. It runs alongside application servers, and uses Envoy's proven HTTP management and scalability features to handle traffic related to prompts and LLMs. * Function calling for fast agentic and RAG apps. Engineered with purpose-built fast LLMs to handle fast, cost-effective, and accurate prompt-based tasks like function/API calling, and parameter extraction from prompts. * Prompt guardrails to prevent jailbreak attempts and ensure safe user interactions without writing a single line of code. * Manages LLM calls, offering smart retries, automatic cutover, and resilient upstream connections for continuous availability. * Uses the W3C Trace Context standard to enable complete request tracing across applications, ensuring compatibility with observability tools, and provides metrics to monitor latency, token usage, and error rates, helping optimize AI application performance. We love arch, love open source and would love to build alongside the community. Please leave a comment or feedback here and I will be happy to answer!
Sarmad Qadri
Congrats on the launch! Really great project -- I believe in the premise of a gateway that consolidates a lot of the infrastructure work needed for any LLM project.
Salman Paracha
@saqadri Thanks Sarmad - really appreciate you taking the time to dig deeper on the project, and believing in the premise
Jai from Worksaga
Congratulations to the team on the launch of Arch! This tool sounds exciting for developers. Building fast, personalized agents in minutes is impressive
Salman Paracha
@jai_singhal Deeply appreciate the kind words. Would love for you to try it out and offer feedback!
Huzaifa Shoukat
Congrats on the launch! Love the speed and ease Arch brings to building personalized agents. Quick question: How does Arch handle data privacy and security?
Salman Paracha
@ihuzaifashoukat We engineered small LLMs that are exceptional for specific tasks around safety. For example, the Arch-Guard model (86M) offers state-of-the-art (SOTA) peformance for jailbreak scenarios.
Wood Peng
This is awesome! Arch is a game-changer for building personalized agents. I love the idea of using Envoy as the foundation, as it's known for its scalability and reliability. The focus on prompt safety and observability is crucial for building trustworthy AI systems. I'm particularly excited about the fast function calling and parameter extraction capabilities – this will be a huge time-saver. I'm definitely going to check out the docs and give Arch a spin!
Salman Paracha
@peng_wood thank you! Great to have you looking at the docs and giving the project a spin. Would love any feedback as you try it out
Mari Bu
Wow grate job! I like it! Could we modeling agent based models on it?
Munna Aziz
Congrats, Salman and the Arch team! 🎉 Arch sounds like a powerful solution for making prompt management and observability in AI-driven applications much more efficient. Love the focus on secure handling, robust traffic management, and fast function calling—perfect for anyone scaling agent-based and RAG apps.
Sarmad Siddiqui
Impressive work - At Meta we have the same core belief that safety of agents is paramount and as much as possible if we can tackle those concerns early in the request path the better - Arch feels like a great fit for responsible and safe AI - not to mention the other super powers it offers developers. One quick question: can you elaborate more about the prompt guard model, I see that you guys fined tuned it over the prompt guard from Meta?"
Salman Paracha
@sarmad_siddiqui Thank you! Yes, Arch uses purpose-built LLMs for guardrails The Arch-Guard collection of models can be found here. https://huggingface.co/collectio.... We fine-tuned over Meta's prompt guard and the optimization was to improve TPR (+4%) without impacting FPR. This was for the jailbreak use case, and the next set of baseline guardrails will include toxicity, harmfulness, etc.
Maximo Hartliness
I’m intrigued by the Prompt Guard feature. Ensuring safe user interactions is crucial, and having something handle that automatically sounds like it could simplify things a lot for devs focused on security. @adil_hafeez
Salman Paracha
@adil_hafeez @maximo_hartliness most certainly. That’s a core capability deeply integrated in Arch
Michael van Dijken
This is a cool project and great to see it come to fruition. We are working on a GenAI-based tool that will involve a front-end of sorts (lets call it an agent), and likely leverage several LLMs on the back-end - depending on the type of request. Is this a good use case for Arch and if so, I'd love to get my team engaged here. Thanks @salman_paracha and team!
Salman Paracha
@michael_pmm_nerd that's exactly one of the core use cases for Arch. Would love to show the team how we can help them move faster with Arch
Puja Sharma
I can see Arch being super helpful for streamlining complex prompt flows. @shuguang_chen
Salman Paracha
@shuguang_chen @puja_sharma11 Thank you! Would love for you to give it a spin. Always open to feedback.
Sheharyar Mehmood
Great work guys. Congrats on launching this.
André J
What's a personalised agent? Web chatbot or personal assistant or? This field is moving so fast, it's hard to know what terms means these days. Thanks 🙏
Salman Paracha
@sentry_co personalized means to customize the agent to be unique. Most agents are just summarizing over some data. With Arch you can build something very tailored like creating ad campaigns via prompts or updating insurance claims - and offer generative summaries in the same experience