pushing docs updated (#508)

* pushing docs updated

* Fixed README.md logo

* Fixed README.md logo

* Fixed README.md spacing

* fixed tag line

* LLM router doc fixes

* minor logo and branding changes

* minor changes to the README

* minor changes to the README

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-329.local>
This commit is contained in:
Salman Paracha 2025-06-17 08:16:42 -07:00 committed by GitHub
parent 7baec20772
commit 1bbd35f1ab
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
7 changed files with 32 additions and 28 deletions

View file

@ -2,18 +2,16 @@
Intro to Arch
=============
AI demos are easy to build. But past the thrill of a quick hack, you are left building, maintaining and scaling low-level plumbing code for agents that slows down AI innovation.
For example:
Arch is an intelligent proxy server designed agentic applications. **Move faster** by letting Arch handle the **pesky heavy lifting** in building agents:
fast input clarification, agent routing, seamless integration of prompts with tools for common tasks, and unified access and observability of LLMs.
- You want to build specialized agents, but get stuck writing **routing and handoff** code.
- You bogged down with prompt engineering work to **clarify user intent and validate inputs**.
- You want to **quickly and safely use new LLMs** but get stuck writing integration code.
- You waste cycles writing and maintaining **observability** code, when it can be transparent.
- You want to **apply guardrails**, but have to write custom code for each prompt and LLM.
Past the thrill of an AI demo, have you found yourself hitting these walls? You know, the all too familiar ones:
- You break a prompt into specialized ones, but **get stuck writing routing** and handoff logic?
- You want use new LLMs, but **struggle to quickly add LLMs** without writing integration logic?
- You're **trapped in tedious prompting work** to clarify inputs and user intents?
- You're **wasting cycles** choosing and integrating **code for observability** instead of it just happening transparently?
And you think to yourself, can't I move faster by focusing on higher-level objectives in a language and framework agnostic way? Well, you can!
Arch is designed to solve these problems by providing a unified, out-of-process architecture that integrates with your existing application stack, enabling you to focus on building high-level features rather than plumbing — all without locking you into a framework.
.. figure:: /_static/img/arch_network_diagram_high_level.png
:width: 100%
@ -22,7 +20,7 @@ And you think to yourself, can't I move faster by focusing on higher-level objec
High-level network flow of where Arch Gateway sits in your agentic stack. Designed for both ingress and egress prompt traffic.
**Arch Gateway was built by the contributors of Envoy Proxy with the belief that:**
Arch is an AI-native proxy server and the universal data plane for AI built by the contributors of Envoy Proxy with the belief that:
*Prompts are nuanced and opaque user requests, which require the same capabilities as traditional HTTP requests
including secure handling, intelligent routing, robust observability, and integration with backend (API)
@ -41,8 +39,8 @@ This gives Arch several advantages:
* Arch can be deployed and upgraded quickly across your infrastructure transparently without the horrid pain of deploying library upgrades in your applications.
**Engineered with Fast LLMs:** Arch is engineered with specialized small LLMs that are designed for fast, cost-effective and accurate handling of prompts.
These LLMs are designed to be best-in-class for critical prompt-related tasks like:
**Engineered with Fast Task-Specific LLMs (TLMs):** Arch is engineered with specialized LLMs that are designed for the fast, cost-effective and accurate handling of prompts.
These LLMs are designed to be best-in-class for critical tasks like:
* **Function Calling:** Arch helps you easily personalize your applications by enabling calls to application-specific (API) operations via user prompts.
This involves any predefined functions or APIs you want to expose to users to perform tasks, gather information, or manipulate data.

View file

@ -3,10 +3,10 @@
Overview
============
Welcome to Arch, The intelligent (edge and LLM) proxy server for agentic applications.
`Arch <https://github.com/katanemo/arch>`_ is an AI-native proxy server and the universal data plane for AI - one that is natively designed to handle and process AI prompts, not just network traffic.
Built by contributors to the widely adopted `Envoy Proxy <https://www.envoyproxy.io/>`_, Arch helps you move faster by handling the pesky *low-level* work in AI agent development—fast input clarification, intelligent agent routing, seamless prompt-to-tool integration, and unified LLM access and observability—all without locking you into a framework.
Move **faster** by letting Arch handle the pesky heavy lifting in building agents: **fast input clarification**, **agent routing**,
seamless integration of prompts with **tools for common tasks**, and **unified access and observability of LLMs**.
In this documentation, you will learn how to quickly set up Arch to trigger API calls via prompts, apply prompt guardrails without writing any application-level logic,
simplify the interaction with upstream LLMs, and improve observability all while simplifying your application development process.