mirror of
https://github.com/katanemo/plano.git
synced 2026-04-26 17:26:26 +02:00
pushing docs updated (#508)
* pushing docs updated * Fixed README.md logo * Fixed README.md logo * Fixed README.md spacing * fixed tag line * LLM router doc fixes * minor logo and branding changes * minor changes to the README * minor changes to the README --------- Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-329.local>
This commit is contained in:
parent
7baec20772
commit
1bbd35f1ab
7 changed files with 32 additions and 28 deletions
|
|
@ -2,18 +2,16 @@
|
|||
|
||||
Intro to Arch
|
||||
=============
|
||||
AI demos are easy to build. But past the thrill of a quick hack, you are left building, maintaining and scaling low-level plumbing code for agents that slows down AI innovation.
|
||||
For example:
|
||||
|
||||
Arch is an intelligent proxy server designed agentic applications. **Move faster** by letting Arch handle the **pesky heavy lifting** in building agents:
|
||||
fast input clarification, agent routing, seamless integration of prompts with tools for common tasks, and unified access and observability of LLMs.
|
||||
- You want to build specialized agents, but get stuck writing **routing and handoff** code.
|
||||
- You bogged down with prompt engineering work to **clarify user intent and validate inputs**.
|
||||
- You want to **quickly and safely use new LLMs** but get stuck writing integration code.
|
||||
- You waste cycles writing and maintaining **observability** code, when it can be transparent.
|
||||
- You want to **apply guardrails**, but have to write custom code for each prompt and LLM.
|
||||
|
||||
Past the thrill of an AI demo, have you found yourself hitting these walls? You know, the all too familiar ones:
|
||||
|
||||
- You break a prompt into specialized ones, but **get stuck writing routing** and handoff logic?
|
||||
- You want use new LLMs, but **struggle to quickly add LLMs** without writing integration logic?
|
||||
- You're **trapped in tedious prompting work** to clarify inputs and user intents?
|
||||
- You're **wasting cycles** choosing and integrating **code for observability** instead of it just happening transparently?
|
||||
|
||||
And you think to yourself, can't I move faster by focusing on higher-level objectives in a language and framework agnostic way? Well, you can!
|
||||
Arch is designed to solve these problems by providing a unified, out-of-process architecture that integrates with your existing application stack, enabling you to focus on building high-level features rather than plumbing — all without locking you into a framework.
|
||||
|
||||
.. figure:: /_static/img/arch_network_diagram_high_level.png
|
||||
:width: 100%
|
||||
|
|
@ -22,7 +20,7 @@ And you think to yourself, can't I move faster by focusing on higher-level objec
|
|||
High-level network flow of where Arch Gateway sits in your agentic stack. Designed for both ingress and egress prompt traffic.
|
||||
|
||||
|
||||
**Arch Gateway was built by the contributors of Envoy Proxy with the belief that:**
|
||||
Arch is an AI-native proxy server and the universal data plane for AI built by the contributors of Envoy Proxy with the belief that:
|
||||
|
||||
*Prompts are nuanced and opaque user requests, which require the same capabilities as traditional HTTP requests
|
||||
including secure handling, intelligent routing, robust observability, and integration with backend (API)
|
||||
|
|
@ -41,8 +39,8 @@ This gives Arch several advantages:
|
|||
|
||||
* Arch can be deployed and upgraded quickly across your infrastructure transparently without the horrid pain of deploying library upgrades in your applications.
|
||||
|
||||
**Engineered with Fast LLMs:** Arch is engineered with specialized small LLMs that are designed for fast, cost-effective and accurate handling of prompts.
|
||||
These LLMs are designed to be best-in-class for critical prompt-related tasks like:
|
||||
**Engineered with Fast Task-Specific LLMs (TLMs):** Arch is engineered with specialized LLMs that are designed for the fast, cost-effective and accurate handling of prompts.
|
||||
These LLMs are designed to be best-in-class for critical tasks like:
|
||||
|
||||
* **Function Calling:** Arch helps you easily personalize your applications by enabling calls to application-specific (API) operations via user prompts.
|
||||
This involves any predefined functions or APIs you want to expose to users to perform tasks, gather information, or manipulate data.
|
||||
|
|
|
|||
|
|
@ -3,10 +3,10 @@
|
|||
|
||||
Overview
|
||||
============
|
||||
Welcome to Arch, The intelligent (edge and LLM) proxy server for agentic applications.
|
||||
`Arch <https://github.com/katanemo/arch>`_ is an AI-native proxy server and the universal data plane for AI - one that is natively designed to handle and process AI prompts, not just network traffic.
|
||||
|
||||
Built by contributors to the widely adopted `Envoy Proxy <https://www.envoyproxy.io/>`_, Arch helps you move faster by handling the pesky *low-level* work in AI agent development—fast input clarification, intelligent agent routing, seamless prompt-to-tool integration, and unified LLM access and observability—all without locking you into a framework.
|
||||
|
||||
Move **faster** by letting Arch handle the pesky heavy lifting in building agents: **fast input clarification**, **agent routing**,
|
||||
seamless integration of prompts with **tools for common tasks**, and **unified access and observability of LLMs**.
|
||||
|
||||
In this documentation, you will learn how to quickly set up Arch to trigger API calls via prompts, apply prompt guardrails without writing any application-level logic,
|
||||
simplify the interaction with upstream LLMs, and improve observability all while simplifying your application development process.
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue