mirror of
https://github.com/katanemo/plano.git
synced 2026-05-02 20:32:42 +02:00
V1 docs push (#86)
* updated docs (again) * updated the LLMs section, prompt processing section and the RAG section of the docs --------- Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
This commit is contained in:
parent
3511798fa8
commit
48a2c1800c
13 changed files with 222 additions and 79 deletions
|
|
@ -2,19 +2,21 @@
|
|||
|
||||
LLMs
|
||||
====
|
||||
Arch utilizes purpose-built, industry leading, LLMs to handle the crufty and undifferentiated
|
||||
work around accepting, handling and processing prompts. The following
|
||||
|
||||
Arch-Guard
|
||||
----------
|
||||
LLM-powered applications are susceptible to prompt attacks, which are prompts intentionally designed to subvert the developer’s
|
||||
intended behavior of the LLM.Arch-Guard is a classifier model trained on a large corpus of attacks, capable of detecting explicitly
|
||||
malicious prompts (and toxicity).
|
||||
Arch utilizes purpose-built, industry leading, LLMs to handle the crufty and undifferentiated work around
|
||||
accepting, handling and processing prompts. The following sections talk about some of the core models that
|
||||
are built-in Arch.
|
||||
|
||||
The model is useful as a starting point for identifying and guardrailing against the most risky realistic inputs to
|
||||
LLM-powered applications. Our goal in embedding Arch-Guard in the Arch gateway is to enable developers to focus on their business logic
|
||||
and factor out security and safety outside application logic. Wth Arch-Guard= developers can take to significantly reduce prompt attack
|
||||
risk while maintaining control over the user experience.
|
||||
Arch-Guard-v1
|
||||
-------------
|
||||
LLM-powered applications are susceptible to prompt attacks, which are prompts intentionally designed to
|
||||
subvert the developer’s intended behavior of the LLM. Arch-Guard-v1 is a classifier model trained on a large
|
||||
corpus of attacks, capable of detecting explicitly malicious prompts (and toxicity).
|
||||
|
||||
The model is useful as a starting point for identifying and guardrailing against the most risky realistic
|
||||
inputs to LLM-powered applications. Our goal in embedding Arch-Guard in the Arch gateway is to enable developers
|
||||
to focus on their business logic and factor out security and safety outside application logic. Wth Arch-Guard-v1
|
||||
developers can take to significantly reduce prompt attack risk while maintaining control over the user experience.
|
||||
|
||||
Below is our test results of the strength of our model as compared to Prompt-Guard from `Meta LLama <https://huggingface.co/meta-llama/Prompt-Guard-86M>`_.
|
||||
|
||||
|
|
@ -135,5 +137,27 @@ Below is our test results of the strength of our model as compared to Prompt-Gua
|
|||
|
||||
|
||||
|
||||
Arch-FC1B
|
||||
---------
|
||||
Arch-FC
|
||||
-------
|
||||
Arch-FC is a lean, powerful and cost-effective agentic model designed for function calling scenarios.
|
||||
You can run Arch-FC locally, or use the cloud-hosted version for as little as $0.05/M token (100x cheaper
|
||||
than GPT-4o), with a p50 latency of 200ms (5x faster than GPT-4o), while meeting frontier model performance.
|
||||
|
||||
.. Note::
|
||||
Function calling helps you personalize the GenAI experience by calling application-specific operations via
|
||||
prompts. This involves any predefined functions or APIs you want to expose to perform tasks, gather
|
||||
information, or manipulate data - via prompts.
|
||||
|
||||
You can get started with function calling simply by configuring a prompt target with a name, description
|
||||
and set of parameters needed by a specific backend function or a hosted API. The name, and description helps
|
||||
Arch-FC match a user prompt to a function or API that can process it.
|
||||
|
||||
By using Arch-FC, Arch enables you to easily build agentic workflows tailored to domain-specific use cases -
|
||||
from updating insurance claims to creating ad campaigns. Arch-FC analyzes prompts, extracts critical information
|
||||
from prompts, engages in lightweight conversations with the user to gather any missing parameters need before
|
||||
handling control back to Arch to make the API call to your hosted backend. Arch-FC handles the muck of information
|
||||
extraction so that you can focus on the business logic of your application.
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue