mirror of
https://github.com/katanemo/plano.git
synced 2026-05-15 11:02:39 +02:00
Add support for local llm (mistral 7b) (#31)
This commit is contained in:
parent
b49fc2f264
commit
445b1ea210
24 changed files with 703 additions and 51 deletions
22
demos/weather-forecast-local-llm/README.md
Normal file
22
demos/weather-forecast-local-llm/README.md
Normal file
|
|
@ -0,0 +1,22 @@
|
|||
# Weather forecasting
|
||||
This demo shows how you can use intelligent prompt gateway to provide realtime weather forecast using Mistral LLM locally hosted using llama.cpp as LLM Hosting Service.
|
||||
|
||||
# Startig the demo
|
||||
1. Ensure that submodule is up to date
|
||||
```sh
|
||||
git submodule sync --recursive
|
||||
```
|
||||
1. Download mistral 7b model using following shell command
|
||||
```sh
|
||||
sh download_mistral_7b.sh
|
||||
```
|
||||
2. Start services
|
||||
```sh
|
||||
docker compose up
|
||||
```
|
||||
3. Navigate to http://localhost:18080/
|
||||
4. You can type in queries like "how is the weather in Seattle"
|
||||
1. You can also ask follow up questions like "show me sunny days"
|
||||
5. To see metrics navigate to "http://localhost:3000/" (use admin/grafana for login)
|
||||
1. Open up dahsboard named "Intelligent Gateway Overview"
|
||||
2. On this dashboard you can see reuqest latency and number of requests
|
||||
Loading…
Add table
Add a link
Reference in a new issue