mirror of
https://github.com/katanemo/plano.git
synced 2026-05-15 02:52:40 +02:00
22 lines
852 B
Markdown
22 lines
852 B
Markdown
# Weather forecasting
|
|
This demo shows how you can use intelligent prompt gateway to provide realtime weather forecast using Mistral LLM locally hosted using llama.cpp as LLM Hosting Service.
|
|
|
|
# Startig the demo
|
|
1. Ensure that submodule is up to date
|
|
```sh
|
|
git submodule sync --recursive
|
|
```
|
|
1. Download mistral 7b model using following shell command
|
|
```sh
|
|
sh download_mistral_7b.sh
|
|
```
|
|
2. Start services
|
|
```sh
|
|
docker compose up
|
|
```
|
|
3. Navigate to http://localhost:18080/
|
|
4. You can type in queries like "how is the weather in Seattle"
|
|
1. You can also ask follow up questions like "show me sunny days"
|
|
5. To see metrics navigate to "http://localhost:3000/" (use admin/grafana for login)
|
|
1. Open up dahsboard named "Intelligent Gateway Overview"
|
|
2. On this dashboard you can see reuqest latency and number of requests
|