plano/tests/e2e
Salman Paracha 48bbc7cce7
fixed reasoning failures (#634)
* fixed reasoning failures

* adding debugging

* made several fixes for transmission isses for SSeEvents, incomplete handling of json types by anthropic, and wrote a bunch of tests

* removed debugging from supervisord.conf

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-342.local>
2025-12-18 11:02:59 -08:00
..
.vscode better model names (#517) 2025-07-11 16:42:16 -07:00
arch_config_memory_state_v1_responses.yaml enable state management for v1/responses (#631) 2025-12-17 12:18:38 -08:00
common.py Use intent model from archfc to pick prompt gateway (#328) 2024-12-20 13:25:01 -08:00
common_scripts.sh Use intent model from archfc to pick prompt gateway (#328) 2024-12-20 13:25:01 -08:00
docker-compose.yaml Use mcp tools for filter chain (#621) 2025-12-17 17:30:14 -08:00
poetry.lock add support for v1/messages and transformations (#558) 2025-09-10 07:40:30 -07:00
pyproject.toml add support for v1/messages and transformations (#558) 2025-09-10 07:40:30 -07:00
README.md Use intent model from archfc to pick prompt gateway (#328) 2024-12-20 13:25:01 -08:00
response.hex Add support for Amazon Bedrock Converse and ConverseStream (#588) 2025-10-22 11:31:21 -07:00
response_with_tools.hex Add support for Amazon Bedrock Converse and ConverseStream (#588) 2025-10-22 11:31:21 -07:00
run_e2e_tests.sh enable state management for v1/responses (#631) 2025-12-17 12:18:38 -08:00
test_model_alias_routing.py fixed test and docs for deployment (#595) 2025-10-22 14:13:16 -07:00
test_openai_responses_api_client.py fixed reasoning failures (#634) 2025-12-18 11:02:59 -08:00
test_openai_responses_api_client_with_state.py enable state management for v1/responses (#631) 2025-12-17 12:18:38 -08:00
test_prompt_gateway.py removing model_server python module to brightstaff (function calling) (#615) 2025-11-22 12:55:00 -08:00

e2e tests

e2e tests for arch llm gateway and prompt gateway

To be able to run e2e tests successfully run_e2e_script prepares environment in following way,

  1. build and start weather_forecast demo (using docker compose)
  2. build, install and start model server async (using poetry)
  3. build and start arch gateway (using docker compose)
  4. wait for model server to be ready
  5. wait for arch gateway to be ready
  6. start e2e tests (using poetry)
    1. runs llm gateway tests for llm routing
    2. runs prompt gateway tests to test function calling, parameter gathering and summarization
  7. cleanup
    1. stops arch gateway
    2. stops model server
    3. stops weather_forecast demo

How to run

To run locally make sure that following requirements are met.

Requirements

  • Python 3.10
  • Poetry
  • Docker

Running tests locally

sh run_e2e_test.sh