mirror of
https://github.com/katanemo/plano.git
synced 2026-05-03 21:02:56 +02:00
fixed issue with groq LLMs that require the openai in the /v1/chat/co… (#460)
* fixed issue with groq LLMs that require the openai in the /v1/chat/completions path. My first change * updated the GH actions with keys for Groq * adding missing groq API keys * add llama-3.2-3b-preview to the model based on addin groq to the demo --------- Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
This commit is contained in:
parent
e7b0de2a72
commit
f31aa59fac
9 changed files with 35 additions and 16 deletions
|
|
@ -62,7 +62,7 @@ def test_prompt_gateway(stream):
|
|||
|
||||
# third..end chunk is summarization (role = assistant)
|
||||
response_json = json.loads(chunks[2])
|
||||
assert response_json.get("model").startswith("gpt-4o-mini")
|
||||
assert response_json.get("model").startswith("llama-3.2-3b-preview")
|
||||
choices = response_json.get("choices", [])
|
||||
assert len(choices) > 0
|
||||
assert "role" in choices[0]["delta"]
|
||||
|
|
@ -71,7 +71,7 @@ def test_prompt_gateway(stream):
|
|||
|
||||
else:
|
||||
response_json = response.json()
|
||||
assert response_json.get("model").startswith("gpt-4o-mini")
|
||||
assert response_json.get("model").startswith("llama-3.2-3b-preview")
|
||||
choices = response_json.get("choices", [])
|
||||
assert len(choices) > 0
|
||||
assert "role" in choices[0]["message"]
|
||||
|
|
@ -231,7 +231,7 @@ def test_prompt_gateway_param_tool_call(stream):
|
|||
|
||||
# third..end chunk is summarization (role = assistant)
|
||||
response_json = json.loads(chunks[2])
|
||||
assert response_json.get("model").startswith("gpt-4o-mini")
|
||||
assert response_json.get("model").startswith("llama-3.2-3b-preview")
|
||||
choices = response_json.get("choices", [])
|
||||
assert len(choices) > 0
|
||||
assert "role" in choices[0]["delta"]
|
||||
|
|
@ -240,7 +240,7 @@ def test_prompt_gateway_param_tool_call(stream):
|
|||
|
||||
else:
|
||||
response_json = response.json()
|
||||
assert response_json.get("model").startswith("gpt-4o-mini")
|
||||
assert response_json.get("model").startswith("llama-3.2-3b-preview")
|
||||
choices = response_json.get("choices", [])
|
||||
assert len(choices) > 0
|
||||
assert "role" in choices[0]["message"]
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue