From cbcba98015cca383f82259bb37037b1ecc078a89 Mon Sep 17 00:00:00 2001 From: Cyber MacGeddon Date: Sat, 28 Mar 2026 11:17:09 +0000 Subject: [PATCH] Fix tests --- tests/integration/test_text_completion_integration.py | 8 ++++---- .../test_text_completion_streaming_integration.py | 2 +- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/tests/integration/test_text_completion_integration.py b/tests/integration/test_text_completion_integration.py index 08e2a995..26fa0ab3 100644 --- a/tests/integration/test_text_completion_integration.py +++ b/tests/integration/test_text_completion_integration.py @@ -93,7 +93,7 @@ class TestTextCompletionIntegration: assert call_args.kwargs['model'] == "gpt-3.5-turbo" assert call_args.kwargs['temperature'] == 0.7 - assert call_args.kwargs['max_tokens'] == 1024 + assert call_args.kwargs['max_completion_tokens'] == 1024 assert len(call_args.kwargs['messages']) == 1 assert call_args.kwargs['messages'][0]['role'] == "user" assert "You are a helpful assistant." in call_args.kwargs['messages'][0]['content'][0]['text'] @@ -134,7 +134,7 @@ class TestTextCompletionIntegration: call_args = mock_openai_client.chat.completions.create.call_args assert call_args.kwargs['model'] == config['model'] assert call_args.kwargs['temperature'] == config['temperature'] - assert call_args.kwargs['max_tokens'] == config['max_output'] + assert call_args.kwargs['max_completion_tokens'] == config['max_output'] # Reset mock for next iteration mock_openai_client.reset_mock() @@ -286,7 +286,7 @@ class TestTextCompletionIntegration: # were removed in #561 as unnecessary parameters assert 'model' in call_args.kwargs assert 'temperature' in call_args.kwargs - assert 'max_tokens' in call_args.kwargs + assert 'max_completion_tokens' in call_args.kwargs # Verify result structure assert hasattr(result, 'text') @@ -362,7 +362,7 @@ class TestTextCompletionIntegration: call_args = mock_openai_client.chat.completions.create.call_args assert call_args.kwargs['model'] == "gpt-4" assert call_args.kwargs['temperature'] == 0.8 - assert call_args.kwargs['max_tokens'] == 2048 + assert call_args.kwargs['max_completion_tokens'] == 2048 # Note: top_p, frequency_penalty, and presence_penalty # were removed in #561 as unnecessary parameters diff --git a/tests/integration/test_text_completion_streaming_integration.py b/tests/integration/test_text_completion_streaming_integration.py index a70afb4c..6968affa 100644 --- a/tests/integration/test_text_completion_streaming_integration.py +++ b/tests/integration/test_text_completion_streaming_integration.py @@ -201,7 +201,7 @@ class TestTextCompletionStreaming: call_args = mock_streaming_openai_client.chat.completions.create.call_args assert call_args.kwargs['model'] == "gpt-4" assert call_args.kwargs['temperature'] == 0.5 - assert call_args.kwargs['max_tokens'] == 2048 + assert call_args.kwargs['max_completion_tokens'] == 2048 assert call_args.kwargs['stream'] is True # Verify chunks have correct model