diff --git a/config/config.yaml b/config/config.yaml index 033ab0af2..711110f97 100644 --- a/config/config.yaml +++ b/config/config.yaml @@ -142,6 +142,4 @@ TIMEOUT: 60 # Timeout for llm invocation #REDIS_PASSWORD: "YOUR_REDIS_PASSWORD" #REDIS_DB: "YOUR_REDIS_DB_INDEX, str, 0-based" -======= # DISABLE_LLM_PROVIDER_CHECK: false ->>>>>>> main diff --git a/metagpt/provider/openai_api.py b/metagpt/provider/openai_api.py index 1d2cdb591..195d2ea16 100644 --- a/metagpt/provider/openai_api.py +++ b/metagpt/provider/openai_api.py @@ -29,7 +29,7 @@ from tenacity import ( from metagpt.config import CONFIG, Config, LLMProviderEnum from metagpt.const import DEFAULT_MAX_TOKENS, DEFAULT_TOKEN_SIZE -from metagpt.logs import logger +from metagpt.logs import log_llm_stream, logger from metagpt.provider.base_gpt_api import BaseGPTAPI from metagpt.provider.constant import GENERAL_FUNCTION_SCHEMA, GENERAL_TOOL_CHOICE from metagpt.provider.llm_provider_registry import register_provider @@ -180,7 +180,7 @@ class OpenAIGPTAPI(BaseGPTAPI, RateLimiter): collected_messages = [] async for i in resp: - print(i, end="") + log_llm_stream(i) collected_messages.append(i) full_reply_content = "".join(collected_messages)