diff --git a/examples/agent_creator.py b/examples/agent_creator.py index 34160d398..bd58840ce 100644 --- a/examples/agent_creator.py +++ b/examples/agent_creator.py @@ -6,13 +6,12 @@ Author: garylin2099 import re from metagpt.actions import Action -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.const import METAGPT_ROOT from metagpt.logs import logger from metagpt.roles import Role from metagpt.schema import Message -config = Config.default() EXAMPLE_CODE_FILE = METAGPT_ROOT / "examples/build_customized_agent.py" MULTI_ACTION_AGENT_CODE_EXAMPLE = EXAMPLE_CODE_FILE.read_text() diff --git a/examples/di/interacting_with_human.py b/examples/di/interacting_with_human.py index c0ce02a40..a02f7c3bc 100644 --- a/examples/di/interacting_with_human.py +++ b/examples/di/interacting_with_human.py @@ -1,4 +1,5 @@ import fire + from metagpt.environment.mgx.mgx_env import MGXEnv from metagpt.logs import logger from metagpt.roles.di.team_leader import TeamLeader diff --git a/examples/rag/rag_bm.py b/examples/rag/rag_bm.py index a6a1145b5..99a546010 100644 --- a/examples/rag/rag_bm.py +++ b/examples/rag/rag_bm.py @@ -1,3 +1,4 @@ +# -*- coding: utf-8 -*- """RAG benchmark pipeline""" import asyncio diff --git a/examples/ui_with_chainlit/app.py b/examples/ui_with_chainlit/app.py index 3b449a12c..1522bd478 100644 --- a/examples/ui_with_chainlit/app.py +++ b/examples/ui_with_chainlit/app.py @@ -1,3 +1,5 @@ +from pathlib import Path + import chainlit as cl from init_setup import ChainlitEnv @@ -67,8 +69,8 @@ async def startup(message: cl.Message) -> None: await company.run(n_round=5) - workdir = company.env.context.git_repo.workdir - files = company.env.context.git_repo.get_files(workdir) + workdir = Path(company.env.context.config.project_path) + files = [file.name for file in workdir.iterdir() if file.is_file()] files = "\n".join([f"{workdir}/{file}" for file in files if not file.startswith(".git")]) await cl.Message( diff --git a/examples/write_design.py b/examples/write_design.py index 7eaa1a87b..fbbe9e1f0 100644 --- a/examples/write_design.py +++ b/examples/write_design.py @@ -1,10 +1,10 @@ import asyncio +from metagpt.environment.mgx.mgx_env import MGXEnv from metagpt.logs import logger from metagpt.roles.architect import Architect -from metagpt.environment.mgx.mgx_env import MGXEnv -from metagpt.schema import Message from metagpt.roles.di.team_leader import TeamLeader +from metagpt.schema import Message async def main(): diff --git a/metagpt/actions/requirement_analysis/framework/__init__.py b/metagpt/actions/requirement_analysis/framework/__init__.py index 5e0653088..968effd86 100644 --- a/metagpt/actions/requirement_analysis/framework/__init__.py +++ b/metagpt/actions/requirement_analysis/framework/__init__.py @@ -16,7 +16,7 @@ from pydantic import BaseModel from metagpt.actions.requirement_analysis.framework.evaluate_framework import EvaluateFramework from metagpt.actions.requirement_analysis.framework.write_framework import WriteFramework -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.utils.common import awrite @@ -54,7 +54,7 @@ async def save_framework( output_dir = ( Path(output_dir) if output_dir - else Config.default().workspace.path / (datetime.now().strftime("%Y%m%d%H%M%ST") + uuid.uuid4().hex[0:8]) + else config.workspace.path / (datetime.now().strftime("%Y%m%d%H%M%ST") + uuid.uuid4().hex[0:8]) ) output_dir.mkdir(parents=True, exist_ok=True) diff --git a/metagpt/config2.py b/metagpt/config2.py index 1942e5ef5..84b6059ef 100644 --- a/metagpt/config2.py +++ b/metagpt/config2.py @@ -178,3 +178,4 @@ def merge_dict(dicts: Iterable[Dict]) -> Dict: _CONFIG_CACHE = {} +config = Config.default() diff --git a/metagpt/environment/android/env_space.py b/metagpt/environment/android/env_space.py index 9580e3a7d..8225f0127 100644 --- a/metagpt/environment/android/env_space.py +++ b/metagpt/environment/android/env_space.py @@ -10,7 +10,7 @@ import numpy.typing as npt from gymnasium import spaces from pydantic import ConfigDict, Field, field_validator -from metagpt.environment.base_env_space import ( +from metagpt.base.base_env_space import ( BaseEnvAction, BaseEnvActionType, BaseEnvObsParams, diff --git a/metagpt/environment/werewolf/env_space.py b/metagpt/environment/werewolf/env_space.py index e6243b10f..dd6ceeabe 100644 --- a/metagpt/environment/werewolf/env_space.py +++ b/metagpt/environment/werewolf/env_space.py @@ -5,7 +5,7 @@ from gymnasium import spaces from pydantic import ConfigDict, Field -from metagpt.environment.base_env_space import BaseEnvAction, BaseEnvActionType +from metagpt.base.base_env_space import BaseEnvAction, BaseEnvActionType from metagpt.environment.werewolf.const import STEP_INSTRUCTIONS diff --git a/metagpt/exp_pool/decorator.py b/metagpt/exp_pool/decorator.py index d49c13e95..bb285d31c 100644 --- a/metagpt/exp_pool/decorator.py +++ b/metagpt/exp_pool/decorator.py @@ -6,7 +6,7 @@ from typing import Any, Callable, Optional, TypeVar from pydantic import BaseModel, ConfigDict, model_validator -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.exp_pool.context_builders import BaseContextBuilder, SimpleContextBuilder from metagpt.exp_pool.manager import ExperienceManager, get_exp_manager from metagpt.exp_pool.perfect_judges import BasePerfectJudge, SimplePerfectJudge @@ -60,8 +60,6 @@ def exp_cache( def decorator(func: Callable[..., ReturnType]) -> Callable[..., ReturnType]: @functools.wraps(func) async def get_or_create(args: Any, kwargs: Any) -> ReturnType: - config = Config.default() - if not config.exp_pool.enabled: rsp = func(*args, **kwargs) return await rsp if asyncio.iscoroutine(rsp) else rsp diff --git a/metagpt/ext/android_assistant/roles/android_assistant.py b/metagpt/ext/android_assistant/roles/android_assistant.py index 45636f519..97d66d30e 100644 --- a/metagpt/ext/android_assistant/roles/android_assistant.py +++ b/metagpt/ext/android_assistant/roles/android_assistant.py @@ -38,7 +38,6 @@ class AndroidAssistant(Role): def __init__(self, **data): super().__init__(**data) - self._watch([UserRequirement, AndroidActionOutput]) extra_config = config.extra self.task_desc = extra_config.get("task_desc", "Just explore any app in this phone!") diff --git a/metagpt/ext/stanford_town/utils/utils.py b/metagpt/ext/stanford_town/utils/utils.py index b4e15f485..4e81298c9 100644 --- a/metagpt/ext/stanford_town/utils/utils.py +++ b/metagpt/ext/stanford_town/utils/utils.py @@ -13,7 +13,7 @@ from typing import Union from openai import OpenAI -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.logs import logger @@ -48,7 +48,6 @@ def read_csv_to_list(curr_file: str, header=False, strip_trail=True): def get_embedding(text, model: str = "text-embedding-ada-002"): - config = Config.default() text = text.replace("\n", " ") embedding = None if not text: diff --git a/metagpt/rag/factories/llm.py b/metagpt/rag/factories/llm.py index e936e3a45..36b17da36 100644 --- a/metagpt/rag/factories/llm.py +++ b/metagpt/rag/factories/llm.py @@ -12,7 +12,7 @@ from llama_index.core.llms import ( from llama_index.core.llms.callbacks import llm_completion_callback from pydantic import Field -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.provider.base_llm import BaseLLM from metagpt.utils.async_helper import NestAsyncio from metagpt.utils.token_counter import TOKEN_MAX @@ -41,7 +41,6 @@ class RAGLLM(CustomLLM): **kwargs ): super().__init__(*args, **kwargs) - config = Config.default() if context_window < 0: context_window = TOKEN_MAX.get(config.llm.model, DEFAULT_CONTEXT_WINDOW) diff --git a/metagpt/rag/schema.py b/metagpt/rag/schema.py index 12e635a1a..5c63b09df 100644 --- a/metagpt/rag/schema.py +++ b/metagpt/rag/schema.py @@ -11,7 +11,7 @@ from llama_index.core.schema import TextNode from llama_index.core.vector_stores.types import VectorStoreQueryMode from pydantic import BaseModel, ConfigDict, Field, PrivateAttr, model_validator -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.configs.embedding_config import EmbeddingType from metagpt.logs import logger from metagpt.rag.interface import RAGObject @@ -47,7 +47,6 @@ class FAISSRetrieverConfig(IndexRetrieverConfig): @model_validator(mode="after") def check_dimensions(self): if self.dimensions == 0: - config = Config.default() self.dimensions = config.embedding.dimensions or self._embedding_type_to_dimensions.get( config.embedding.api_type, 1536 ) @@ -89,7 +88,6 @@ class MilvusRetrieverConfig(IndexRetrieverConfig): @model_validator(mode="after") def check_dimensions(self): if self.dimensions == 0: - config = Config.default() self.dimensions = config.embedding.dimensions or self._embedding_type_to_dimensions.get( config.embedding.api_type, 1536 ) diff --git a/metagpt/software_company.py b/metagpt/software_company.py index 508a6a5f3..73dcdd5db 100644 --- a/metagpt/software_company.py +++ b/metagpt/software_company.py @@ -27,7 +27,7 @@ def generate_repo( recover_path=None, ): """Run the startup logic. Can be called from CLI or other Python scripts.""" - from metagpt.config2 import Config + from metagpt.config2 import config from metagpt.context import Context from metagpt.roles import ( Architect, @@ -38,8 +38,6 @@ def generate_repo( ) from metagpt.team import Team - config = Config.default() - config.update_via_cli(project_path, project_name, inc, reqa_file, max_auto_summarize_code) ctx = Context(config=config) diff --git a/metagpt/tools/libs/index_repo.py b/metagpt/tools/libs/index_repo.py index 4c4e6c59b..5f95f9c02 100644 --- a/metagpt/tools/libs/index_repo.py +++ b/metagpt/tools/libs/index_repo.py @@ -11,7 +11,7 @@ from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.core.schema import NodeWithScore from pydantic import BaseModel, Field, model_validator -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.context import Context from metagpt.logs import logger from metagpt.rag.engines import SimpleEngine @@ -142,7 +142,6 @@ class IndexRepo(BaseModel): return flat_nodes if not self.embedding: - config = Config.default() if self.model: config.embedding.model = self.model factory = RAGEmbeddingFactory(config) diff --git a/metagpt/tools/ut_writer.py b/metagpt/tools/ut_writer.py index 9e67a3585..243871aff 100644 --- a/metagpt/tools/ut_writer.py +++ b/metagpt/tools/ut_writer.py @@ -4,7 +4,7 @@ import json from pathlib import Path -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.provider.openai_api import OpenAILLM as GPTAPI from metagpt.utils.common import awrite @@ -282,7 +282,6 @@ class UTGenerator: """Choose based on different calling methods""" result = "" if self.chatgpt_method == "API": - config = Config.default() result = await GPTAPI(config.get_openai_llm()).aask_code(messages=messages) return result diff --git a/metagpt/utils/embedding.py b/metagpt/utils/embedding.py index 3fcf1f25b..3d53a314c 100644 --- a/metagpt/utils/embedding.py +++ b/metagpt/utils/embedding.py @@ -7,11 +7,10 @@ """ from llama_index.embeddings.openai import OpenAIEmbedding -from metagpt.config2 import Config +from metagpt.config2 import config def get_embedding() -> OpenAIEmbedding: - config = Config.default() llm = config.get_openai_llm() if llm is None: raise ValueError("To use OpenAIEmbedding, please ensure that config.llm.api_type is correctly set to 'openai'.") diff --git a/metagpt/utils/file.py b/metagpt/utils/file.py index 75107c8be..d4cfc4d0a 100644 --- a/metagpt/utils/file.py +++ b/metagpt/utils/file.py @@ -13,7 +13,7 @@ from typing import Optional, Tuple, Union import aiofiles from fsspec.implementations.memory import MemoryFileSystem as _MemoryFileSystem -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.logs import logger from metagpt.utils import read_docx from metagpt.utils.common import aread, aread_bin, awrite_bin, check_http_endpoint @@ -190,7 +190,6 @@ class File: @staticmethod async def _read_omniparse_config() -> Tuple[str, int]: - config = Config.default() if config.omniparse and config.omniparse.base_url: return config.omniparse.base_url, config.omniparse.timeout return "", 0 diff --git a/metagpt/utils/make_sk_kernel.py b/metagpt/utils/make_sk_kernel.py index f0c55b07c..283a682d6 100644 --- a/metagpt/utils/make_sk_kernel.py +++ b/metagpt/utils/make_sk_kernel.py @@ -13,11 +13,10 @@ from semantic_kernel.connectors.ai.open_ai.services.open_ai_chat_completion impo OpenAIChatCompletion, ) -from metagpt.config2 import Config +from metagpt.config2 import config def make_sk_kernel(): - config = Config.default() kernel = sk.Kernel() if llm := config.get_azure_llm(): kernel.add_chat_service( diff --git a/setup.py b/setup.py index 2ffc09ee8..658c82219 100644 --- a/setup.py +++ b/setup.py @@ -43,7 +43,7 @@ extras_require = { "llama-index-postprocessor-cohere-rerank==0.1.4", "llama-index-postprocessor-colbert-rerank==0.1.1", "llama-index-postprocessor-flag-embedding-reranker==0.1.2", - # "llama-index-vector-stores-milvus==0.1.23", + "llama-index-vector-stores-milvus==0.1.23", "docx2txt==0.8", ], } diff --git a/tests/metagpt/learn/test_text_to_embedding.py b/tests/metagpt/learn/test_text_to_embedding.py index f50f6a7aa..3b5486c5d 100644 --- a/tests/metagpt/learn/test_text_to_embedding.py +++ b/tests/metagpt/learn/test_text_to_embedding.py @@ -11,7 +11,7 @@ from pathlib import Path import pytest -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.learn.text_to_embedding import text_to_embedding from metagpt.utils.common import aread @@ -19,7 +19,6 @@ from metagpt.utils.common import aread @pytest.mark.asyncio async def test_text_to_embedding(mocker): # mock - config = Config.default() mock_post = mocker.patch("aiohttp.ClientSession.post") mock_response = mocker.AsyncMock() mock_response.status = 200 diff --git a/tests/metagpt/learn/test_text_to_image.py b/tests/metagpt/learn/test_text_to_image.py index d3272dadd..eb252589b 100644 --- a/tests/metagpt/learn/test_text_to_image.py +++ b/tests/metagpt/learn/test_text_to_image.py @@ -12,7 +12,7 @@ import openai import pytest from pydantic import BaseModel -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.learn.text_to_image import text_to_image from metagpt.tools.metagpt_text_to_image import MetaGPTText2Image from metagpt.tools.openai_text_to_image import OpenAIText2Image @@ -26,7 +26,6 @@ async def test_text_to_image(mocker): mocker.patch.object(OpenAIText2Image, "text_2_image", return_value=b"mock OpenAIText2Image") mocker.patch.object(S3, "cache", return_value="http://mock/s3") - config = Config.default() assert config.metagpt_tti_url data = await text_to_image("Panda emoji", size_type="512x512", config=config) @@ -51,7 +50,6 @@ async def test_openai_text_to_image(mocker): mock_post.return_value.__aenter__.return_value = mock_response mocker.patch.object(S3, "cache", return_value="http://mock.s3.com/0.png") - config = Config.default() config.metagpt_tti_url = None assert config.get_openai_llm() diff --git a/tests/metagpt/learn/test_text_to_speech.py b/tests/metagpt/learn/test_text_to_speech.py index f01e5d132..480e35f7a 100644 --- a/tests/metagpt/learn/test_text_to_speech.py +++ b/tests/metagpt/learn/test_text_to_speech.py @@ -10,7 +10,7 @@ import pytest from azure.cognitiveservices.speech import ResultReason, SpeechSynthesizer -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.learn.text_to_speech import text_to_speech from metagpt.tools.iflytek_tts import IFlyTekTTS from metagpt.utils.s3 import S3 @@ -19,7 +19,6 @@ from metagpt.utils.s3 import S3 @pytest.mark.asyncio async def test_azure_text_to_speech(mocker): # mock - config = Config.default() config.iflytek_api_key = None config.iflytek_api_secret = None config.iflytek_app_id = None @@ -47,7 +46,6 @@ async def test_azure_text_to_speech(mocker): @pytest.mark.asyncio async def test_iflytek_text_to_speech(mocker): # mock - config = Config.default() config.azure_tts_subscription_key = None config.azure_tts_region = None mocker.patch.object(IFlyTekTTS, "synthesize_speech", return_value=None) diff --git a/tests/metagpt/roles/di/run_swe_agent_for_benchmark.py b/tests/metagpt/roles/di/run_swe_agent_for_benchmark.py index 5ceba6dcc..ce4ef94a4 100644 --- a/tests/metagpt/roles/di/run_swe_agent_for_benchmark.py +++ b/tests/metagpt/roles/di/run_swe_agent_for_benchmark.py @@ -7,7 +7,7 @@ import sys from datetime import datetime from pathlib import Path -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.const import DEFAULT_WORKSPACE_ROOT, METAGPT_ROOT from metagpt.logs import logger from metagpt.roles.di.engineer2 import Engineer2 @@ -15,7 +15,6 @@ from metagpt.tools.libs.editor import Editor from metagpt.tools.libs.terminal import Terminal from metagpt.tools.swe_agent_commands.swe_agent_utils import load_hf_dataset -config = Config.default() # Specify by yourself TEST_REPO_DIR = METAGPT_ROOT / "data" / "test_repo" DATA_DIR = METAGPT_ROOT / "data/hugging_face" diff --git a/tests/metagpt/roles/di/test_swe_agent.py b/tests/metagpt/roles/di/test_swe_agent.py index 325e7bed8..3dbfa910b 100644 --- a/tests/metagpt/roles/di/test_swe_agent.py +++ b/tests/metagpt/roles/di/test_swe_agent.py @@ -1,22 +1,17 @@ import pytest +from metagpt.environment.mgx.mgx_env import MGXEnv from metagpt.roles.di.swe_agent import SWEAgent +from metagpt.roles.di.team_leader import TeamLeader from metagpt.schema import Message from metagpt.tools.libs.terminal import Bash -from metagpt.environment.mgx.mgx_env import MGXEnv -from metagpt.roles.di.team_leader import TeamLeader @pytest.fixture def env(): test_env = MGXEnv() tl = TeamLeader() - test_env.add_roles( - [ - tl, - SWEAgent() - ] - ) + test_env.add_roles([tl, SWEAgent()]) return test_env diff --git a/tests/metagpt/roles/test_architect.py b/tests/metagpt/roles/test_architect.py index 9a834ffbe..cb636b6a1 100644 --- a/tests/metagpt/roles/test_architect.py +++ b/tests/metagpt/roles/test_architect.py @@ -8,18 +8,19 @@ distribution feature for message handling. """ import uuid +from pathlib import Path import pytest from metagpt.actions import WritePRD +from metagpt.actions.di.run_command import RunCommand from metagpt.const import PRDS_FILE_REPO from metagpt.logs import logger from metagpt.roles import Architect from metagpt.schema import Message from metagpt.utils.common import any_to_str, awrite from tests.metagpt.roles.mock import MockMessages -from pathlib import Path -from metagpt.actions.di.run_command import RunCommand + @pytest.mark.asyncio async def test_architect(context): diff --git a/tests/metagpt/roles/test_engineer.py b/tests/metagpt/roles/test_engineer.py index c340e910f..18b297ae5 100644 --- a/tests/metagpt/roles/test_engineer.py +++ b/tests/metagpt/roles/test_engineer.py @@ -9,6 +9,7 @@ """ import json from pathlib import Path +from types import SimpleNamespace import pytest @@ -19,9 +20,8 @@ from metagpt.roles.engineer import Engineer from metagpt.schema import CodingContext, Message from metagpt.utils.common import CodeParser, any_to_name, any_to_str, aread, awrite from metagpt.utils.git_repository import ChangeType -from tests.metagpt.roles.mock import STRS_FOR_PARSING, TASKS, MockMessages from metagpt.utils.project_repo import ProjectRepo -from types import SimpleNamespace +from tests.metagpt.roles.mock import STRS_FOR_PARSING, TASKS, MockMessages @pytest.mark.asyncio diff --git a/tests/metagpt/test_document.py b/tests/metagpt/test_document.py index 29393bb13..9c076f4e6 100644 --- a/tests/metagpt/test_document.py +++ b/tests/metagpt/test_document.py @@ -5,12 +5,10 @@ @Author : alexanderwu @File : test_document.py """ -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.document import Repo from metagpt.logs import logger -config = Config.default() - def set_existing_repo(path): repo1 = Repo.from_path(path) diff --git a/tests/metagpt/tools/test_azure_tts.py b/tests/metagpt/tools/test_azure_tts.py index ee55616d2..f72b5663b 100644 --- a/tests/metagpt/tools/test_azure_tts.py +++ b/tests/metagpt/tools/test_azure_tts.py @@ -12,11 +12,9 @@ from pathlib import Path import pytest from azure.cognitiveservices.speech import ResultReason, SpeechSynthesizer -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.tools.azure_tts import AzureTTS -config = Config.default() - @pytest.mark.asyncio async def test_azure_tts(mocker): diff --git a/tests/metagpt/tools/test_iflytek_tts.py b/tests/metagpt/tools/test_iflytek_tts.py index c51f62b8e..b4bcadb89 100644 --- a/tests/metagpt/tools/test_iflytek_tts.py +++ b/tests/metagpt/tools/test_iflytek_tts.py @@ -7,14 +7,13 @@ """ import pytest -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.tools.iflytek_tts import IFlyTekTTS, oas3_iflytek_tts @pytest.mark.asyncio async def test_iflytek_tts(mocker): # mock - config = Config.default() config.azure_tts_subscription_key = None config.azure_tts_region = None mocker.patch.object(IFlyTekTTS, "synthesize_speech", return_value=None) diff --git a/tests/metagpt/tools/test_metagpt_text_to_image.py b/tests/metagpt/tools/test_metagpt_text_to_image.py index bd0fcaf8b..d3797a460 100644 --- a/tests/metagpt/tools/test_metagpt_text_to_image.py +++ b/tests/metagpt/tools/test_metagpt_text_to_image.py @@ -10,11 +10,9 @@ from unittest.mock import AsyncMock import pytest -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.tools.metagpt_text_to_image import oas3_metagpt_text_to_image -config = Config.default() - @pytest.mark.asyncio async def test_draw(mocker): diff --git a/tests/metagpt/tools/test_moderation.py b/tests/metagpt/tools/test_moderation.py index 0f921887f..8dc9e9d5e 100644 --- a/tests/metagpt/tools/test_moderation.py +++ b/tests/metagpt/tools/test_moderation.py @@ -8,12 +8,10 @@ import pytest -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.llm import LLM from metagpt.tools.moderation import Moderation -config = Config.default() - @pytest.mark.asyncio @pytest.mark.parametrize( diff --git a/tests/metagpt/tools/test_openai_text_to_embedding.py b/tests/metagpt/tools/test_openai_text_to_embedding.py index 81b3895c3..047206d48 100644 --- a/tests/metagpt/tools/test_openai_text_to_embedding.py +++ b/tests/metagpt/tools/test_openai_text_to_embedding.py @@ -10,7 +10,7 @@ from pathlib import Path import pytest -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.tools.openai_text_to_embedding import oas3_openai_text_to_embedding from metagpt.utils.common import aread @@ -18,7 +18,6 @@ from metagpt.utils.common import aread @pytest.mark.asyncio async def test_embedding(mocker): # mock - config = Config.default() mock_post = mocker.patch("aiohttp.ClientSession.post") mock_response = mocker.AsyncMock() mock_response.status = 200 diff --git a/tests/metagpt/tools/test_openai_text_to_image.py b/tests/metagpt/tools/test_openai_text_to_image.py index 4856342d1..3f9169ddd 100644 --- a/tests/metagpt/tools/test_openai_text_to_image.py +++ b/tests/metagpt/tools/test_openai_text_to_image.py @@ -11,7 +11,7 @@ import openai import pytest from pydantic import BaseModel -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.llm import LLM from metagpt.tools.openai_text_to_image import ( OpenAIText2Image, @@ -19,8 +19,6 @@ from metagpt.tools.openai_text_to_image import ( ) from metagpt.utils.s3 import S3 -config = Config.default() - @pytest.mark.asyncio async def test_draw(mocker): diff --git a/tests/metagpt/tools/test_ut_writer.py b/tests/metagpt/tools/test_ut_writer.py index ebb8c8aa2..557067191 100644 --- a/tests/metagpt/tools/test_ut_writer.py +++ b/tests/metagpt/tools/test_ut_writer.py @@ -20,12 +20,10 @@ from openai.types.chat.chat_completion_message_tool_call import ( Function, ) -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.const import API_QUESTIONS_PATH, UT_PY_PATH from metagpt.tools.ut_writer import YFT_PROMPT_PREFIX, UTGenerator -config = Config.default() - class TestUTWriter: @pytest.mark.asyncio diff --git a/tests/metagpt/utils/test_repair_llm_raw_output.py b/tests/metagpt/utils/test_repair_llm_raw_output.py index 75bd9f165..7a29ea3ee 100644 --- a/tests/metagpt/utils/test_repair_llm_raw_output.py +++ b/tests/metagpt/utils/test_repair_llm_raw_output.py @@ -2,9 +2,7 @@ # -*- coding: utf-8 -*- # @Desc : unittest of repair_llm_raw_output -from metagpt.config2 import Config - -config = Config.default() +from metagpt.config2 import config """ CONFIG.repair_llm_output should be True before retry_parse_json_text imported. diff --git a/tests/mock/mock_llm.py b/tests/mock/mock_llm.py index e58ce4120..704403e64 100644 --- a/tests/mock/mock_llm.py +++ b/tests/mock/mock_llm.py @@ -1,7 +1,7 @@ import json from typing import Optional, Union -from metagpt.config2 import Config +from metagpt.config2 import config from metagpt.configs.llm_config import LLMType from metagpt.const import LLM_API_TIMEOUT from metagpt.logs import logger @@ -10,8 +10,6 @@ from metagpt.provider.constant import GENERAL_FUNCTION_SCHEMA from metagpt.provider.openai_api import OpenAILLM from metagpt.schema import Message -config = Config.default() - OriginalLLM = OpenAILLM if config.llm.api_type == LLMType.OPENAI else AzureOpenAILLM