diff --git a/examples/example.faiss b/examples/example.faiss new file mode 100644 index 000000000..a5a539dc4 Binary files /dev/null and b/examples/example.faiss differ diff --git a/examples/example.json b/examples/example.json new file mode 100644 index 000000000..996cbec3b --- /dev/null +++ b/examples/example.json @@ -0,0 +1,10 @@ +[ + { + "source": "Which facial cleanser is good for oily skin?", + "output": "ABC cleanser is preferred by many with oily skin." + }, + { + "source": "Is L'Oreal good to use?", + "output": "L'Oreal is a popular brand with many positive reviews." + } +] \ No newline at end of file diff --git a/examples/example.pkl b/examples/example.pkl new file mode 100644 index 000000000..a0e839763 Binary files /dev/null and b/examples/example.pkl differ diff --git a/examples/search_kb.py b/examples/search_kb.py index 01267943b..37b229f25 100644 --- a/examples/search_kb.py +++ b/examples/search_kb.py @@ -9,28 +9,15 @@ import asyncio from langchain.embeddings import OpenAIEmbeddings from metagpt.config import CONFIG -from metagpt.const import DATA_PATH +from metagpt.const import EXAMPLE_PATH from metagpt.document_store import FaissStore from metagpt.logs import logger from metagpt.roles import Sales -""" example.json, e.g. -[ - { - "source": "Which facial cleanser is good for oily skin?", - "output": "ABC cleanser is preferred by many with oily skin." - }, - { - "source": "Is L'Oreal good to use?", - "output": "L'Oreal is a popular brand with many positive reviews." - } -] -""" - def get_store(): embedding = OpenAIEmbeddings(openai_api_key=CONFIG.openai_api_key, openai_api_base=CONFIG.openai_base_url) - return FaissStore(DATA_PATH / "example.json", embedding=embedding) + return FaissStore(EXAMPLE_PATH / "example.json", embedding=embedding) async def search(): diff --git a/metagpt/config.py b/metagpt/config.py index 0109f4b1d..222254ac7 100644 --- a/metagpt/config.py +++ b/metagpt/config.py @@ -138,7 +138,7 @@ class Config(metaclass=Singleton): self.gemini_api_key = self._get("GEMINI_API_KEY") self.ollama_api_base = self._get("OLLAMA_API_BASE") self.ollama_api_model = self._get("OLLAMA_API_MODEL") - _ = self.get_default_llm_provider_enum() + # _ = self.get_default_llm_provider_enum() # self.openai_base_url = self._get("OPENAI_BASE_URL") self.openai_proxy = self._get("OPENAI_PROXY") or self.global_proxy diff --git a/metagpt/const.py b/metagpt/const.py index 012c84542..5e149ed72 100644 --- a/metagpt/const.py +++ b/metagpt/const.py @@ -51,6 +51,7 @@ def get_metagpt_root(): METAGPT_ROOT = get_metagpt_root() # Dependent on METAGPT_PROJECT_ROOT DEFAULT_WORKSPACE_ROOT = METAGPT_ROOT / "workspace" +EXAMPLE_PATH = METAGPT_ROOT / "examples" DATA_PATH = METAGPT_ROOT / "data" RESEARCH_PATH = DATA_PATH / "research" TUTORIAL_PATH = DATA_PATH / "tutorial_docx"