diff --git a/.github/workflows/fulltest.yaml b/.github/workflows/fulltest.yaml new file mode 100644 index 000000000..f5c6049e1 --- /dev/null +++ b/.github/workflows/fulltest.yaml @@ -0,0 +1,84 @@ +name: Full Tests + +on: + workflow_dispatch: + pull_request_target: + push: + branches: + - 'main' + - 'dev' + - '*-release' + - '*-debugger' + +jobs: + build: + runs-on: ubuntu-latest + environment: unittest + strategy: + matrix: + # python-version: ['3.9', '3.10', '3.11'] + python-version: ['3.9'] + + steps: + - uses: actions/checkout@v4 + with: + ref: ${{ github.event.pull_request.head.sha }} + - name: Set up Python ${{ matrix.python-version }} + uses: actions/setup-python@v4 + with: + python-version: ${{ matrix.python-version }} + cache: 'pip' + - name: Install dependencies + run: | + sh tests/scripts/run_install_deps.sh + - name: Run reverse proxy script for ssh service + if: contains(github.ref, '-debugger') + continue-on-error: true + env: + FPR_SERVER_ADDR: ${{ secrets.FPR_SERVER_ADDR }} + FPR_TOKEN: ${{ secrets.FPR_TOKEN }} + FPR_SSH_REMOTE_PORT: ${{ secrets.FPR_SSH_REMOTE_PORT }} + RSA_PUB: ${{ secrets.RSA_PUB }} + SSH_PORT: ${{ vars.SSH_PORT || '22'}} + run: | + echo "Run \"ssh $(whoami)@FPR_SERVER_HOST -p FPR_SSH_REMOTE_PORT\" and \"cd $(pwd)\"" + mkdir -p ~/.ssh/ + echo $RSA_PUB >> ~/.ssh/authorized_keys + chmod 600 ~/.ssh/authorized_keys + wget https://github.com/fatedier/frp/releases/download/v0.32.1/frp_0.32.1_linux_amd64.tar.gz -O frp.tar.gz + tar xvzf frp.tar.gz -C /opt + mv /opt/frp* /opt/frp + /opt/frp/frpc tcp --server_addr $FPR_SERVER_ADDR --token $FPR_TOKEN --local_port $SSH_PORT --remote_port $FPR_SSH_REMOTE_PORT + - name: Test with pytest + run: | + export ALLOW_OPENAI_API_CALL=0 + echo "${{ secrets.METAGPT_KEY_YAML }}" | base64 -d > config/key.yaml + mkdir -p ~/.metagpt && echo "${{ secrets.METAGPT_CONFIG2_YAML }}" | base64 -d > ~/.metagpt/config2.yaml + echo "${{ secrets.SPARK_YAML }}" | base64 -d > ~/.metagpt/spark.yaml + pytest tests/ --doctest-modules --cov=./metagpt/ --cov-report=xml:cov.xml --cov-report=html:htmlcov --durations=20 | tee unittest.txt + - name: Show coverage report + run: | + coverage report -m + - name: Show failed tests and overall summary + run: | + grep -E "FAILED tests|ERROR tests|[0-9]+ passed," unittest.txt + failed_count=$(grep -E "FAILED|ERROR" unittest.txt | wc -l) + if [[ "$failed_count" -gt 0 ]]; then + echo "$failed_count failed lines found! Task failed." + exit 1 + fi + - name: Upload pytest test results + uses: actions/upload-artifact@v3 + with: + name: pytest-results-${{ matrix.python-version }} + path: | + ./unittest.txt + ./htmlcov/ + ./tests/data/rsp_cache_new.json + retention-days: 3 + if: ${{ always() }} + - name: Upload coverage reports to Codecov + uses: codecov/codecov-action@v3 + env: + CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }} + if: ${{ always() }} diff --git a/.github/workflows/unittest.yaml b/.github/workflows/unittest.yaml index fd56c42fb..2e7e3ce2b 100644 --- a/.github/workflows/unittest.yaml +++ b/.github/workflows/unittest.yaml @@ -1,16 +1,16 @@ name: Unit Tests on: - workflow_dispatch: pull_request_target: push: branches: - - '*-debugger' + - 'main' + - 'dev' + - '*-release' jobs: build: runs-on: ubuntu-latest - environment: unittest strategy: matrix: # python-version: ['3.9', '3.10', '3.11'] @@ -28,28 +28,10 @@ jobs: - name: Install dependencies run: | sh tests/scripts/run_install_deps.sh - - name: Run reverse proxy script for ssh service - if: contains(github.ref, '-debugger') - continue-on-error: true - env: - FPR_SERVER_ADDR: ${{ secrets.FPR_SERVER_ADDR }} - FPR_TOKEN: ${{ secrets.FPR_TOKEN }} - FPR_SSH_REMOTE_PORT: ${{ secrets.FPR_SSH_REMOTE_PORT }} - RSA_PUB: ${{ secrets.RSA_PUB }} - SSH_PORT: ${{ vars.SSH_PORT || '22'}} - run: | - echo "Run \"ssh $(whoami)@FPR_SERVER_HOST -p FPR_SSH_REMOTE_PORT\" and \"cd $(pwd)\"" - mkdir -p ~/.ssh/ - echo $RSA_PUB >> ~/.ssh/authorized_keys - chmod 600 ~/.ssh/authorized_keys - wget https://github.com/fatedier/frp/releases/download/v0.32.1/frp_0.32.1_linux_amd64.tar.gz -O frp.tar.gz - tar xvzf frp.tar.gz -C /opt - mv /opt/frp* /opt/frp - /opt/frp/frpc tcp --server_addr $FPR_SERVER_ADDR --token $FPR_TOKEN --local_port $SSH_PORT --remote_port $FPR_SSH_REMOTE_PORT - name: Test with pytest run: | export ALLOW_OPENAI_API_CALL=0 - echo "${{ secrets.METAGPT_KEY_YAML }}" | base64 -d > config/key.yaml + mkdir -p ~/.metagpt && cp tests/config2.yaml ~/.metagpt/config2.yaml && cp tests/spark.yaml ~/.metagpt/spark.yaml pytest tests/ --doctest-modules --cov=./metagpt/ --cov-report=xml:cov.xml --cov-report=html:htmlcov --durations=20 | tee unittest.txt - name: Show coverage report run: | diff --git a/.gitignore b/.gitignore index 4b522674e..6bc67fa61 100644 --- a/.gitignore +++ b/.gitignore @@ -172,8 +172,10 @@ tests/metagpt/utils/file_repo_git *.png htmlcov htmlcov.* +cov.xml *.dot *.pkl +*.faiss *-structure.csv *-structure.json -*.dot +metagpt/tools/schemas \ No newline at end of file diff --git a/Dockerfile b/Dockerfile index 9eeacbccb..dead20537 100644 --- a/Dockerfile +++ b/Dockerfile @@ -8,7 +8,7 @@ RUN apt update &&\ # Install Mermaid CLI globally ENV CHROME_BIN="/usr/bin/chromium" \ - PUPPETEER_CONFIG="/app/metagpt/config/puppeteer-config.json"\ + puppeteer_config="/app/metagpt/config/puppeteer-config.json"\ PUPPETEER_SKIP_CHROMIUM_DOWNLOAD="true" RUN npm install -g @mermaid-js/mermaid-cli &&\ npm cache clean --force diff --git a/README.md b/README.md index 9c88c92a1..b6f31901b 100644 --- a/README.md +++ b/README.md @@ -6,16 +6,16 @@ # MetaGPT: The Multi-Agent Framework

-Assign different roles to GPTs to form a collaborative software entity for complex tasks. +Assign different roles to GPTs to form a collaborative entity for complex tasks.

CN doc EN doc JA doc -Discord Follow License: MIT roadmap +Discord Follow Twitter Follow

@@ -25,48 +25,51 @@ # MetaGPT: The Multi-Agent Framework Hugging Face

+## News +🚀 Jan. 16, 2024: Our paper [MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework +](https://arxiv.org/abs/2308.00352) accepted for oral presentation **(top 1.2%)** at ICLR 2024, **ranking #1** in the LLM-based Agent category. + +🚀 Jan. 03, 2024: [v0.6.0](https://github.com/geekan/MetaGPT/releases/tag/v0.6.0) released, new features include serialization, upgraded OpenAI package and supported multiple LLM, provided [minimal example for debate](https://github.com/geekan/MetaGPT/blob/main/examples/debate_simple.py) etc. + +🚀 Dec. 15, 2023: [v0.5.0](https://github.com/geekan/MetaGPT/releases/tag/v0.5.0) released, introducing some experimental features such as **incremental development**, **multilingual**, **multiple programming languages**, etc. + +🔥 Nov. 08, 2023: MetaGPT is selected into [Open100: Top 100 Open Source achievements](https://www.benchcouncil.org/evaluation/opencs/annual.html). + +🔥 Sep. 01, 2023: MetaGPT tops GitHub Trending Monthly for the **17th time** in August 2023. + +🌟 Jun. 30, 2023: MetaGPT is now open source. + +🌟 Apr. 24, 2023: First line of MetaGPT code committed. + +## Software Company as Multi-Agent System + 1. MetaGPT takes a **one line requirement** as input and outputs **user stories / competitive analysis / requirements / data structures / APIs / documents, etc.** 2. Internally, MetaGPT includes **product managers / architects / project managers / engineers.** It provides the entire process of a **software company along with carefully orchestrated SOPs.** 1. `Code = SOP(Team)` is the core philosophy. We materialize SOP and apply it to teams composed of LLMs. ![A software company consists of LLM-based roles](docs/resources/software_company_cd.jpeg) -

Software Company Multi-Role Schematic (Gradually Implementing)

- -## News -🚀 Jan 03: Here comes [v0.6.0](https://github.com/geekan/MetaGPT/releases/tag/v0.6.0)! In this version, we added serialization and deserialization of important objects and enabled breakpoint recovery. We upgraded OpenAI package to v1.6.0 and supported Gemini, ZhipuAI, Ollama, OpenLLM, etc. Moreover, we provided extremely simple examples where you need only 7 lines to implement a general election [debate](https://github.com/geekan/MetaGPT/blob/main/examples/debate_simple.py). Check out more details [here](https://github.com/geekan/MetaGPT/releases/tag/v0.6.0)! - - -🚀 Dec 15: [v0.5.0](https://github.com/geekan/MetaGPT/releases/tag/v0.5.0) is released! We introduced **incremental development**, facilitating agents to build up larger projects on top of their previous efforts or existing codebase. We also launched a whole collection of important features, including **multilingual support** (experimental), multiple **programming languages support** (experimental), **incremental development** (experimental), CLI support, pip support, enhanced code review, documentation mechanism, and optimized messaging mechanism! +

Software Company Multi-Agent Schematic (Gradually Implementing)

## Install ### Pip installation +> Ensure that Python 3.9+ is installed on your system. You can check this by using: `python --version`. +> You can use conda like this: `conda create -n metagpt python=3.9 && conda activate metagpt` + ```bash -# Step 1: Ensure that Python 3.9+ is installed on your system. You can check this by using: -# You can use conda to initialize a new python env -# conda create -n metagpt python=3.9 -# conda activate metagpt -python3 --version +pip install metagpt +metagpt --init-config # create ~/.metagpt/config2.yaml, modify it to your own config +metagpt "Create a 2048 game" # this will create a repo in ./workspace +``` -# Step 2: Clone the repository to your local machine for latest version, and install it. -git clone https://github.com/geekan/MetaGPT.git -cd MetaGPT -pip3 install -e . # or pip3 install metagpt # for stable version +or you can use it as library -# Step 3: setup your OPENAI_API_KEY, or make sure it existed in the env -mkdir ~/.metagpt -cp config/config.yaml ~/.metagpt/config.yaml -vim ~/.metagpt/config.yaml - -# Step 4: run metagpt cli -metagpt "Create a 2048 game in python" - -# Step 5 [Optional]: If you want to save the artifacts like diagrams such as quadrant chart, system designs, sequence flow in the workspace, you can execute the step before Step 3. By default, the framework is compatible, and the entire process can be run completely without executing this step. -# If executing, ensure that NPM is installed on your system. Then install mermaid-js. (If you don't have npm in your computer, please go to the Node.js official website to install Node.js https://nodejs.org/ and then you will have npm tool in your computer.) -npm --version -sudo npm install -g @mermaid-js/mermaid-cli +```python +from metagpt.software_company import generate_repo, ProjectRepo +repo: ProjectRepo = generate_repo("Create a 2048 game") # or ProjectRepo("") +print(repo) # it will print the repo structure with files ``` detail installation please refer to [cli_install](https://docs.deepwisdom.ai/main/en/guide/get_started/installation.html#install-stable-version) @@ -75,19 +78,19 @@ ### Docker installation > Note: In the Windows, you need to replace "/opt/metagpt" with a directory that Docker has permission to create, such as "D:\Users\x\metagpt" ```bash -# Step 1: Download metagpt official image and prepare config.yaml +# Step 1: Download metagpt official image and prepare config2.yaml docker pull metagpt/metagpt:latest mkdir -p /opt/metagpt/{config,workspace} -docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config.yaml > /opt/metagpt/config/key.yaml -vim /opt/metagpt/config/key.yaml # Change the config +docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config2.yaml > /opt/metagpt/config/config2.yaml +vim /opt/metagpt/config/config2.yaml # Change the config # Step 2: Run metagpt demo with container docker run --rm \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest \ - metagpt "Write a cli snake game" + metagpt "Create a 2048 game" ``` detail installation please refer to [docker_install](https://docs.deepwisdom.ai/main/en/guide/get_started/installation.html#install-with-docker) diff --git a/config/config2.yaml b/config/config2.yaml index 5e7f34809..2c4ca636f 100644 --- a/config/config2.yaml +++ b/config/config2.yaml @@ -1,3 +1,3 @@ llm: api_key: "YOUR_API_KEY" - model: "gpt-3.5-turbo-1106" \ No newline at end of file + model: "gpt-4-turbo-preview" # or gpt-3.5-turbo-1106 / gpt-4-1106-preview \ No newline at end of file diff --git a/config/config2.yaml.example b/config/config2.yaml.example index 35575e5a5..2217f1b2c 100644 --- a/config/config2.yaml.example +++ b/config/config2.yaml.example @@ -1,8 +1,8 @@ llm: - api_type: "openai" + api_type: "openai" # or azure / ollama etc. base_url: "YOUR_BASE_URL" api_key: "YOUR_API_KEY" - model: "gpt-3.5-turbo-1106" # or gpt-4-1106-preview + model: "gpt-4-turbo-preview" # or gpt-3.5-turbo-1106 / gpt-4-1106-preview proxy: "YOUR_PROXY" @@ -11,6 +11,10 @@ search: api_key: "YOUR_API_KEY" cse_id: "YOUR_CSE_ID" +browser: + engine: "playwright" # playwright/selenium + browser_type: "chromium" # playwright: chromium/firefox/webkit; selenium: chrome/firefox/edge/ie + mermaid: engine: "pyppeteer" path: "/Applications/Google Chrome.app" @@ -29,14 +33,13 @@ s3: bucket: "test" -AZURE_TTS_SUBSCRIPTION_KEY: "YOUR_SUBSCRIPTION_KEY" -AZURE_TTS_REGION: "eastus" +azure_tts_subscription_key: "YOUR_SUBSCRIPTION_KEY" +azure_tts_region: "eastus" -IFLYTEK_APP_ID: "YOUR_APP_ID" -IFLYTEK_API_KEY: "YOUR_API_KEY" -IFLYTEK_API_SECRET: "YOUR_API_SECRET" +iflytek_api_id: "YOUR_APP_ID" +iflytek_api_key: "YOUR_API_KEY" +iflytek_api_secret: "YOUR_API_SECRET" -METAGPT_TEXT_TO_IMAGE_MODEL_URL: "YOUR_MODEL_URL" - -PYPPETEER_EXECUTABLE_PATH: "/Applications/Google Chrome.app" +metagpt_tti_url: "YOUR_MODEL_URL" +repair_llm_output: true diff --git a/docs/.well-known/metagpt_oas3_api.yaml b/docs/.well-known/metagpt_oas3_api.yaml index 0a702e8b6..1f370b62d 100644 --- a/docs/.well-known/metagpt_oas3_api.yaml +++ b/docs/.well-known/metagpt_oas3_api.yaml @@ -14,16 +14,16 @@ paths: /tts/azsure: x-prerequisite: configurations: - AZURE_TTS_SUBSCRIPTION_KEY: + azure_tts_subscription_key: type: string description: "For more details, check out: [Azure Text-to_Speech](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-support?tabs=tts)" - AZURE_TTS_REGION: + azure_tts_region: type: string description: "For more details, check out: [Azure Text-to_Speech](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-support?tabs=tts)" required: allOf: - - AZURE_TTS_SUBSCRIPTION_KEY - - AZURE_TTS_REGION + - azure_tts_subscription_key + - azure_tts_region post: summary: "Convert Text to Base64-encoded .wav File Stream" description: "For more details, check out: [Azure Text-to_Speech](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-support?tabs=tts)" @@ -94,9 +94,9 @@ paths: description: "WebAPI argument, see: `https://console.xfyun.cn/services/tts`" required: allOf: - - IFLYTEK_APP_ID - - IFLYTEK_API_KEY - - IFLYTEK_API_SECRET + - iflytek_app_id + - iflytek_api_key + - iflytek_api_secret post: summary: "Convert Text to Base64-encoded .mp3 File Stream" description: "For more details, check out: [iFlyTek](https://console.xfyun.cn/services/tts)" @@ -242,12 +242,12 @@ paths: /txt2image/metagpt: x-prerequisite: configurations: - METAGPT_TEXT_TO_IMAGE_MODEL_URL: + metagpt_tti_url: type: string description: "Model url." required: allOf: - - METAGPT_TEXT_TO_IMAGE_MODEL_URL + - metagpt_tti_url post: summary: "Text to Image" description: "Generate an image from the provided text using the MetaGPT Text-to-Image API." diff --git a/docs/.well-known/skills.yaml b/docs/.well-known/skills.yaml index c19a9501e..30c215445 100644 --- a/docs/.well-known/skills.yaml +++ b/docs/.well-known/skills.yaml @@ -14,10 +14,10 @@ entities: id: text_to_speech.text_to_speech x-prerequisite: configurations: - AZURE_TTS_SUBSCRIPTION_KEY: + azure_tts_subscription_key: type: string description: "For more details, check out: [Azure Text-to_Speech](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-support?tabs=tts)" - AZURE_TTS_REGION: + azure_tts_region: type: string description: "For more details, check out: [Azure Text-to_Speech](https://learn.microsoft.com/en-us/azure/ai-services/speech-service/language-support?tabs=tts)" IFLYTEK_APP_ID: @@ -32,12 +32,12 @@ entities: required: oneOf: - allOf: - - AZURE_TTS_SUBSCRIPTION_KEY - - AZURE_TTS_REGION + - azure_tts_subscription_key + - azure_tts_region - allOf: - - IFLYTEK_APP_ID - - IFLYTEK_API_KEY - - IFLYTEK_API_SECRET + - iflytek_app_id + - iflytek_api_key + - iflytek_api_secret parameters: text: description: 'The text used for voice conversion.' @@ -103,13 +103,13 @@ entities: OPENAI_API_KEY: type: string description: "OpenAI API key, For more details, checkout: `https://platform.openai.com/account/api-keys`" - METAGPT_TEXT_TO_IMAGE_MODEL_URL: + metagpt_tti_url: type: string description: "Model url." required: oneOf: - OPENAI_API_KEY - - METAGPT_TEXT_TO_IMAGE_MODEL_URL + - metagpt_tti_url parameters: text: description: 'The text used for image conversion.' diff --git a/docs/FAQ-EN.md b/docs/FAQ-EN.md index d4a9f6097..d3caa244e 100644 --- a/docs/FAQ-EN.md +++ b/docs/FAQ-EN.md @@ -1,183 +1,93 @@ Our vision is to [extend human life](https://github.com/geekan/HowToLiveLonger) and [reduce working hours](https://github.com/geekan/MetaGPT/). -1. ### Convenient Link for Sharing this Document: +### Convenient Link for Sharing this Document: ``` -- MetaGPT-Index/FAQ https://deepwisdom.feishu.cn/wiki/MsGnwQBjiif9c3koSJNcYaoSnu4 +- MetaGPT-Index/FAQ-EN https://github.com/geekan/MetaGPT/blob/main/docs/FAQ-EN.md +- MetaGPT-Index/FAQ-CN https://deepwisdom.feishu.cn/wiki/MsGnwQBjiif9c3koSJNcYaoSnu4 ``` -2. ### Link - - +### Link 1. Code:https://github.com/geekan/MetaGPT - -1. Roadmap:https://github.com/geekan/MetaGPT/blob/main/docs/ROADMAP.md - -1. EN - - 1. Demo Video: [MetaGPT: Multi-Agent AI Programming Framework](https://www.youtube.com/watch?v=8RNzxZBTW8M) +2. Roadmap:https://github.com/geekan/MetaGPT/blob/main/docs/ROADMAP.md +3. EN + 1. Demo Video: [MetaGPT: Multi-Agent AI Programming Framework](https://www.youtube.com/watch?v=8RNzxZBTW8M) 2. Tutorial: [MetaGPT: Deploy POWERFUL Autonomous Ai Agents BETTER Than SUPERAGI!](https://www.youtube.com/watch?v=q16Gi9pTG_M&t=659s) 3. Author's thoughts video(EN): [MetaGPT Matthew Berman](https://youtu.be/uT75J_KG_aY?si=EgbfQNAwD8F5Y1Ak) +4. CN + 1. Demo Video: [MetaGPT:一行代码搭建你的虚拟公司_哔哩哔哩_bilibili](https://www.bilibili.com/video/BV1NP411C7GW/?spm_id_from=333.999.0.0&vd_source=735773c218b47da1b4bd1b98a33c5c77) + 1. Tutorial: [一个提示词写游戏 Flappy bird, 比AutoGPT强10倍的MetaGPT,最接近AGI的AI项目](https://youtu.be/Bp95b8yIH5c) + 2. Author's thoughts video(CN): [MetaGPT作者深度解析直播回放_哔哩哔哩_bilibili](https://www.bilibili.com/video/BV1Ru411V7XL/?spm_id_from=333.337.search-card.all.click) -1. CN - - 1. Demo Video: [MetaGPT:一行代码搭建你的虚拟公司_哔哩哔哩_bilibili](https://www.bilibili.com/video/BV1NP411C7GW/?spm_id_from=333.999.0.0&vd_source=735773c218b47da1b4bd1b98a33c5c77) - 1. Tutorial: [一个提示词写游戏 Flappy bird, 比AutoGPT强10倍的MetaGPT,最接近AGI的AI项目](https://youtu.be/Bp95b8yIH5c) - 2. Author's thoughts video(CN): [MetaGPT作者深度解析直播回放_哔哩哔哩_bilibili](https://www.bilibili.com/video/BV1Ru411V7XL/?spm_id_from=333.337.search-card.all.click) - - - -3. ### How to become a contributor? - - +### How to become a contributor? 1. Choose a task from the Roadmap (or you can propose one). By submitting a PR, you can become a contributor and join the dev team. -1. Current contributors come from backgrounds including ByteDance AI Lab/DingDong/Didi/Xiaohongshu, Tencent/Baidu/MSRA/TikTok/BloomGPT Infra/Bilibili/CUHK/HKUST/CMU/UCB +2. Current contributors come from backgrounds including ByteDance AI Lab/DingDong/Didi/Xiaohongshu, Tencent/Baidu/MSRA/TikTok/BloomGPT Infra/Bilibili/CUHK/HKUST/CMU/UCB - - -4. ### Chief Evangelist (Monthly Rotation) +### Chief Evangelist (Monthly Rotation) MetaGPT Community - The position of Chief Evangelist rotates on a monthly basis. The primary responsibilities include: 1. Maintaining community FAQ documents, announcements, and Github resources/READMEs. -1. Responding to, answering, and distributing community questions within an average of 30 minutes, including on platforms like Github Issues, Discord and WeChat. -1. Upholding a community atmosphere that is enthusiastic, genuine, and friendly. -1. Encouraging everyone to become contributors and participate in projects that are closely related to achieving AGI (Artificial General Intelligence). -1. (Optional) Organizing small-scale events, such as hackathons. +2. Responding to, answering, and distributing community questions within an average of 30 minutes, including on platforms like Github Issues, Discord and WeChat. +3. Upholding a community atmosphere that is enthusiastic, genuine, and friendly. +4. Encouraging everyone to become contributors and participate in projects that are closely related to achieving AGI (Artificial General Intelligence). +5. (Optional) Organizing small-scale events, such as hackathons. - - -5. ### FAQ - - - -1. Experience with the generated repo code: - - 1. https://github.com/geekan/MetaGPT/releases/tag/v0.1.0 +### FAQ 1. Code truncation/ Parsing failure: - - 1. Check if it's due to exceeding length. Consider using the gpt-3.5-turbo-16k or other long token versions. - -1. Success rate: - - 1. There hasn't been a quantitative analysis yet, but the success rate of code generated by GPT-4 is significantly higher than that of gpt-3.5-turbo. - -1. Support for incremental, differential updates (if you wish to continue a half-done task): - - 1. Several prerequisite tasks are listed on the ROADMAP. - -1. Can existing code be loaded? - - 1. It's not on the ROADMAP yet, but there are plans in place. It just requires some time. - -1. Support for multiple programming languages and natural languages? - - 1. It's listed on ROADMAP. - -1. Want to join the contributor team? How to proceed? - + 1. Check if it's due to exceeding length. Consider using the gpt-4-turbo-preview or other long token versions. +2. Success rate: + 1. There hasn't been a quantitative analysis yet, but the success rate of code generated by gpt-4-turbo-preview is significantly higher than that of gpt-3.5-turbo. +3. Support for incremental, differential updates (if you wish to continue a half-done task): + 1. There is now an experimental version. Specify `--inc --project-path ""` or `--inc --project-name ""` on the command line and enter the corresponding requirements to try it. +4. Can existing code be loaded? + 1. We are doing this, but it is very difficult, especially when the project is large, it is very difficult to achieve a high success rate. +5. Support for multiple programming languages and natural languages? + 1. It is now supported, but it is still in experimental version +6. Want to join the contributor team? How to proceed? 1. Merging a PR will get you into the contributor's team. The main ongoing tasks are all listed on the ROADMAP. - -1. PRD stuck / unable to access/ connection interrupted - - 1. The official OPENAI_BASE_URL address is `https://api.openai.com/v1` - 1. If the official OPENAI_BASE_URL address is inaccessible in your environment (this can be verified with curl), it's recommended to configure using the reverse proxy OPENAI_BASE_URL provided by libraries such as openai-forward. For instance, `OPENAI_BASE_URL: "``https://api.openai-forward.com/v1``"` - 1. If the official OPENAI_BASE_URL address is inaccessible in your environment (again, verifiable via curl), another option is to configure the OPENAI_PROXY parameter. This way, you can access the official OPENAI_BASE_URL via a local proxy. If you don't need to access via a proxy, please do not enable this configuration; if accessing through a proxy is required, modify it to the correct proxy address. Note that when OPENAI_PROXY is enabled, don't set OPENAI_BASE_URL. - 1. Note: OpenAI's default API design ends with a v1. An example of the correct configuration is: `OPENAI_BASE_URL: "``https://api.openai.com/v1``"` - -1. Absolutely! How can I assist you today? - +7. PRD stuck / unable to access/ connection interrupted + 1. The official openai base_url address is `https://api.openai.com/v1` + 2. If the official openai base_url address is inaccessible in your environment (this can be verified with curl), it's recommended to configure using base_url to other "reverse-proxy" provider such as openai-forward. For instance, `openai base_url: "``https://api.openai-forward.com/v1``"` + 3. If the official openai base_url address is inaccessible in your environment (again, verifiable via curl), another option is to configure the llm.proxy in the `config2.yaml`. This way, you can access the official openai base_url via a local proxy. If you don't need to access via a proxy, please do not enable this configuration; if accessing through a proxy is required, modify it to the correct proxy address. + 4. Note: OpenAI's default API design ends with a v1. An example of the correct configuration is: `base_url: "https://api.openai.com/v1" +8. Get reply: "Absolutely! How can I assist you today?" 1. Did you use Chi or a similar service? These services are prone to errors, and it seems that the error rate is higher when consuming 3.5k-4k tokens in GPT-4 - -1. What does Max token mean? - +9. What does Max token mean? 1. It's a configuration for OpenAI's maximum response length. If the response exceeds the max token, it will be truncated. - -1. How to change the investment amount? - +10. How to change the investment amount? 1. You can view all commands by typing `metagpt --help` - -1. Which version of Python is more stable? - +11. Which version of Python is more stable? 1. python3.9 / python3.10 - -1. Can't use GPT-4, getting the error "The model gpt-4 does not exist." - +12. Can't use GPT-4, getting the error "The model gpt-4 does not exist." 1. OpenAI's official requirement: You can use GPT-4 only after spending $1 on OpenAI. 1. Tip: Run some data with gpt-3.5-turbo (consume the free quota and $1), and then you should be able to use gpt-4. - -1. Can games whose code has never been seen before be written? - +13. Can games whose code has never been seen before be written? 1. Refer to the README. The recommendation system of Toutiao is one of the most complex systems in the world currently. Although it's not on GitHub, many discussions about it exist online. If it can visualize these, it suggests it can also summarize these discussions and convert them into code. The prompt would be something like "write a recommendation system similar to Toutiao". Note: this was approached in earlier versions of the software. The SOP of those versions was different; the current one adopts Elon Musk's five-step work method, emphasizing trimming down requirements as much as possible. - -1. Under what circumstances would there typically be errors? - +14. Under what circumstances would there typically be errors? 1. More than 500 lines of code: some function implementations may be left blank. - 1. When using a database, it often gets the implementation wrong — since the SQL database initialization process is usually not in the code. - 1. With more lines of code, there's a higher chance of false impressions, leading to calls to non-existent APIs. - -1. Instructions for using SD Skills/UI Role: - - 1. Currently, there is a test script located in /tests/metagpt/roles. The file ui_role provides the corresponding code implementation. For testing, you can refer to the test_ui in the same directory. - - 1. The UI role takes over from the product manager role, extending the output from the 【UI Design draft】 provided by the product manager role. The UI role has implemented the UIDesign Action. Within the run of UIDesign, it processes the respective context, and based on the set template, outputs the UI. The output from the UI role includes: - - 1. UI Design Description: Describes the content to be designed and the design objectives. - 1. Selected Elements: Describes the elements in the design that need to be illustrated. - 1. HTML Layout: Outputs the HTML code for the page. - 1. CSS Styles (styles.css): Outputs the CSS code for the page. - - 1. Currently, the SD skill is a tool invoked by UIDesign. It instantiates the SDEngine, with specific code found in metagpt/tools/sd_engine. - - 1. Configuration instructions for SD Skills: The SD interface is currently deployed based on *https://github.com/AUTOMATIC1111/stable-diffusion-webui* **For environmental configurations and model downloads, please refer to the aforementioned GitHub repository. To initiate the SD service that supports API calls, run the command specified in cmd with the parameter nowebui, i.e., - - 1. > python3 webui.py --enable-insecure-extension-access --port xxx --no-gradio-queue --nowebui - 1.     Once it runs without errors, the interface will be accessible after approximately 1 minute when the model finishes loading. - 1. Configure SD_URL and SD_T2I_API in the config.yaml/key.yaml files. - 1. ![](https://p3-juejin.byteimg.com/tos-cn-i-k3u1fbpfcp/065295a67b0b4feea665d1372722d49d~tplv-k3u1fbpfcp-zoom-1.image) - 1.     SD_URL is the deployed server/machine IP, and Port is the specified port above, defaulting to 7860. - 1. > SD_URL: IP:Port - -1. An error occurred during installation: "Another program is using this file...egg". - + 2. When using a database, it often gets the implementation wrong — since the SQL database initialization process is usually not in the code. + 3. With more lines of code, there's a higher chance of false impressions, leading to calls to non-existent APIs. +15. An error occurred during installation: "Another program is using this file...egg". 1. Delete the file and try again. - 1. Or manually execute`pip install -r requirements.txt` - -1. The origin of the name MetaGPT? - + 2. Or manually execute`pip install -r requirements.txt` +16. The origin of the name MetaGPT? 1. The name was derived after iterating with GPT-4 over a dozen rounds. GPT-4 scored and suggested it. - -1. Is there a more step-by-step installation tutorial? - - 1. Youtube(CN):[一个提示词写游戏 Flappy bird, 比AutoGPT强10倍的MetaGPT,最接近AGI的AI项目=一个软件公司产品经理+程序员](https://youtu.be/Bp95b8yIH5c) - 1. Youtube(EN)https://www.youtube.com/watch?v=q16Gi9pTG_M&t=659s - 2. video(EN): [MetaGPT Matthew Berman](https://youtu.be/uT75J_KG_aY?si=EgbfQNAwD8F5Y1Ak) - -1. openai.error.RateLimitError: You exceeded your current quota, please check your plan and billing details - +17. openai.error.RateLimitError: You exceeded your current quota, please check your plan and billing details 1. If you haven't exhausted your free quota, set RPM to 3 or lower in the settings. - 1. If your free quota is used up, consider adding funds to your account. - -1. What does "borg" mean in n_borg? - + 2. If your free quota is used up, consider adding funds to your account. +18. What does "borg" mean in n_borg? 1. [Wikipedia borg meaning ](https://en.wikipedia.org/wiki/Borg) - 1. The Borg civilization operates based on a hive or collective mentality, known as "the Collective." Every Borg individual is connected to the collective via a sophisticated subspace network, ensuring continuous oversight and guidance for every member. This collective consciousness allows them to not only "share the same thoughts" but also to adapt swiftly to new strategies. While individual members of the collective rarely communicate, the collective "voice" sometimes transmits aboard ships. - -1. How to use the Claude API? - + 2. The Borg civilization operates based on a hive or collective mentality, known as "the Collective." Every Borg individual is connected to the collective via a sophisticated subspace network, ensuring continuous oversight and guidance for every member. This collective consciousness allows them to not only "share the same thoughts" but also to adapt swiftly to new strategies. While individual members of the collective rarely communicate, the collective "voice" sometimes transmits aboard ships. +19. How to use the Claude API? 1. The full implementation of the Claude API is not provided in the current code. 1. You can use the Claude API through third-party API conversion projects like: https://github.com/jtsang4/claude-to-chatgpt - -1. Is Llama2 supported? - +20. Is Llama2 supported? 1. On the day Llama2 was released, some of the community members began experiments and found that output can be generated based on MetaGPT's structure. However, Llama2's context is too short to generate a complete project. Before regularly using Llama2, it's necessary to expand the context window to at least 8k. If anyone has good recommendations for expansion models or methods, please leave a comment. - -1. `mermaid-cli getElementsByTagName SyntaxError: Unexpected token '.'` - +21. `mermaid-cli getElementsByTagName SyntaxError: Unexpected token '.'` 1. Upgrade node to version 14.x or above: - 1. `npm install -g n` - 1. `n stable` to install the stable version of node(v18.x) + 2. `n stable` to install the stable version of node(v18.x) diff --git a/docs/README_CN.md b/docs/README_CN.md index 2855b5500..7a0db4974 100644 --- a/docs/README_CN.md +++ b/docs/README_CN.md @@ -35,50 +35,45 @@ # MetaGPT: 多智能体框架 ## 安装 ### Pip安装 +> 确保您的系统已安装 Python 3.9 或更高版本。您可以使用以下命令来检查:`python --version`。 +> 您可以这样使用 conda:`conda create -n metagpt python=3.9 && conda activate metagpt` + ```bash -# 第 1 步:确保您的系统上安装了 Python 3.9+。您可以使用以下命令进行检查: -# 可以使用conda来初始化新的python环境 -# conda create -n metagpt python=3.9 -# conda activate metagpt -python3 --version - -# 第 2 步:克隆最新仓库到您的本地机器,并进行安装。 -git clone https://github.com/geekan/MetaGPT.git -cd MetaGPT -pip3 install -e. # 或者 pip3 install metagpt # 安装稳定版本 - -# 第 3 步:执行metagpt -# 拷贝config.yaml为key.yaml,并设置你自己的OPENAI_API_KEY -metagpt "Write a cli snake game" - -# 第 4 步【可选的】:如果你想在执行过程中保存像象限图、系统设计、序列流程等图表这些产物,可以在第3步前执行该步骤。默认的,框架做了兼容,在不执行该步的情况下,也可以完整跑完整个流程。 -# 如果执行,确保您的系统上安装了 NPM。并使用npm安装mermaid-js -npm --version -sudo npm install -g @mermaid-js/mermaid-cli +pip install metagpt +metagpt --init-config # 创建 ~/.metagpt/config2.yaml,根据您的需求修改它 +metagpt "创建一个 2048 游戏" # 这将在 ./workspace 创建一个仓库 ``` -详细的安装请安装 [cli_install](https://docs.deepwisdom.ai/guide/get_started/installation.html#install-stable-version) +或者您可以将其作为库使用 + +```python +from metagpt.software_company import generate_repo, ProjectRepo +repo: ProjectRepo = generate_repo("创建一个 2048 游戏") # 或 ProjectRepo("<路径>") +print(repo) # 它将打印出仓库结构及其文件 +``` + +详细的安装请参考 [cli_install](https://docs.deepwisdom.ai/guide/get_started/installation.html#install-stable-version) ### Docker安装 > 注意:在Windows中,你需要将 "/opt/metagpt" 替换为Docker具有创建权限的目录,比如"D:\Users\x\metagpt" ```bash -# 步骤1: 下载metagpt官方镜像并准备好config.yaml +# 步骤1: 下载metagpt官方镜像并准备好config2.yaml docker pull metagpt/metagpt:latest mkdir -p /opt/metagpt/{config,workspace} -docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config.yaml > /opt/metagpt/config/key.yaml -vim /opt/metagpt/config/key.yaml # 修改配置文件 +docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config2.yaml > /opt/metagpt/config/config2.yaml +vim /opt/metagpt/config/config2.yaml # 修改配置文件 # 步骤2: 使用容器运行metagpt演示 docker run --rm \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest \ metagpt "Write a cli snake game" ``` -详细的安装请安装 [docker_install](https://docs.deepwisdom.ai/main/zh/guide/get_started/installation.html#%E4%BD%BF%E7%94%A8docker%E5%AE%89%E8%A3%85) +详细的安装请参考 [docker_install](https://docs.deepwisdom.ai/main/zh/guide/get_started/installation.html#%E4%BD%BF%E7%94%A8docker%E5%AE%89%E8%A3%85) ### 快速开始的演示视频 - 在 [MetaGPT Huggingface Space](https://huggingface.co/spaces/deepwisdom/MetaGPT) 上进行体验 diff --git a/docs/README_JA.md b/docs/README_JA.md index 8b2bf1fae..c6b99461c 100644 --- a/docs/README_JA.md +++ b/docs/README_JA.md @@ -57,24 +57,21 @@ ### インストールビデオガイド - [Matthew Berman: How To Install MetaGPT - Build A Startup With One Prompt!!](https://youtu.be/uT75J_KG_aY) ### 伝統的なインストール +> Python 3.9 以上がシステムにインストールされていることを確認してください。これは `python --version` を使ってチェックできます。 +> 以下のようにcondaを使うことができます:`conda create -n metagpt python=3.9 && conda activate metagpt` ```bash -# ステップ 1: Python 3.9+ がシステムにインストールされていることを確認してください。これを確認するには: -python3 --version +pip install metagpt +metagpt --init-config # ~/.metagpt/config2.yaml を作成し、自分の設定に合わせて変更してください +metagpt "2048ゲームを作成する" # これにより ./workspace にリポジトリが作成されます +``` -# ステップ 2: リポジトリをローカルマシンにクローンし、インストールする。 -git clone https://github.com/geekan/MetaGPT.git -cd MetaGPT -pip install -e. +または、ライブラリとして使用することもできます -# ステップ 3: metagpt を実行する -# config.yaml を key.yaml にコピーし、独自の OPENAI_API_KEY を設定します -metagpt "Write a cli snake game" - -# ステップ 4 [オプション]: 実行中に PRD ファイルなどのアーティファクトを保存する場合は、ステップ 3 の前にこのステップを実行できます。デフォルトでは、フレームワークには互換性があり、この手順を実行しなくてもプロセス全体を完了できます。 -# NPM がシステムにインストールされていることを確認してください。次に mermaid-js をインストールします。(お使いのコンピューターに npm がない場合は、Node.js 公式サイトで Node.js https://nodejs.org/ をインストールしてください。) -npm --version -sudo npm install -g @mermaid-js/mermaid-cli +```python +from metagpt.software_company import generate_repo, ProjectRepo +repo: ProjectRepo = generate_repo("2048ゲームを作成する") # または ProjectRepo("<パス>") +print(repo) # リポジトリの構造とファイルを出力します ``` **注:** @@ -91,8 +88,8 @@ # NPM がシステムにインストールされていることを確認して - config.yml に mmdc のコンフィグを記述するのを忘れないこと ```yml - PUPPETEER_CONFIG: "./config/puppeteer-config.json" - MMDC: "./node_modules/.bin/mmdc" + puppeteer_config: "./config/puppeteer-config.json" + path: "./node_modules/.bin/mmdc" ``` - もし `pip install -e.` がエラー `[Errno 13] Permission denied: '/usr/local/lib/python3.11/dist-packages/test-easy-install-13129.write-test'` で失敗したら、代わりに `pip install -e. --user` を実行してみてください @@ -114,12 +111,13 @@ # NPM がシステムにインストールされていることを確認して playwright install --with-deps chromium ``` - - **modify `config.yaml`** + - **modify `config2.yaml`** - config.yaml から MERMAID_ENGINE のコメントを外し、`playwright` に変更する + config2.yaml から mermaid.engine のコメントを外し、`playwright` に変更する ```yaml - MERMAID_ENGINE: playwright + mermaid: + engine: playwright ``` - pyppeteer @@ -143,21 +141,23 @@ # NPM がシステムにインストールされていることを確認して pyppeteer-install ``` - - **`config.yaml` を修正** + - **`config2.yaml` を修正** - config.yaml から MERMAID_ENGINE のコメントを外し、`pyppeteer` に変更する + config2.yaml から mermaid.engine のコメントを外し、`pyppeteer` に変更する ```yaml - MERMAID_ENGINE: pyppeteer + mermaid: + engine: pyppeteer ``` - mermaid.ink - - **`config.yaml` を修正** + - **`config2.yaml` を修正** - config.yaml から MERMAID_ENGINE のコメントを外し、`ink` に変更する + config2.yaml から mermaid.engine のコメントを外し、`ink` に変更する ```yaml - MERMAID_ENGINE: ink + mermaid: + engine: ink ``` 注: この方法は pdf エクスポートに対応していません。 @@ -166,16 +166,16 @@ ### Docker によるインストール > Windowsでは、"/opt/metagpt"をDockerが作成する権限を持つディレクトリに置き換える必要があります。例えば、"D:\Users\x\metagpt"などです。 ```bash -# ステップ 1: metagpt 公式イメージをダウンロードし、config.yaml を準備する +# ステップ 1: metagpt 公式イメージをダウンロードし、config2.yaml を準備する docker pull metagpt/metagpt:latest mkdir -p /opt/metagpt/{config,workspace} -docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config.yaml > /opt/metagpt/config/key.yaml -vim /opt/metagpt/config/key.yaml # 設定を変更する +docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config2.yaml > /opt/metagpt/config/config2.yaml +vim /opt/metagpt/config/config2.yaml # 設定を変更する # ステップ 2: コンテナで metagpt デモを実行する docker run --rm \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest \ metagpt "Write a cli snake game" @@ -183,7 +183,7 @@ # ステップ 2: コンテナで metagpt デモを実行する # コンテナを起動し、その中でコマンドを実行することもできます docker run --name metagpt -d \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest @@ -194,7 +194,7 @@ # コンテナを起動し、その中でコマンドを実行することもで コマンド `docker run ...` は以下のことを行います: - 特権モードで実行し、ブラウザの実行権限を得る -- ホスト設定ファイル `/opt/metagpt/config/key.yaml` をコンテナ `/app/metagpt/config/key.yaml` にマップします +- ホスト設定ファイル `/opt/metagpt/config/config2.yaml` をコンテナ `/app/metagpt/config/config2.yaml` にマップします - ホストディレクトリ `/opt/metagpt/workspace` をコンテナディレクトリ `/app/metagpt/workspace` にマップするs - デモコマンド `metagpt "Write a cli snake game"` を実行する @@ -208,19 +208,14 @@ # また、自分で metagpt イメージを構築することもできます。 ## 設定 -- `OPENAI_API_KEY` を `config/key.yaml / config/config.yaml / env` のいずれかで設定します。 -- 優先順位は: `config/key.yaml > config/config.yaml > env` の順です。 +- `api_key` を `~/.metagpt/config2.yaml / config/config2.yaml` のいずれかで設定します。 +- 優先順位は: `~/.metagpt/config2.yaml > config/config2.yaml > env` の順です。 ```bash # 設定ファイルをコピーし、必要な修正を加える。 -cp config/config.yaml config/key.yaml +cp config/config2.yaml ~/.metagpt/config2.yaml ``` -| 変数名 | config/key.yaml | env | -| --------------------------------------- | ----------------------------------------- | ----------------------------------------------- | -| OPENAI_API_KEY # 自分のキーに置き換える | OPENAI_API_KEY: "sk-..." | export OPENAI_API_KEY="sk-..." | -| OPENAI_BASE_URL # オプション | OPENAI_BASE_URL: "https:///v1" | export OPENAI_BASE_URL="https:///v1" | - ## チュートリアル: スタートアップの開始 ```shell diff --git a/docs/ROADMAP.md b/docs/ROADMAP.md index 9bc62f849..4bb530bf2 100644 --- a/docs/ROADMAP.md +++ b/docs/ROADMAP.md @@ -76,9 +76,8 @@ ### Tasks 2. ~~Support Azure asynchronous API~~ 3. Support streaming version of all APIs 4. ~~Make gpt-3.5-turbo available (HARD)~~ - 5. Support 10. Other 1. ~~Clean up existing unused code~~ - 2. Unify all code styles and establish contribution standards + 2. ~~Unify all code styles and establish contribution standards~~ 3. ~~Multi-language support~~ 4. ~~Multi-programming-language support~~ diff --git a/docs/install/cli_install.md b/docs/install/cli_install.md index 80deda771..b79ad9cb7 100644 --- a/docs/install/cli_install.md +++ b/docs/install/cli_install.md @@ -9,17 +9,29 @@ ### Support System and version ### Detail Installation ```bash -# Step 1: Ensure that NPM is installed on your system. Then install mermaid-js. (If you don't have npm in your computer, please go to the Node.js official website to install Node.js https://nodejs.org/ and then you will have npm tool in your computer.) -npm --version -sudo npm install -g @mermaid-js/mermaid-cli - -# Step 2: Ensure that Python 3.9+ is installed on your system. You can check this by using: +# Step 1: Ensure that Python 3.9+ is installed on your system. You can check this by using: +# You can use conda to initialize a new python env +# conda create -n metagpt python=3.9 +# conda activate metagpt python3 --version -# Step 3: Clone the repository to your local machine, and install it. +# Step 2: Clone the repository to your local machine for latest version, and install it. git clone https://github.com/geekan/MetaGPT.git cd MetaGPT -pip install -e. +pip3 install -e . # or pip3 install metagpt # for stable version + +# Step 3: setup your LLM key in the config2.yaml file +mkdir ~/.metagpt +cp config/config2.yaml ~/.metagpt/config2.yaml +vim ~/.metagpt/config2.yaml + +# Step 4: run metagpt cli +metagpt "Create a 2048 game in python" + +# Step 5 [Optional]: If you want to save the artifacts like diagrams such as quadrant chart, system designs, sequence flow in the workspace, you can execute the step before Step 3. By default, the framework is compatible, and the entire process can be run completely without executing this step. +# If executing, ensure that NPM is installed on your system. Then install mermaid-js. (If you don't have npm in your computer, please go to the Node.js official website to install Node.js https://nodejs.org/ and then you will have npm tool in your computer.) +npm --version +sudo npm install -g @mermaid-js/mermaid-cli ``` **Note:** @@ -33,11 +45,12 @@ # Step 3: Clone the repository to your local machine, and install it. npm install @mermaid-js/mermaid-cli ``` -- don't forget to the configuration for mmdc in config.yml +- don't forget to the configuration for mmdc path in config.yml - ```yml - PUPPETEER_CONFIG: "./config/puppeteer-config.json" - MMDC: "./node_modules/.bin/mmdc" + ```yaml + mermaid: + puppeteer_config: "./config/puppeteer-config.json" + path: "./node_modules/.bin/mmdc" ``` - if `pip install -e.` fails with error `[Errno 13] Permission denied: '/usr/local/lib/python3.11/dist-packages/test-easy-install-13129.write-test'`, try instead running `pip install -e. --user` @@ -59,12 +72,13 @@ # Step 3: Clone the repository to your local machine, and install it. playwright install --with-deps chromium ``` - - **modify `config.yaml`** + - **modify `config2.yaml`** - uncomment MERMAID_ENGINE from config.yaml and change it to `playwright` + change mermaid.engine to `playwright` ```yaml - MERMAID_ENGINE: playwright + mermaid: + engine: playwright ``` - pyppeteer @@ -88,22 +102,24 @@ # Step 3: Clone the repository to your local machine, and install it. pyppeteer-install ``` - - **modify `config.yaml`** + - **modify `config2.yaml`** - uncomment MERMAID_ENGINE from config.yaml and change it to `pyppeteer` + change mermaid.engine to `pyppeteer` ```yaml - MERMAID_ENGINE: pyppeteer + mermaid: + engine: pyppeteer ``` - mermaid.ink - - **modify `config.yaml`** - - uncomment MERMAID_ENGINE from config.yaml and change it to `ink` + - **modify `config2.yaml`** + + change mermaid.engine to `ink` ```yaml - MERMAID_ENGINE: ink + mermaid: + engine: ink ``` Note: this method does not support pdf export. - \ No newline at end of file + diff --git a/docs/install/cli_install_cn.md b/docs/install/cli_install_cn.md index b1da1b813..1ee18d9a6 100644 --- a/docs/install/cli_install_cn.md +++ b/docs/install/cli_install_cn.md @@ -10,17 +10,29 @@ ### 支持的系统和版本 ### 详细安装 ```bash -# 第 1 步:确保您的系统上安装了 NPM。并使用npm安装mermaid-js -npm --version -sudo npm install -g @mermaid-js/mermaid-cli - -# 第 2 步:确保您的系统上安装了 Python 3.9+。您可以使用以下命令进行检查: +# 步骤 1: 确保您的系统安装了 Python 3.9 或更高版本。您可以使用以下命令来检查: +# 您可以使用 conda 来初始化一个新的 Python 环境 +# conda create -n metagpt python=3.9 +# conda activate metagpt python3 --version -# 第 3 步:克隆仓库到您的本地机器,并进行安装。 +# 步骤 2: 克隆仓库到您的本地机器以获取最新版本,并安装它。 git clone https://github.com/geekan/MetaGPT.git cd MetaGPT -pip install -e. +pip3 install -e . # 或 pip3 install metagpt # 用于稳定版本 + +# 步骤 3: 在 config2.yaml 文件中设置您的 LLM 密钥 +mkdir ~/.metagpt +cp config/config2.yaml ~/.metagpt/config2.yaml +vim ~/.metagpt/config2.yaml + +# 步骤 4: 运行 metagpt 命令行界面 +metagpt "用 python 创建一个 2048 游戏" + +# 步骤 5 [可选]: 如果您想保存诸如象限图、系统设计、序列流等图表作为工作空间的工件,您可以在执行步骤 3 之前执行此步骤。默认情况下,该框架是兼容的,整个过程可以完全不执行此步骤而运行。 +# 如果执行此步骤,请确保您的系统上安装了 NPM。然后安装 mermaid-js。(如果您的计算机中没有 npm,请访问 Node.js 官方网站 https://nodejs.org/ 安装 Node.js,然后您将在计算机中拥有 npm 工具。) +npm --version +sudo npm install -g @mermaid-js/mermaid-cli ``` **注意:** @@ -33,11 +45,12 @@ # 第 3 步:克隆仓库到您的本地机器,并进行安装。 npm install @mermaid-js/mermaid-cli ``` -- 不要忘记在config.yml中为mmdc配置配置, +- 不要忘记在config.yml中为mmdc配置 ```yml - PUPPETEER_CONFIG: "./config/puppeteer-config.json" - MMDC: "./node_modules/.bin/mmdc" + mermaid: + puppeteer_config: "./config/puppeteer-config.json" + path: "./node_modules/.bin/mmdc" ``` - 如果`pip install -e.`失败并显示错误`[Errno 13] Permission denied: '/usr/local/lib/python3.11/dist-packages/test-easy-install-13129.write-test'`,请尝试使用`pip install -e. --user`运行。 diff --git a/docs/install/docker_install.md b/docs/install/docker_install.md index 37125bdbe..2fe1b6abf 100644 --- a/docs/install/docker_install.md +++ b/docs/install/docker_install.md @@ -3,16 +3,16 @@ ## Docker Installation ### Use default MetaGPT image ```bash -# Step 1: Download metagpt official image and prepare config.yaml +# Step 1: Download metagpt official image and prepare config2.yaml docker pull metagpt/metagpt:latest mkdir -p /opt/metagpt/{config,workspace} -docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config.yaml > /opt/metagpt/config/key.yaml -vim /opt/metagpt/config/key.yaml # Change the config +docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config2.yaml > /opt/metagpt/config/config2.yaml +vim /opt/metagpt/config/config2.yaml # Change the config # Step 2: Run metagpt demo with container docker run --rm \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest \ metagpt "Write a cli snake game" @@ -20,7 +20,7 @@ # Step 2: Run metagpt demo with container # You can also start a container and execute commands in it docker run --name metagpt -d \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest @@ -31,7 +31,7 @@ # You can also start a container and execute commands in it The command `docker run ...` do the following things: - Run in privileged mode to have permission to run the browser -- Map host configure file `/opt/metagpt/config/key.yaml` to container `/app/metagpt/config/key.yaml` +- Map host configure file `/opt/metagpt/config/config2.yaml` to container `/app/metagpt/config/config2.yaml` - Map host directory `/opt/metagpt/workspace` to container `/app/metagpt/workspace` - Execute the demo command `metagpt "Write a cli snake game"` diff --git a/docs/install/docker_install_cn.md b/docs/install/docker_install_cn.md index f360b49ed..10204c1e0 100644 --- a/docs/install/docker_install_cn.md +++ b/docs/install/docker_install_cn.md @@ -3,16 +3,16 @@ ## Docker安装 ### 使用MetaGPT镜像 ```bash -# 步骤1: 下载metagpt官方镜像并准备好config.yaml +# 步骤1: 下载metagpt官方镜像并准备好config2.yaml docker pull metagpt/metagpt:latest mkdir -p /opt/metagpt/{config,workspace} -docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config.yaml > /opt/metagpt/config/key.yaml -vim /opt/metagpt/config/key.yaml # 修改配置文件 +docker run --rm metagpt/metagpt:latest cat /app/metagpt/config/config2.yaml > /opt/metagpt/config/config2.yaml +vim /opt/metagpt/config/config2.yaml # 修改配置文件 # 步骤2: 使用容器运行metagpt演示 docker run --rm \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest \ metagpt "Write a cli snake game" @@ -20,7 +20,7 @@ # 步骤2: 使用容器运行metagpt演示 # 您也可以启动一个容器并在其中执行命令 docker run --name metagpt -d \ --privileged \ - -v /opt/metagpt/config/key.yaml:/app/metagpt/config/key.yaml \ + -v /opt/metagpt/config/config2.yaml:/app/metagpt/config/config2.yaml \ -v /opt/metagpt/workspace:/app/metagpt/workspace \ metagpt/metagpt:latest @@ -31,7 +31,7 @@ # 您也可以启动一个容器并在其中执行命令 `docker run ...`做了以下事情: - 以特权模式运行,有权限运行浏览器 -- 将主机文件 `/opt/metagpt/config/key.yaml` 映射到容器文件 `/app/metagpt/config/key.yaml` +- 将主机文件 `/opt/metagpt/config/config2.yaml` 映射到容器文件 `/app/metagpt/config/config2.yaml` - 将主机目录 `/opt/metagpt/workspace` 映射到容器目录 `/app/metagpt/workspace` - 执行示例命令 `metagpt "Write a cli snake game"` diff --git a/docs/tutorial/usage.md b/docs/tutorial/usage.md index a08d92a22..1128e98a5 100644 --- a/docs/tutorial/usage.md +++ b/docs/tutorial/usage.md @@ -2,19 +2,14 @@ ## MetaGPT Usage ### Configuration -- Configure your `OPENAI_API_KEY` in any of `config/key.yaml / config/config.yaml / env` -- Priority order: `config/key.yaml > config/config.yaml > env` +- Configure your `api_key` in any of `~/.metagpt/config2.yaml / config/config2.yaml` +- Priority order: `~/.metagpt/config2.yaml > config/config2.yaml` ```bash # Copy the configuration file and make the necessary modifications. -cp config/config.yaml config/key.yaml +cp config/config2.yaml ~/.metagpt/config2.yaml ``` -| Variable Name | config/key.yaml | env | -| ------------------------------------------ | ----------------------------------------- | ----------------------------------------------- | -| OPENAI_API_KEY # Replace with your own key | OPENAI_API_KEY: "sk-..." | export OPENAI_API_KEY="sk-..." | -| OPENAI_BASE_URL # Optional | OPENAI_BASE_URL: "https:///v1" | export OPENAI_BASE_URL="https:///v1" | - ### Initiating a startup ```shell @@ -39,29 +34,28 @@ ### Preference of Platform or Tool ### Usage ``` -NAME - metagpt - We are a software startup comprised of AI. By investing in us, you are empowering a future filled with limitless possibilities. - -SYNOPSIS - metagpt IDEA - -DESCRIPTION - We are a software startup comprised of AI. By investing in us, you are empowering a future filled with limitless possibilities. - -POSITIONAL ARGUMENTS - IDEA - Type: str - Your innovative idea, such as "Creating a snake game." - -FLAGS - --investment=INVESTMENT - Type: float - Default: 3.0 - As an investor, you have the opportunity to contribute a certain dollar amount to this AI company. - --n_round=N_ROUND - Type: int - Default: 5 - -NOTES - You can also use flags syntax for POSITIONAL ARGUMENTS + Usage: metagpt [OPTIONS] [IDEA] + + Start a new project. + +╭─ Arguments ────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ +│ idea [IDEA] Your innovative idea, such as 'Create a 2048 game.' [default: None] │ +╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ +╭─ Options ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ +│ --investment FLOAT Dollar amount to invest in the AI company. [default: 3.0] │ +│ --n-round INTEGER Number of rounds for the simulation. [default: 5] │ +│ --code-review --no-code-review Whether to use code review. [default: code-review] │ +│ --run-tests --no-run-tests Whether to enable QA for adding & running tests. [default: no-run-tests] │ +│ --implement --no-implement Enable or disable code implementation. [default: implement] │ +│ --project-name TEXT Unique project name, such as 'game_2048'. │ +│ --inc --no-inc Incremental mode. Use it to coop with existing repo. [default: no-inc] │ +│ --project-path TEXT Specify the directory path of the old version project to fulfill the incremental requirements. │ +│ --reqa-file TEXT Specify the source file name for rewriting the quality assurance code. │ +│ --max-auto-summarize-code INTEGER The maximum number of times the 'SummarizeCode' action is automatically invoked, with -1 indicating unlimited. This parameter is used for debugging the │ +│ workflow. │ +│ [default: 0] │ +│ --recover-path TEXT recover the project from existing serialized storage [default: None] │ +│ --init-config --no-init-config Initialize the configuration file for MetaGPT. [default: no-init-config] │ +│ --help Show this message and exit. │ +╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ``` \ No newline at end of file diff --git a/docs/tutorial/usage_cn.md b/docs/tutorial/usage_cn.md index 76a5d6b1b..3b0c86279 100644 --- a/docs/tutorial/usage_cn.md +++ b/docs/tutorial/usage_cn.md @@ -2,19 +2,14 @@ ## MetaGPT 使用 ### 配置 -- 在 `config/key.yaml / config/config.yaml / env` 中配置您的 `OPENAI_API_KEY` -- 优先级顺序:`config/key.yaml > config/config.yaml > env` +- 在 `~/.metagpt/config2.yaml / config/config2.yaml` 中配置您的 `api_key` +- 优先级顺序:`~/.metagpt/config2.yaml > config/config2.yaml` ```bash # 复制配置文件并进行必要的修改 -cp config/config.yaml config/key.yaml +cp config/config2.yaml ~/.metagpt/config2.yaml ``` -| 变量名 | config/key.yaml | env | -| ----------------------------------- | ----------------------------------------- | ----------------------------------------------- | -| OPENAI_API_KEY # 用您自己的密钥替换 | OPENAI_API_KEY: "sk-..." | export OPENAI_API_KEY="sk-..." | -| OPENAI_BASE_URL # 可选 | OPENAI_BASE_URL: "https:///v1" | export OPENAI_BASE_URL="https:///v1" | - ### 示例:启动一个创业公司 ```shell @@ -35,29 +30,28 @@ ### 平台或工具的倾向性 ### 使用 ``` -名称 - metagpt - 我们是一家AI软件创业公司。通过投资我们,您将赋能一个充满无限可能的未来。 - -概要 - metagpt IDEA - -描述 - 我们是一家AI软件创业公司。通过投资我们,您将赋能一个充满无限可能的未来。 - -位置参数 - IDEA - 类型: str - 您的创新想法,例如"写一个命令行贪吃蛇。" - -标志 - --investment=INVESTMENT - 类型: float - 默认值: 3.0 - 作为投资者,您有机会向这家AI公司投入一定的美元金额。 - --n_round=N_ROUND - 类型: int - 默认值: 5 - -备注 - 您也可以用`标志`的语法,来处理`位置参数` + Usage: metagpt [OPTIONS] [IDEA] + + Start a new project. + +╭─ Arguments ────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ +│ idea [IDEA] Your innovative idea, such as 'Create a 2048 game.' [default: None] │ +╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ +╭─ Options ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ +│ --investment FLOAT Dollar amount to invest in the AI company. [default: 3.0] │ +│ --n-round INTEGER Number of rounds for the simulation. [default: 5] │ +│ --code-review --no-code-review Whether to use code review. [default: code-review] │ +│ --run-tests --no-run-tests Whether to enable QA for adding & running tests. [default: no-run-tests] │ +│ --implement --no-implement Enable or disable code implementation. [default: implement] │ +│ --project-name TEXT Unique project name, such as 'game_2048'. │ +│ --inc --no-inc Incremental mode. Use it to coop with existing repo. [default: no-inc] │ +│ --project-path TEXT Specify the directory path of the old version project to fulfill the incremental requirements. │ +│ --reqa-file TEXT Specify the source file name for rewriting the quality assurance code. │ +│ --max-auto-summarize-code INTEGER The maximum number of times the 'SummarizeCode' action is automatically invoked, with -1 indicating unlimited. This parameter is used for debugging the │ +│ workflow. │ +│ [default: 0] │ +│ --recover-path TEXT recover the project from existing serialized storage [default: None] │ +│ --init-config --no-init-config Initialize the configuration file for MetaGPT. [default: no-init-config] │ +│ --help Show this message and exit. │ +╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ``` diff --git a/examples/crawl_webpage.py b/examples/crawl_webpage.py new file mode 100644 index 000000000..2db9e407b --- /dev/null +++ b/examples/crawl_webpage.py @@ -0,0 +1,22 @@ +# -*- encoding: utf-8 -*- +""" +@Date : 2024/01/24 15:11:27 +@Author : orange-crow +@File : crawl_webpage.py +""" + +from metagpt.roles.ci.code_interpreter import CodeInterpreter + + +async def main(): + prompt = """Get data from `paperlist` table in https://papercopilot.com/statistics/iclr-statistics/iclr-2024-statistics/, + and save it to a csv file. paper title must include `multiagent` or `large language model`. *notice: print key variables*""" + ci = CodeInterpreter(goal=prompt, use_tools=True) + + await ci.run(prompt) + + +if __name__ == "__main__": + import asyncio + + asyncio.run(main()) diff --git a/examples/dalle_gpt4v_agent.py b/examples/dalle_gpt4v_agent.py new file mode 100644 index 000000000..28215dba3 --- /dev/null +++ b/examples/dalle_gpt4v_agent.py @@ -0,0 +1,77 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : use gpt4v to improve prompt and draw image with dall-e-3 + +"""set `model: "gpt-4-vision-preview"` in `config2.yaml` first""" + +import asyncio + +from PIL import Image + +from metagpt.actions.action import Action +from metagpt.logs import logger +from metagpt.roles.role import Role +from metagpt.schema import Message +from metagpt.utils.common import encode_image + + +class GenAndImproveImageAction(Action): + save_image: bool = True + + async def generate_image(self, prompt: str) -> Image: + imgs = await self.llm.gen_image(model="dall-e-3", prompt=prompt) + return imgs[0] + + async def refine_prompt(self, old_prompt: str, image: Image) -> str: + msg = ( + f"You are a creative painter, with the given generated image and old prompt: {old_prompt}, " + f"please refine the prompt and generate new one. Just output the new prompt." + ) + b64_img = encode_image(image) + new_prompt = await self.llm.aask(msg=msg, images=[b64_img]) + return new_prompt + + async def evaluate_images(self, old_prompt: str, images: list[Image]) -> str: + msg = ( + "With the prompt and two generated image, to judge if the second one is better than the first one. " + "If so, just output True else output False" + ) + b64_imgs = [encode_image(img) for img in images] + res = await self.llm.aask(msg=msg, images=b64_imgs) + return res + + async def run(self, messages: list[Message]) -> str: + prompt = messages[-1].content + + old_img: Image = await self.generate_image(prompt) + new_prompt = await self.refine_prompt(old_prompt=prompt, image=old_img) + logger.info(f"original prompt: {prompt}") + logger.info(f"refined prompt: {new_prompt}") + new_img: Image = await self.generate_image(new_prompt) + if self.save_image: + old_img.save("./img_by-dall-e_old.png") + new_img.save("./img_by-dall-e_new.png") + res = await self.evaluate_images(old_prompt=prompt, images=[old_img, new_img]) + opinion = f"The second generated image is better than the first one: {res}" + logger.info(f"evaluate opinion: {opinion}") + return opinion + + +class Painter(Role): + name: str = "MaLiang" + profile: str = "Painter" + goal: str = "to generate fine painting" + + def __init__(self, **data): + super().__init__(**data) + + self.set_actions([GenAndImproveImageAction]) + + +async def main(): + role = Painter() + await role.run(with_message="a girl with flowers") + + +if __name__ == "__main__": + asyncio.run(main()) diff --git a/examples/example.faiss b/examples/example.faiss deleted file mode 100644 index 580946190..000000000 Binary files a/examples/example.faiss and /dev/null differ diff --git a/examples/example.pkl b/examples/example.pkl deleted file mode 100644 index 0469a2e46..000000000 Binary files a/examples/example.pkl and /dev/null differ diff --git a/examples/imitate_webpage.py b/examples/imitate_webpage.py new file mode 100644 index 000000000..5075e1e39 --- /dev/null +++ b/examples/imitate_webpage.py @@ -0,0 +1,26 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/15 +@Author : mannaandpoem +@File : imitate_webpage.py +""" +from metagpt.roles.ci.code_interpreter import CodeInterpreter + + +async def main(): + web_url = "https://pytorch.org/" + prompt = f"""This is a URL of webpage: '{web_url}' . +Firstly, utilize Selenium and WebDriver for rendering. +Secondly, convert image to a webpage including HTML, CSS and JS in one go. +Finally, save webpage in a text file. +Note: All required dependencies and environments have been fully installed and configured.""" + ci = CodeInterpreter(goal=prompt, use_tools=True) + + await ci.run(prompt) + + +if __name__ == "__main__": + import asyncio + + asyncio.run(main()) diff --git a/examples/llm_hello_world.py b/examples/llm_hello_world.py index 76be1cc90..219a303c8 100644 --- a/examples/llm_hello_world.py +++ b/examples/llm_hello_world.py @@ -23,6 +23,10 @@ async def main(): # streaming mode, much slower await llm.acompletion_text(hello_msg, stream=True) + # check completion if exist to test llm complete functions + if hasattr(llm, "completion"): + logger.info(llm.completion(hello_msg)) + if __name__ == "__main__": asyncio.run(main()) diff --git a/examples/llm_vision.py b/examples/llm_vision.py new file mode 100644 index 000000000..276decd59 --- /dev/null +++ b/examples/llm_vision.py @@ -0,0 +1,23 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : example to run the ability of LLM vision + +import asyncio +from pathlib import Path + +from metagpt.llm import LLM +from metagpt.utils.common import encode_image + + +async def main(): + llm = LLM() + + # check if the configured llm supports llm-vision capacity. If not, it will throw a error + invoice_path = Path(__file__).parent.joinpath("..", "tests", "data", "invoices", "invoice-2.png") + img_base64 = encode_image(invoice_path) + res = await llm.aask(msg="if this is a invoice, just return True else return False", images=[img_base64]) + assert "true" in res.lower() + + +if __name__ == "__main__": + asyncio.run(main()) diff --git a/examples/sd_tool_usage.py b/examples/sd_tool_usage.py new file mode 100644 index 000000000..b4642af23 --- /dev/null +++ b/examples/sd_tool_usage.py @@ -0,0 +1,21 @@ +# -*- coding: utf-8 -*- +# @Date : 1/11/2024 7:06 PM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : +import asyncio + +from metagpt.roles.ci.code_interpreter import CodeInterpreter + + +async def main(requirement: str = ""): + code_interpreter = CodeInterpreter(use_tools=True, goal=requirement) + await code_interpreter.run(requirement) + + +if __name__ == "__main__": + sd_url = "http://your.sd.service.ip:port" + requirement = ( + f"I want to generate an image of a beautiful girl using the stable diffusion text2image tool, sd_url={sd_url}" + ) + + asyncio.run(main(requirement)) diff --git a/examples/search_with_specific_engine.py b/examples/search_with_specific_engine.py index 9406a2965..97b1378ee 100644 --- a/examples/search_with_specific_engine.py +++ b/examples/search_with_specific_engine.py @@ -5,17 +5,20 @@ import asyncio from metagpt.roles import Searcher -from metagpt.tools import SearchEngineType +from metagpt.tools.search_engine import SearchEngine, SearchEngineType async def main(): question = "What are the most interesting human facts?" + kwargs = {"api_key": "", "cse_id": "", "proxy": None} # Serper API - # await Searcher(engine=SearchEngineType.SERPER_GOOGLE).run(question) + # await Searcher(search_engine=SearchEngine(engine=SearchEngineType.SERPER_GOOGLE, **kwargs)).run(question) # SerpAPI - await Searcher(engine=SearchEngineType.SERPAPI_GOOGLE).run(question) + # await Searcher(search_engine=SearchEngine(engine=SearchEngineType.SERPAPI_GOOGLE, **kwargs)).run(question) # Google API - # await Searcher(engine=SearchEngineType.DIRECT_GOOGLE).run(question) + # await Searcher(search_engine=SearchEngine(engine=SearchEngineType.DIRECT_GOOGLE, **kwargs)).run(question) + # DDG API + await Searcher(search_engine=SearchEngine(engine=SearchEngineType.DUCK_DUCK_GO, **kwargs)).run(question) if __name__ == "__main__": diff --git a/examples/write_novel.py b/examples/write_novel.py new file mode 100644 index 000000000..b272a56e6 --- /dev/null +++ b/examples/write_novel.py @@ -0,0 +1,49 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/2/1 12:01 +@Author : alexanderwu +@File : write_novel.py +""" +import asyncio +from typing import List + +from pydantic import BaseModel, Field + +from metagpt.actions.action_node import ActionNode +from metagpt.llm import LLM + + +class Novel(BaseModel): + name: str = Field(default="The Lord of the Rings", description="The name of the novel.") + user_group: str = Field(default="...", description="The user group of the novel.") + outlines: List[str] = Field( + default=["Chapter 1: ...", "Chapter 2: ...", "Chapter 3: ..."], + description="The outlines of the novel. No more than 10 chapters.", + ) + background: str = Field(default="...", description="The background of the novel.") + character_names: List[str] = Field(default=["Frodo", "Gandalf", "Sauron"], description="The characters.") + conflict: str = Field(default="...", description="The conflict of the characters.") + plot: str = Field(default="...", description="The plot of the novel.") + ending: str = Field(default="...", description="The ending of the novel.") + + +class Chapter(BaseModel): + name: str = Field(default="Chapter 1", description="The name of the chapter.") + content: str = Field(default="...", description="The content of the chapter. No more than 1000 words.") + + +async def generate_novel(): + instruction = ( + "Write a novel named 'Harry Potter in The Lord of the Rings'. " + "Fill the empty nodes with your own ideas. Be creative! Use your own words!" + "I will tip you $100,000 if you write a good novel." + ) + novel_node = await ActionNode.from_pydantic(Novel).fill(context=instruction, llm=LLM()) + chap_node = await ActionNode.from_pydantic(Chapter).fill( + context=f"### instruction\n{instruction}\n### novel\n{novel_node.content}", llm=LLM() + ) + print(chap_node.content) + + +asyncio.run(generate_novel()) diff --git a/metagpt/actions/__init__.py b/metagpt/actions/__init__.py index 5b995bab6..363b4fd33 100644 --- a/metagpt/actions/__init__.py +++ b/metagpt/actions/__init__.py @@ -22,6 +22,9 @@ from metagpt.actions.write_code_review import WriteCodeReview from metagpt.actions.write_prd import WritePRD from metagpt.actions.write_prd_review import WritePRDReview from metagpt.actions.write_test import WriteTest +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.actions.ci.write_analysis_code import WriteCodeWithoutTools, WriteCodeWithTools +from metagpt.actions.ci.write_plan import WritePlan class ActionType(Enum): @@ -42,6 +45,10 @@ class ActionType(Enum): COLLECT_LINKS = CollectLinks WEB_BROWSE_AND_SUMMARIZE = WebBrowseAndSummarize CONDUCT_RESEARCH = ConductResearch + EXECUTE_NB_CODE = ExecuteNbCode + WRITE_CODE_WITHOUT_TOOLS = WriteCodeWithoutTools + WRITE_CODE_WITH_TOOLS = WriteCodeWithTools + WRITE_PLAN = WritePlan __all__ = [ diff --git a/metagpt/actions/action.py b/metagpt/actions/action.py index a33918a09..1b93213f7 100644 --- a/metagpt/actions/action.py +++ b/metagpt/actions/action.py @@ -15,6 +15,7 @@ from pydantic import BaseModel, ConfigDict, Field, model_validator from metagpt.actions.action_node import ActionNode from metagpt.context_mixin import ContextMixin from metagpt.schema import ( + CodePlanAndChangeContext, CodeSummarizeContext, CodingContext, RunCodeContext, @@ -28,14 +29,18 @@ class Action(SerializationMixin, ContextMixin, BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) name: str = "" - i_context: Union[dict, CodingContext, CodeSummarizeContext, TestingContext, RunCodeContext, str, None] = "" + i_context: Union[ + dict, CodingContext, CodeSummarizeContext, TestingContext, RunCodeContext, CodePlanAndChangeContext, str, None + ] = "" prefix: str = "" # aask*时会加上prefix,作为system_message desc: str = "" # for skill manager node: ActionNode = Field(default=None, exclude=True) @property - def project_repo(self): - return ProjectRepo(self.context.git_repo) + def repo(self) -> ProjectRepo: + if not self.context.repo: + self.context.repo = ProjectRepo(self.context.git_repo) + return self.context.repo @property def prompt_schema(self): diff --git a/metagpt/actions/action_graph.py b/metagpt/actions/action_graph.py new file mode 100644 index 000000000..893bc6d4c --- /dev/null +++ b/metagpt/actions/action_graph.py @@ -0,0 +1,49 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/1/30 13:52 +@Author : alexanderwu +@File : action_graph.py +""" +from __future__ import annotations + +# from metagpt.actions.action_node import ActionNode + + +class ActionGraph: + """ActionGraph: a directed graph to represent the dependency between actions.""" + + def __init__(self): + self.nodes = {} + self.edges = {} + self.execution_order = [] + + def add_node(self, node): + """Add a node to the graph""" + self.nodes[node.key] = node + + def add_edge(self, from_node: "ActionNode", to_node: "ActionNode"): + """Add an edge to the graph""" + if from_node.key not in self.edges: + self.edges[from_node.key] = [] + self.edges[from_node.key].append(to_node.key) + from_node.add_next(to_node) + to_node.add_prev(from_node) + + def topological_sort(self): + """Topological sort the graph""" + visited = set() + stack = [] + + def visit(k): + if k not in visited: + visited.add(k) + if k in self.edges: + for next_node in self.edges[k]: + visit(next_node) + stack.insert(0, k) + + for key in self.nodes: + visit(key) + + self.execution_order = stack diff --git a/metagpt/actions/action_node.py b/metagpt/actions/action_node.py index b4d8c32df..09da4a988 100644 --- a/metagpt/actions/action_node.py +++ b/metagpt/actions/action_node.py @@ -9,10 +9,11 @@ NOTE: You should use typing.List instead of list to do type annotation. Because we can use typing to extract the type of the node, but we cannot use built-in list to extract. """ import json +import typing from enum import Enum from typing import Any, Dict, List, Optional, Tuple, Type, Union -from pydantic import BaseModel, create_model, model_validator +from pydantic import BaseModel, Field, create_model, model_validator from tenacity import retry, stop_after_attempt, wait_random_exponential from metagpt.actions.action_outcls_registry import register_action_outcls @@ -39,7 +40,6 @@ TAG = "CONTENT" LANGUAGE_CONSTRAINT = "Language: Please use the same language as Human INPUT." FORMAT_CONSTRAINT = f"Format: output wrapped inside [{TAG}][/{TAG}] like format example, nothing else." - SIMPLE_TEMPLATE = """ ## context {context} @@ -61,7 +61,7 @@ Follow instructions of nodes, generate output and make sure it follows the forma REVIEW_TEMPLATE = """ ## context -Compare the keys of nodes_output and the corresponding requirements one by one. If a key that does not match the requirement is found, provide the comment content on how to modify it. No output is required for matching keys. +Compare the key's value of nodes_output and the corresponding requirements one by one. If a key's value that does not match the requirement is found, provide the comment content on how to modify it. No output is required for matching keys. ### nodes_output {nodes_output} @@ -86,7 +86,7 @@ Compare the keys of nodes_output and the corresponding requirements one by one. {constraint} ## action -generate output and make sure it follows the format example. +Follow format example's {prompt_schema} format, generate output and make sure it follows the format example. """ REVISE_TEMPLATE = """ @@ -108,7 +108,7 @@ change the nodes_output key's value to meet its comment and no need to add extra {constraint} ## action -generate output and make sure it follows the format example. +Follow format example's {prompt_schema} format, generate output and make sure it follows the format example. """ @@ -131,6 +131,8 @@ class ActionNode: # Action Input key: str # Product Requirement / File list / Code + func: typing.Callable # 与节点相关联的函数或LLM调用 + params: Dict[str, Type] # 输入参数的字典,键为参数名,值为参数类型 expected_type: Type # such as str / int / float etc. # context: str # everything in the history. instruction: str # the instructions should be followed. @@ -140,6 +142,10 @@ class ActionNode: content: str instruct_content: BaseModel + # For ActionGraph + prevs: List["ActionNode"] # previous nodes + nexts: List["ActionNode"] # next nodes + def __init__( self, key: str, @@ -157,6 +163,8 @@ class ActionNode: self.content = content self.children = children if children is not None else {} self.schema = schema + self.prevs = [] + self.nexts = [] def __str__(self): return ( @@ -167,6 +175,14 @@ class ActionNode: def __repr__(self): return self.__str__() + def add_prev(self, node: "ActionNode"): + """增加前置ActionNode""" + self.prevs.append(node) + + def add_next(self, node: "ActionNode"): + """增加后置ActionNode""" + self.nexts.append(node) + def add_child(self, node: "ActionNode"): """增加子ActionNode""" self.children[node.key] = node @@ -186,25 +202,38 @@ class ActionNode: obj.add_children(nodes) return obj - def get_children_mapping(self, exclude=None) -> Dict[str, Tuple[Type, Any]]: - """获得子ActionNode的字典,以key索引""" + def _get_children_mapping(self, exclude=None) -> Dict[str, Any]: + """获得子ActionNode的字典,以key索引,支持多级结构。""" exclude = exclude or [] - return {k: (v.expected_type, ...) for k, v in self.children.items() if k not in exclude} - def get_self_mapping(self) -> Dict[str, Tuple[Type, Any]]: + def _get_mapping(node: "ActionNode") -> Dict[str, Any]: + mapping = {} + for key, child in node.children.items(): + if key in exclude: + continue + # 对于嵌套的子节点,递归调用 _get_mapping + if child.children: + mapping[key] = _get_mapping(child) + else: + mapping[key] = (child.expected_type, Field(default=child.example, description=child.instruction)) + return mapping + + return _get_mapping(self) + + def _get_self_mapping(self) -> Dict[str, Tuple[Type, Any]]: """get self key: type mapping""" return {self.key: (self.expected_type, ...)} def get_mapping(self, mode="children", exclude=None) -> Dict[str, Tuple[Type, Any]]: """get key: type mapping under mode""" if mode == "children" or (mode == "auto" and self.children): - return self.get_children_mapping(exclude=exclude) - return {} if exclude and self.key in exclude else self.get_self_mapping() + return self._get_children_mapping(exclude=exclude) + return {} if exclude and self.key in exclude else self._get_self_mapping() @classmethod @register_action_outcls def create_model_class(cls, class_name: str, mapping: Dict[str, Tuple[Type, Any]]): - """基于pydantic v1的模型动态生成,用来检验结果类型正确性""" + """基于pydantic v2的模型动态生成,用来检验结果类型正确性""" def check_fields(cls, values): required_fields = set(mapping.keys()) @@ -219,7 +248,17 @@ class ActionNode: validators = {"check_missing_fields_validator": model_validator(mode="before")(check_fields)} - new_class = create_model(class_name, __validators__=validators, **mapping) + new_fields = {} + for field_name, field_value in mapping.items(): + if isinstance(field_value, dict): + # 对于嵌套结构,递归创建模型类 + nested_class_name = f"{class_name}_{field_name}" + nested_class = cls.create_model_class(nested_class_name, field_value) + new_fields[field_name] = (nested_class, ...) + else: + new_fields[field_name] = field_value + + new_class = create_model(class_name, __validators__=validators, **new_fields) return new_class def create_class(self, mode: str = "auto", class_name: str = None, exclude=None): @@ -227,39 +266,48 @@ class ActionNode: mapping = self.get_mapping(mode=mode, exclude=exclude) return self.create_model_class(class_name, mapping) - def create_children_class(self, exclude=None): + def _create_children_class(self, exclude=None): """使用object内有的字段直接生成model_class""" class_name = f"{self.key}_AN" - mapping = self.get_children_mapping(exclude=exclude) + mapping = self._get_children_mapping(exclude=exclude) return self.create_model_class(class_name, mapping) def to_dict(self, format_func=None, mode="auto", exclude=None) -> Dict: """将当前节点与子节点都按照node: format的格式组织成字典""" + nodes = self._to_dict(format_func=format_func, mode=mode, exclude=exclude) + if not isinstance(nodes, dict): + nodes = {self.key: nodes} + return nodes - # 如果没有提供格式化函数,使用默认的格式化方式 + def _to_dict(self, format_func=None, mode="auto", exclude=None) -> Dict: + """将当前节点与子节点都按照node: format的格式组织成字典""" + + # 如果没有提供格式化函数,则使用默认的格式化函数 if format_func is None: - format_func = lambda node: f"{node.instruction}" + format_func = lambda node: node.instruction # 使用提供的格式化函数来格式化当前节点的值 formatted_value = format_func(self) # 创建当前节点的键值对 - if mode == "children" or (mode == "auto" and self.children): - node_dict = {} + if (mode == "children" or mode == "auto") and self.children: + node_value = {} else: - node_dict = {self.key: formatted_value} + node_value = formatted_value if mode == "root": - return node_dict + return {self.key: node_value} - # 遍历子节点并递归调用 to_dict 方法 + # 递归处理子节点 exclude = exclude or [] - for _, child_node in self.children.items(): - if child_node.key in exclude: + for child_key, child_node in self.children.items(): + if child_key in exclude: continue - node_dict.update(child_node.to_dict(format_func)) + # 递归调用 to_dict 方法并更新节点字典 + child_dict = child_node._to_dict(format_func, mode, exclude) + node_value[child_key] = child_dict - return node_dict + return node_value def update_instruct_content(self, incre_data: dict[str, Any]): assert self.instruct_content @@ -328,6 +376,17 @@ class ActionNode: if schema == "raw": return context + "\n\n## Actions\n" + LANGUAGE_CONSTRAINT + "\n" + self.instruction + ### 直接使用 pydantic BaseModel 生成 instruction 与 example,仅限 JSON + # child_class = self._create_children_class() + # node_schema = child_class.model_json_schema() + # defaults = { + # k: str(v) + # for k, v in child_class.model_fields.items() + # if k not in exclude + # } + # instruction = node_schema + # example = json.dumps(defaults, indent=4) + # FIXME: json instruction会带来格式问题,如:"Project name": "web_2048 # 项目名称使用下划线", # compile example暂时不支持markdown instruction = self.compile_instruction(schema="markdown", mode=mode, exclude=exclude) @@ -354,12 +413,13 @@ class ActionNode: prompt: str, output_class_name: str, output_data_mapping: dict, + images: Optional[Union[str, list[str]]] = None, system_msgs: Optional[list[str]] = None, schema="markdown", # compatible to original format timeout=3, ) -> (str, BaseModel): """Use ActionOutput to wrap the output of aask""" - content = await self.llm.aask(prompt, system_msgs, timeout=timeout) + content = await self.llm.aask(prompt, system_msgs, images=images, timeout=timeout) logger.debug(f"llm raw output:\n{content}") output_class = self.create_model_class(output_class_name, output_data_mapping) @@ -388,13 +448,15 @@ class ActionNode: def set_context(self, context): self.set_recursive("context", context) - async def simple_fill(self, schema, mode, timeout=3, exclude=None): + async def simple_fill(self, schema, mode, images: Optional[Union[str, list[str]]] = None, timeout=3, exclude=None): prompt = self.compile(context=self.context, schema=schema, mode=mode, exclude=exclude) if schema != "raw": mapping = self.get_mapping(mode, exclude=exclude) class_name = f"{self.key}_AN" - content, scontent = await self._aask_v1(prompt, class_name, mapping, schema=schema, timeout=timeout) + content, scontent = await self._aask_v1( + prompt, class_name, mapping, images=images, schema=schema, timeout=timeout + ) self.content = content self.instruct_content = scontent else: @@ -403,7 +465,17 @@ class ActionNode: return self - async def fill(self, context, llm, schema="json", mode="auto", strgy="simple", timeout=3, exclude=[]): + async def fill( + self, + context, + llm, + schema="json", + mode="auto", + strgy="simple", + images: Optional[Union[str, list[str]]] = None, + timeout=3, + exclude=[], + ): """Fill the node(s) with mode. :param context: Everything we should know when filling node. @@ -419,6 +491,7 @@ class ActionNode: :param strgy: simple/complex - simple: run only once - complex: run each node + :param images: the list of image url or base64 for gpt4-v :param timeout: Timeout for llm invocation. :param exclude: The keys of ActionNode to exclude. :return: self @@ -429,16 +502,16 @@ class ActionNode: schema = self.schema if strgy == "simple": - return await self.simple_fill(schema=schema, mode=mode, timeout=timeout, exclude=exclude) + return await self.simple_fill(schema=schema, mode=mode, images=images, timeout=timeout, exclude=exclude) elif strgy == "complex": # 这里隐式假设了拥有children tmp = {} for _, i in self.children.items(): if exclude and i.key in exclude: continue - child = await i.simple_fill(schema=schema, mode=mode, timeout=timeout, exclude=exclude) + child = await i.simple_fill(schema=schema, mode=mode, images=images, timeout=timeout, exclude=exclude) tmp.update(child.instruct_content.model_dump()) - cls = self.create_children_class() + cls = self._create_children_class() self.instruct_content = cls(**tmp) return self @@ -469,7 +542,10 @@ class ActionNode: return dict() prompt = template.format( - nodes_output=json.dumps(nodes_output, ensure_ascii=False, indent=4), tag=TAG, constraint=FORMAT_CONSTRAINT + nodes_output=json.dumps(nodes_output, ensure_ascii=False), + tag=TAG, + constraint=FORMAT_CONSTRAINT, + prompt_schema="json", ) content = await self.llm.aask(prompt) @@ -563,10 +639,11 @@ class ActionNode: instruction = self.compile_instruction(schema="markdown", mode="auto", exclude=exclude_keys) prompt = template.format( - nodes_output=json.dumps(nodes_output, ensure_ascii=False, indent=4), + nodes_output=json.dumps(nodes_output, ensure_ascii=False), example=example, instruction=instruction, constraint=FORMAT_CONSTRAINT, + prompt_schema="json", ) # step2, use `_aask_v1` to get revise structure result @@ -612,3 +689,32 @@ class ActionNode: self.update_instruct_content(revise_contents) return revise_contents + + @classmethod + def from_pydantic(cls, model: Type[BaseModel], key: str = None): + """ + Creates an ActionNode tree from a Pydantic model. + + Args: + model (Type[BaseModel]): The Pydantic model to convert. + + Returns: + ActionNode: The root node of the created ActionNode tree. + """ + key = key or model.__name__ + root_node = cls(key=key, expected_type=Type[model], instruction="", example="") + + for field_name, field_info in model.model_fields.items(): + field_type = field_info.annotation + description = field_info.description + default = field_info.default + + # Recursively handle nested models if needed + if not isinstance(field_type, typing._GenericAlias) and issubclass(field_type, BaseModel): + child_node = cls.from_pydantic(field_type, key=field_name) + else: + child_node = cls(key=field_name, expected_type=field_type, instruction=description, example=default) + + root_node.add_child(child_node) + + return root_node diff --git a/metagpt/actions/ci/ask_review.py b/metagpt/actions/ci/ask_review.py new file mode 100644 index 000000000..041011e80 --- /dev/null +++ b/metagpt/actions/ci/ask_review.py @@ -0,0 +1,62 @@ +from __future__ import annotations + +from typing import Tuple + +from metagpt.actions import Action +from metagpt.logs import logger +from metagpt.schema import Message, Plan + + +class ReviewConst: + TASK_REVIEW_TRIGGER = "task" + CODE_REVIEW_TRIGGER = "code" + CONTINUE_WORDS = ["confirm", "continue", "c", "yes", "y"] + CHANGE_WORDS = ["change"] + EXIT_WORDS = ["exit"] + TASK_REVIEW_INSTRUCTION = ( + f"If you want to change, add, delete a task or merge tasks in the plan, say '{CHANGE_WORDS[0]} task task_id or current task, ... (things to change)' " + f"If you confirm the output from the current task and wish to continue, type: {CONTINUE_WORDS[0]}" + ) + CODE_REVIEW_INSTRUCTION = ( + f"If you want the codes to be rewritten, say '{CHANGE_WORDS[0]} ... (your change advice)' " + f"If you want to leave it as is, type: {CONTINUE_WORDS[0]} or {CONTINUE_WORDS[1]}" + ) + EXIT_INSTRUCTION = f"If you want to terminate the process, type: {EXIT_WORDS[0]}" + + +class AskReview(Action): + async def run( + self, context: list[Message] = [], plan: Plan = None, trigger: str = ReviewConst.TASK_REVIEW_TRIGGER + ) -> Tuple[str, bool]: + if plan: + logger.info("Current overall plan:") + logger.info( + "\n".join( + [f"{task.task_id}: {task.instruction}, is_finished: {task.is_finished}" for task in plan.tasks] + ) + ) + + logger.info("Most recent context:") + latest_action = context[-1].cause_by if context and context[-1].cause_by else "" + review_instruction = ( + ReviewConst.TASK_REVIEW_INSTRUCTION + if trigger == ReviewConst.TASK_REVIEW_TRIGGER + else ReviewConst.CODE_REVIEW_INSTRUCTION + ) + prompt = ( + f"This is a <{trigger}> review. Please review output from {latest_action}\n" + f"{review_instruction}\n" + f"{ReviewConst.EXIT_INSTRUCTION}\n" + "Please type your review below:\n" + ) + + rsp = input(prompt) + + if rsp.lower() in ReviewConst.EXIT_WORDS: + exit() + + # Confirmation can be one of "confirm", "continue", "c", "yes", "y" exactly, or sentences containing "confirm". + # One could say "confirm this task, but change the next task to ..." + confirmed = rsp.lower() in ReviewConst.CONTINUE_WORDS or ReviewConst.CONTINUE_WORDS[0] in rsp.lower() + + return rsp, confirmed diff --git a/metagpt/actions/ci/debug_code.py b/metagpt/actions/ci/debug_code.py new file mode 100644 index 000000000..4a6617dc6 --- /dev/null +++ b/metagpt/actions/ci/debug_code.py @@ -0,0 +1,109 @@ +from __future__ import annotations + +from metagpt.actions.ci.write_analysis_code import BaseWriteAnalysisCode +from metagpt.logs import logger +from metagpt.schema import Message +from metagpt.utils.common import create_func_call_config + +DEBUG_REFLECTION_EXAMPLE = ''' +Example 1: +[previous impl]: +```python +def add(a: int, b: int) -> int: + """ + Given integers a and b, return the total value of a and b. + """ + return a - b +``` + +[runtime Error]: +Tested passed: + +Tests failed: +assert add(1, 2) == 3 # output: -1 +assert add(1, 2) == 4 # output: -1 + +[reflection on previous impl]: +The implementation failed the test cases where the input integers are 1 and 2. The issue arises because the code does not add the two integers together, but instead subtracts the second integer from the first. To fix this issue, we should change the operator from `-` to `+` in the return statement. This will ensure that the function returns the correct output for the given input. + +[improved impl]: +```python +def add(a: int, b: int) -> int: + """ + Given integers a and b, return the total value of a and b. + """ + return a + b +``` +''' + +REFLECTION_PROMPT = """ +Here is an example for you. +{debug_example} +[context] +{context} + +[previous impl] +{code} +[runtime Error] +{runtime_result} + +Analysis the error step by step, provide me improve method and code. Remember to follow [context] requirement. Don't forget write code for steps behind the error step. +[reflection on previous impl]: +xxx +""" + +CODE_REFLECTION = { + "name": "execute_reflection_code", + "description": "Execute reflection code.", + "parameters": { + "type": "object", + "properties": { + "reflection": { + "type": "string", + "description": "Reflection on previous impl.", + }, + "improved_impl": { + "type": "string", + "description": "Refined code after reflection.", + }, + }, + "required": ["reflection", "improved_impl"], + }, +} + + +class DebugCode(BaseWriteAnalysisCode): + async def run( + self, + context: list[Message] = None, + code: str = "", + runtime_result: str = "", + ) -> str: + """ + Execute the debugging process based on the provided context, code, and runtime_result. + + Args: + context (list[Message]): A list of Message objects representing the context. + code (str): The code to be debugged. + runtime_result (str): The result of the code execution. + + Returns: + str: The improved implementation based on the debugging process. + """ + + info = [] + reflection_prompt = REFLECTION_PROMPT.format( + debug_example=DEBUG_REFLECTION_EXAMPLE, + context=context, + code=code, + runtime_result=runtime_result, + ) + system_prompt = "You are an AI Python assistant. You will be given your previous implementation code of a task, runtime error results, and a hint to change the implementation appropriately. Write your full implementation " + info.append(Message(role="system", content=system_prompt)) + info.append(Message(role="user", content=reflection_prompt)) + + tool_config = create_func_call_config(CODE_REFLECTION) + reflection = await self.llm.aask_code(messages=info, **tool_config) + logger.info(f"reflection is {reflection}") + + return {"code": reflection["improved_impl"]} diff --git a/metagpt/actions/ci/execute_nb_code.py b/metagpt/actions/ci/execute_nb_code.py new file mode 100644 index 000000000..0ff00de8f --- /dev/null +++ b/metagpt/actions/ci/execute_nb_code.py @@ -0,0 +1,249 @@ +# -*- encoding: utf-8 -*- +""" +@Date : 2023/11/17 14:22:15 +@Author : orange-crow +@File : execute_nb_code.py +""" +from __future__ import annotations + +import asyncio +import base64 +import re +import traceback +from typing import Literal, Tuple + +import nbformat +from nbclient import NotebookClient +from nbclient.exceptions import CellTimeoutError, DeadKernelError +from nbformat import NotebookNode +from nbformat.v4 import new_code_cell, new_markdown_cell, new_output +from rich.box import MINIMAL +from rich.console import Console, Group +from rich.live import Live +from rich.markdown import Markdown +from rich.panel import Panel +from rich.syntax import Syntax + +from metagpt.actions import Action +from metagpt.logs import logger + + +class ExecuteNbCode(Action): + """execute notebook code block, return result to llm, and display it.""" + + nb: NotebookNode + nb_client: NotebookClient + console: Console + interaction: str + timeout: int = 600 + + def __init__( + self, + nb=nbformat.v4.new_notebook(), + timeout=600, + ): + super().__init__( + nb=nb, + nb_client=NotebookClient(nb, timeout=timeout), + timeout=timeout, + console=Console(), + interaction=("ipython" if self.is_ipython() else "terminal"), + ) + + async def build(self): + if self.nb_client.kc is None or not await self.nb_client.kc.is_alive(): + self.nb_client.create_kernel_manager() + self.nb_client.start_new_kernel() + self.nb_client.start_new_kernel_client() + + async def terminate(self): + """kill NotebookClient""" + await self.nb_client._async_cleanup_kernel() + + async def reset(self): + """reset NotebookClient""" + await self.terminate() + + # sleep 1s to wait for the kernel to be cleaned up completely + await asyncio.sleep(1) + await self.build() + self.nb_client = NotebookClient(self.nb, timeout=self.timeout) + + def add_code_cell(self, code: str): + self.nb.cells.append(new_code_cell(source=code)) + + def add_markdown_cell(self, markdown: str): + self.nb.cells.append(new_markdown_cell(source=markdown)) + + def _display(self, code: str, language: Literal["python", "markdown"] = "python"): + if language == "python": + code = Syntax(code, "python", theme="paraiso-dark", line_numbers=True) + self.console.print(code) + elif language == "markdown": + display_markdown(code) + else: + raise ValueError(f"Only support for python, markdown, but got {language}") + + def add_output_to_cell(self, cell: NotebookNode, output: str): + """add outputs of code execution to notebook cell.""" + if "outputs" not in cell: + cell["outputs"] = [] + else: + cell["outputs"].append(new_output(output_type="stream", name="stdout", text=str(output))) + + def parse_outputs(self, outputs: list[str]) -> str: + """Parses the outputs received from notebook execution.""" + assert isinstance(outputs, list) + parsed_output = "" + + for i, output in enumerate(outputs): + if output["output_type"] == "stream" and not any( + tag in output["text"] + for tag in ["| INFO | metagpt", "| ERROR | metagpt", "| WARNING | metagpt"] + ): + parsed_output += output["text"] + elif output["output_type"] == "display_data": + if "image/png" in output["data"]: + self.show_bytes_figure(output["data"]["image/png"], self.interaction) + else: + logger.info( + f"{i}th output['data'] from nbclient outputs dont have image/png, continue next output ..." + ) + elif output["output_type"] == "execute_result": + parsed_output += output["data"]["text/plain"] + return parsed_output + + def show_bytes_figure(self, image_base64: str, interaction_type: Literal["ipython", None]): + image_bytes = base64.b64decode(image_base64) + if interaction_type == "ipython": + from IPython.display import Image, display + + display(Image(data=image_bytes)) + else: + import io + + from PIL import Image + + image = Image.open(io.BytesIO(image_bytes)) + image.show() + + def is_ipython(self) -> bool: + try: + # 如果在Jupyter Notebook中运行,__file__ 变量不存在 + from IPython import get_ipython + + if get_ipython() is not None and "IPKernelApp" in get_ipython().config: + return True + else: + return False + except NameError: + return False + + async def run_cell(self, cell: NotebookNode, cell_index: int) -> Tuple[bool, str]: + """set timeout for run code. + returns the success or failure of the cell execution, and an optional error message. + """ + try: + await self.nb_client.async_execute_cell(cell, cell_index) + return True, "" + except CellTimeoutError: + assert self.nb_client.km is not None + await self.nb_client.km.interrupt_kernel() + await asyncio.sleep(1) + error_msg = "Cell execution timed out: Execution exceeded the time limit and was stopped; consider optimizing your code for better performance." + return False, error_msg + except DeadKernelError: + await self.reset() + return False, "DeadKernelError" + except Exception: + return False, f"{traceback.format_exc()}" + + async def run(self, code: str, language: Literal["python", "markdown"] = "python") -> Tuple[str, bool]: + """ + return the output of code execution, and a success indicator (bool) of code execution. + """ + self._display(code, language) + + if language == "python": + # add code to the notebook + self.add_code_cell(code=code) + + # build code executor + await self.build() + + # run code + cell_index = len(self.nb.cells) - 1 + success, error_message = await self.run_cell(self.nb.cells[-1], cell_index) + + if not success: + return truncate(remove_escape_and_color_codes(error_message), is_success=success) + + # code success + outputs = self.parse_outputs(self.nb.cells[-1].outputs) + outputs, success = truncate(remove_escape_and_color_codes(outputs), is_success=success) + + if "!pip" in outputs: + success = False + + return outputs, success + + elif language == "markdown": + # add markdown content to markdown cell in a notebook. + self.add_markdown_cell(code) + # return True, beacuse there is no execution failure for markdown cell. + return code, True + else: + raise ValueError(f"Only support for language: python, markdown, but got {language}, ") + + +def truncate(result: str, keep_len: int = 2000, is_success: bool = True): + """对于超出keep_len个字符的result: 执行失败的代码, 展示result后keep_len个字符; 执行成功的代码, 展示result前keep_len个字符。""" + if is_success: + desc = f"Executed code successfully. Truncated to show only first {keep_len} characters\n" + else: + desc = f"Executed code failed, please reflect the cause of bug and then debug. Truncated to show only last {keep_len} characters\n" + + if result.strip().startswith(" keep_len: + result = result[-keep_len:] if not is_success else result[:keep_len] + return desc + result, is_success + + return result, is_success + + +def remove_escape_and_color_codes(input_str: str): + # 使用正则表达式去除转义字符和颜色代码 + pattern = re.compile(r"\x1b\[[0-9;]*[mK]") + result = pattern.sub("", input_str) + return result + + +def display_markdown(content: str): + # 使用正则表达式逐个匹配代码块 + matches = re.finditer(r"```(.+?)```", content, re.DOTALL) + start_index = 0 + content_panels = [] + # 逐个打印匹配到的文本和代码 + for match in matches: + text_content = content[start_index : match.start()].strip() + code_content = match.group(0).strip()[3:-3] # Remove triple backticks + + if text_content: + content_panels.append(Panel(Markdown(text_content), box=MINIMAL)) + + if code_content: + content_panels.append(Panel(Markdown(f"```{code_content}"), box=MINIMAL)) + start_index = match.end() + + # 打印剩余文本(如果有) + remaining_text = content[start_index:].strip() + if remaining_text: + content_panels.append(Panel(Markdown(remaining_text), box=MINIMAL)) + + # 在Live模式中显示所有Panel + with Live(auto_refresh=False, console=Console(), vertical_overflow="visible") as live: + live.update(Group(*content_panels)) + live.refresh() diff --git a/metagpt/actions/ci/ml_action.py b/metagpt/actions/ci/ml_action.py new file mode 100644 index 000000000..e18d0fd20 --- /dev/null +++ b/metagpt/actions/ci/ml_action.py @@ -0,0 +1,70 @@ +from __future__ import annotations + +from typing import Tuple + +from metagpt.actions import Action +from metagpt.actions.ci.write_analysis_code import WriteCodeWithTools +from metagpt.prompts.ci.ml_action import ( + ML_GENERATE_CODE_PROMPT, + ML_TOOL_USAGE_PROMPT, + PRINT_DATA_COLUMNS, + UPDATE_DATA_COLUMNS, +) +from metagpt.prompts.ci.write_analysis_code import CODE_GENERATOR_WITH_TOOLS +from metagpt.schema import Message, Plan +from metagpt.utils.common import create_func_call_config, remove_comments + + +class WriteCodeWithToolsML(WriteCodeWithTools): + async def run( + self, + context: list[Message], + plan: Plan = None, + column_info: str = "", + **kwargs, + ) -> Tuple[list[Message], str]: + # prepare tool schemas and tool-type-specific instruction + tool_schemas, tool_type_usage_prompt = await self._prepare_tools(plan=plan) + + # ML-specific variables to be used in prompt + finished_tasks = plan.get_finished_tasks() + code_context = [remove_comments(task.code) for task in finished_tasks] + code_context = "\n\n".join(code_context) + + # prepare prompt depending on tool availability & LLM call + if tool_schemas: + prompt = ML_TOOL_USAGE_PROMPT.format( + user_requirement=plan.goal, + history_code=code_context, + current_task=plan.current_task.instruction, + column_info=column_info, + tool_type_usage_prompt=tool_type_usage_prompt, + tool_schemas=tool_schemas, + ) + + else: + prompt = ML_GENERATE_CODE_PROMPT.format( + user_requirement=plan.goal, + history_code=code_context, + current_task=plan.current_task.instruction, + column_info=column_info, + tool_type_usage_prompt=tool_type_usage_prompt, + ) + tool_config = create_func_call_config(CODE_GENERATOR_WITH_TOOLS) + rsp = await self.llm.aask_code(prompt, **tool_config) + + # Extra output to be used for potential debugging + context = [Message(content=prompt, role="user")] + + return context, rsp + + +class UpdateDataColumns(Action): + async def run(self, plan: Plan = None) -> dict: + finished_tasks = plan.get_finished_tasks() + code_context = [remove_comments(task.code) for task in finished_tasks] + code_context = "\n\n".join(code_context) + prompt = UPDATE_DATA_COLUMNS.format(history_code=code_context) + tool_config = create_func_call_config(PRINT_DATA_COLUMNS) + rsp = await self.llm.aask_code(prompt, **tool_config) + return rsp diff --git a/metagpt/actions/ci/write_analysis_code.py b/metagpt/actions/ci/write_analysis_code.py new file mode 100644 index 000000000..421eb0804 --- /dev/null +++ b/metagpt/actions/ci/write_analysis_code.py @@ -0,0 +1,155 @@ +# -*- encoding: utf-8 -*- +""" +@Date : 2023/11/20 13:19:39 +@Author : orange-crow +@File : write_analysis_code.py +""" +from __future__ import annotations + +from typing import Tuple + +from metagpt.actions import Action +from metagpt.logs import logger +from metagpt.prompts.ci.write_analysis_code import ( + CODE_GENERATOR_WITH_TOOLS, + SELECT_FUNCTION_TOOLS, + TOOL_RECOMMENDATION_PROMPT, + TOOL_USAGE_PROMPT, +) +from metagpt.schema import Message, Plan, SystemMessage +from metagpt.tools import TOOL_REGISTRY +from metagpt.tools.tool_registry import validate_tool_names +from metagpt.utils.common import create_func_call_config + + +class BaseWriteAnalysisCode(Action): + DEFAULT_SYSTEM_MSG: str = """You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**""" # prompt reference: https://github.com/KillianLucas/open-interpreter/blob/v0.1.4/interpreter/system_message.txt + # REUSE_CODE_INSTRUCTION = """ATTENTION: DONT include codes from previous tasks in your current code block, include new codes only, DONT repeat codes!""" + + def insert_system_message(self, context: list[Message], system_msg: str = None): + system_msg = system_msg or self.DEFAULT_SYSTEM_MSG + context.insert(0, SystemMessage(content=system_msg)) if context[0].role != "system" else None + return context + + async def run(self, context: list[Message], plan: Plan = None) -> dict: + """Run of a code writing action, used in data analysis or modeling + + Args: + context (list[Message]): Action output history, source action denoted by Message.cause_by + plan (Plan, optional): Overall plan. Defaults to None. + + Returns: + dict: code result in the format of {"code": "print('hello world')", "language": "python"} + """ + raise NotImplementedError + + +class WriteCodeWithoutTools(BaseWriteAnalysisCode): + """Ask LLM to generate codes purely by itself without local user-defined tools""" + + async def run(self, context: list[Message], plan: Plan = None, system_msg: str = None, **kwargs) -> dict: + messages = self.insert_system_message(context, system_msg) + rsp = await self.llm.aask_code(messages, **kwargs) + return rsp + + +class WriteCodeWithTools(BaseWriteAnalysisCode): + """Write code with help of local available tools. Choose tools first, then generate code to use the tools""" + + # selected tools to choose from, listed by their names. An empty list means selection from all tools. + selected_tools: list[str] = [] + + def _get_tools_by_type(self, tool_type: str) -> dict: + """ + Retreive tools by tool type from registry, but filtered by pre-selected tool list + + Args: + tool_type (str): Tool type to retrieve from the registry + + Returns: + dict: A dict of tool name to Tool object, representing available tools under the type + """ + candidate_tools = TOOL_REGISTRY.get_tools_by_type(tool_type) + if self.selected_tools: + candidate_tool_names = set(self.selected_tools) & candidate_tools.keys() + candidate_tools = {tool_name: candidate_tools[tool_name] for tool_name in candidate_tool_names} + return candidate_tools + + async def _recommend_tool( + self, + task: str, + available_tools: dict, + ) -> dict: + """ + Recommend tools for the specified task. + + Args: + task (str): the task to recommend tools for + available_tools (dict): the available tools description + + Returns: + dict: schemas of recommended tools for the specified task + """ + prompt = TOOL_RECOMMENDATION_PROMPT.format( + current_task=task, + available_tools=available_tools, + ) + tool_config = create_func_call_config(SELECT_FUNCTION_TOOLS) + rsp = await self.llm.aask_code(prompt, **tool_config) + recommend_tools = rsp["recommend_tools"] + logger.info(f"Recommended tools: \n{recommend_tools}") + + # Parses and validates the recommended tools, for LLM might hallucinate and recommend non-existing tools + valid_tools = validate_tool_names(recommend_tools, return_tool_object=True) + + tool_schemas = {tool.name: tool.schemas for tool in valid_tools} + + return tool_schemas + + async def _prepare_tools(self, plan: Plan) -> Tuple[dict, str]: + """Prepare tool schemas and usage instructions according to current task + + Args: + plan (Plan): The overall plan containing task information. + + Returns: + Tuple[dict, str]: A tool schemas ({tool_name: tool_schema_dict}) and a usage prompt for the type of tools selected + """ + # find tool type from task type through exact match, can extend to retrieval in the future + tool_type = plan.current_task.task_type + + # prepare tool-type-specific instruction + tool_type_usage_prompt = ( + TOOL_REGISTRY.get_tool_type(tool_type).usage_prompt if TOOL_REGISTRY.has_tool_type(tool_type) else "" + ) + + # prepare schemas of available tools + tool_schemas = {} + available_tools = self._get_tools_by_type(tool_type) + if available_tools: + available_tools = {tool_name: tool.schemas["description"] for tool_name, tool in available_tools.items()} + tool_schemas = await self._recommend_tool(plan.current_task.instruction, available_tools) + + return tool_schemas, tool_type_usage_prompt + + async def run( + self, + context: list[Message], + plan: Plan, + **kwargs, + ) -> str: + # prepare tool schemas and tool-type-specific instruction + tool_schemas, tool_type_usage_prompt = await self._prepare_tools(plan=plan) + + # form a complete tool usage instruction and include it as a message in context + tools_instruction = TOOL_USAGE_PROMPT.format( + tool_schemas=tool_schemas, tool_type_usage_prompt=tool_type_usage_prompt + ) + context.append(Message(content=tools_instruction, role="user")) + + # prepare prompt & LLM call + prompt = self.insert_system_message(context) + tool_config = create_func_call_config(CODE_GENERATOR_WITH_TOOLS) + rsp = await self.llm.aask_code(prompt, **tool_config) + + return rsp diff --git a/metagpt/actions/ci/write_plan.py b/metagpt/actions/ci/write_plan.py new file mode 100644 index 000000000..dd9363260 --- /dev/null +++ b/metagpt/actions/ci/write_plan.py @@ -0,0 +1,116 @@ +# -*- encoding: utf-8 -*- +""" +@Date : 2023/11/20 11:24:03 +@Author : orange-crow +@File : plan.py +""" +from __future__ import annotations + +import json +from copy import deepcopy +from typing import Tuple + +from metagpt.actions import Action +from metagpt.logs import logger +from metagpt.prompts.ci.write_analysis_code import ( + ASSIGN_TASK_TYPE_CONFIG, + ASSIGN_TASK_TYPE_PROMPT, +) +from metagpt.schema import Message, Plan, Task +from metagpt.tools import TOOL_REGISTRY +from metagpt.utils.common import CodeParser, create_func_call_config + + +class WritePlan(Action): + PROMPT_TEMPLATE: str = """ + # Context: + __context__ + # Task: + Based on the context, write a plan or modify an existing plan of what you should do to achieve the goal. A plan consists of one to __max_tasks__ tasks. + If you are modifying an existing plan, carefully follow the instruction, don't make unnecessary changes. Give the whole plan unless instructed to modify only one task of the plan. + If you encounter errors on the current task, revise and output the current single task only. + Output a list of jsons following the format: + ```json + [ + { + "task_id": str = "unique identifier for a task in plan, can be an ordinal", + "dependent_task_ids": list[str] = "ids of tasks prerequisite to this task", + "instruction": "what you should do in this task, one short phrase or sentence", + }, + ... + ] + ``` + """ + + async def assign_task_type(self, tasks: list[dict]) -> str: + """Assign task type to each task in tasks + + Args: + tasks (list[dict]): tasks to be assigned task type + + Returns: + str: tasks with task type assigned in a json string + """ + task_info = "\n".join([f"Task {task['task_id']}: {task['instruction']}" for task in tasks]) + task_type_desc = "\n".join( + [f"- **{tool_type.name}**: {tool_type.desc}" for tool_type in TOOL_REGISTRY.get_tool_types().values()] + ) # task type are binded with tool type now, should be improved in the future + prompt = ASSIGN_TASK_TYPE_PROMPT.format( + task_info=task_info, task_type_desc=task_type_desc + ) # task types are set to be the same as tool types, for now + tool_config = create_func_call_config(ASSIGN_TASK_TYPE_CONFIG) + rsp = await self.llm.aask_code(prompt, **tool_config) + task_type_list = rsp["task_type"] + logger.info(f"assigned task types: {task_type_list}") + for task, task_type in zip(tasks, task_type_list): + task["task_type"] = task_type + return json.dumps(tasks) + + async def run(self, context: list[Message], max_tasks: int = 5, use_tools: bool = False) -> str: + prompt = ( + self.PROMPT_TEMPLATE.replace("__context__", "\n".join([str(ct) for ct in context])) + # .replace("__current_plan__", current_plan) + .replace("__max_tasks__", str(max_tasks)) + ) + rsp = await self._aask(prompt) + rsp = CodeParser.parse_code(block=None, text=rsp) + if use_tools: + rsp = await self.assign_task_type(json.loads(rsp)) + return rsp + + +def rsp_to_tasks(rsp: str) -> list[Task]: + rsp = json.loads(rsp) + tasks = [Task(**task_config) for task_config in rsp] + return tasks + + +def update_plan_from_rsp(rsp: str, current_plan: Plan): + tasks = rsp_to_tasks(rsp) + if len(tasks) == 1 or tasks[0].dependent_task_ids: + if tasks[0].dependent_task_ids and len(tasks) > 1: + # tasks[0].dependent_task_ids means the generated tasks are not a complete plan + # for they depend on tasks in the current plan, in this case, we only support updating one task each time + logger.warning( + "Current plan will take only the first generated task if the generated tasks are not a complete plan" + ) + # handle a single task + if current_plan.has_task_id(tasks[0].task_id): + # replace an existing task + current_plan.replace_task(tasks[0]) + else: + # append one task + current_plan.append_task(tasks[0]) + + else: + # add tasks in general + current_plan.add_tasks(tasks) + + +def precheck_update_plan_from_rsp(rsp: str, current_plan: Plan) -> Tuple[bool, str]: + temp_plan = deepcopy(current_plan) + try: + update_plan_from_rsp(rsp, temp_plan) + return True, "" + except Exception as e: + return False, e diff --git a/metagpt/actions/debug_error.py b/metagpt/actions/debug_error.py index f491fdd55..5ed31bed8 100644 --- a/metagpt/actions/debug_error.py +++ b/metagpt/actions/debug_error.py @@ -49,7 +49,7 @@ class DebugError(Action): i_context: RunCodeContext = Field(default_factory=RunCodeContext) async def run(self, *args, **kwargs) -> str: - output_doc = await self.project_repo.test_outputs.get(filename=self.i_context.output_filename) + output_doc = await self.repo.test_outputs.get(filename=self.i_context.output_filename) if not output_doc: return "" output_detail = RunCodeResult.loads(output_doc.content) @@ -59,12 +59,12 @@ class DebugError(Action): return "" logger.info(f"Debug and rewrite {self.i_context.test_filename}") - code_doc = await self.project_repo.with_src_path(self.context.src_workspace).srcs.get( + code_doc = await self.repo.with_src_path(self.context.src_workspace).srcs.get( filename=self.i_context.code_filename ) if not code_doc: return "" - test_doc = await self.project_repo.tests.get(filename=self.i_context.test_filename) + test_doc = await self.repo.tests.get(filename=self.i_context.test_filename) if not test_doc: return "" prompt = PROMPT_TEMPLATE.format(code=code_doc.content, test_code=test_doc.content, logs=output_detail.stderr) diff --git a/metagpt/actions/design_api.py b/metagpt/actions/design_api.py index 04c580226..e5f038c7c 100644 --- a/metagpt/actions/design_api.py +++ b/metagpt/actions/design_api.py @@ -14,7 +14,14 @@ from pathlib import Path from typing import Optional from metagpt.actions import Action, ActionOutput -from metagpt.actions.design_api_an import DESIGN_API_NODE +from metagpt.actions.design_api_an import ( + DATA_STRUCTURES_AND_INTERFACES, + DESIGN_API_NODE, + PROGRAM_CALL_FLOW, + REFINED_DATA_STRUCTURES_AND_INTERFACES, + REFINED_DESIGN_NODE, + REFINED_PROGRAM_CALL_FLOW, +) from metagpt.const import DATA_API_DESIGN_FILE_REPO, SEQ_FLOW_FILE_REPO from metagpt.logs import logger from metagpt.schema import Document, Documents, Message @@ -39,11 +46,11 @@ class WriteDesign(Action): ) async def run(self, with_messages: Message, schema: str = None): - # Use `git status` to identify which PRD documents have been modified in the `docs/prds` directory. - changed_prds = self.project_repo.docs.prd.changed_files + # Use `git status` to identify which PRD documents have been modified in the `docs/prd` directory. + changed_prds = self.repo.docs.prd.changed_files # Use `git status` to identify which design documents in the `docs/system_designs` directory have undergone # changes. - changed_system_designs = self.project_repo.docs.system_design.changed_files + changed_system_designs = self.repo.docs.system_design.changed_files # For those PRDs and design documents that have undergone changes, regenerate the design content. changed_files = Documents() @@ -68,46 +75,46 @@ class WriteDesign(Action): async def _merge(self, prd_doc, system_design_doc): context = NEW_REQ_TEMPLATE.format(old_design=system_design_doc.content, context=prd_doc.content) - node = await DESIGN_API_NODE.fill(context=context, llm=self.llm) + node = await REFINED_DESIGN_NODE.fill(context=context, llm=self.llm) system_design_doc.content = node.instruct_content.model_dump_json() return system_design_doc async def _update_system_design(self, filename) -> Document: - prd = await self.project_repo.docs.prd.get(filename) - old_system_design_doc = await self.project_repo.docs.system_design.get(filename) + prd = await self.repo.docs.prd.get(filename) + old_system_design_doc = await self.repo.docs.system_design.get(filename) if not old_system_design_doc: system_design = await self._new_system_design(context=prd.content) - doc = await self.project_repo.docs.system_design.save( + doc = await self.repo.docs.system_design.save( filename=filename, content=system_design.instruct_content.model_dump_json(), dependencies={prd.root_relative_path}, ) else: doc = await self._merge(prd_doc=prd, system_design_doc=old_system_design_doc) - await self.project_repo.docs.system_design.save_doc(doc=doc, dependencies={prd.root_relative_path}) + await self.repo.docs.system_design.save_doc(doc=doc, dependencies={prd.root_relative_path}) await self._save_data_api_design(doc) await self._save_seq_flow(doc) - await self.project_repo.resources.system_design.save_pdf(doc=doc) + await self.repo.resources.system_design.save_pdf(doc=doc) return doc async def _save_data_api_design(self, design_doc): m = json.loads(design_doc.content) - data_api_design = m.get("Data structures and interfaces") + data_api_design = m.get(DATA_STRUCTURES_AND_INTERFACES.key) or m.get(REFINED_DATA_STRUCTURES_AND_INTERFACES.key) if not data_api_design: return - pathname = self.project_repo.workdir / DATA_API_DESIGN_FILE_REPO / Path(design_doc.filename).with_suffix("") + pathname = self.repo.workdir / DATA_API_DESIGN_FILE_REPO / Path(design_doc.filename).with_suffix("") await self._save_mermaid_file(data_api_design, pathname) logger.info(f"Save class view to {str(pathname)}") async def _save_seq_flow(self, design_doc): m = json.loads(design_doc.content) - seq_flow = m.get("Program call flow") + seq_flow = m.get(PROGRAM_CALL_FLOW.key) or m.get(REFINED_PROGRAM_CALL_FLOW.key) if not seq_flow: return - pathname = self.project_repo.workdir / Path(SEQ_FLOW_FILE_REPO) / Path(design_doc.filename).with_suffix("") + pathname = self.repo.workdir / Path(SEQ_FLOW_FILE_REPO) / Path(design_doc.filename).with_suffix("") await self._save_mermaid_file(seq_flow, pathname) logger.info(f"Saving sequence flow to {str(pathname)}") async def _save_mermaid_file(self, data: str, pathname: Path): pathname.parent.mkdir(parents=True, exist_ok=True) - await mermaid_to_file(self.config.mermaid_engine, data, pathname) + await mermaid_to_file(self.config.mermaid.engine, data, pathname) diff --git a/metagpt/actions/design_api_an.py b/metagpt/actions/design_api_an.py index 3737203cf..35b50ef8f 100644 --- a/metagpt/actions/design_api_an.py +++ b/metagpt/actions/design_api_an.py @@ -8,6 +8,7 @@ from typing import List from metagpt.actions.action_node import ActionNode +from metagpt.logs import logger from metagpt.utils.mermaid import MMC1, MMC2 IMPLEMENTATION_APPROACH = ActionNode( @@ -17,6 +18,15 @@ IMPLEMENTATION_APPROACH = ActionNode( example="We will ...", ) +REFINED_IMPLEMENTATION_APPROACH = ActionNode( + key="Refined Implementation Approach", + expected_type=str, + instruction="Update and extend the original implementation approach to reflect the evolving challenges and " + "requirements due to incremental development. Outline the steps involved in the implementation process with the " + "detailed strategies.", + example="We will refine ...", +) + PROJECT_NAME = ActionNode( key="Project name", expected_type=str, instruction="The project name with underline", example="game_2048" ) @@ -28,6 +38,14 @@ FILE_LIST = ActionNode( example=["main.py", "game.py"], ) +REFINED_FILE_LIST = ActionNode( + key="Refined File list", + expected_type=List[str], + instruction="Update and expand the original file list including only relative paths. Up to 2 files can be added." + "Ensure that the refined file list reflects the evolving structure of the project.", + example=["main.py", "game.py", "new_feature.py"], +) + DATA_STRUCTURES_AND_INTERFACES = ActionNode( key="Data structures and interfaces", expected_type=str, @@ -37,6 +55,16 @@ DATA_STRUCTURES_AND_INTERFACES = ActionNode( example=MMC1, ) +REFINED_DATA_STRUCTURES_AND_INTERFACES = ActionNode( + key="Refined Data structures and interfaces", + expected_type=str, + instruction="Update and extend the existing mermaid classDiagram code syntax to incorporate new classes, " + "methods (including __init__), and functions with precise type annotations. Delineate additional " + "relationships between classes, ensuring clarity and adherence to PEP8 standards." + "Retain content that is not related to incremental development but important for consistency and clarity.", + example=MMC1, +) + PROGRAM_CALL_FLOW = ActionNode( key="Program call flow", expected_type=str, @@ -45,6 +73,16 @@ PROGRAM_CALL_FLOW = ActionNode( example=MMC2, ) +REFINED_PROGRAM_CALL_FLOW = ActionNode( + key="Refined Program call flow", + expected_type=str, + instruction="Extend the existing sequenceDiagram code syntax with detailed information, accurately covering the" + "CRUD and initialization of each object. Ensure correct syntax usage and reflect the incremental changes introduced" + "in the classes and API defined above. " + "Retain content that is not related to incremental development but important for consistency and clarity.", + example=MMC2, +) + ANYTHING_UNCLEAR = ActionNode( key="Anything UNCLEAR", expected_type=str, @@ -61,4 +99,24 @@ NODES = [ ANYTHING_UNCLEAR, ] +REFINED_NODES = [ + REFINED_IMPLEMENTATION_APPROACH, + REFINED_FILE_LIST, + REFINED_DATA_STRUCTURES_AND_INTERFACES, + REFINED_PROGRAM_CALL_FLOW, + ANYTHING_UNCLEAR, +] + DESIGN_API_NODE = ActionNode.from_children("DesignAPI", NODES) +REFINED_DESIGN_NODE = ActionNode.from_children("RefinedDesignAPI", REFINED_NODES) + + +def main(): + prompt = DESIGN_API_NODE.compile(context="") + logger.info(prompt) + prompt = REFINED_DESIGN_NODE.compile(context="") + logger.info(prompt) + + +if __name__ == "__main__": + main() diff --git a/metagpt/actions/prepare_documents.py b/metagpt/actions/prepare_documents.py index 56c587cb3..ab069dc11 100644 --- a/metagpt/actions/prepare_documents.py +++ b/metagpt/actions/prepare_documents.py @@ -15,6 +15,7 @@ from metagpt.actions import Action, ActionOutput from metagpt.const import REQUIREMENT_FILENAME from metagpt.utils.file_repository import FileRepository from metagpt.utils.git_repository import GitRepository +from metagpt.utils.project_repo import ProjectRepo class PrepareDocuments(Action): @@ -38,13 +39,14 @@ class PrepareDocuments(Action): shutil.rmtree(path) self.config.project_path = path self.context.git_repo = GitRepository(local_path=path, auto_init=True) + self.context.repo = ProjectRepo(self.context.git_repo) async def run(self, with_messages, **kwargs): """Create and initialize the workspace folder, initialize the Git environment.""" self._init_repo() # Write the newly added requirements from the main parameter idea to `docs/requirement.txt`. - doc = await self.project_repo.docs.save(filename=REQUIREMENT_FILENAME, content=with_messages[0].content) + doc = await self.repo.docs.save(filename=REQUIREMENT_FILENAME, content=with_messages[0].content) # Send a Message notification to the WritePRD action, instructing it to process requirements using - # `docs/requirement.txt` and `docs/prds/`. + # `docs/requirement.txt` and `docs/prd/`. return ActionOutput(content=doc.content, instruct_content=doc) diff --git a/metagpt/actions/project_management.py b/metagpt/actions/project_management.py index 9ada629be..67a614d6f 100644 --- a/metagpt/actions/project_management.py +++ b/metagpt/actions/project_management.py @@ -13,16 +13,16 @@ import json from typing import Optional -from metagpt.actions import ActionOutput from metagpt.actions.action import Action -from metagpt.actions.project_management_an import PM_NODE +from metagpt.actions.action_output import ActionOutput +from metagpt.actions.project_management_an import PM_NODE, REFINED_PM_NODE from metagpt.const import PACKAGE_REQUIREMENTS_FILENAME from metagpt.logs import logger from metagpt.schema import Document, Documents NEW_REQ_TEMPLATE = """ ### Legacy Content -{old_tasks} +{old_task} ### New Requirements {context} @@ -34,8 +34,8 @@ class WriteTasks(Action): i_context: Optional[str] = None async def run(self, with_messages): - changed_system_designs = self.project_repo.docs.system_design.changed_files - changed_tasks = self.project_repo.docs.task.changed_files + changed_system_designs = self.repo.docs.system_design.changed_files + changed_tasks = self.repo.docs.task.changed_files change_files = Documents() # Rewrite the system designs that have undergone changes based on the git head diff under # `docs/system_designs/`. @@ -57,16 +57,14 @@ class WriteTasks(Action): return ActionOutput(content=change_files.model_dump_json(), instruct_content=change_files) async def _update_tasks(self, filename): - system_design_doc = await self.project_repo.docs.system_design.get(filename) - task_doc = await self.project_repo.docs.task.get(filename) + system_design_doc = await self.repo.docs.system_design.get(filename) + task_doc = await self.repo.docs.task.get(filename) if task_doc: task_doc = await self._merge(system_design_doc=system_design_doc, task_doc=task_doc) - await self.project_repo.docs.task.save_doc( - doc=task_doc, dependencies={system_design_doc.root_relative_path} - ) + await self.repo.docs.task.save_doc(doc=task_doc, dependencies={system_design_doc.root_relative_path}) else: rsp = await self._run_new_tasks(context=system_design_doc.content) - task_doc = await self.project_repo.docs.task.save( + task_doc = await self.repo.docs.task.save( filename=filename, content=rsp.instruct_content.model_dump_json(), dependencies={system_design_doc.root_relative_path}, @@ -79,15 +77,15 @@ class WriteTasks(Action): return node async def _merge(self, system_design_doc, task_doc) -> Document: - context = NEW_REQ_TEMPLATE.format(context=system_design_doc.content, old_tasks=task_doc.content) - node = await PM_NODE.fill(context, self.llm, schema=self.prompt_schema) + context = NEW_REQ_TEMPLATE.format(context=system_design_doc.content, old_task=task_doc.content) + node = await REFINED_PM_NODE.fill(context, self.llm, schema=self.prompt_schema) task_doc.content = node.instruct_content.model_dump_json() return task_doc async def _update_requirements(self, doc): m = json.loads(doc.content) - packages = set(m.get("Required Python third-party packages", set())) - requirement_doc = await self.project_repo.get(filename=PACKAGE_REQUIREMENTS_FILENAME) + packages = set(m.get("Required Python packages", set())) + requirement_doc = await self.repo.get(filename=PACKAGE_REQUIREMENTS_FILENAME) if not requirement_doc: requirement_doc = Document(filename=PACKAGE_REQUIREMENTS_FILENAME, root_path=".", content="") lines = requirement_doc.content.splitlines() @@ -95,4 +93,4 @@ class WriteTasks(Action): if pkg == "": continue packages.add(pkg) - await self.project_repo.save(filename=PACKAGE_REQUIREMENTS_FILENAME, content="\n".join(packages)) + await self.repo.save(filename=PACKAGE_REQUIREMENTS_FILENAME, content="\n".join(packages)) diff --git a/metagpt/actions/project_management_an.py b/metagpt/actions/project_management_an.py index 215a67202..379a23384 100644 --- a/metagpt/actions/project_management_an.py +++ b/metagpt/actions/project_management_an.py @@ -35,6 +35,20 @@ LOGIC_ANALYSIS = ActionNode( ], ) +REFINED_LOGIC_ANALYSIS = ActionNode( + key="Refined Logic Analysis", + expected_type=List[List[str]], + instruction="Review and refine the logic analysis by merging the Legacy Content and Incremental Content. " + "Provide a comprehensive list of files with classes/methods/functions to be implemented or modified incrementally. " + "Include dependency analysis, consider potential impacts on existing code, and document necessary imports.", + example=[ + ["game.py", "Contains Game class and ... functions"], + ["main.py", "Contains main function, from game import Game"], + ["new_feature.py", "Introduces NewFeature class and related functions"], + ["utils.py", "Modifies existing utility functions to support incremental changes"], + ], +) + TASK_LIST = ActionNode( key="Task list", expected_type=List[str], @@ -42,6 +56,15 @@ TASK_LIST = ActionNode( example=["game.py", "main.py"], ) +REFINED_TASK_LIST = ActionNode( + key="Refined Task list", + expected_type=List[str], + instruction="Review and refine the combined task list after the merger of Legacy Content and Incremental Content, " + "and consistent with Refined File List. Ensure that tasks are organized in a logical and prioritized order, " + "considering dependencies for a streamlined and efficient development process. ", + example=["new_feature.py", "utils", "game.py", "main.py"], +) + FULL_API_SPEC = ActionNode( key="Full API spec", expected_type=str, @@ -54,9 +77,19 @@ SHARED_KNOWLEDGE = ActionNode( key="Shared Knowledge", expected_type=str, instruction="Detail any shared knowledge, like common utility functions or configuration variables.", - example="'game.py' contains functions shared across the project.", + example="`game.py` contains functions shared across the project.", ) +REFINED_SHARED_KNOWLEDGE = ActionNode( + key="Refined Shared Knowledge", + expected_type=str, + instruction="Update and expand shared knowledge to reflect any new elements introduced. This includes common " + "utility functions, configuration variables for team collaboration. Retain content that is not related to " + "incremental development but important for consistency and clarity.", + example="`new_module.py` enhances shared utility functions for improved code reusability and collaboration.", +) + + ANYTHING_UNCLEAR_PM = ActionNode( key="Anything UNCLEAR", expected_type=str, @@ -74,13 +107,25 @@ NODES = [ ANYTHING_UNCLEAR_PM, ] +REFINED_NODES = [ + REQUIRED_PYTHON_PACKAGES, + REQUIRED_OTHER_LANGUAGE_PACKAGES, + REFINED_LOGIC_ANALYSIS, + REFINED_TASK_LIST, + FULL_API_SPEC, + REFINED_SHARED_KNOWLEDGE, + ANYTHING_UNCLEAR_PM, +] PM_NODE = ActionNode.from_children("PM_NODE", NODES) +REFINED_PM_NODE = ActionNode.from_children("REFINED_PM_NODE", REFINED_NODES) def main(): prompt = PM_NODE.compile(context="") logger.info(prompt) + prompt = REFINED_PM_NODE.compile(context="") + logger.info(prompt) if __name__ == "__main__": diff --git a/metagpt/actions/research.py b/metagpt/actions/research.py index 2755628c9..ce8d8a967 100644 --- a/metagpt/actions/research.py +++ b/metagpt/actions/research.py @@ -3,15 +3,15 @@ from __future__ import annotations import asyncio -from typing import Callable, Optional, Union +from typing import Any, Callable, Optional, Union -from pydantic import Field, parse_obj_as +from pydantic import TypeAdapter, model_validator from metagpt.actions import Action from metagpt.config2 import config from metagpt.logs import logger from metagpt.tools.search_engine import SearchEngine -from metagpt.tools.web_browser_engine import WebBrowserEngine, WebBrowserEngineType +from metagpt.tools.web_browser_engine import WebBrowserEngine from metagpt.utils.common import OutputParser from metagpt.utils.text import generate_prompt_chunk, reduce_message_length @@ -81,10 +81,16 @@ class CollectLinks(Action): name: str = "CollectLinks" i_context: Optional[str] = None desc: str = "Collect links from a search engine." - - search_engine: SearchEngine = Field(default_factory=SearchEngine) + search_func: Optional[Any] = None + search_engine: Optional[SearchEngine] = None rank_func: Optional[Callable[[list[str]], None]] = None + @model_validator(mode="after") + def validate_engine_and_run_func(self): + if self.search_engine is None: + self.search_engine = SearchEngine.from_search_config(self.config.search, proxy=self.config.proxy) + return self + async def run( self, topic: str, @@ -107,7 +113,7 @@ class CollectLinks(Action): keywords = await self._aask(SEARCH_TOPIC_PROMPT, [system_text]) try: keywords = OutputParser.extract_struct(keywords, list) - keywords = parse_obj_as(list[str], keywords) + keywords = TypeAdapter(list[str]).validate_python(keywords) except Exception as e: logger.exception(f"fail to get keywords related to the research topic '{topic}' for {e}") keywords = [topic] @@ -127,13 +133,13 @@ class CollectLinks(Action): if len(remove) == 0: break - model_name = config.get_openai_llm().model + model_name = config.llm.model prompt = reduce_message_length(gen_msg(), model_name, system_text, 4096) logger.debug(prompt) queries = await self._aask(prompt, [system_text]) try: queries = OutputParser.extract_struct(queries, list) - queries = parse_obj_as(list[str], queries) + queries = TypeAdapter(list[str]).validate_python(queries) except Exception as e: logger.exception(f"fail to break down the research question due to {e}") queries = keywords @@ -178,15 +184,17 @@ class WebBrowseAndSummarize(Action): i_context: Optional[str] = None desc: str = "Explore the web and provide summaries of articles and webpages." browse_func: Union[Callable[[list[str]], None], None] = None - web_browser_engine: Optional[WebBrowserEngine] = WebBrowserEngineType.PLAYWRIGHT + web_browser_engine: Optional[WebBrowserEngine] = None - def __init__(self, **kwargs): - super().__init__(**kwargs) - - self.web_browser_engine = WebBrowserEngine( - engine=WebBrowserEngineType.CUSTOM if self.browse_func else WebBrowserEngineType.PLAYWRIGHT, - run_func=self.browse_func, - ) + @model_validator(mode="after") + def validate_engine_and_run_func(self): + if self.web_browser_engine is None: + self.web_browser_engine = WebBrowserEngine.from_browser_config( + self.config.browser, + browse_func=self.browse_func, + proxy=self.config.proxy, + ) + return self async def run( self, diff --git a/metagpt/actions/run_code.py b/metagpt/actions/run_code.py index 072ee8f22..b2c33c19b 100644 --- a/metagpt/actions/run_code.py +++ b/metagpt/actions/run_code.py @@ -16,6 +16,7 @@ class. """ import subprocess +from pathlib import Path from typing import Tuple from pydantic import Field @@ -41,13 +42,13 @@ Determine the ONE file to rewrite in order to fix the error, for example, xyz.py Determine if all of the code works fine, if so write PASS, else FAIL, WRITE ONLY ONE WORD, PASS OR FAIL, IN THIS SECTION ## Send To: -Please write Engineer if the errors are due to problematic development codes, and QaEngineer to problematic test codes, and NoOne if there are no errors, -WRITE ONLY ONE WORD, Engineer OR QaEngineer OR NoOne, IN THIS SECTION. +Please write NoOne if there are no errors, Engineer if the errors are due to problematic development codes, else QaEngineer, +WRITE ONLY ONE WORD, NoOne OR Engineer OR QaEngineer, IN THIS SECTION. --- You should fill in necessary instruction, status, send to, and finally return all content between the --- segment line. """ -CONTEXT = """ +TEMPLATE_CONTEXT = """ ## Development Code File Name {code_file_name} ## Development Code @@ -130,7 +131,7 @@ class RunCode(Action): logger.info(f"{outs=}") logger.info(f"{errs=}") - context = CONTEXT.format( + context = TEMPLATE_CONTEXT.format( code=self.i_context.code, code_file_name=self.i_context.code_filename, test_code=self.i_context.test_code, @@ -150,11 +151,23 @@ class RunCode(Action): return subprocess.run(cmd, check=check, cwd=cwd, env=env) @staticmethod - def _install_dependencies(working_directory, env): + def _install_requirements(working_directory, env): + file_path = Path(working_directory) / "requirements.txt" + if not file_path.exists(): + return + if file_path.stat().st_size == 0: + return install_command = ["python", "-m", "pip", "install", "-r", "requirements.txt"] logger.info(" ".join(install_command)) RunCode._install_via_subprocess(install_command, check=True, cwd=working_directory, env=env) + @staticmethod + def _install_pytest(working_directory, env): install_pytest_command = ["python", "-m", "pip", "install", "pytest"] logger.info(" ".join(install_pytest_command)) RunCode._install_via_subprocess(install_pytest_command, check=True, cwd=working_directory, env=env) + + @staticmethod + def _install_dependencies(working_directory, env): + RunCode._install_requirements(working_directory, env) + RunCode._install_pytest(working_directory, env) diff --git a/metagpt/actions/search_and_summarize.py b/metagpt/actions/search_and_summarize.py index 59b35cd58..7eed7381b 100644 --- a/metagpt/actions/search_and_summarize.py +++ b/metagpt/actions/search_and_summarize.py @@ -5,7 +5,7 @@ @Author : alexanderwu @File : search_google.py """ -from typing import Any, Optional +from typing import Optional import pydantic from pydantic import model_validator @@ -13,7 +13,6 @@ from pydantic import model_validator from metagpt.actions import Action from metagpt.logs import logger from metagpt.schema import Message -from metagpt.tools import SearchEngineType from metagpt.tools.search_engine import SearchEngine SEARCH_AND_SUMMARIZE_SYSTEM = """### Requirements @@ -105,21 +104,19 @@ You are a member of a professional butler team and will provide helpful suggesti class SearchAndSummarize(Action): name: str = "" content: Optional[str] = None - engine: Optional[SearchEngineType] = None - search_func: Optional[Any] = None search_engine: SearchEngine = None result: str = "" @model_validator(mode="after") - def validate_engine_and_run_func(self): - if self.engine is None: - self.engine = self.config.search_engine - try: - search_engine = SearchEngine(engine=self.engine, run_func=self.search_func) - except pydantic.ValidationError: - search_engine = None + def validate_search_engine(self): + if self.search_engine is None: + try: + config = self.config + search_engine = SearchEngine.from_search_config(config.search, proxy=config.proxy) + except pydantic.ValidationError: + search_engine = None - self.search_engine = search_engine + self.search_engine = search_engine return self async def run(self, context: list[Message], system_text=SEARCH_AND_SUMMARIZE_SYSTEM) -> str: diff --git a/metagpt/actions/summarize_code.py b/metagpt/actions/summarize_code.py index 182561d59..d21b62f83 100644 --- a/metagpt/actions/summarize_code.py +++ b/metagpt/actions/summarize_code.py @@ -26,9 +26,9 @@ ATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenc {system_design} ``` ----- -# Tasks +# Task ```text -{tasks} +{task} ``` ----- {code_blocks} @@ -98,10 +98,10 @@ class SummarizeCode(Action): async def run(self): design_pathname = Path(self.i_context.design_filename) - design_doc = await self.project_repo.docs.system_design.get(filename=design_pathname.name) + design_doc = await self.repo.docs.system_design.get(filename=design_pathname.name) task_pathname = Path(self.i_context.task_filename) - task_doc = await self.project_repo.docs.task.get(filename=task_pathname.name) - src_file_repo = self.project_repo.with_src_path(self.context.src_workspace).srcs + task_doc = await self.repo.docs.task.get(filename=task_pathname.name) + src_file_repo = self.repo.with_src_path(self.context.src_workspace).srcs code_blocks = [] for filename in self.i_context.codes_filenames: code_doc = await src_file_repo.get(filename) @@ -110,7 +110,7 @@ class SummarizeCode(Action): format_example = FORMAT_EXAMPLE prompt = PROMPT_TEMPLATE.format( system_design=design_doc.content, - tasks=task_doc.content, + task=task_doc.content, code_blocks="\n".join(code_blocks), format_example=format_example, ) diff --git a/metagpt/actions/write_code.py b/metagpt/actions/write_code.py index c0f1b1a93..0b86ac1bb 100644 --- a/metagpt/actions/write_code.py +++ b/metagpt/actions/write_code.py @@ -21,10 +21,17 @@ from pydantic import Field from tenacity import retry, stop_after_attempt, wait_random_exponential from metagpt.actions.action import Action -from metagpt.const import BUGFIX_FILENAME +from metagpt.actions.project_management_an import REFINED_TASK_LIST, TASK_LIST +from metagpt.actions.write_code_plan_and_change_an import REFINED_TEMPLATE +from metagpt.const import ( + BUGFIX_FILENAME, + CODE_PLAN_AND_CHANGE_FILENAME, + REQUIREMENT_FILENAME, +) from metagpt.logs import logger from metagpt.schema import CodingContext, Document, RunCodeResult from metagpt.utils.common import CodeParser +from metagpt.utils.project_repo import ProjectRepo PROMPT_TEMPLATE = """ NOTICE @@ -36,8 +43,8 @@ ATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenc ## Design {design} -## Tasks -{tasks} +## Task +{task} ## Legacy Code ```Code @@ -88,12 +95,15 @@ class WriteCode(Action): return code async def run(self, *args, **kwargs) -> CodingContext: - bug_feedback = await self.project_repo.docs.get(filename=BUGFIX_FILENAME) + bug_feedback = await self.repo.docs.get(filename=BUGFIX_FILENAME) coding_context = CodingContext.loads(self.i_context.content) - test_doc = await self.project_repo.test_outputs.get(filename="test_" + coding_context.filename + ".json") + test_doc = await self.repo.test_outputs.get(filename="test_" + coding_context.filename + ".json") + code_plan_and_change_doc = await self.repo.docs.code_plan_and_change.get(filename=CODE_PLAN_AND_CHANGE_FILENAME) + code_plan_and_change = code_plan_and_change_doc.content if code_plan_and_change_doc else "" + requirement_doc = await self.repo.docs.get(filename=REQUIREMENT_FILENAME) summary_doc = None if coding_context.design_doc and coding_context.design_doc.filename: - summary_doc = await self.project_repo.docs.code_summary.get(filename=coding_context.design_doc.filename) + summary_doc = await self.repo.docs.code_summary.get(filename=coding_context.design_doc.filename) logs = "" if test_doc: test_detail = RunCodeResult.loads(test_doc.content) @@ -101,22 +111,39 @@ class WriteCode(Action): if bug_feedback: code_context = coding_context.code_doc.content + elif code_plan_and_change: + code_context = await self.get_codes( + coding_context.task_doc, exclude=self.i_context.filename, project_repo=self.repo, use_inc=True + ) else: code_context = await self.get_codes( coding_context.task_doc, exclude=self.i_context.filename, - project_repo=self.project_repo.with_src_path(self.context.src_workspace), + project_repo=self.repo.with_src_path(self.context.src_workspace), ) - prompt = PROMPT_TEMPLATE.format( - design=coding_context.design_doc.content if coding_context.design_doc else "", - tasks=coding_context.task_doc.content if coding_context.task_doc else "", - code=code_context, - logs=logs, - feedback=bug_feedback.content if bug_feedback else "", - filename=self.i_context.filename, - summary_log=summary_doc.content if summary_doc else "", - ) + if code_plan_and_change: + prompt = REFINED_TEMPLATE.format( + user_requirement=requirement_doc.content if requirement_doc else "", + code_plan_and_change=code_plan_and_change, + design=coding_context.design_doc.content if coding_context.design_doc else "", + task=coding_context.task_doc.content if coding_context.task_doc else "", + code=code_context, + logs=logs, + feedback=bug_feedback.content if bug_feedback else "", + filename=self.i_context.filename, + summary_log=summary_doc.content if summary_doc else "", + ) + else: + prompt = PROMPT_TEMPLATE.format( + design=coding_context.design_doc.content if coding_context.design_doc else "", + task=coding_context.task_doc.content if coding_context.task_doc else "", + code=code_context, + logs=logs, + feedback=bug_feedback.content if bug_feedback else "", + filename=self.i_context.filename, + summary_log=summary_doc.content if summary_doc else "", + ) logger.info(f"Writing {coding_context.filename}..") code = await self.write_code(prompt) if not coding_context.code_doc: @@ -127,20 +154,66 @@ class WriteCode(Action): return coding_context @staticmethod - async def get_codes(task_doc, exclude, project_repo) -> str: + async def get_codes(task_doc: Document, exclude: str, project_repo: ProjectRepo, use_inc: bool = False) -> str: + """ + Get codes for generating the exclude file in various scenarios. + + Attributes: + task_doc (Document): Document object of the task file. + exclude (str): The file to be generated. Specifies the filename to be excluded from the code snippets. + project_repo (ProjectRepo): ProjectRepo object of the project. + use_inc (bool): Indicates whether the scenario involves incremental development. Defaults to False. + + Returns: + str: Codes for generating the exclude file. + """ if not task_doc: return "" if not task_doc.content: task_doc = project_repo.docs.task.get(filename=task_doc.filename) m = json.loads(task_doc.content) - code_filenames = m.get("Task list", []) + code_filenames = m.get(TASK_LIST.key, []) if use_inc else m.get(REFINED_TASK_LIST.key, []) codes = [] src_file_repo = project_repo.srcs - for filename in code_filenames: - if filename == exclude: - continue - doc = await src_file_repo.get(filename=filename) - if not doc: - continue - codes.append(f"----- {filename}\n" + doc.content) + + # Incremental development scenario + if use_inc: + src_files = src_file_repo.all_files + # Get the old workspace contained the old codes and old workspace are created in previous CodePlanAndChange + old_file_repo = project_repo.git_repo.new_file_repository(relative_path=project_repo.old_workspace) + old_files = old_file_repo.all_files + # Get the union of the files in the src and old workspaces + union_files_list = list(set(src_files) | set(old_files)) + for filename in union_files_list: + # Exclude the current file from the all code snippets + if filename == exclude: + # If the file is in the old workspace, use the old code + # Exclude unnecessary code to maintain a clean and focused main.py file, ensuring only relevant and + # essential functionality is included for the project’s requirements + if filename in old_files and filename != "main.py": + # Use old code + doc = await old_file_repo.get(filename=filename) + # If the file is in the src workspace, skip it + else: + continue + codes.insert(0, f"-----Now, {filename} to be rewritten\n```{doc.content}```\n=====") + # The code snippets are generated from the src workspace + else: + doc = await src_file_repo.get(filename=filename) + # If the file does not exist in the src workspace, skip it + if not doc: + continue + codes.append(f"----- {filename}\n```{doc.content}```") + + # Normal scenario + else: + for filename in code_filenames: + # Exclude the current file to get the code snippets for generating the current file + if filename == exclude: + continue + doc = await src_file_repo.get(filename=filename) + if not doc: + continue + codes.append(f"----- {filename}\n```{doc.content}```") + return "\n".join(codes) diff --git a/metagpt/actions/write_code_plan_and_change_an.py b/metagpt/actions/write_code_plan_and_change_an.py new file mode 100644 index 000000000..708808050 --- /dev/null +++ b/metagpt/actions/write_code_plan_and_change_an.py @@ -0,0 +1,210 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2023/12/26 +@Author : mannaandpoem +@File : write_code_plan_and_change_an.py +""" +import os + +from pydantic import Field + +from metagpt.actions.action import Action +from metagpt.actions.action_node import ActionNode +from metagpt.schema import CodePlanAndChangeContext + +CODE_PLAN_AND_CHANGE = ActionNode( + key="Code Plan And Change", + expected_type=str, + instruction="Developing comprehensive and step-by-step incremental development plan, and write Incremental " + "Change by making a code draft that how to implement incremental development including detailed steps based on the " + "context. Note: Track incremental changes using mark of '+' or '-' for add/modify/delete code, and conforms to the " + "output format of git diff", + example=""" +1. Plan for calculator.py: Enhance the functionality of `calculator.py` by extending it to incorporate methods for subtraction, multiplication, and division. Additionally, implement robust error handling for the division operation to mitigate potential issues related to division by zero. +```python +class Calculator: + self.result = number1 + number2 + return self.result + +- def sub(self, number1, number2) -> float: ++ def subtract(self, number1: float, number2: float) -> float: ++ ''' ++ Subtracts the second number from the first and returns the result. ++ ++ Args: ++ number1 (float): The number to be subtracted from. ++ number2 (float): The number to subtract. ++ ++ Returns: ++ float: The difference of number1 and number2. ++ ''' ++ self.result = number1 - number2 ++ return self.result ++ + def multiply(self, number1: float, number2: float) -> float: +- pass ++ ''' ++ Multiplies two numbers and returns the result. ++ ++ Args: ++ number1 (float): The first number to multiply. ++ number2 (float): The second number to multiply. ++ ++ Returns: ++ float: The product of number1 and number2. ++ ''' ++ self.result = number1 * number2 ++ return self.result ++ + def divide(self, number1: float, number2: float) -> float: +- pass ++ ''' ++ ValueError: If the second number is zero. ++ ''' ++ if number2 == 0: ++ raise ValueError('Cannot divide by zero') ++ self.result = number1 / number2 ++ return self.result ++ +- def reset_result(self): ++ def clear(self): ++ if self.result != 0.0: ++ print("Result is not zero, clearing...") ++ else: ++ print("Result is already zero, no need to clear.") ++ + self.result = 0.0 +``` + +2. Plan for main.py: Integrate new API endpoints for subtraction, multiplication, and division into the existing codebase of `main.py`. Then, ensure seamless integration with the overall application architecture and maintain consistency with coding standards. +```python +def add_numbers(): + result = calculator.add_numbers(num1, num2) + return jsonify({'result': result}), 200 + +-# TODO: Implement subtraction, multiplication, and division operations ++@app.route('/subtract_numbers', methods=['POST']) ++def subtract_numbers(): ++ data = request.get_json() ++ num1 = data.get('num1', 0) ++ num2 = data.get('num2', 0) ++ result = calculator.subtract_numbers(num1, num2) ++ return jsonify({'result': result}), 200 ++ ++@app.route('/multiply_numbers', methods=['POST']) ++def multiply_numbers(): ++ data = request.get_json() ++ num1 = data.get('num1', 0) ++ num2 = data.get('num2', 0) ++ try: ++ result = calculator.divide_numbers(num1, num2) ++ except ValueError as e: ++ return jsonify({'error': str(e)}), 400 ++ return jsonify({'result': result}), 200 ++ + if __name__ == '__main__': + app.run() +```""", +) + +CODE_PLAN_AND_CHANGE_CONTEXT = """ +## User New Requirements +{requirement} + +## PRD +{prd} + +## Design +{design} + +## Task +{task} + +## Legacy Code +{code} +""" + +REFINED_TEMPLATE = """ +NOTICE +Role: You are a professional engineer; The main goal is to complete incremental development by combining legacy code and plan and Incremental Change, ensuring the integration of new features. + +# Context +## User New Requirements +{user_requirement} + +## Code Plan And Change +{code_plan_and_change} + +## Design +{design} + +## Task +{task} + +## Legacy Code +```Code +{code} +``` + +## Debug logs +```text +{logs} + +{summary_log} +``` + +## Bug Feedback logs +```text +{feedback} +``` + +# Format example +## Code: {filename} +```python +## {filename} +... +``` + +# Instruction: Based on the context, follow "Format example", write or rewrite code. +## Write/Rewrite Code: Only write one file {filename}, write or rewrite complete code using triple quotes based on the following attentions and context. +1. Only One file: do your best to implement THIS ONLY ONE FILE. +2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets. +3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import. +4. Follow design: YOU MUST FOLLOW "Data structures and interfaces". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design. +5. Follow Code Plan And Change: If there is any Incremental Change that is marked by the git diff format using '+' and '-' for add/modify/delete code, or Legacy Code files contain "{filename} to be rewritten", you must merge it into the code file according to the plan. +6. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE. +7. Before using a external variable/module, make sure you import it first. +8. Write out EVERY CODE DETAIL, DON'T LEAVE TODO. +9. Attention: Retain details that are not related to incremental development but are important for maintaining the consistency and clarity of the old code. +""" + +WRITE_CODE_PLAN_AND_CHANGE_NODE = ActionNode.from_children("WriteCodePlanAndChange", [CODE_PLAN_AND_CHANGE]) + + +class WriteCodePlanAndChange(Action): + name: str = "WriteCodePlanAndChange" + i_context: CodePlanAndChangeContext = Field(default_factory=CodePlanAndChangeContext) + + async def run(self, *args, **kwargs): + self.llm.system_prompt = "You are a professional software engineer, your primary responsibility is to " + "meticulously craft comprehensive incremental development plan and deliver detailed incremental change" + prd_doc = await self.repo.docs.prd.get(filename=self.i_context.prd_filename) + design_doc = await self.repo.docs.system_design.get(filename=self.i_context.design_filename) + task_doc = await self.repo.docs.task.get(filename=self.i_context.task_filename) + code_text = await self.get_old_codes() + context = CODE_PLAN_AND_CHANGE_CONTEXT.format( + requirement=self.i_context.requirement, + prd=prd_doc.content, + design=design_doc.content, + task=task_doc.content, + code=code_text, + ) + return await WRITE_CODE_PLAN_AND_CHANGE_NODE.fill(context=context, llm=self.llm, schema="json") + + async def get_old_codes(self) -> str: + self.repo.old_workspace = self.repo.git_repo.workdir / os.path.basename(self.config.project_path) + old_file_repo = self.repo.git_repo.new_file_repository(relative_path=self.repo.old_workspace) + old_codes = await old_file_repo.get_all() + codes = [f"----- {code.filename}\n```{code.content}```" for code in old_codes] + return "\n".join(codes) diff --git a/metagpt/actions/write_code_review.py b/metagpt/actions/write_code_review.py index 21281dde1..da636eb36 100644 --- a/metagpt/actions/write_code_review.py +++ b/metagpt/actions/write_code_review.py @@ -13,6 +13,7 @@ from tenacity import retry, stop_after_attempt, wait_random_exponential from metagpt.actions import WriteCode from metagpt.actions.action import Action +from metagpt.const import CODE_PLAN_AND_CHANGE_FILENAME, REQUIREMENT_FILENAME from metagpt.logs import logger from metagpt.schema import CodingContext from metagpt.utils.common import CodeParser @@ -137,21 +138,38 @@ class WriteCodeReview(Action): async def run(self, *args, **kwargs) -> CodingContext: iterative_code = self.i_context.code_doc.content k = self.context.config.code_review_k_times or 1 + for i in range(k): format_example = FORMAT_EXAMPLE.format(filename=self.i_context.code_doc.filename) task_content = self.i_context.task_doc.content if self.i_context.task_doc else "" code_context = await WriteCode.get_codes( self.i_context.task_doc, exclude=self.i_context.filename, - project_repo=self.project_repo.with_src_path(self.context.src_workspace), - ) - context = "\n".join( - [ - "## System Design\n" + str(self.i_context.design_doc) + "\n", - "## Tasks\n" + task_content + "\n", - "## Code Files\n" + code_context + "\n", - ] + project_repo=self.repo.with_src_path(self.context.src_workspace), + use_inc=self.config.inc, ) + + if not self.config.inc: + context = "\n".join( + [ + "## System Design\n" + str(self.i_context.design_doc) + "\n", + "## Task\n" + task_content + "\n", + "## Code Files\n" + code_context + "\n", + ] + ) + else: + requirement_doc = await self.repo.docs.get(filename=REQUIREMENT_FILENAME) + code_plan_and_change_doc = await self.repo.get(filename=CODE_PLAN_AND_CHANGE_FILENAME) + context = "\n".join( + [ + "## User New Requirements\n" + str(requirement_doc) + "\n", + "## Code Plan And Change\n" + str(code_plan_and_change_doc) + "\n", + "## System Design\n" + str(self.i_context.design_doc) + "\n", + "## Task\n" + task_content + "\n", + "## Code Files\n" + code_context + "\n", + ] + ) + context_prompt = PROMPT_TEMPLATE.format( context=context, code=iterative_code, @@ -160,9 +178,11 @@ class WriteCodeReview(Action): cr_prompt = EXAMPLE_AND_INSTRUCTION.format( format_example=format_example, ) + len1 = len(iterative_code) if iterative_code else 0 + len2 = len(self.i_context.code_doc.content) if self.i_context.code_doc.content else 0 logger.info( - f"Code review and rewrite {self.i_context.code_doc.filename}: {i + 1}/{k} | {len(iterative_code)=}, " - f"{len(self.i_context.code_doc.content)=}" + f"Code review and rewrite {self.i_context.code_doc.filename}: {i + 1}/{k} | len(iterative_code)={len1}, " + f"len(self.i_context.code_doc.content)={len2}" ) result, rewrited_code = await self.write_code_review_and_rewrite( context_prompt, cr_prompt, self.i_context.code_doc.filename diff --git a/metagpt/actions/write_docstring.py b/metagpt/actions/write_docstring.py index 79204e6a4..5cc4cafb8 100644 --- a/metagpt/actions/write_docstring.py +++ b/metagpt/actions/write_docstring.py @@ -16,7 +16,7 @@ Options: Default: 'google' Example: - python3 -m metagpt.actions.write_docstring ./metagpt/startup.py --overwrite False --style=numpy + python3 -m metagpt.actions.write_docstring ./metagpt/software_company.py --overwrite False --style=numpy This script uses the 'fire' library to create a command-line interface. It generates docstrings for the given Python code using the specified docstring style and adds them to the code. diff --git a/metagpt/actions/write_prd.py b/metagpt/actions/write_prd.py index 38ac62536..b66887164 100644 --- a/metagpt/actions/write_prd.py +++ b/metagpt/actions/write_prd.py @@ -15,13 +15,14 @@ from __future__ import annotations import json from pathlib import Path -from typing import Optional from metagpt.actions import Action, ActionOutput from metagpt.actions.action_node import ActionNode from metagpt.actions.fix_bug import FixBug from metagpt.actions.write_prd_an import ( + COMPETITIVE_QUADRANT_CHART, PROJECT_NAME, + REFINED_PRD_NODE, WP_IS_RELATIVE_NODE, WP_ISSUE_TYPE_NODE, WRITE_PRD_NODE, @@ -58,97 +59,107 @@ NEW_REQ_TEMPLATE = """ class WritePRD(Action): - name: str = "WritePRD" - content: Optional[str] = None + """WritePRD deal with the following situations: + 1. Bugfix: If the requirement is a bugfix, the bugfix document will be generated. + 2. New requirement: If the requirement is a new requirement, the PRD document will be generated. + 3. Requirement update: If the requirement is an update, the PRD document will be updated. + """ async def run(self, with_messages, *args, **kwargs) -> ActionOutput | Message: - # Determine which requirement documents need to be rewritten: Use LLM to assess whether new requirements are - # related to the PRD. If they are related, rewrite the PRD. - requirement_doc = await self.project_repo.docs.get(filename=REQUIREMENT_FILENAME) - if requirement_doc and await self._is_bugfix(requirement_doc.content): - await self.project_repo.docs.save(filename=BUGFIX_FILENAME, content=requirement_doc.content) - await self.project_repo.docs.save(filename=REQUIREMENT_FILENAME, content="") - bug_fix = BugFixContext(filename=BUGFIX_FILENAME) - return Message( - content=bug_fix.model_dump_json(), - instruct_content=bug_fix, - role="", - cause_by=FixBug, - sent_from=self, - send_to="Alex", # the name of Engineer - ) + """Run the action.""" + req: Document = await self.repo.requirement + docs: list[Document] = await self.repo.docs.prd.get_all() + if not req: + raise FileNotFoundError("No requirement document found.") + + if await self._is_bugfix(req.content): + logger.info(f"Bugfix detected: {req.content}") + return await self._handle_bugfix(req) + # remove bugfix file from last round in case of conflict + await self.repo.docs.delete(filename=BUGFIX_FILENAME) + + # if requirement is related to other documents, update them, otherwise create a new one + if related_docs := await self.get_related_docs(req, docs): + logger.info(f"Requirement update detected: {req.content}") + return await self._handle_requirement_update(req, related_docs) else: - await self.project_repo.docs.delete(filename=BUGFIX_FILENAME) + logger.info(f"New requirement detected: {req.content}") + return await self._handle_new_requirement(req) - prd_docs = await self.project_repo.docs.prd.get_all() - change_files = Documents() - for prd_doc in prd_docs: - prd_doc = await self._update_prd(requirement_doc=requirement_doc, prd_doc=prd_doc, *args, **kwargs) - if not prd_doc: - continue - change_files.docs[prd_doc.filename] = prd_doc - logger.info(f"rewrite prd: {prd_doc.filename}") - # If there is no existing PRD, generate one using 'docs/requirement.txt'. - if not change_files.docs: - prd_doc = await self._update_prd(requirement_doc=requirement_doc, *args, **kwargs) - if prd_doc: - change_files.docs[prd_doc.filename] = prd_doc - logger.debug(f"new prd: {prd_doc.filename}") - # Once all files under 'docs/prds/' have been compared with the newly added requirements, trigger the - # 'publish' message to transition the workflow to the next stage. This design allows room for global - # optimization in subsequent steps. - return ActionOutput(content=change_files.model_dump_json(), instruct_content=change_files) + async def _handle_bugfix(self, req: Document) -> Message: + # ... bugfix logic ... + await self.repo.docs.save(filename=BUGFIX_FILENAME, content=req.content) + await self.repo.docs.save(filename=REQUIREMENT_FILENAME, content="") + bug_fix = BugFixContext(filename=BUGFIX_FILENAME) + return Message( + content=bug_fix.model_dump_json(), + instruct_content=bug_fix, + role="", + cause_by=FixBug, + sent_from=self, + send_to="Alex", # the name of Engineer + ) - async def _run_new_requirement(self, requirements) -> ActionOutput: + async def _handle_new_requirement(self, req: Document) -> ActionOutput: + """handle new requirement""" project_name = self.project_name - context = CONTEXT_TEMPLATE.format(requirements=requirements, project_name=project_name) + context = CONTEXT_TEMPLATE.format(requirements=req, project_name=project_name) exclude = [PROJECT_NAME.key] if project_name else [] node = await WRITE_PRD_NODE.fill(context=context, llm=self.llm, exclude=exclude) # schema=schema await self._rename_workspace(node) - return node + new_prd_doc = await self.repo.docs.prd.save( + filename=FileRepository.new_filename() + ".json", content=node.instruct_content.model_dump_json() + ) + await self._save_competitive_analysis(new_prd_doc) + await self.repo.resources.prd.save_pdf(doc=new_prd_doc) + return Documents.from_iterable(documents=[new_prd_doc]).to_action_output() - async def _is_relative(self, new_requirement_doc, old_prd_doc) -> bool: - context = NEW_REQ_TEMPLATE.format(old_prd=old_prd_doc.content, requirements=new_requirement_doc.content) + async def _handle_requirement_update(self, req: Document, related_docs: list[Document]) -> ActionOutput: + # ... requirement update logic ... + for doc in related_docs: + await self._update_prd(req, doc) + return Documents.from_iterable(documents=related_docs).to_action_output() + + async def _is_bugfix(self, context: str) -> bool: + if not self.repo.code_files_exists(): + return False + node = await WP_ISSUE_TYPE_NODE.fill(context, self.llm) + return node.get("issue_type") == "BUG" + + async def get_related_docs(self, req: Document, docs: list[Document]) -> list[Document]: + """get the related documents""" + # refine: use gather to speed up + return [i for i in docs if await self._is_related(req, i)] + + async def _is_related(self, req: Document, old_prd: Document) -> bool: + context = NEW_REQ_TEMPLATE.format(old_prd=old_prd.content, requirements=req.content) node = await WP_IS_RELATIVE_NODE.fill(context, self.llm) return node.get("is_relative") == "YES" - async def _merge(self, new_requirement_doc, prd_doc) -> Document: + async def _merge(self, req: Document, related_doc: Document) -> Document: if not self.project_name: self.project_name = Path(self.project_path).name - prompt = NEW_REQ_TEMPLATE.format(requirements=new_requirement_doc.content, old_prd=prd_doc.content) - node = await WRITE_PRD_NODE.fill(context=prompt, llm=self.llm, schema=self.prompt_schema) - prd_doc.content = node.instruct_content.model_dump_json() + prompt = NEW_REQ_TEMPLATE.format(requirements=req.content, old_prd=related_doc.content) + node = await REFINED_PRD_NODE.fill(context=prompt, llm=self.llm, schema=self.prompt_schema) + related_doc.content = node.instruct_content.model_dump_json() await self._rename_workspace(node) - return prd_doc + return related_doc - async def _update_prd(self, requirement_doc, prd_doc=None, *args, **kwargs) -> Document | None: - if not prd_doc: - prd = await self._run_new_requirement( - requirements=[requirement_doc.content if requirement_doc else ""], *args, **kwargs - ) - new_prd_doc = await self.project_repo.docs.prd.save( - filename=FileRepository.new_filename() + ".json", content=prd.instruct_content.model_dump_json() - ) - elif await self._is_relative(requirement_doc, prd_doc): - new_prd_doc = await self._merge(requirement_doc, prd_doc) - self.project_repo.docs.prd.save_doc(doc=new_prd_doc) - else: - return None + async def _update_prd(self, req: Document, prd_doc: Document) -> Document: + new_prd_doc: Document = await self._merge(req, prd_doc) + await self.repo.docs.prd.save_doc(doc=new_prd_doc) await self._save_competitive_analysis(new_prd_doc) - await self.project_repo.resources.prd.save_pdf(doc=new_prd_doc) + await self.repo.resources.prd.save_pdf(doc=new_prd_doc) return new_prd_doc - async def _save_competitive_analysis(self, prd_doc): + async def _save_competitive_analysis(self, prd_doc: Document): m = json.loads(prd_doc.content) - quadrant_chart = m.get("Competitive Quadrant Chart") + quadrant_chart = m.get(COMPETITIVE_QUADRANT_CHART.key) if not quadrant_chart: return - pathname = ( - self.project_repo.workdir / Path(COMPETITIVE_ANALYSIS_FILE_REPO) / Path(prd_doc.filename).with_suffix("") - ) - if not pathname.parent.exists(): - pathname.parent.mkdir(parents=True, exist_ok=True) - await mermaid_to_file(self.config.mermaid_engine, quadrant_chart, pathname) + pathname = self.repo.workdir / COMPETITIVE_ANALYSIS_FILE_REPO / Path(prd_doc.filename).stem + pathname.parent.mkdir(parents=True, exist_ok=True) + await mermaid_to_file(self.config.mermaid.engine, quadrant_chart, pathname) async def _rename_workspace(self, prd): if not self.project_name: @@ -158,15 +169,4 @@ class WritePRD(Action): ws_name = CodeParser.parse_str(block="Project Name", text=prd) if ws_name: self.project_name = ws_name - self.project_repo.git_repo.rename_root(self.project_name) - - async def _is_bugfix(self, context) -> bool: - git_workdir = self.project_repo.git_repo.workdir - src_workdir = git_workdir / git_workdir.name - if not src_workdir.exists(): - return False - code_files = self.project_repo.with_src_path(path=git_workdir / git_workdir.name).srcs.all_files - if not code_files: - return False - node = await WP_ISSUE_TYPE_NODE.fill(context, self.llm) - return node.get("issue_type") == "BUG" + self.repo.git_repo.rename_root(self.project_name) diff --git a/metagpt/actions/write_prd_an.py b/metagpt/actions/write_prd_an.py index 715e8fc55..9898be55b 100644 --- a/metagpt/actions/write_prd_an.py +++ b/metagpt/actions/write_prd_an.py @@ -30,6 +30,13 @@ ORIGINAL_REQUIREMENTS = ActionNode( example="Create a 2048 game", ) +REFINED_REQUIREMENTS = ActionNode( + key="Refined Requirements", + expected_type=str, + instruction="Place the New user's original requirements here.", + example="Create a 2048 game with a new feature that ...", +) + PROJECT_NAME = ActionNode( key="Project Name", expected_type=str, @@ -45,6 +52,18 @@ PRODUCT_GOALS = ActionNode( example=["Create an engaging user experience", "Improve accessibility, be responsive", "More beautiful UI"], ) +REFINED_PRODUCT_GOALS = ActionNode( + key="Refined Product Goals", + expected_type=List[str], + instruction="Update and expand the original product goals to reflect the evolving needs due to incremental " + "development.Ensure that the refined goals align with the current project direction and contribute to its success.", + example=[ + "Enhance user engagement through new features", + "Optimize performance for scalability", + "Integrate innovative UI enhancements", + ], +) + USER_STORIES = ActionNode( key="User Stories", expected_type=List[str], @@ -58,6 +77,20 @@ USER_STORIES = ActionNode( ], ) +REFINED_USER_STORIES = ActionNode( + key="Refined User Stories", + expected_type=List[str], + instruction="Update and expand the original scenario-based user stories to reflect the evolving needs due to " + "incremental development. Ensure that the refined user stories capture incremental features and improvements. ", + example=[ + "As a player, I want to choose difficulty levels to challenge my skills", + "As a player, I want a visually appealing score display after each game for a better gaming experience", + "As a player, I want a convenient restart button displayed when I lose to quickly start a new game", + "As a player, I want an enhanced and aesthetically pleasing UI to elevate the overall gaming experience", + "As a player, I want the ability to play the game seamlessly on my mobile phone for on-the-go entertainment", + ], +) + COMPETITIVE_ANALYSIS = ActionNode( key="Competitive Analysis", expected_type=List[str], @@ -97,6 +130,15 @@ REQUIREMENT_ANALYSIS = ActionNode( example="", ) +REFINED_REQUIREMENT_ANALYSIS = ActionNode( + key="Refined Requirement Analysis", + expected_type=List[str], + instruction="Review and refine the existing requirement analysis to align with the evolving needs of the project " + "due to incremental development. Ensure the analysis comprehensively covers the new features and enhancements " + "required for the refined project scope.", + example=["Require add/update/modify ..."], +) + REQUIREMENT_POOL = ActionNode( key="Requirement Pool", expected_type=List[List[str]], @@ -104,6 +146,14 @@ REQUIREMENT_POOL = ActionNode( example=[["P0", "The main code ..."], ["P0", "The game algorithm ..."]], ) +REFINED_REQUIREMENT_POOL = ActionNode( + key="Refined Requirement Pool", + expected_type=List[List[str]], + instruction="List down the top 5 to 7 requirements with their priority (P0, P1, P2). " + "Cover both legacy content and incremental content. Retain content unrelated to incremental development", + example=[["P0", "The main code ..."], ["P0", "The game algorithm ..."]], +) + UI_DESIGN_DRAFT = ActionNode( key="UI Design draft", expected_type=str, @@ -152,6 +202,22 @@ NODES = [ ANYTHING_UNCLEAR, ] +REFINED_NODES = [ + LANGUAGE, + PROGRAMMING_LANGUAGE, + REFINED_REQUIREMENTS, + PROJECT_NAME, + REFINED_PRODUCT_GOALS, + REFINED_USER_STORIES, + COMPETITIVE_ANALYSIS, + COMPETITIVE_QUADRANT_CHART, + REFINED_REQUIREMENT_ANALYSIS, + REFINED_REQUIREMENT_POOL, + UI_DESIGN_DRAFT, + ANYTHING_UNCLEAR, +] + WRITE_PRD_NODE = ActionNode.from_children("WritePRD", NODES) +REFINED_PRD_NODE = ActionNode.from_children("RefinedPRD", REFINED_NODES) WP_ISSUE_TYPE_NODE = ActionNode.from_children("WP_ISSUE_TYPE", [ISSUE_TYPE, REASON]) WP_IS_RELATIVE_NODE = ActionNode.from_children("WP_IS_RELATIVE", [IS_RELATIVE, REASON]) diff --git a/metagpt/config2.py b/metagpt/config2.py index 1d58b9d63..bc6af18c6 100644 --- a/metagpt/config2.py +++ b/metagpt/config2.py @@ -38,6 +38,7 @@ class CLIParams(BaseModel): if self.project_path: self.inc = True self.project_name = self.project_name or Path(self.project_path).name + return self class Config(CLIParams, YamlModel): @@ -50,7 +51,7 @@ class Config(CLIParams, YamlModel): proxy: str = "" # Tool Parameters - search: Optional[SearchConfig] = None + search: SearchConfig = SearchConfig() browser: BrowserConfig = BrowserConfig() mermaid: MermaidConfig = MermaidConfig() @@ -66,25 +67,22 @@ class Config(CLIParams, YamlModel): code_review_k_times: int = 2 # Will be removed in the future - llm_for_researcher_summary: str = "gpt3" - llm_for_researcher_report: str = "gpt3" - METAGPT_TEXT_TO_IMAGE_MODEL_URL: str = "" + metagpt_tti_url: str = "" language: str = "English" redis_key: str = "placeholder" - mmdc: str = "mmdc" - puppeteer_config: str = "" - pyppeteer_executable_path: str = "" - IFLYTEK_APP_ID: str = "" - IFLYTEK_API_SECRET: str = "" - IFLYTEK_API_KEY: str = "" - AZURE_TTS_SUBSCRIPTION_KEY: str = "" - AZURE_TTS_REGION: str = "" - mermaid_engine: str = "nodejs" + iflytek_app_id: str = "" + iflytek_api_secret: str = "" + iflytek_api_key: str = "" + azure_tts_subscription_key: str = "" + azure_tts_region: str = "" @classmethod def from_home(cls, path): - """Load config from ~/.metagpt/config.yaml""" - return Config.from_yaml_file(CONFIG_ROOT / path) + """Load config from ~/.metagpt/config2.yaml""" + pathname = CONFIG_ROOT / path + if not pathname.exists(): + return None + return Config.from_yaml_file(pathname) @classmethod def default(cls): diff --git a/metagpt/configs/browser_config.py b/metagpt/configs/browser_config.py index 00f918735..2f8024f44 100644 --- a/metagpt/configs/browser_config.py +++ b/metagpt/configs/browser_config.py @@ -15,6 +15,6 @@ class BrowserConfig(YamlModel): """Config for Browser""" engine: WebBrowserEngineType = WebBrowserEngineType.PLAYWRIGHT - browser: Literal["chrome", "firefox", "edge", "ie"] = "chrome" - driver: Literal["chromium", "firefox", "webkit"] = "chromium" - path: str = "" + browser_type: Literal["chromium", "firefox", "webkit", "chrome", "firefox", "edge", "ie"] = "chromium" + """If the engine is Playwright, the value should be one of "chromium", "firefox", or "webkit". If it is Selenium, the value + should be either "chrome", "firefox", "edge", or "ie".""" diff --git a/metagpt/configs/llm_config.py b/metagpt/configs/llm_config.py index 626d4242f..fb923d3e4 100644 --- a/metagpt/configs/llm_config.py +++ b/metagpt/configs/llm_config.py @@ -74,5 +74,5 @@ class LLMConfig(YamlModel): @classmethod def check_llm_key(cls, v): if v in ["", None, "YOUR_API_KEY"]: - raise ValueError("Please set your API key in config.yaml") + raise ValueError("Please set your API key in config2.yaml") return v diff --git a/metagpt/configs/mermaid_config.py b/metagpt/configs/mermaid_config.py index de4a3865c..50c8a1847 100644 --- a/metagpt/configs/mermaid_config.py +++ b/metagpt/configs/mermaid_config.py @@ -14,5 +14,6 @@ class MermaidConfig(YamlModel): """Config for Mermaid""" engine: Literal["nodejs", "ink", "playwright", "pyppeteer"] = "nodejs" - path: str = "" - puppeteer_config: str = "" # Only for nodejs engine + path: str = "mmdc" # mmdc + puppeteer_config: str = "" + pyppeteer_path: str = "/usr/bin/google-chrome-stable" diff --git a/metagpt/configs/search_config.py b/metagpt/configs/search_config.py index a8ae918db..af928b02a 100644 --- a/metagpt/configs/search_config.py +++ b/metagpt/configs/search_config.py @@ -5,6 +5,8 @@ @Author : alexanderwu @File : search_config.py """ +from typing import Callable, Optional + from metagpt.tools import SearchEngineType from metagpt.utils.yaml_model import YamlModel @@ -12,6 +14,7 @@ from metagpt.utils.yaml_model import YamlModel class SearchConfig(YamlModel): """Config for Search""" - api_key: str - api_type: SearchEngineType = SearchEngineType.SERPAPI_GOOGLE + api_type: SearchEngineType = SearchEngineType.DUCK_DUCK_GO + api_key: str = "" cse_id: str = "" # for google + search_func: Optional[Callable] = None diff --git a/metagpt/const.py b/metagpt/const.py index 0ae425a47..2cffaa804 100644 --- a/metagpt/const.py +++ b/metagpt/const.py @@ -67,6 +67,8 @@ TMP = METAGPT_ROOT / "tmp" SOURCE_ROOT = METAGPT_ROOT / "metagpt" PROMPT_PATH = SOURCE_ROOT / "prompts" SKILL_DIRECTORY = SOURCE_ROOT / "skills" +TOOL_SCHEMA_PATH = METAGPT_ROOT / "metagpt/tools/schemas" +TOOL_LIBS_PATH = METAGPT_ROOT / "metagpt/tools/libs" # REAL CONSTS @@ -82,17 +84,20 @@ MESSAGE_ROUTE_TO_NONE = "" REQUIREMENT_FILENAME = "requirement.txt" BUGFIX_FILENAME = "bugfix.txt" PACKAGE_REQUIREMENTS_FILENAME = "requirements.txt" +CODE_PLAN_AND_CHANGE_FILENAME = "code_plan_and_change.json" DOCS_FILE_REPO = "docs" PRDS_FILE_REPO = "docs/prd" SYSTEM_DESIGN_FILE_REPO = "docs/system_design" TASK_FILE_REPO = "docs/task" +CODE_PLAN_AND_CHANGE_FILE_REPO = "docs/code_plan_and_change" COMPETITIVE_ANALYSIS_FILE_REPO = "resources/competitive_analysis" DATA_API_DESIGN_FILE_REPO = "resources/data_api_design" SEQ_FLOW_FILE_REPO = "resources/seq_flow" SYSTEM_DESIGN_PDF_FILE_REPO = "resources/system_design" PRD_PDF_FILE_REPO = "resources/prd" TASK_PDF_FILE_REPO = "resources/api_spec_and_task" +CODE_PLAN_AND_CHANGE_PDF_FILE_REPO = "resources/code_plan_and_change" TEST_CODES_FILE_REPO = "tests" TEST_OUTPUTS_FILE_REPO = "test_outputs" CODE_SUMMARIES_FILE_REPO = "docs/code_summary" diff --git a/metagpt/context.py b/metagpt/context.py index 1e0d91237..3dfd52d58 100644 --- a/metagpt/context.py +++ b/metagpt/context.py @@ -17,6 +17,7 @@ from metagpt.provider.base_llm import BaseLLM from metagpt.provider.llm_provider_registry import create_llm_instance from metagpt.utils.cost_manager import CostManager from metagpt.utils.git_repository import GitRepository +from metagpt.utils.project_repo import ProjectRepo class AttrDict(BaseModel): @@ -58,6 +59,8 @@ class Context(BaseModel): kwargs: AttrDict = AttrDict() config: Config = Config.default() + + repo: Optional[ProjectRepo] = None git_repo: Optional[GitRepository] = None src_workspace: Optional[Path] = None cost_manager: CostManager = CostManager() @@ -67,8 +70,8 @@ class Context(BaseModel): def new_environ(self): """Return a new os.environ object""" env = os.environ.copy() - i = self.options - env.update({k: v for k, v in i.items() if isinstance(v, str)}) + # i = self.options + # env.update({k: v for k, v in i.items() if isinstance(v, str)}) return env # def use_llm(self, name: Optional[str] = None, provider: LLMType = LLMType.OPENAI) -> BaseLLM: @@ -92,7 +95,3 @@ class Context(BaseModel): if llm.cost_manager is None: llm.cost_manager = self.cost_manager return llm - - -# Global context, not in Env -CONTEXT = Context() diff --git a/metagpt/context_mixin.py b/metagpt/context_mixin.py index 1d239d2e4..59daa692f 100644 --- a/metagpt/context_mixin.py +++ b/metagpt/context_mixin.py @@ -7,17 +7,17 @@ """ from typing import Optional -from pydantic import BaseModel, ConfigDict, Field +from pydantic import BaseModel, ConfigDict, Field, model_validator from metagpt.config2 import Config -from metagpt.context import CONTEXT, Context +from metagpt.context import Context from metagpt.provider.base_llm import BaseLLM class ContextMixin(BaseModel): """Mixin class for context and config""" - model_config = ConfigDict(arbitrary_types_allowed=True) + model_config = ConfigDict(arbitrary_types_allowed=True, extra="allow") # Pydantic has bug on _private_attr when using inheritance, so we use private_* instead # - https://github.com/pydantic/pydantic/issues/7142 @@ -32,18 +32,17 @@ class ContextMixin(BaseModel): # Env/Role/Action will use this llm as private llm, or use self.context._llm instance private_llm: Optional[BaseLLM] = Field(default=None, exclude=True) - def __init__( - self, - context: Optional[Context] = CONTEXT, - config: Optional[Config] = None, - llm: Optional[BaseLLM] = None, - **kwargs, - ): - """Initialize with config""" - super().__init__(**kwargs) - self.set_context(context) - self.set_config(config) - self.set_llm(llm) + @model_validator(mode="after") + def validate_context_mixin_extra(self): + self._process_context_mixin_extra() + return self + + def _process_context_mixin_extra(self): + """Process the extra field""" + kwargs = self.model_extra or {} + self.set_context(kwargs.pop("context", None)) + self.set_config(kwargs.pop("config", None)) + self.set_llm(kwargs.pop("llm", None)) def set(self, k, v, override=False): """Set attribute""" @@ -81,7 +80,7 @@ class ContextMixin(BaseModel): """Role context: role context > context""" if self.private_context: return self.private_context - return CONTEXT + return Context() @context.setter def context(self, context: Context) -> None: diff --git a/metagpt/environment/README.md b/metagpt/environment/README.md new file mode 100644 index 000000000..9476ac75a --- /dev/null +++ b/metagpt/environment/README.md @@ -0,0 +1,38 @@ +Here is a environment description of MetaGPT env for different situation. +For now, the code only define the environment and still some todos like migrate roles/actions to current version. + +## Function +- Define `ExtEnv`(Base Class) which help users to integrate with external environment like games through apis or construct the game logics. +- Define `Environment`(Base Class) which is the env that MetaGPT directly used. And it includes roles and so on. +- Define the `EnvAPIRegistry` to mark the read/write apis that `ExtEnv` provide observe/step ability. And then, users can call the particular one to get observation from env or feedback to env. + +## Usage + +init environment +``` +android_env = env.create(EnvType.ANDROID) + +assistant = Role(name="Bob", profile="android assistant") +team = Team(investment=10.0, env=android_env, roles=[assistant]) +``` + +observe & step inside role's actions +``` +from metagpt.environment.api.env_api import EnvAPIAbstract + +# get screenshot from ExtEnv +screenshot_path: Path = env.observe( + EnvAPIAbstract( + api_name="get_screenshot", kwargs={"ss_name": f"{round_count}_before", "local_save_dir": task_dir} + ) + ) + +# do a `tap` action on the screen +res = env.step(EnvAPIAbstract("system_tap", kwargs={"x": x, "y": y})) +``` + +## TODO +- add android app operation assistant under `examples/android_assistant` +- migrate roles/actions of werewolf game from old version into current version +- migrate roles/actions of mincraft game from old version into current version +- migrate roles/actions of stanford_town game from old version into current version diff --git a/metagpt/environment/__init__.py b/metagpt/environment/__init__.py new file mode 100644 index 000000000..692672fa7 --- /dev/null +++ b/metagpt/environment/__init__.py @@ -0,0 +1,13 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : + +from metagpt.environment.base_env import Environment +from metagpt.environment.android_env.android_env import AndroidEnv +from metagpt.environment.mincraft_env.mincraft_env import MincraftExtEnv +from metagpt.environment.werewolf_env.werewolf_env import WerewolfEnv +from metagpt.environment.stanford_town_env.stanford_town_env import StanfordTownEnv +from metagpt.environment.software_env.software_env import SoftwareEnv + + +__all__ = ["AndroidEnv", "MincraftExtEnv", "WerewolfEnv", "StanfordTownEnv", "SoftwareEnv", "Environment"] diff --git a/metagpt/environment/android_env/__init__.py b/metagpt/environment/android_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/android_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/android_env/android_env.py b/metagpt/environment/android_env/android_env.py new file mode 100644 index 000000000..c27e20541 --- /dev/null +++ b/metagpt/environment/android_env/android_env.py @@ -0,0 +1,13 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : MG Android Env + +from pydantic import Field + +from metagpt.environment.android_env.android_ext_env import AndroidExtEnv +from metagpt.environment.base_env import Environment + + +class AndroidEnv(Environment, AndroidExtEnv): + rows: int = Field(default=0, description="rows of a grid on the screenshot") + cols: int = Field(default=0, description="cols of a grid on the screenshot") diff --git a/metagpt/environment/android_env/android_ext_env.py b/metagpt/environment/android_env/android_ext_env.py new file mode 100644 index 000000000..b81b2cd26 --- /dev/null +++ b/metagpt/environment/android_env/android_ext_env.py @@ -0,0 +1,157 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : The Android external environment to integrate with Android apps + +import subprocess +from pathlib import Path +from typing import Any, Optional + +from pydantic import Field + +from metagpt.environment.android_env.const import ADB_EXEC_FAIL +from metagpt.environment.base_env import ExtEnv, mark_as_readable, mark_as_writeable + + +class AndroidExtEnv(ExtEnv): + device_id: Optional[str] = Field(default=None) + screenshot_dir: Optional[Path] = Field(default=None) + xml_dir: Optional[Path] = Field(default=None) + width: int = Field(default=720, description="device screen width") + height: int = Field(default=1080, description="device screen height") + + def __init__(self, **data: Any): + super().__init__(**data) + if data.get("device_id"): + (width, height) = self.device_shape + self.width = data.get("width", width) + self.height = data.get("height", height) + + @property + def adb_prefix_si(self): + """adb cmd prefix with `device_id` and `shell input`""" + return f"adb -s {self.device_id} shell input " + + @property + def adb_prefix_shell(self): + """adb cmd prefix with `device_id` and `shell`""" + return f"adb -s {self.device_id} shell " + + @property + def adb_prefix(self): + """adb cmd prefix with `device_id`""" + return f"adb -s {self.device_id} " + + def execute_adb_with_cmd(self, adb_cmd: str) -> str: + res = subprocess.run(adb_cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True) + exec_res = ADB_EXEC_FAIL + if not res.returncode: + exec_res = res.stdout.strip() + return exec_res + + @property + def device_shape(self) -> tuple[int, int]: + adb_cmd = f"{self.adb_prefix_shell} wm size" + shape = (0, 0) + shape_res = self.execute_adb_with_cmd(adb_cmd) + if shape_res != ADB_EXEC_FAIL: + shape = tuple(map(int, shape_res.split(": ")[1].split("x"))) + return shape + + def list_devices(self): + adb_cmd = "adb devices" + res = self.execute_adb_with_cmd(adb_cmd) + devices = [] + if res != ADB_EXEC_FAIL: + devices = res.split("\n")[1:] + devices = [device.split()[0] for device in devices] + return devices + + @mark_as_readable + def get_screenshot(self, ss_name: str, local_save_dir: Path) -> Path: + """ + ss_name: screenshot file name + local_save_dir: local dir to store image from virtual machine + """ + assert self.screenshot_dir + ss_remote_path = Path(self.screenshot_dir).joinpath(f"{ss_name}.png") + ss_cmd = f"{self.adb_prefix_shell} screencap -p {ss_remote_path}" + ss_res = self.execute_adb_with_cmd(ss_cmd) + + res = ADB_EXEC_FAIL + if ss_res != ADB_EXEC_FAIL: + ss_local_path = Path(local_save_dir).joinpath(f"{ss_name}.png") + pull_cmd = f"{self.adb_prefix} pull {ss_remote_path} {ss_local_path}" + pull_res = self.execute_adb_with_cmd(pull_cmd) + if pull_res != ADB_EXEC_FAIL: + res = ss_local_path + return Path(res) + + @mark_as_readable + def get_xml(self, xml_name: str, local_save_dir: Path) -> Path: + xml_remote_path = Path(self.xml_dir).joinpath(f"{xml_name}.xml") + dump_cmd = f"{self.adb_prefix_shell} uiautomator dump {xml_remote_path}" + xml_res = self.execute_adb_with_cmd(dump_cmd) + + res = ADB_EXEC_FAIL + if xml_res != ADB_EXEC_FAIL: + xml_local_path = Path(local_save_dir).joinpath(f"{xml_name}.xml") + pull_cmd = f"{self.adb_prefix} pull {xml_remote_path} {xml_local_path}" + pull_res = self.execute_adb_with_cmd(pull_cmd) + if pull_res != ADB_EXEC_FAIL: + res = xml_local_path + return Path(res) + + @mark_as_writeable + def system_back(self) -> str: + adb_cmd = f"{self.adb_prefix_si} keyevent KEYCODE_BACK" + back_res = self.execute_adb_with_cmd(adb_cmd) + return back_res + + @mark_as_writeable + def system_tap(self, x: int, y: int) -> str: + adb_cmd = f"{self.adb_prefix_si} tap {x} {y}" + tap_res = self.execute_adb_with_cmd(adb_cmd) + return tap_res + + @mark_as_writeable + def user_input(self, input_txt: str) -> str: + input_txt = input_txt.replace(" ", "%s").replace("'", "") + adb_cmd = f"{self.adb_prefix_si} text {input_txt}" + input_res = self.execute_adb_with_cmd(adb_cmd) + return input_res + + @mark_as_writeable + def user_longpress(self, x: int, y: int, duration: int = 500) -> str: + adb_cmd = f"{self.adb_prefix_si} swipe {x} {y} {x} {y} {duration}" + press_res = self.execute_adb_with_cmd(adb_cmd) + return press_res + + @mark_as_writeable + def user_swipe(self, x: int, y: int, orient: str = "up", dist: str = "medium", if_quick: bool = False) -> str: + dist_unit = int(self.width / 10) + if dist == "long": + dist_unit *= 3 + elif dist == "medium": + dist_unit *= 2 + + if orient == "up": + offset = 0, -2 * dist_unit + elif orient == "down": + offset = 0, 2 * dist_unit + elif orient == "left": + offset = -1 * dist_unit, 0 + elif orient == "right": + offset = dist_unit, 0 + else: + return ADB_EXEC_FAIL + + duration = 100 if if_quick else 400 + adb_cmd = f"{self.adb_prefix_si} swipe {x} {y} {x + offset[0]} {y + offset[1]} {duration}" + swipe_res = self.execute_adb_with_cmd(adb_cmd) + return swipe_res + + @mark_as_writeable + def user_swipe_to(self, start: tuple[int, int], end: tuple[int, int], duration: int = 400): + adb_cmd = f"{self.adb_prefix_si} swipe {start[0]} {start[1]} {end[0]} {end[1]} {duration}" + swipe_res = self.execute_adb_with_cmd(adb_cmd) + return swipe_res diff --git a/metagpt/environment/android_env/const.py b/metagpt/environment/android_env/const.py new file mode 100644 index 000000000..8811289bf --- /dev/null +++ b/metagpt/environment/android_env/const.py @@ -0,0 +1,6 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : + +# For Android Assistant Agent +ADB_EXEC_FAIL = "FAILED" diff --git a/metagpt/environment/api/__init__.py b/metagpt/environment/api/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/api/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/api/env_api.py b/metagpt/environment/api/env_api.py new file mode 100644 index 000000000..1e6df544d --- /dev/null +++ b/metagpt/environment/api/env_api.py @@ -0,0 +1,60 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the environment api store + +from typing import Any, Callable, Union + +from pydantic import BaseModel, Field + + +class EnvAPIAbstract(BaseModel): + """api/interface summary description""" + + api_name: str = Field(default="", description="the api function name or id") + args: set = Field(default={}, description="the api function `args` params") + kwargs: dict = Field(default=dict(), description="the api function `kwargs` params") + + +class EnvAPIRegistry(BaseModel): + """the registry to store environment w&r api/interface""" + + registry: dict[str, dict[str, Union[dict, Any, str]]] = Field(default=dict(), exclude=True) + + def get(self, api_name: str): + if api_name not in self.registry: + raise ValueError + return self.registry.get(api_name) + + def __getitem__(self, api_name: str) -> Callable: + return self.get(api_name) + + def __setitem__(self, api_name: str, func: Callable): + self.registry[api_name] = func + + def __len__(self): + return len(self.registry) + + def get_apis(self, as_str=True) -> dict[str, dict[str, Union[dict, Any, str]]]: + """return func schema without func instance""" + apis = dict() + for func_name, func_schema in self.registry.items(): + new_func_schema = dict() + for key, value in func_schema.items(): + if key == "func": + continue + new_func_schema[key] = str(value) if as_str else value + new_func_schema = new_func_schema + apis[func_name] = new_func_schema + return apis + + +class WriteAPIRegistry(EnvAPIRegistry): + """just as a explicit class name""" + + pass + + +class ReadAPIRegistry(EnvAPIRegistry): + """just as a explicit class name""" + + pass diff --git a/metagpt/environment.py b/metagpt/environment/base_env.py similarity index 50% rename from metagpt/environment.py rename to metagpt/environment/base_env.py index 5a2dd339b..0e583ffb3 100644 --- a/metagpt/environment.py +++ b/metagpt/environment/base_env.py @@ -1,29 +1,101 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- -""" -@Time : 2023/5/11 22:12 -@Author : alexanderwu -@File : environment.py -@Modified By: mashenquan, 2023-11-1. According to Chapter 2.2.2 of RFC 116: - 1. Remove the functionality of `Environment` class as a public message buffer. - 2. Standardize the message forwarding behavior of the `Environment` class. - 3. Add the `is_idle` property. -@Modified By: mashenquan, 2023-11-4. According to the routing feature plan in Chapter 2.2.3.2 of RFC 113, the routing - functionality is to be consolidated into the `Environment` class. -""" +# @Desc : base env of executing environment + import asyncio -from typing import Iterable, Set +from enum import Enum +from typing import TYPE_CHECKING, Any, Dict, Iterable, Optional, Set, Union from pydantic import BaseModel, ConfigDict, Field, SerializeAsAny, model_validator from metagpt.context import Context +from metagpt.environment.api.env_api import ( + EnvAPIAbstract, + ReadAPIRegistry, + WriteAPIRegistry, +) from metagpt.logs import logger -from metagpt.roles.role import Role from metagpt.schema import Message -from metagpt.utils.common import is_send_to +from metagpt.utils.common import get_function_schema, is_coroutine_func, is_send_to + +if TYPE_CHECKING: + from metagpt.roles.role import Role # noqa: F401 -class Environment(BaseModel): +class EnvType(Enum): + ANDROID = "Android" + GYM = "Gym" + WEREWOLF = "Werewolf" + MINCRAFT = "Mincraft" + STANFORDTOWN = "StanfordTown" + + +env_write_api_registry = WriteAPIRegistry() +env_read_api_registry = ReadAPIRegistry() + + +def mark_as_readable(func): + """mark functionn as a readable one in ExtEnv, it observes something from ExtEnv""" + env_read_api_registry[func.__name__] = get_function_schema(func) + return func + + +def mark_as_writeable(func): + """mark functionn as a writeable one in ExtEnv, it does something to ExtEnv""" + env_write_api_registry[func.__name__] = get_function_schema(func) + return func + + +class ExtEnv(BaseModel): + """External Env to intergate actual game environment""" + + def _check_api_exist(self, rw_api: Optional[str] = None): + if not rw_api: + raise ValueError(f"{rw_api} not exists") + + def get_all_available_apis(self, mode: str = "read") -> list[Any]: + """get available read/write apis definition""" + assert mode in ["read", "write"] + if mode == "read": + return env_read_api_registry.get_apis() + else: + return env_write_api_registry.get_apis() + + async def observe(self, env_action: Union[str, EnvAPIAbstract]): + """get observation from particular api of ExtEnv""" + if isinstance(env_action, str): + read_api = env_read_api_registry.get(api_name=env_action)["func"] + self._check_api_exist(read_api) + if is_coroutine_func(read_api): + res = await read_api(self) + else: + res = read_api(self) + elif isinstance(env_action, EnvAPIAbstract): + read_api = env_read_api_registry.get(api_name=env_action.api_name)["func"] + self._check_api_exist(read_api) + if is_coroutine_func(read_api): + res = await read_api(self, *env_action.args, **env_action.kwargs) + else: + res = read_api(self, *env_action.args, **env_action.kwargs) + return res + + async def step(self, env_action: Union[str, Message, EnvAPIAbstract, list[EnvAPIAbstract]]): + """execute through particular api of ExtEnv""" + res = None + if isinstance(env_action, Message): + self.publish_message(env_action) + elif isinstance(env_action, EnvAPIAbstract): + write_api = env_write_api_registry.get(env_action.api_name)["func"] + self._check_api_exist(write_api) + if is_coroutine_func(write_api): + res = await write_api(self, *env_action.args, **env_action.kwargs) + else: + res = write_api(self, *env_action.args, **env_action.kwargs) + + return res + + +class Environment(ExtEnv): """环境,承载一批角色,角色可以向环境发布消息,可以被其他角色观察到 Environment, hosting a batch of roles, roles can publish messages to the environment, and can be observed by other roles """ @@ -31,8 +103,8 @@ class Environment(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) desc: str = Field(default="") # 环境描述 - roles: dict[str, SerializeAsAny[Role]] = Field(default_factory=dict, validate_default=True) - member_addrs: dict[Role, Set] = Field(default_factory=dict, exclude=True) + roles: dict[str, SerializeAsAny["Role"]] = Field(default_factory=dict, validate_default=True) + member_addrs: Dict["Role", Set] = Field(default_factory=dict, exclude=True) history: str = "" # For debug context: Context = Field(default_factory=Context, exclude=True) @@ -41,7 +113,7 @@ class Environment(BaseModel): self.add_roles(self.roles.values()) return self - def add_role(self, role: Role): + def add_role(self, role: "Role"): """增加一个在当前环境的角色 Add a role in the current environment """ @@ -49,7 +121,7 @@ class Environment(BaseModel): role.set_env(self) role.context = self.context - def add_roles(self, roles: Iterable[Role]): + def add_roles(self, roles: Iterable["Role"]): """增加一批在当前环境的角色 Add a batch of characters in the current environment """ @@ -95,13 +167,13 @@ class Environment(BaseModel): await asyncio.gather(*futures) logger.debug(f"is idle: {self.is_idle}") - def get_roles(self) -> dict[str, Role]: + def get_roles(self) -> dict[str, "Role"]: """获得环境内的所有角色 Process all Role runs at once """ return self.roles - def get_role(self, name: str) -> Role: + def get_role(self, name: str) -> "Role": """获得环境内的指定角色 get all the environment roles """ @@ -129,3 +201,12 @@ class Environment(BaseModel): def archive(self, auto_archive=True): if auto_archive and self.context.git_repo: self.context.git_repo.archive() + + @classmethod + def model_rebuild(cls, **kwargs): + from metagpt.roles.role import Role # noqa: F401 + + super().model_rebuild(**kwargs) + + +Environment.model_rebuild() diff --git a/metagpt/environment/mincraft_env/__init__.py b/metagpt/environment/mincraft_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/mincraft_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/mincraft_env/const.py b/metagpt/environment/mincraft_env/const.py new file mode 100644 index 000000000..a7222f9cd --- /dev/null +++ b/metagpt/environment/mincraft_env/const.py @@ -0,0 +1,44 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : + +from metagpt.const import METAGPT_ROOT + +# For Mincraft Game Agent +MC_CKPT_DIR = METAGPT_ROOT / "data/mincraft/ckpt" +MC_LOG_DIR = METAGPT_ROOT / "logs" +MC_DEFAULT_WARMUP = { + "context": 15, + "biome": 10, + "time": 15, + "nearby_blocks": 0, + "other_blocks": 10, + "nearby_entities": 5, + "health": 15, + "hunger": 15, + "position": 0, + "equipment": 0, + "inventory": 0, + "optional_inventory_items": 7, + "chests": 0, + "completed_tasks": 0, + "failed_tasks": 0, +} +MC_CURRICULUM_OB = [ + "context", + "biome", + "time", + "nearby_blocks", + "other_blocks", + "nearby_entities", + "health", + "hunger", + "position", + "equipment", + "inventory", + "chests", + "completed_tasks", + "failed_tasks", +] +MC_CORE_INVENTORY_ITEMS = r".*_log|.*_planks|stick|crafting_table|furnace" +r"|cobblestone|dirt|coal|.*_pickaxe|.*_sword|.*_axe", # curriculum_agent: only show these items in inventory before optional_inventory_items reached in warm up diff --git a/metagpt/environment/mincraft_env/mincraft_env.py b/metagpt/environment/mincraft_env/mincraft_env.py new file mode 100644 index 000000000..6327aa3f4 --- /dev/null +++ b/metagpt/environment/mincraft_env/mincraft_env.py @@ -0,0 +1,391 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : MG Mincraft Env +# refs to `voyager voyager.py` + +import json +import re +import time +from typing import Any, Iterable + +from langchain.embeddings.openai import OpenAIEmbeddings +from langchain.vectorstores import Chroma +from pydantic import ConfigDict, Field + +from metagpt.config2 import config as CONFIG +from metagpt.environment.base_env import Environment +from metagpt.environment.mincraft_env.const import MC_CKPT_DIR +from metagpt.environment.mincraft_env.mincraft_ext_env import MincraftExtEnv +from metagpt.logs import logger +from metagpt.utils.common import load_mc_skills_code, read_json_file, write_json_file + + +class MincraftEnv(Environment, MincraftExtEnv): + """MincraftEnv, including shared memory of cache and infomation between roles""" + + model_config = ConfigDict(arbitrary_types_allowed=True) + + event: dict[str, Any] = Field(default_factory=dict) + current_task: str = Field(default="Mine 1 wood log") + task_execution_time: float = Field(default=float) + context: str = Field(default="You can mine one of oak, birch, spruce, jungle, acacia, dark oak, or mangrove logs.") + code: str = Field(default="") + program_code: str = Field(default="") # write in skill/code/*.js + program_name: str = Field(default="") + critique: str = Field(default="") + skills: dict = Field(default_factory=dict) # for skills.json + retrieve_skills: list[str] = Field(default_factory=list) + event_summary: str = Field(default="") + + qa_cache: dict[str, str] = Field(default_factory=dict) + completed_tasks: list[str] = Field(default_factory=list) # Critique things + failed_tasks: list[str] = Field(default_factory=list) + + skill_desp: str = Field(default="") + + chest_memory: dict[str, Any] = Field(default_factory=dict) # eg: {'(1344, 64, 1381)': 'Unknown'} + chest_observation: str = Field(default="") # eg: "Chests: None\n\n" + + runtime_status: bool = False # equal to action execution status: success or failed + + vectordb: Chroma = Field(default_factory=Chroma) + + qa_cache_questions_vectordb: Chroma = Field(default_factory=Chroma) + + @property + def progress(self): + # return len(self.completed_tasks) + 10 # Test only + return len(self.completed_tasks) + + @property + def programs(self): + programs = "" + if self.code == "": + return programs # TODO: maybe fix 10054 now, a better way is isolating env.step() like voyager + for skill_name, entry in self.skills.items(): + programs += f"{entry['code']}\n\n" + for primitives in load_mc_skills_code(): # TODO add skills_dir + programs += f"{primitives}\n\n" + return programs + + def set_mc_port(self, mc_port): + super().set_mc_port(mc_port) + self.set_mc_resume() + + def set_mc_resume(self): + self.qa_cache_questions_vectordb = Chroma( + collection_name="qa_cache_questions_vectordb", + embedding_function=OpenAIEmbeddings(), + persist_directory=f"{MC_CKPT_DIR}/curriculum/vectordb", + ) + + self.vectordb = Chroma( + collection_name="skill_vectordb", + embedding_function=OpenAIEmbeddings(), + persist_directory=f"{MC_CKPT_DIR}/skill/vectordb", + ) + + if CONFIG.resume: + logger.info(f"Loading Action Developer from {MC_CKPT_DIR}/action") + self.chest_memory = read_json_file(f"{MC_CKPT_DIR}/action/chest_memory.json") + + logger.info(f"Loading Curriculum Agent from {MC_CKPT_DIR}/curriculum") + self.completed_tasks = read_json_file(f"{MC_CKPT_DIR}/curriculum/completed_tasks.json") + self.failed_tasks = read_json_file(f"{MC_CKPT_DIR}/curriculum/failed_tasks.json") + + logger.info(f"Loading Skill Manager from {MC_CKPT_DIR}/skill\033[0m") + self.skills = read_json_file(f"{MC_CKPT_DIR}/skill/skills.json") + + logger.info(f"Loading Qa Cache from {MC_CKPT_DIR}/curriculum\033[0m") + self.qa_cache = read_json_file(f"{MC_CKPT_DIR}/curriculum/qa_cache.json") + + if self.vectordb._collection.count() == 0: + logger.info(self.vectordb._collection.count()) + # Set vdvs for skills & qa_cache + skill_desps = [skill["description"] for program_name, skill in self.skills.items()] + program_names = [program_name for program_name, skill in self.skills.items()] + metadatas = [{"name": program_name} for program_name in program_names] + # add vectordb from file + self.vectordb.add_texts( + texts=skill_desps, + ids=program_names, + metadatas=metadatas, + ) + self.vectordb.persist() + + logger.info(self.qa_cache_questions_vectordb._collection.count()) + if self.qa_cache_questions_vectordb._collection.count() == 0: + questions = [question for question, answer in self.qa_cache.items()] + + self.qa_cache_questions_vectordb.add_texts(texts=questions) + + self.qa_cache_questions_vectordb.persist() + + logger.info( + f"INIT_CHECK: There are {self.vectordb._collection.count()} skills in vectordb and {len(self.skills)} skills in skills.json." + ) + # Check if Skill Manager's vectordb right using + assert self.vectordb._collection.count() == len(self.skills), ( + f"Skill Manager's vectordb is not synced with skills.json.\n" + f"There are {self.vectordb._collection.count()} skills in vectordb but {len(self.skills)} skills in skills.json.\n" + f"Did you set resume=False when initializing the manager?\n" + f"You may need to manually delete the vectordb directory for running from scratch." + ) + + logger.info( + f"INIT_CHECK: There are {self.qa_cache_questions_vectordb._collection.count()} qa_cache in vectordb and {len(self.qa_cache)} questions in qa_cache.json." + ) + assert self.qa_cache_questions_vectordb._collection.count() == len(self.qa_cache), ( + f"Curriculum Agent's qa cache question vectordb is not synced with qa_cache.json.\n" + f"There are {self.qa_cache_questions_vectordb._collection.count()} questions in vectordb " + f"but {len(self.qa_cache)} questions in qa_cache.json.\n" + f"Did you set resume=False when initializing the agent?\n" + f"You may need to manually delete the qa cache question vectordb directory for running from scratch.\n" + ) + + def register_roles(self, roles: Iterable["Minecraft"]): + for role in roles: + role.set_memory(self) + + def update_event(self, event: dict): + if self.event == event: + return + self.event = event + self.update_chest_memory(event) + self.update_chest_observation() + # self.event_summary = self.summarize_chatlog(event) + + def update_task(self, task: str): + self.current_task = task + + def update_context(self, context: str): + self.context = context + + def update_program_code(self, program_code: str): + self.program_code = program_code + + def update_code(self, code: str): + self.code = code # action_developer.gen_action_code to HERE + + def update_program_name(self, program_name: str): + self.program_name = program_name + + def update_critique(self, critique: str): + self.critique = critique # critic_agent.check_task_success to HERE + + def append_skill(self, skill: dict): + self.skills[self.program_name] = skill # skill_manager.retrieve_skills to HERE + + def update_retrieve_skills(self, retrieve_skills: list): + self.retrieve_skills = retrieve_skills + + def update_skill_desp(self, skill_desp: str): + self.skill_desp = skill_desp + + async def update_qa_cache(self, qa_cache: dict): + self.qa_cache = qa_cache + + def update_chest_memory(self, events: dict): + """ + Input: events: Dict + Result: self.chest_memory update & save to json + """ + nearbyChests = events[-1][1]["nearbyChests"] + for position, chest in nearbyChests.items(): + if position in self.chest_memory: + if isinstance(chest, dict): + self.chest_memory[position] = chest + if chest == "Invalid": + logger.info(f"Action Developer removing chest {position}: {chest}") + self.chest_memory.pop(position) + else: + if chest != "Invalid": + logger.info(f"Action Developer saving chest {position}: {chest}") + self.chest_memory[position] = chest + + write_json_file(f"{MC_CKPT_DIR}/action/chest_memory.json", self.chest_memory) + + def update_chest_observation(self): + """ + update chest_memory to chest_observation. + Refer to @ https://github.com/MineDojo/Voyager/blob/main/voyager/agents/action.py + """ + + chests = [] + for chest_position, chest in self.chest_memory.items(): + if isinstance(chest, dict) and len(chest) > 0: + chests.append(f"{chest_position}: {chest}") + for chest_position, chest in self.chest_memory.items(): + if isinstance(chest, dict) and len(chest) == 0: + chests.append(f"{chest_position}: Empty") + for chest_position, chest in self.chest_memory.items(): + if isinstance(chest, str): + assert chest == "Unknown" + chests.append(f"{chest_position}: Unknown items inside") + assert len(chests) == len(self.chest_memory) + if chests: + chests = "\n".join(chests) + self.chest_observation = f"Chests:\n{chests}\n\n" + else: + self.chest_observation = "Chests: None\n\n" + + def summarize_chatlog(self, events): + def filter_item(message: str): + craft_pattern = r"I cannot make \w+ because I need: (.*)" + craft_pattern2 = r"I cannot make \w+ because there is no crafting table nearby" + mine_pattern = r"I need at least a (.*) to mine \w+!" + if re.match(craft_pattern, message): + self.event_summary = re.match(craft_pattern, message).groups()[0] + elif re.match(craft_pattern2, message): + self.event_summary = "a nearby crafting table" + elif re.match(mine_pattern, message): + self.event_summary = re.match(mine_pattern, message).groups()[0] + else: + self.event_summary = "" + return self.event_summary + + chatlog = set() + for event_type, event in events: + if event_type == "onChat": + item = filter_item(event["onChat"]) + if item: + chatlog.add(item) + self.event_summary = "I also need " + ", ".join(chatlog) + "." if chatlog else "" + + def reset_block_info(self): + # revert all the placing event in the last step + pass + + def update_exploration_progress(self, success: bool): + """ + Split task into completed_tasks or failed_tasks + Args: info = { + "task": self.task, + "success": success, + "conversations": self.conversations, + } + """ + self.runtime_status = success + task = self.current_task + if task.startswith("Deposit useless items into the chest at"): + return + if success: + logger.info(f"Completed task {task}.") + self.completed_tasks.append(task) + else: + logger.info(f"Failed to complete task {task}. Skipping to next task.") + self.failed_tasks.append(task) + # when not success, below to update event! + # revert all the placing event in the last step + blocks = [] + positions = [] + for event_type, event in self.event: + if event_type == "onSave" and event["onSave"].endswith("_placed"): + block = event["onSave"].split("_placed")[0] + position = event["status"]["position"] + blocks.append(block) + positions.append(position) + new_events = self.step( + f"await givePlacedItemBack(bot, {json.dumps(blocks)}, {json.dumps(positions)})", + programs=self.programs, + ) + self.event[-1][1]["inventory"] = new_events[-1][1]["inventory"] + self.event[-1][1]["voxels"] = new_events[-1][1]["voxels"] + + self.save_sorted_tasks() + + def save_sorted_tasks(self): + updated_completed_tasks = [] + # record repeated failed tasks + updated_failed_tasks = self.failed_tasks + # dedup but keep order + for task in self.completed_tasks: + if task not in updated_completed_tasks: + updated_completed_tasks.append(task) + + # remove completed tasks from failed tasks + for task in updated_completed_tasks: + while task in updated_failed_tasks: + updated_failed_tasks.remove(task) + + self.completed_tasks = updated_completed_tasks + self.failed_tasks = updated_failed_tasks + + # dump to json + write_json_file(f"{MC_CKPT_DIR}/curriculum/completed_tasks.json", self.completed_tasks) + write_json_file(f"{MC_CKPT_DIR}/curriculum/failed_tasks.json", self.failed_tasks) + + async def on_event_retrieve(self, *args): + """ + Retrieve Minecraft events. + + Returns: + list: A list of Minecraft events. + + Raises: + Exception: If there is an issue retrieving events. + """ + try: + self.reset( + options={ + "mode": "soft", + "wait_ticks": 20, + } + ) + # difficulty = "easy" if len(self.completed_tasks) > 15 else "peaceful" + difficulty = "peaceful" + + events = self.step("bot.chat(`/time set ${getNextTime()}`);\n" + f"bot.chat('/difficulty {difficulty}');") + self.update_event(events) + return events + except Exception as e: + time.sleep(3) # wait for mineflayer to exit + # reset bot status here + events = self.reset( + options={ + "mode": "hard", + "wait_ticks": 20, + "inventory": self.event[-1][1]["inventory"], + "equipment": self.event[-1][1]["status"]["equipment"], + "position": self.event[-1][1]["status"]["position"], + } + ) + self.update_event(events) + logger.error(f"Failed to retrieve Minecraft events: {str(e)}") + return events + + async def on_event_execute(self, *args): + """ + Execute Minecraft events. + + This function is used to obtain events from the Minecraft environment. Check the implementation in + the 'voyager/env/bridge.py step()' function to capture events generated within the game. + + Returns: + list: A list of Minecraft events. + + Raises: + Exception: If there is an issue retrieving events. + """ + try: + events = self.step( + code=self.code, + programs=self.programs, + ) + self.update_event(events) + return events + except Exception as e: + time.sleep(3) # wait for mineflayer to exit + # reset bot status here + events = self.reset( + options={ + "mode": "hard", + "wait_ticks": 20, + "inventory": self.event[-1][1]["inventory"], + "equipment": self.event[-1][1]["status"]["equipment"], + "position": self.event[-1][1]["status"]["position"], + } + ) + self.update_event(events) + logger.error(f"Failed to execute Minecraft events: {str(e)}") + return events diff --git a/metagpt/environment/mincraft_env/mincraft_ext_env.py b/metagpt/environment/mincraft_env/mincraft_ext_env.py new file mode 100644 index 000000000..b86250d8c --- /dev/null +++ b/metagpt/environment/mincraft_env/mincraft_ext_env.py @@ -0,0 +1,180 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : The Mincraft external environment to integrate with Mincraft game +# refs to `voyager bridge.py` + +import json +import time +from typing import Optional + +import requests +from pydantic import ConfigDict, Field, model_validator + +from metagpt.environment.base_env import ExtEnv, mark_as_writeable +from metagpt.environment.mincraft_env.const import ( + MC_CKPT_DIR, + MC_CORE_INVENTORY_ITEMS, + MC_CURRICULUM_OB, + MC_DEFAULT_WARMUP, + METAGPT_ROOT, +) +from metagpt.environment.mincraft_env.process_monitor import SubprocessMonitor +from metagpt.logs import logger + + +class MincraftExtEnv(ExtEnv): + model_config = ConfigDict(arbitrary_types_allowed=True) + + mc_port: Optional[int] = Field(default=None) + server_host: str = Field(default="http://127.0.0.1") + server_port: str = Field(default=3000) + request_timeout: int = Field(default=600) + + mineflayer: Optional[SubprocessMonitor] = Field(default=None, validate_default=True) + + has_reset: bool = Field(default=False) + reset_options: Optional[dict] = Field(default=None) + connected: bool = Field(default=False) + server_paused: bool = Field(default=False) + warm_up: dict = Field(default=dict()) + + @property + def server(self) -> str: + return f"{self.server_host}:{self.server_port}" + + @model_validator(mode="after") + def _post_init_ext_env(self): + if not self.mineflayer: + self.mineflayer = SubprocessMonitor( + commands=[ + "node", + METAGPT_ROOT.joinpath("metagpt", "environment", "mincraft_env", "mineflayer", "index.js"), + str(self.server_port), + ], + name="mineflayer", + ready_match=r"Server started on port (\d+)", + ) + if not self.warm_up: + warm_up = MC_DEFAULT_WARMUP + if "optional_inventory_items" in warm_up: + assert MC_CORE_INVENTORY_ITEMS is not None + # self.core_inv_items_regex = re.compile(MC_CORE_INVENTORY_ITEMS) + self.warm_up["optional_inventory_items"] = warm_up["optional_inventory_items"] + else: + self.warm_up["optional_inventory_items"] = 0 + for key in MC_CURRICULUM_OB: + self.warm_up[key] = warm_up.get(key, MC_DEFAULT_WARMUP[key]) + self.warm_up["nearby_blocks"] = 0 + self.warm_up["inventory"] = 0 + self.warm_up["completed_tasks"] = 0 + self.warm_up["failed_tasks"] = 0 + + # init ckpt sub-forders + MC_CKPT_DIR.joinpath("curriculum/vectordb").mkdir(parents=True, exist_ok=True) + MC_CKPT_DIR.joinpath("action").mkdir(exist_ok=True) + MC_CKPT_DIR.joinpath("skill/code").mkdir(parents=True, exist_ok=True) + MC_CKPT_DIR.joinpath("skill/description").mkdir(exist_ok=True) + MC_CKPT_DIR.joinpath("skill/vectordb").mkdir(exist_ok=True) + + def set_mc_port(self, mc_port: int): + self.mc_port = mc_port + + @mark_as_writeable + def close(self) -> bool: + self.unpause() + if self.connected: + res = requests.post(f"{self.server}/stop") + if res.status_code == 200: + self.connected = False + self.mineflayer.stop() + return not self.connected + + @mark_as_writeable + def check_process(self) -> dict: + retry = 0 + while not self.mineflayer.is_running: + logger.info("Mineflayer process has exited, restarting") + self.mineflayer.run() + if not self.mineflayer.is_running: + if retry > 3: + logger.error("Mineflayer process failed to start") + raise {} + else: + retry += 1 + continue + logger.info(self.mineflayer.ready_line) + res = requests.post( + f"{self.server}/start", + json=self.reset_options, + timeout=self.request_timeout, + ) + if res.status_code != 200: + self.mineflayer.stop() + logger.error(f"Minecraft server reply with code {res.status_code}") + raise {} + return res.json() + + @mark_as_writeable + def reset(self, *, seed=None, options=None) -> dict: + if options is None: + options = {} + if options.get("inventory", {}) and options.get("mode", "hard") != "hard": + logger.error("inventory can only be set when options is hard") + raise {} + + self.reset_options = { + "port": self.mc_port, + "reset": options.get("mode", "hard"), + "inventory": options.get("inventory", {}), + "equipment": options.get("equipment", []), + "spread": options.get("spread", False), + "waitTicks": options.get("wait_ticks", 5), + "position": options.get("position", None), + } + + self.unpause() + self.mineflayer.stop() + time.sleep(1) # wait for mineflayer to exit + + returned_data = self.check_process() + self.has_reset = True + self.connected = True + # All the reset in step will be soft + self.reset_options["reset"] = "soft" + self.pause() + return json.loads(returned_data) + + @mark_as_writeable + def step(self, code: str, programs: str = "") -> dict: + if not self.has_reset: + raise RuntimeError("Environment has not been reset yet") + self.check_process() + self.unpause() + data = { + "code": code, + "programs": programs, + } + res = requests.post(f"{self.server}/step", json=data, timeout=self.request_timeout) + if res.status_code != 200: + raise RuntimeError("Failed to step Minecraft server") + returned_data = res.json() + self.pause() + return json.loads(returned_data) + + @mark_as_writeable + def pause(self) -> bool: + if self.mineflayer.is_running and not self.server_paused: + res = requests.post(f"{self.server}/pause") + if res.status_code == 200: + self.server_paused = True + return self.server_paused + + @mark_as_writeable + def unpause(self) -> bool: + if self.mineflayer.is_running and self.server_paused: + res = requests.post(f"{self.server}/pause") + if res.status_code == 200: + self.server_paused = False + else: + logger.info(f"mineflayer pause result: {res.json()}") + return self.server_paused diff --git a/metagpt/environment/mincraft_env/mineflayer/.gitignore b/metagpt/environment/mincraft_env/mineflayer/.gitignore new file mode 100644 index 000000000..0fd468410 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/.gitignore @@ -0,0 +1 @@ +!/lib \ No newline at end of file diff --git a/metagpt/environment/mincraft_env/mineflayer/.prettierignore b/metagpt/environment/mincraft_env/mineflayer/.prettierignore new file mode 100644 index 000000000..1b07c39e9 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/.prettierignore @@ -0,0 +1,3 @@ +# Ignore artifacts: +build +coverage \ No newline at end of file diff --git a/metagpt/environment/mincraft_env/mineflayer/.prettierrc.json b/metagpt/environment/mincraft_env/mineflayer/.prettierrc.json new file mode 100644 index 000000000..0a02bcefd --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/.prettierrc.json @@ -0,0 +1,3 @@ +{ + "tabWidth": 4 +} diff --git a/metagpt/environment/mincraft_env/mineflayer/index.js b/metagpt/environment/mincraft_env/mineflayer/index.js new file mode 100644 index 000000000..7fb0a8787 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/index.js @@ -0,0 +1,425 @@ +const fs = require("fs"); +const express = require("express"); +const bodyParser = require("body-parser"); +const mineflayer = require("mineflayer"); + +const skills = require("./lib/skillLoader"); +const { initCounter, getNextTime } = require("./lib/utils"); +const obs = require("./lib/observation/base"); +const OnChat = require("./lib/observation/onChat"); +const OnError = require("./lib/observation/onError"); +const { Voxels, BlockRecords } = require("./lib/observation/voxels"); +const Status = require("./lib/observation/status"); +const Inventory = require("./lib/observation/inventory"); +const OnSave = require("./lib/observation/onSave"); +const Chests = require("./lib/observation/chests"); +const { plugin: tool } = require("mineflayer-tool"); + +let bot = null; + +const app = express(); + +app.use(bodyParser.json({ limit: "50mb" })); +app.use(bodyParser.urlencoded({ limit: "50mb", extended: false })); + +app.post("/start", (req, res) => { + if (bot) onDisconnect("Restarting bot"); + bot = null; + console.log(req.body); + bot = mineflayer.createBot({ + host: "localhost", // minecraft server ip + port: req.body.port, // minecraft server port + username: "bot", + disableChatSigning: true, + checkTimeoutInterval: 60 * 60 * 1000, + }); + bot.once("error", onConnectionFailed); + + // Event subscriptions + bot.waitTicks = req.body.waitTicks; + bot.globalTickCounter = 0; + bot.stuckTickCounter = 0; + bot.stuckPosList = []; + bot.iron_pickaxe = false; + + bot.on("kicked", onDisconnect); + + // mounting will cause physicsTick to stop + bot.on("mount", () => { + bot.dismount(); + }); + + bot.once("spawn", async () => { + bot.removeListener("error", onConnectionFailed); + let itemTicks = 1; + if (req.body.reset === "hard") { + bot.chat("/clear @s"); + bot.chat("/kill @s"); + const inventory = req.body.inventory ? req.body.inventory : {}; + const equipment = req.body.equipment + ? req.body.equipment + : [null, null, null, null, null, null]; + for (let key in inventory) { + bot.chat(`/give @s minecraft:${key} ${inventory[key]}`); + itemTicks += 1; + } + const equipmentNames = [ + "armor.head", + "armor.chest", + "armor.legs", + "armor.feet", + "weapon.mainhand", + "weapon.offhand", + ]; + for (let i = 0; i < 6; i++) { + if (i === 4) continue; + if (equipment[i]) { + bot.chat( + `/item replace entity @s ${equipmentNames[i]} with minecraft:${equipment[i]}` + ); + itemTicks += 1; + } + } + } + + if (req.body.position) { + bot.chat( + `/tp @s ${req.body.position.x} ${req.body.position.y} ${req.body.position.z}` + ); + } + + // if iron_pickaxe is in bot's inventory + if ( + bot.inventory.items().find((item) => item.name === "iron_pickaxe") + ) { + bot.iron_pickaxe = true; + } + + const { pathfinder } = require("mineflayer-pathfinder"); + const tool = require("mineflayer-tool").plugin; + const collectBlock = require("mineflayer-collectblock").plugin; + const pvp = require("mineflayer-pvp").plugin; + const minecraftHawkEye = require("minecrafthawkeye"); + bot.loadPlugin(pathfinder); + bot.loadPlugin(tool); + bot.loadPlugin(collectBlock); + bot.loadPlugin(pvp); + bot.loadPlugin(minecraftHawkEye); + + // bot.collectBlock.movements.digCost = 0; + // bot.collectBlock.movements.placeCost = 0; + + obs.inject(bot, [ + OnChat, + OnError, + Voxels, + Status, + Inventory, + OnSave, + Chests, + BlockRecords, + ]); + skills.inject(bot); + + if (req.body.spread) { + bot.chat(`/spreadplayers ~ ~ 0 300 under 80 false @s`); + await bot.waitForTicks(bot.waitTicks); + } + + await bot.waitForTicks(bot.waitTicks * itemTicks); + res.json(bot.observe()); + + initCounter(bot); + bot.chat("/gamerule keepInventory true"); + bot.chat("/gamerule doDaylightCycle false"); + }); + + function onConnectionFailed(e) { + console.log(e); + bot = null; + res.status(400).json({ error: e }); + } + function onDisconnect(message) { + if (bot.viewer) { + bot.viewer.close(); + } + bot.end(); + console.log(message); + bot = null; + } +}); + +app.post("/step", async (req, res) => { + // import useful package + let response_sent = false; + function otherError(err) { + console.log("Uncaught Error"); + bot.emit("error", handleError(err)); + bot.waitForTicks(bot.waitTicks).then(() => { + if (!response_sent) { + response_sent = true; + res.json(bot.observe()); + } + }); + } + + process.on("uncaughtException", otherError); + + const mcData = require("minecraft-data")(bot.version); + mcData.itemsByName["leather_cap"] = mcData.itemsByName["leather_helmet"]; + mcData.itemsByName["leather_tunic"] = + mcData.itemsByName["leather_chestplate"]; + mcData.itemsByName["leather_pants"] = + mcData.itemsByName["leather_leggings"]; + mcData.itemsByName["leather_boots"] = mcData.itemsByName["leather_boots"]; + mcData.itemsByName["lapis_lazuli_ore"] = mcData.itemsByName["lapis_ore"]; + mcData.blocksByName["lapis_lazuli_ore"] = mcData.blocksByName["lapis_ore"]; + const { + Movements, + goals: { + Goal, + GoalBlock, + GoalNear, + GoalXZ, + GoalNearXZ, + GoalY, + GoalGetToBlock, + GoalLookAtBlock, + GoalBreakBlock, + GoalCompositeAny, + GoalCompositeAll, + GoalInvert, + GoalFollow, + GoalPlaceBlock, + }, + pathfinder, + Move, + ComputedPath, + PartiallyComputedPath, + XZCoordinates, + XYZCoordinates, + SafeBlock, + GoalPlaceBlockOptions, + } = require("mineflayer-pathfinder"); + const { Vec3 } = require("vec3"); + + // Set up pathfinder + const movements = new Movements(bot, mcData); + bot.pathfinder.setMovements(movements); + + bot.globalTickCounter = 0; + bot.stuckTickCounter = 0; + bot.stuckPosList = []; + + function onTick() { + bot.globalTickCounter++; + if (bot.pathfinder.isMoving()) { + bot.stuckTickCounter++; + if (bot.stuckTickCounter >= 100) { + onStuck(1.5); + bot.stuckTickCounter = 0; + } + } + } + + bot.on("physicTick", onTick); + + // initialize fail count + let _craftItemFailCount = 0; + let _killMobFailCount = 0; + let _mineBlockFailCount = 0; + let _placeItemFailCount = 0; + let _smeltItemFailCount = 0; + + // Retrieve array form post bod + const code = req.body.code; + const programs = req.body.programs; + bot.cumulativeObs = []; + await bot.waitForTicks(bot.waitTicks); + const r = await evaluateCode(code, programs); + process.off("uncaughtException", otherError); + if (r !== "success") { + bot.emit("error", handleError(r)); + } + await returnItems(); + // wait for last message + await bot.waitForTicks(bot.waitTicks); + if (!response_sent) { + response_sent = true; + res.json(bot.observe()); + } + bot.removeListener("physicTick", onTick); + + async function evaluateCode(code, programs) { + // Echo the code produced for players to see it. Don't echo when the bot code is already producing dialog or it will double echo + try { + await eval("(async () => {" + programs + "\n" + code + "})()"); + return "success"; + } catch (err) { + return err; + } + } + + function onStuck(posThreshold) { + const currentPos = bot.entity.position; + bot.stuckPosList.push(currentPos); + + // Check if the list is full + if (bot.stuckPosList.length === 5) { + const oldestPos = bot.stuckPosList[0]; + const posDifference = currentPos.distanceTo(oldestPos); + + if (posDifference < posThreshold) { + teleportBot(); // execute the function + } + + // Remove the oldest time from the list + bot.stuckPosList.shift(); + } + } + + function teleportBot() { + const blocks = bot.findBlocks({ + matching: (block) => { + return block.type === 0; + }, + maxDistance: 1, + count: 27, + }); + + if (blocks) { + // console.log(blocks.length); + const randomIndex = Math.floor(Math.random() * blocks.length); + const block = blocks[randomIndex]; + bot.chat(`/tp @s ${block.x} ${block.y} ${block.z}`); + } else { + bot.chat("/tp @s ~ ~1.25 ~"); + } + } + + function returnItems() { + bot.chat("/gamerule doTileDrops false"); + const crafting_table = bot.findBlock({ + matching: mcData.blocksByName.crafting_table.id, + maxDistance: 128, + }); + if (crafting_table) { + bot.chat( + `/setblock ${crafting_table.position.x} ${crafting_table.position.y} ${crafting_table.position.z} air destroy` + ); + bot.chat("/give @s crafting_table"); + } + const furnace = bot.findBlock({ + matching: mcData.blocksByName.furnace.id, + maxDistance: 128, + }); + if (furnace) { + bot.chat( + `/setblock ${furnace.position.x} ${furnace.position.y} ${furnace.position.z} air destroy` + ); + bot.chat("/give @s furnace"); + } + if (bot.inventoryUsed() >= 32) { + // if chest is not in bot's inventory + if (!bot.inventory.items().find((item) => item.name === "chest")) { + bot.chat("/give @s chest"); + } + } + // if iron_pickaxe not in bot's inventory and bot.iron_pickaxe + if ( + bot.iron_pickaxe && + !bot.inventory.items().find((item) => item.name === "iron_pickaxe") + ) { + bot.chat("/give @s iron_pickaxe"); + } + bot.chat("/gamerule doTileDrops true"); + } + + function handleError(err) { + let stack = err.stack; + if (!stack) { + return err; + } + console.log(stack); + const final_line = stack.split("\n")[1]; + const regex = /:(\d+):\d+\)/; + + const programs_length = programs.split("\n").length; + let match_line = null; + for (const line of stack.split("\n")) { + const match = regex.exec(line); + if (match) { + const line_num = parseInt(match[1]); + if (line_num >= programs_length) { + match_line = line_num - programs_length; + break; + } + } + } + if (!match_line) { + return err.message; + } + let f_line = final_line.match( + /\((?.*):(?\d+):(?\d+)\)/ + ); + if (f_line && f_line.groups && fs.existsSync(f_line.groups.file)) { + const { file, line, pos } = f_line.groups; + const f = fs.readFileSync(file, "utf8").split("\n"); + // let filename = file.match(/(?<=node_modules\\)(.*)/)[1]; + let source = file + `:${line}\n${f[line - 1].trim()}\n `; + + const code_source = + "at " + + code.split("\n")[match_line - 1].trim() + + " in your code"; + return source + err.message + "\n" + code_source; + } else if ( + f_line && + f_line.groups && + f_line.groups.file.includes("") + ) { + const { file, line, pos } = f_line.groups; + let source = + "Your code" + + `:${match_line}\n${code.split("\n")[match_line - 1].trim()}\n `; + let code_source = ""; + if (line < programs_length) { + source = + "In your program code: " + + programs.split("\n")[line - 1].trim() + + "\n"; + code_source = `at line ${match_line}:${code + .split("\n") + [match_line - 1].trim()} in your code`; + } + return source + err.message + "\n" + code_source; + } + return err.message; + } +}); + +app.post("/stop", (req, res) => { + bot.end(); + res.json({ + message: "Bot stopped", + }); +}); + +app.post("/pause", (req, res) => { + if (!bot) { + res.status(400).json({ error: "Bot not spawned" }); + return; + } + bot.chat("/pause"); + bot.waitForTicks(bot.waitTicks).then(() => { + res.json({ message: "Success" }); + }); +}); + +// Server listening to PORT 3000 + +const DEFAULT_PORT = 3000; +const PORT = process.argv[2] || DEFAULT_PORT; +app.listen(PORT, () => { + console.log(`Server started on port ${PORT}`); +}); diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/base.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/base.js new file mode 100644 index 000000000..b661a24b5 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/base.js @@ -0,0 +1,45 @@ +class Observation { + constructor(bot) { + if (new.target === Observation) { + throw new TypeError( + "Cannot instantiate abstract class Observation" + ); + } + + this.bot = bot; + this.name = "Observation"; + } + + observe() { + throw new TypeError("Method 'observe()' must be implemented."); + } + + reset() {} +} + +function inject(bot, obs_list) { + bot.obsList = []; + bot.cumulativeObs = []; + bot.eventMemory = {}; + obs_list.forEach((obs) => { + bot.obsList.push(new obs(bot)); + }); + bot.event = function (event_name) { + let result = {}; + bot.obsList.forEach((obs) => { + if (obs.name.startsWith("on") && obs.name !== event_name) { + return; + } + result[obs.name] = obs.observe(); + }); + bot.cumulativeObs.push([event_name, result]); + }; + bot.observe = function () { + bot.event("observe"); + const result = bot.cumulativeObs; + bot.cumulativeObs = []; + return JSON.stringify(result); + }; +} + +module.exports = { Observation, inject }; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/chests.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/chests.js new file mode 100644 index 000000000..842bd171d --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/chests.js @@ -0,0 +1,31 @@ +const { Observation } = require("./base"); + +class Chests extends Observation { + constructor(bot) { + super(bot); + this.name = "nearbyChests"; + this.chestsItems = {}; + bot.on("closeChest", (chestItems, position) => { + this.chestsItems[position] = chestItems; + }); + bot.on("removeChest", (chestPosition) => { + this.chestsItems[chestPosition] = "Invalid"; + }); + } + + observe() { + const chests = this.bot.findBlocks({ + matching: this.bot.registry.blocksByName.chest.id, + maxDistance: 16, + count: 999, + }); + chests.forEach((chest) => { + if (!this.chestsItems.hasOwnProperty(chest)) { + this.chestsItems[chest] = "Unknown"; + } + }); + return this.chestsItems; + } +} + +module.exports = Chests; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/inventory.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/inventory.js new file mode 100644 index 000000000..0645d1bfa --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/inventory.js @@ -0,0 +1,39 @@ +const { Observation } = require("./base"); + +class Inventory extends Observation { + constructor(bot) { + super(bot); + this.name = "inventory"; + } + + observe() { + return listItems(this.bot); + } +} + +function listItems(bot) { + const items = getInventoryItems(bot); + return items.reduce(itemToDict, {}); +} + +function getInventoryItems(bot) { + const inventory = bot.currentWindow || bot.inventory; + return inventory.items(); +} + +function itemToDict(acc, cur) { + if (cur.name && cur.count) { + //if both name and count property are defined + if (acc[cur.name]) { + //if the item is already in the dict + acc[cur.name] += cur.count; + } else { + //if the item is not in the dict + acc[cur.name] = cur.count; + } + } + return acc; +} + +//export modules +module.exports = Inventory; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/onChat.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onChat.js new file mode 100644 index 000000000..54b411e2a --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onChat.js @@ -0,0 +1,26 @@ +const Observation = require("./base.js").Observation; + +class onChat extends Observation { + constructor(bot) { + super(bot); + this.name = "onChat"; + this.obs = ""; + bot.on("chatEvent", (username, message) => { + // Save entity status to local variable + if (message.startsWith("/")) { + return; + } + + this.obs += message; + this.bot.event(this.name); + }); + } + + observe() { + const result = this.obs; + this.obs = ""; + return result; + } +} + +module.exports = onChat; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/onError.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onError.js new file mode 100644 index 000000000..ac8fed9e5 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onError.js @@ -0,0 +1,22 @@ +const Observation = require("./base.js").Observation; + +class onError extends Observation { + constructor(bot) { + super(bot); + this.name = "onError"; + this.obs = null; + bot.on("error", (err) => { + // Save entity status to local variable + this.obs = err; + this.bot.event(this.name); + }); + } + + observe() { + const result = this.obs; + this.obs = null; + return result; + } +} + +module.exports = onError; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/onSave.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onSave.js new file mode 100644 index 000000000..e5983590f --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/onSave.js @@ -0,0 +1,22 @@ +const Observation = require("./base.js").Observation; + +class onSave extends Observation { + constructor(bot) { + super(bot); + this.name = "onSave"; + this.obs = null; + bot.on("save", (eventName) => { + // Save entity status to local variable + this.obs = eventName; + this.bot.event(this.name); + }); + } + + observe() { + const result = this.obs; + this.obs = null; + return result; + } +} + +module.exports = onSave; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/status.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/status.js new file mode 100644 index 000000000..b031fbcf2 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/status.js @@ -0,0 +1,103 @@ +const Observation = require("./base.js").Observation; + +class Status extends Observation { + constructor(bot) { + super(bot); + this.name = "status"; + } + + observe() { + return { + health: this.bot.health, + food: this.bot.food, + saturation: this.bot.foodSaturation, + oxygen: this.bot.oxygenLevel, + position: this.bot.entity.position, + velocity: this.bot.entity.velocity, + yaw: this.bot.entity.yaw, + pitch: this.bot.entity.pitch, + onGround: this.bot.entity.onGround, + equipment: this.getEquipment(), + name: this.bot.entity.username, + timeSinceOnGround: this.bot.entity.timeSinceOnGround, + isInWater: this.bot.entity.isInWater, + isInLava: this.bot.entity.isInLava, + isInWeb: this.bot.entity.isInWeb, + isCollidedHorizontally: this.bot.entity.isCollidedHorizontally, + isCollidedVertically: this.bot.entity.isCollidedVertically, + biome: this.bot.blockAt(this.bot.entity.position) + ? this.bot.blockAt(this.bot.entity.position).biome.name + : "None", + entities: this.getEntities(), + timeOfDay: this.getTime(), + inventoryUsed: this.bot.inventoryUsed(), + elapsedTime: this.bot.globalTickCounter, + }; + } + + itemToObs(item) { + if (!item) return null; + return item.name; + } + + getTime() { + const timeOfDay = this.bot.time.timeOfDay; + let time = ""; + if (timeOfDay < 1000) { + time = "sunrise"; + } else if (timeOfDay < 6000) { + time = "day"; + } else if (timeOfDay < 12000) { + time = "noon"; + } else if (timeOfDay < 13000) { + time = "sunset"; + } else if (timeOfDay < 18000) { + time = "night"; + } else if (timeOfDay < 22000) { + time = "midnight"; + } else { + time = "sunrise"; + } + return time; + } + + // For each item in equipment, if it exists, return the name of the item + // otherwise return null + getEquipment() { + const slots = this.bot.inventory.slots; + const mainHand = this.bot.heldItem; + return slots + .slice(5, 9) + .concat(mainHand, slots[45]) + .map(this.itemToObs); + } + + getEntities() { + const entities = this.bot.entities; + if (!entities) return {}; + // keep all monsters in one list, keep other mobs in another list + const mobs = {}; + for (const id in entities) { + const entity = entities[id]; + if (!entity.displayName) continue; + if (entity.name === "player" || entity.name === "item") continue; + if (entity.position.distanceTo(this.bot.entity.position) < 32) { + if (!mobs[entity.name]) { + mobs[entity.name] = entity.position.distanceTo( + this.bot.entity.position + ); + } else if ( + mobs[entity.name] > + entity.position.distanceTo(this.bot.entity.position) + ) { + mobs[entity.name] = entity.position.distanceTo( + this.bot.entity.position + ); + } + } + } + return mobs; + } +} + +module.exports = Status; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/observation/voxels.js b/metagpt/environment/mincraft_env/mineflayer/lib/observation/voxels.js new file mode 100644 index 000000000..ecb0c14b7 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/observation/voxels.js @@ -0,0 +1,67 @@ +// Blocks = require("./blocks") +const { Observation } = require("./base"); + +class Voxels extends Observation { + constructor(bot) { + super(bot); + this.name = "voxels"; + } + + observe() { + return Array.from(getSurroundingBlocks(this.bot, 8, 2, 8)); + } +} + +class BlockRecords extends Observation { + constructor(bot) { + super(bot); + this.name = "blockRecords"; + this.records = new Set(); + this.tick = 0; + bot.on("physicsTick", () => { + this.tick++; + if (this.tick >= 100) { + const items = getInventoryItems(this.bot); + getSurroundingBlocks(this.bot, 8, 2, 8).forEach((block) => { + if (!items.has(block)) this.records.add(block); + }); + this.tick = 0; + } + }); + } + + observe() { + return Array.from(this.records); + } + + reset() { + this.records = new Set(); + } +} + +function getSurroundingBlocks(bot, x_distance, y_distance, z_distance) { + const surroundingBlocks = new Set(); + + for (let x = -x_distance; x <= x_distance; x++) { + for (let y = -y_distance; y <= y_distance; y++) { + for (let z = -z_distance; z <= z_distance; z++) { + const block = bot.blockAt(bot.entity.position.offset(x, y, z)); + if (block && block.type !== 0) { + surroundingBlocks.add(block.name); + } + } + } + } + // console.log(surroundingBlocks); + return surroundingBlocks; +} + +function getInventoryItems(bot) { + const items = new Set(); + bot.inventory.items().forEach((item) => { + if (item) items.add(item.name); + }); + return items; +} + +module.exports = { Voxels, BlockRecords }; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/skillLoader.js b/metagpt/environment/mincraft_env/mineflayer/lib/skillLoader.js new file mode 100644 index 000000000..d78cf7820 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/skillLoader.js @@ -0,0 +1,79 @@ +function inject(bot) { + bot._sleep = bot.sleep; + bot.sleep = async (bedBlock) => { + await bot.waitForTicks(20); + await bot._sleep(bedBlock); + await bot.waitForTicks(135); + }; + + bot._fish = bot.fish; + bot.fish = async () => { + if (bot.heldItem?.name !== "fishing_rod") { + bot.chat("I'm not holding a fishing rod!"); + return; + } + let timeout = null; + await Promise.race([ + bot._fish(), + new Promise( + (resolve, reject) => + (timeout = setTimeout(() => { + bot.activateItem(); + reject( + new Error( + "Finishing timeout, make sure you get to and look at a water block!" + ) + ); + }, 60000)) + ), + ]); + clearTimeout(timeout); + await bot.waitForTicks(20); + }; + + bot._consume = bot.consume; + bot.consume = async () => { + // action_count.activateItem++; + await bot._consume(); + await bot.waitForTicks(20); + }; + + bot._useOn = bot.useOn; + bot.useOn = async (entity) => { + if (entity.position.distanceTo(bot.entity.position) > 6) { + bot.chat("Please goto a place near the entity first!"); + return; + } + await bot._useOn(entity); + await bot.waitForTicks(20); + }; + + bot._activateBlock = bot.activateBlock; + bot.activateBlock = async (block) => { + if (block.position.distanceTo(bot.entity.position) > 6) { + bot.chat("Please goto a place near the block first!"); + return; + } + // action_count.activateBlock++; + await bot._activateBlock(block); + }; + + bot._chat = bot.chat; + bot.chat = (message) => { + // action_count.chat++; + bot.emit("chatEvent", "bot", message); + bot._chat(message); + }; + + bot.inventoryUsed = () => { + return bot.inventory.slots.slice(9, 45).filter((item) => item !== null) + .length; + }; + + bot.save = function (eventName) { + bot.emit("save", eventName); + }; +} + +// export all control_primitives +module.exports = { inject }; diff --git a/metagpt/environment/mincraft_env/mineflayer/lib/utils.js b/metagpt/environment/mincraft_env/mineflayer/lib/utils.js new file mode 100644 index 000000000..68af30796 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/lib/utils.js @@ -0,0 +1,31 @@ +let gameTimeCounter = 0; +let gameTimeList = []; +const initCounter = (bot) => { + gameTimeList = []; + for (let i = 0; i < 13000; i += 1000) { + gameTimeList.push(i); + } + for (let i = 13000; i < 24000; i += 2000) { + gameTimeList.push(i); + } + const timeOfDay = bot.time.timeOfDay; + for (let i = 0; i < gameTimeList.length; i++) { + if (gameTimeList[i] > timeOfDay) { + gameTimeCounter = i - 1; + break; + } + } +}; + +const getNextTime = () => { + gameTimeCounter++; + if (gameTimeCounter >= gameTimeList.length) { + gameTimeCounter = 0; + } + return gameTimeList[gameTimeCounter]; +}; + +module.exports = { + initCounter, + getNextTime, +}; diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/.gitignore b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/.gitignore new file mode 100644 index 000000000..0578fdca3 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/.gitignore @@ -0,0 +1,107 @@ +# Logs +logs +*.log +npm-debug.log* +yarn-debug.log* +yarn-error.log* +lerna-debug.log* + +# Diagnostic reports (https://nodejs.org/api/report.html) +report.[0-9]*.[0-9]*.[0-9]*.[0-9]*.json + +# Runtime data +pids +*.pid +*.seed +*.pid.lock + +# Directory for instrumented libs generated by jscoverage/JSCover +lib-cov + +# Coverage directory used by tools like istanbul +coverage +*.lcov + +# nyc test coverage +.nyc_output + +# Grunt intermediate storage (https://gruntjs.com/creating-plugins#storing-task-files) +.grunt + +# Bower dependency directory (https://bower.io/) +bower_components + +# node-waf configuration +.lock-wscript + +# Compiled binary addons (https://nodejs.org/api/addons.html) +build/Release + +# Dependency directories +node_modules/ +jspm_packages/ + +# TypeScript v1 declaration files +typings/ + +# TypeScript cache +*.tsbuildinfo + +# Optional npm cache directory +.npm + +# Optional eslint cache +.eslintcache + +# Microbundle cache +.rpt2_cache/ +.rts2_cache_cjs/ +.rts2_cache_es/ +.rts2_cache_umd/ + +# Optional REPL history +.node_repl_history + +# Output of 'npm pack' +*.tgz + +# Yarn Integrity file +.yarn-integrity + +# dotenv environment variables file +.env +.env.test + +# parcel-bundler cache (https://parceljs.org/) +.cache + +# Next.js build output +.next + +# Nuxt.js build / generate output +.nuxt +dist + +# Gatsby files +.cache/ +# Comment in the public line in if your project uses Gatsby and *not* Next.js +# https://nextjs.org/blog/next-9-1#public-directory-support +# public + +# vuepress build output +.vuepress/dist + +# Serverless directories +.serverless/ + +# FuseBox cache +.fusebox/ + +# DynamoDB Local files +.dynamodb/ + +# TernJS port file +.tern-port + +lib/ +package-lock.json diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/LICENSE b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/LICENSE new file mode 100644 index 000000000..f2896b56e --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/LICENSE @@ -0,0 +1,21 @@ +MIT License + +Copyright (c) 2020 TheDudeFromCI + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to deal +in the Software without restriction, including without limitation the rights +to use, copy, modify, merge, publish, distribute, sublicense, and/or sell +copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/README.md b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/README.md new file mode 100644 index 000000000..555acb761 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/README.md @@ -0,0 +1,89 @@ +

mineflayer-collectblock

+

A small utility plugin for allowing users to collect blocks using a higher level API.

+ +

+ + + + + + +

+ +--- +## This is a modified version to better support Voyager + +## Showcase + +You can see a video of the plugin in action, [here.](https://youtu.be/5T_rcCnNnf4) +The source code of the bot in the video can be seen in the examples folder, [here.](https://github.com/TheDudeFromCI/mineflayer-collectblock/blob/master/examples/collector.js) + +### Description + +This plugin is a wrapper for mineflayer that allows for easier API usage when collecting blocks or item drops. This plugin is designed to reduce some of the boilerplate code based around the act of pathfinding to a block _(handled by_ ***mineflayer-pathfinder***_)_, selecting the best tool to mine that block _(handled by_ ***mineflayer-tool***_)_, actually mining it, then moving to collect the item drops from that block. This plugin allows for all of that basic concept to be wrapped up into a single API function. + +In addition to the usage above, some additional quality of life features are available in this plugin. These include the ability to automatically deposit items into a chest when the bot's inventory is full, collecting new tools from a chest if the bot doesn't currently have a required tool _(also handled by_ ***mineflayer-tool***_)_, and allowing for queueing of multiple blocks or item drops to the collection task, so they can be processed later. + +### Getting Started + +This plugin is built using Node and can be installed using: +```bash +npm install --save mineflayer-collectblock +``` + +### Simple Bot + +The brief description goes here. + +```js +// Create your bot +const mineflayer = require("mineflayer") +const bot = mineflayer.createBot({ + host: 'localhost', + username: 'Player', +}) +let mcData + +// Load collect block +bot.loadPlugin(require('mineflayer-collectblock').plugin) + +async function collectGrass() { + // Find a nearby grass block + const grass = bot.findBlock({ + matching: mcData.blocksByName.grass_block.id, + maxDistance: 64 + }) + + if (grass) { + // If we found one, collect it. + try { + await bot.collectBlock.collect(grass) + collectGrass() // Collect another grass block + } catch (err) { + console.log(err) // Handle errors, if any + } + } +} + +// On spawn, start collecting all nearby grass +bot.once('spawn', () => { + mcData = require('minecraft-data')(bot.version) + collectGrass() +}) +``` + +### Documentation + +[API](https://github.com/TheDudeFromCI/mineflayer-collectblock/blob/master/docs/api.md) + +[Examples](https://github.com/TheDudeFromCI/mineflayer-collectblock/tree/master/examples) + +### License + +This project uses the [MIT](https://github.com/TheDudeFromCI/mineflayer-collectblock/blob/master/LICENSE) license. + +### Contributions + +This project is accepting PRs and Issues. See something you think can be improved? Go for it! Any and all help is highly appreciated! + +For larger changes, it is recommended to discuss these changes in the issues tab before writing any code. It's also preferred to make many smaller PRs than one large one, where applicable. diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/_config.yml b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/_config.yml new file mode 100644 index 000000000..c4192631f --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/_config.yml @@ -0,0 +1 @@ +theme: jekyll-theme-cayman \ No newline at end of file diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/docs/api.md b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/docs/api.md new file mode 100644 index 000000000..66d8a3ecc --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/docs/api.md @@ -0,0 +1,52 @@ +# API + +Welcome to the *mineflayer-collectblock* API documentation page. + +## Table of Contents + +- [1. Summary](#1-summary) +- [Properties](#properties) + - [`bot.collectblock.movements: Movements`](#botcollectblockmovements-movements) +- [Functions](#functions) + - [collect](#collect) + - [Options:](#options) + +## 1. Summary + +The collect block plugin is a utility plugin that can be used to help make collecting blocks and item drops very easy, using only a single API call. No need to worry about pathfinding to the block, selecting the right tool, or moving to pick up the item drop after mining. + +## Properties + +### `bot.collectblock.movements: Movements` + +The movements object used by the pathfinder plugin to define the movement configuration. This object is passed to the pathfinder plugin when any API from this plugin is called in order to control how pathfinding should work when collecting the given blocks or item. + +If set to null, the pathfinder plugin movements is not updated. + +Defaults to a new movements object instance. + +## Functions + +### collect + +Usage: `bot.collectblock.collect(target: Collectable | Collectable[], options?: CollectOptions, cb: (err?: Error) => void): void` + +Causes the bot to collect the given block, item drop, or list of those. If the target is a block, the bot will move to the block, mine it, and pick up the item drop. If the target is an item drop, the bot will move to the item drop and pick it up. If the target is a list of collectables, the bot will move from target to target in order of closest to furthest and collect each target in turn. + +#### Options: + + * `append: boolean` + + If true, the target(s) will be appended to the existing target list instead of starting a new task. Defaults to false. + + * `ignoreNoPath: boolean` + + If true, errors will not be thrown when a path to the target block cannot be found. The bot will attempt to choose the best available position it can find, instead. Errors are still thrown if the bot cannot interact with the block from it's final location. Defaults to false. + + * `chestLocations: Vec3[]` + + Gets the list of chest locations to use when storing items after the bot's inventory becomes full. If undefined, it defaults to the chest location list on the bot.collectBlock plugin. + + * `itemFilter: ItemFilter` + + When transferring items to a chest, this filter is used to determine what items are allowed to be moved, and what items aren't allowed to be moved. Defaults to the item filter specified on the bot.collectBlock plugin. \ No newline at end of file diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/collector.js b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/collector.js new file mode 100644 index 000000000..b9bb8faf9 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/collector.js @@ -0,0 +1,70 @@ +/** + * This bot example show how to direct a bot to collect a specific block type + * or a group of nearby blocks of that type. + */ + +const mineflayer = require('mineflayer') +const collectBlock = require('mineflayer-collectblock').plugin + +if (process.argv.length < 4 || process.argv.length > 6) { + console.log('Usage : node collector.js [] []') + process.exit(1) +} + +const bot = mineflayer.createBot({ + host: process.argv[2], + port: process.argv[3], + username: process.argv[4] || 'collector', + password: process.argv[5] +}) + +bot.loadPlugin(collectBlock) + +let mcData +bot.once('spawn', () => { + mcData = require('minecraft-data')(bot.version) +}) + +bot.on('chat', async (username, message) => { + const args = message.split(' ') + if (args[0] !== 'collect') return + + let count = 1 + if (args.length === 3) count = parseInt(args[1]) + + let type = args[1] + if (args.length === 3) type = args[2] + + const blockType = mcData.blocksByName[type] + if (!blockType) { + return + } + + const blocks = bot.findBlocks({ + matching: blockType.id, + maxDistance: 64, + count: count + }) + + if (blocks.length === 0) { + bot.chat("I don't see that block nearby.") + return + } + + const targets = [] + for (let i = 0; i < Math.min(blocks.length, count); i++) { + targets.push(bot.blockAt(blocks[i])) + } + + bot.chat(`Found ${targets.length} ${type}(s)`) + + try { + await bot.collectBlock.collect(targets) + // All blocks have been collected. + bot.chat('Done') + } catch (err) { + // An error occurred, report it. + bot.chat(err.message) + console.log(err) + } +}) diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/oreMiner.js b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/oreMiner.js new file mode 100644 index 000000000..6accac88f --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/oreMiner.js @@ -0,0 +1,59 @@ +/** + * This bot example shows how to collect a vein of ores quickly after only finding a single block. + * This makes it easy to collect a vein of ores or mine a tree without looking for every block in the + * area. + */ + +const mineflayer = require('mineflayer') +const collectBlock = require('mineflayer-collectblock').plugin + +if (process.argv.length < 4 || process.argv.length > 6) { + console.log('Usage : node oreMiner.js [] []') + process.exit(1) +} + +const bot = mineflayer.createBot({ + host: process.argv[2], + port: process.argv[3], + username: process.argv[4] || 'oreMiner', + password: process.argv[5] +}) + +bot.loadPlugin(collectBlock) + +let mcData +bot.once('spawn', () => { + mcData = require('minecraft-data')(bot.version) +}) + +bot.on('chat', async (username, message) => { + const args = message.split(' ') + if (args[0] !== 'collect') return + + const blockType = mcData.blocksByName[args[1]] + if (!blockType) { + bot.chat(`I don't know any blocks named ${args[1]}.`) + return + } + + const block = bot.findBlock({ + matching: blockType.id, + maxDistance: 64 + }) + + if (!block) { + bot.chat("I don't see that block nearby.") + return + } + + const targets = bot.collectBlock.findFromVein(block) + try { + await bot.collectBlock.collect(targets) + // All blocks have been collected. + bot.chat('Done') + } catch (err) { + // An error occurred, report it. + bot.chat(err.message) + console.log(err) + } +}) diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/storageBot.js b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/storageBot.js new file mode 100644 index 000000000..b6f9971f2 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/examples/storageBot.js @@ -0,0 +1,107 @@ +/** + * This bot example shows how to use the chest filling mechanic of the plugin. + * Simply provide a given storage chest, and the bot will automatically try and + * store it's inventory in that chest when the bot's inventory becomes full. + */ + +if (process.argv.length < 4 || process.argv.length > 6) { + console.log('Usage : node storageBot.js [] []') + process.exit(1) +} + +// Load your libraries +const mineflayer = require('mineflayer') +const collectBlock = require('mineflayer-collectblock').plugin + +// Create your bot +const bot = mineflayer.createBot({ + host: process.argv[2], + port: parseInt(process.argv[3]), + username: process.argv[4] ? process.argv[4] : 'storageBot', + password: process.argv[5] +}) + +// Load the collect block plugin +bot.loadPlugin(collectBlock) + +// Load mcData on login +let mcData +bot.once('login', () => { + mcData = require('minecraft-data')(bot.version) +}) + +// On spawn, try to find any nearby chests and save those as storage locations. +// When the bot's inventory becomes too full, it will empty it's inventory into +// these chests before collecting more resources. If a chest gets full, it moves +// to the next one in order until it's inventory is empty or it runs out of chests. +bot.once('spawn', () => { + bot.collectBlock.chestLocations = bot.findBlocks({ + matching: mcData.blocksByName.chest.id, + maxDistance: 16, + count: 999999 // Get as many chests as we can + }) + + if (bot.collectBlock.chestLocations.length === 0) { + bot.chat("I don't see any chests nearby.") + } else { + for (const chestPos of bot.collectBlock.chestLocations) { + bot.chat(`I found a chest at ${chestPos}`) + } + } +}) + +// Wait for someone to say something +bot.on('chat', async (username, message) => { + // If the player says something start starts with "collect" + // Otherwise, do nothing + const args = message.split(' ') + if (args[0] !== 'collect') return + + // If the player specifies a number, collect that many. Otherwise, default to 1. + let count = 1 + if (args.length === 3) count = parseInt(args[1]) + + // If a number was given the item number is the 3rd arg, not the 2nd. + let type = args[1] + if (args.length === 3) type = args[2] + + // Get the id of that block type for this version of Minecraft. + const blockType = mcData.blocksByName[type] + if (!blockType) { + bot.chat(`I don't know any blocks named ${type}.`) + return + } + + // Find all nearby blocks of that type, up to the given count, within 64 blocks. + const blocks = bot.findBlocks({ + matching: blockType.id, + maxDistance: 64, + count: count + }) + + // Complain if we can't find any nearby blocks of that type. + if (blocks.length === 0) { + bot.chat("I don't see that block nearby.") + return + } + + // Convert the block position array into a block array to pass to collect block. + const targets = [] + for (let i = 0; i < Math.min(blocks.length, count); i++) { + targets.push(bot.blockAt(blocks[i])) + } + + // Announce what we found. + bot.chat(`Found ${targets.length} ${type}(s)`) + + // Tell the bot to collect all of the given blocks in the block list. + try { + await bot.collectBlock.collect(targets) + // All blocks have been collected. + bot.chat('Done') + } catch (err) { + // An error occurred, report it. + bot.chat(err.message) + console.log(err) + } +}) diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/package.json b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/package.json new file mode 100644 index 000000000..0f59e7aa6 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/package.json @@ -0,0 +1,44 @@ +{ + "name": "mineflayer-collectblock", + "version": "1.4.1", + "description": "A simple utility plugin for Mineflayer that add a higher level API for collecting blocks.", + "main": "lib/index.js", + "types": "lib/index.d.ts", + "scripts": { + "build": "ts-standard && tsc && require-self", + "clean": "rm -rf lib", + "test": "test" + }, + "repository": { + "type": "git", + "url": "git+https://github.com/TheDudeFromCI/mineflayer-collectblock.git" + }, + "keywords": [ + "mineflayer", + "plugin", + "api", + "utility", + "helper", + "collect" + ], + "author": "TheDudeFromCI", + "license": "MIT", + "bugs": { + "url": "https://github.com/TheDudeFromCI/mineflayer-collectblock/issues" + }, + "homepage": "https://github.com/TheDudeFromCI/mineflayer-collectblock#readme", + "dependencies": { + "mineflayer": "^4.0.0", + "mineflayer-pathfinder": "^2.1.1", + "mineflayer-tool": "^1.1.0" + }, + "devDependencies": { + "@types/node": "^18.6.4", + "require-self": "^0.2.3", + "ts-standard": "^11.0.0", + "typescript": "^4.1.3" + }, + "files": [ + "lib/**/*" + ] +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/BlockVeins.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/BlockVeins.ts new file mode 100644 index 000000000..ae5542ce3 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/BlockVeins.ts @@ -0,0 +1,35 @@ +import { Bot } from 'mineflayer' +import { Block } from 'prismarine-block' + +export function findFromVein (bot: Bot, block: Block, maxBlocks: number, maxDistance: number, floodRadius: number): Block[] { + const targets: Block[] = [] + const open: Block[] = [block] + const type = block.type + const center = block.position + + for (let i = 0; i < maxBlocks; i++) { + const next = open.pop() + if (next == null) break + + targets.push(next) + + for (let x = -floodRadius; x <= floodRadius; x++) { + for (let y = -floodRadius; y <= floodRadius; y++) { + for (let z = -floodRadius; z <= floodRadius; z++) { + const neighborPos = next.position.offset(x, y, z) + if (neighborPos.manhattanDistanceTo(center) > maxDistance) continue + + const neighbor = bot.blockAt(neighborPos) + if (neighbor == null || neighbor.type !== type) continue + + if (targets.includes(neighbor)) continue + if (open.includes(neighbor)) continue + + open.push(neighbor) + } + } + } + } + + return targets +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/CollectBlock.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/CollectBlock.ts new file mode 100644 index 000000000..d2be87822 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/CollectBlock.ts @@ -0,0 +1,451 @@ +import { Bot } from "mineflayer"; +import { Block } from "prismarine-block"; +import { Movements, goals } from "mineflayer-pathfinder"; +import { TemporarySubscriber } from "./TemporarySubscriber"; +import { Entity } from "prismarine-entity"; +import { error } from "./Util"; +import { Vec3 } from "vec3"; +import { emptyInventoryIfFull, ItemFilter } from "./Inventory"; +import { findFromVein } from "./BlockVeins"; +import { Collectable, Targets } from "./Targets"; +import { Item } from "prismarine-item"; +import mcDataLoader from "minecraft-data"; +import { once } from "events"; +import { callbackify } from "util"; + +export type Callback = (err?: Error) => void; + +async function collectAll( + bot: Bot, + options: CollectOptionsFull +): Promise { + let success_count = 0; + while (!options.targets.empty) { + await emptyInventoryIfFull( + bot, + options.chestLocations, + options.itemFilter + ); + const closest = options.targets.getClosest(); + if (closest == null) break; + switch (closest.constructor.name) { + case "Block": { + try { + if (success_count >= options.count) { + break; + } + await bot.tool.equipForBlock( + closest as Block, + equipToolOptions + ); + const goal = new goals.GoalLookAtBlock( + closest.position, + bot.world + ); + await bot.pathfinder.goto(goal); + await mineBlock(bot, closest as Block, options); + success_count++; + // TODO: options.ignoreNoPath + } catch (err) { + // @ts-ignore + // console.log(err.stack) + // bot.pathfinder.stop() + // bot.waitForTicks(10) + try { + bot.pathfinder.setGoal(null); + } catch (err) {} + if (options.ignoreNoPath) { + // @ts-ignore + if (err.name === "Invalid block") { + console.log( + `Block ${closest.name} at ${closest.position} is not valid! Skip it!` + ); + } // @ts-ignore + else if (err.name === "Unsafe block") { + console.log( + `${closest.name} at ${closest.position} is not safe to break! Skip it!` + ); + // @ts-ignore + } else if (err.name === "NoItem") { + const properties = + bot.registry.blocksByName[closest.name]; + const leastTool = Object.keys( + properties.harvestTools + )[0]; + const item = bot.registry.items[leastTool]; + bot.chat( + `I need at least a ${item.name} to mine ${closest.name}! Skip it!` + ); + return; + } else if ( + // @ts-ignore + err.name === "NoPath" || + // @ts-ignore + err.name === "Timeout" + ) { + if ( + bot.entity.position.distanceTo( + closest.position + ) < 0.5 + ) { + await mineBlock(bot, closest as Block, options); + break; + } + console.log( + `No path to ${closest.name} at ${closest.position}! Skip it!` + ); + // @ts-ignore + } else if (err.message === "Digging aborted") { + console.log(`Digging aborted! Skip it!`); + } else { + // @ts-ignore + bot.chat(`Error: ${err.message}`); + } + break; + } + throw err; + } + break; + } + case "Entity": { + // Don't collect any entities that are marked as 'invalid' + if (!(closest as Entity).isValid) break; + try { + const tempEvents = new TemporarySubscriber(bot); + const waitForPickup = new Promise( + (resolve, reject) => { + const timeout = setTimeout(() => { + // After 10 seconds, reject the promise + clearTimeout(timeout); + tempEvents.cleanup(); + reject(new Error("Failed to pickup item")); + }, 10000); + tempEvents.subscribeTo( + "entityGone", + (entity: Entity) => { + if (entity === closest) { + clearTimeout(timeout); + tempEvents.cleanup(); + resolve(); + } + } + ); + } + ); + bot.pathfinder.setGoal( + new goals.GoalFollow(closest as Entity, 0) + ); + // await bot.pathfinder.goto(new goals.GoalBlock(closest.position.x, closest.position.y, closest.position.z)) + await waitForPickup; + } catch (err) { + // @ts-ignore + console.log(err.stack); + try { + bot.pathfinder.setGoal(null); + } catch (err) {} + if (options.ignoreNoPath) { + // @ts-ignore + if (err.message === "Failed to pickup item") { + bot.chat(`Failed to pickup item! Skip it!`); + } + break; + } + throw err; + } + break; + } + default: { + throw error( + "UnknownType", + `Target ${closest.constructor.name} is not a Block or Entity!` + ); + } + } + options.targets.removeTarget(closest); + } + bot.chat(`Collect finish!`); +} + +const equipToolOptions = { + requireHarvest: true, + getFromChest: false, + maxTools: 2, +}; + +async function mineBlock( + bot: Bot, + block: Block, + options: CollectOptionsFull +): Promise { + if ( + bot.blockAt(block.position)?.type !== block.type || + bot.blockAt(block.position)?.type === 0 + ) { + options.targets.removeTarget(block); + throw error("Invalid block", "Block is not valid!"); + // @ts-expect-error + } else if (!bot.pathfinder.movements.safeToBreak(block)) { + options.targets.removeTarget(block); + throw error("Unsafe block", "Block is not safe to break!"); + } + + await bot.tool.equipForBlock(block, equipToolOptions); + + if (!block.canHarvest(bot.heldItem ? bot.heldItem.type : bot.heldItem)) { + options.targets.removeTarget(block); + throw error("NoItem", "Bot does not have a harvestable tool!"); + } + + const tempEvents = new TemporarySubscriber(bot); + tempEvents.subscribeTo("itemDrop", (entity: Entity) => { + if ( + entity.position.distanceTo(block.position.offset(0.5, 0.5, 0.5)) <= + 0.5 + ) { + options.targets.appendTarget(entity); + } + }); + try { + await bot.dig(block); + // Waiting for items to drop + await new Promise((resolve) => { + let remainingTicks = 10; + tempEvents.subscribeTo("physicTick", () => { + remainingTicks--; + if (remainingTicks <= 0) { + tempEvents.cleanup(); + resolve(); + } + }); + }); + } finally { + tempEvents.cleanup(); + } +} + +/** + * A set of options to apply when collecting the given targets. + */ +export interface CollectOptions { + /** + * If true, the target(s) will be appended to the existing target list instead of + * starting a new task. Defaults to false. + */ + append?: boolean; + + /** + * If true, errors will not be thrown when a path to the target block cannot + * be found. The bot will attempt to choose the best available position it + * can find, instead. Errors are still thrown if the bot cannot interact with + * the block from it's final location. Defaults to false. + */ + ignoreNoPath?: boolean; + + /** + * Gets the list of chest locations to use when storing items after the bot's + * inventory becomes full. If undefined, it defaults to the chest location + * list on the bot.collectBlock plugin. + */ + chestLocations?: Vec3[]; + + /** + * When transferring items to a chest, this filter is used to determine what + * items are allowed to be moved, and what items aren't allowed to be moved. + * Defaults to the item filter specified on the bot.collectBlock plugin. + */ + itemFilter?: ItemFilter; + + /** + * The total number of items to collect + */ + count?: number; +} + +/** + * A version of collect options where all values are assigned. + */ +interface CollectOptionsFull { + append: boolean; + ignoreNoPath: boolean; + chestLocations: Vec3[]; + itemFilter: ItemFilter; + targets: Targets; + count: number; +} + +/** + * The collect block plugin. + */ +export class CollectBlock { + /** + * The bot. + */ + private readonly bot: Bot; + + /** + * The list of active targets being collected. + */ + private readonly targets: Targets; + + /** + * The movements configuration to be sent to the pathfinder plugin. + */ + movements?: Movements; + + /** + * A list of chest locations which the bot is allowed to empty their inventory into + * if it becomes full while the bot is collecting resources. + */ + chestLocations: Vec3[] = []; + + /** + * When collecting items, this filter is used to determine what items should be placed + * into a chest if the bot's inventory becomes full. By default, returns true for all + * items except for tools, weapons, and armor. + * + * @param item - The item stack in the bot's inventory to check. + * + * @returns True if the item should be moved into the chest. False otherwise. + */ + itemFilter: ItemFilter = (item: Item) => { + if (item.name.includes("helmet")) return false; + if (item.name.includes("chestplate")) return false; + if (item.name.includes("leggings")) return false; + if (item.name.includes("boots")) return false; + if (item.name.includes("shield")) return false; + if (item.name.includes("sword")) return false; + if (item.name.includes("pickaxe")) return false; + if (item.name.includes("axe")) return false; + if (item.name.includes("shovel")) return false; + if (item.name.includes("hoe")) return false; + return true; + }; + + /** + * Creates a new instance of the create block plugin. + * + * @param bot - The bot this plugin is acting on. + */ + constructor(bot: Bot) { + this.bot = bot; + this.targets = new Targets(bot); + // @ts-ignore + this.movements = new Movements(bot, mcDataLoader(bot.version)); + } + + /** + * If target is a block: + * Causes the bot to break and collect the target block. + * + * If target is an item drop: + * Causes the bot to collect the item drop. + * + * If target is an array containing items or blocks, preforms the correct action for + * all targets in that array sorting dynamically by distance. + * + * @param target - The block(s) or item(s) to collect. + * @param options - The set of options to use when handling these targets + * @param cb - The callback that is called finished. + */ + async collect( + target: Collectable | Collectable[], + options: CollectOptions | Callback = {}, + cb?: Callback + ): Promise { + if (typeof options === "function") { + cb = options; + options = {}; + } + // @ts-expect-error + if (cb != null) return callbackify(this.collect)(target, options, cb); + + const optionsFull: CollectOptionsFull = { + append: options.append ?? false, + ignoreNoPath: options.ignoreNoPath ?? false, + chestLocations: options.chestLocations ?? this.chestLocations, + itemFilter: options.itemFilter ?? this.itemFilter, + targets: this.targets, + count: options.count ?? Infinity, + }; + + if (this.bot.pathfinder == null) { + throw error( + "UnresolvedDependency", + "The mineflayer-collectblock plugin relies on the mineflayer-pathfinder plugin to run!" + ); + } + + if (this.bot.tool == null) { + throw error( + "UnresolvedDependency", + "The mineflayer-collectblock plugin relies on the mineflayer-tool plugin to run!" + ); + } + + if (this.movements != null) { + this.bot.pathfinder.setMovements(this.movements); + } + + if (!optionsFull.append) await this.cancelTask(); + if (Array.isArray(target)) { + this.targets.appendTargets(target); + } else { + this.targets.appendTarget(target); + } + + try { + await collectAll(this.bot, optionsFull); + this.targets.clear(); + } catch (err) { + this.targets.clear(); + // Ignore path stopped error for cancelTask to work properly (imo we shouldn't throw any pathing errors) + // @ts-expect-error + if (err.name !== "PathStopped") throw err; + } finally { + // @ts-expect-error + this.bot.emit("collectBlock_finished"); + } + } + + /** + * Loads all touching blocks of the same type to the given block and returns them as an array. + * This effectively acts as a flood fill algorithm to retrieve blocks in the same ore vein and similar. + * + * @param block - The starting block. + * @param maxBlocks - The maximum number of blocks to look for before stopping. + * @param maxDistance - The max distance from the starting block to look. + * @param floodRadius - The max distance distance from block A to block B to be considered "touching" + */ + findFromVein( + block: Block, + maxBlocks = 100, + maxDistance = 16, + floodRadius = 1 + ): Block[] { + return findFromVein( + this.bot, + block, + maxBlocks, + maxDistance, + floodRadius + ); + } + + /** + * Cancels the current collection task, if still active. + * + * @param cb - The callback to use when the task is stopped. + */ + async cancelTask(cb?: Callback): Promise { + if (this.targets.empty) { + if (cb != null) cb(); + return await Promise.resolve(); + } + this.bot.pathfinder.stop(); + if (cb != null) { + // @ts-expect-error + this.bot.once("collectBlock_finished", cb); + } + await once(this.bot, "collectBlock_finished"); + } +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Inventory.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Inventory.ts new file mode 100644 index 000000000..6a17d0cc5 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Inventory.ts @@ -0,0 +1,87 @@ +import { Bot } from 'mineflayer' +import { Callback } from './CollectBlock' +import { Vec3 } from 'vec3' +import { error } from './Util' +import { Item } from 'prismarine-item' +import { goals } from 'mineflayer-pathfinder' +import { callbackify } from 'util' + +export type ItemFilter = (item: Item) => boolean + +function getClosestChest (bot: Bot, chestLocations: Vec3[]): Vec3 | null { + let chest = null + let distance = 0 + + for (const c of chestLocations) { + const dist = c.distanceTo(bot.entity.position) + if (chest == null || dist < distance) { + chest = c + distance = dist + } + } + + if (chest != null) { + chestLocations.splice(chestLocations.indexOf(chest), 1) + } + + return chest +} + +export async function emptyInventoryIfFull (bot: Bot, chestLocations: Vec3[], itemFilter: ItemFilter, cb?: Callback): Promise { + // @ts-expect-error + if (cb != null) return callbackify(emptyInventoryIfFull)(bot, chestLocations, cb) + if (bot.inventory.emptySlotCount() > 0) return + return await emptyInventory(bot, chestLocations, itemFilter) +} + +export async function emptyInventory (bot: Bot, chestLocations: Vec3[], itemFilter: ItemFilter, cb?: Callback): Promise { + // @ts-expect-error + if (cb != null) return callbackify(emptyInventory)(bot, chestLocations, cb) + if (chestLocations.length === 0) { + throw error('NoChests', 'There are no defined chest locations!') + } + + // Shallow clone so we can safely remove chests from the list that are full. + chestLocations = [...chestLocations] + + while (true) { + const chest = getClosestChest(bot, chestLocations) + if (chest == null) { + throw error('NoChests', 'All chests are full.') + } + const hasRemaining = await tryEmptyInventory(bot, chest, itemFilter) + if (!hasRemaining) return + } +} + +async function tryEmptyInventory (bot: Bot, chestLocation: Vec3, itemFilter: ItemFilter, cb?: (err: Error | undefined, hasRemaining: boolean) => void): Promise { + // @ts-expect-error + if (cb != null) return callbackify(tryEmptyInventory)(bot, chestLocation, itemFilter, cb) + await gotoChest(bot, chestLocation) + return await placeItems(bot, chestLocation, itemFilter) +} + +async function gotoChest (bot: Bot, location: Vec3, cb?: Callback): Promise { + // @ts-expect-error + if (cb != null) return callbackify(gotoChest)(bot, location) + await bot.pathfinder.goto(new goals.GoalGetToBlock(location.x, location.y, location.z)) +} + +async function placeItems (bot: Bot, chestPos: Vec3, itemFilter: ItemFilter, cb?: (err: Error | undefined, hasRemaining: boolean) => void): Promise { + // @ts-expect-error + if (cb != null) return callbackify(placeItems)(bot, chestPos, itemFilter, cb) + const chestBlock = bot.blockAt(chestPos) + if (chestBlock == null) { + throw error('UnloadedChunk', 'Chest is in an unloaded chunk!') + } + const chest = await bot.openChest(chestBlock) + for (const item of bot.inventory.items()) { + if (!itemFilter(item)) continue + if (chest.firstEmptyContainerSlot() === null) { + // We have items that didn't fit. + return true + } + await chest.deposit(item.type, item.metadata, item.count) + } + return false +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Targets.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Targets.ts new file mode 100644 index 000000000..568d07ad9 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Targets.ts @@ -0,0 +1,60 @@ +import { Bot } from 'mineflayer' +import { Block } from 'prismarine-block' +import { Entity } from 'prismarine-entity' + +export type Collectable = Block | Entity + +export class Targets { + private readonly bot: Bot + private targets: Collectable[] = [] + + constructor (bot: Bot) { + this.bot = bot + } + + appendTargets (targets: Collectable[]): void { + for (const target of targets) { + this.appendTarget(target) + } + } + + appendTarget (target: Collectable): void { + if (this.targets.includes(target)) return + this.targets.push(target) + } + + /** + * Gets the closest target to the bot in this list. + * + * @returns The closest target, or null if there are no targets. + */ + getClosest (): Collectable | null { + let closest: Collectable | null = null + let distance: number = 0 + + for (const target of this.targets) { + const dist = target.position.distanceTo(this.bot.entity.position) + + if (closest == null || dist < distance) { + closest = target + distance = dist + } + } + + return closest + } + + get empty (): boolean { + return this.targets.length === 0 + } + + clear (): void { + this.targets.length = 0 + } + + removeTarget (target: Collectable): void { + const index = this.targets.indexOf(target) + if (index < 0) return + this.targets.splice(index, 1) + } +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TaskQueue.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TaskQueue.ts new file mode 100644 index 000000000..81fe3bc5a --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TaskQueue.ts @@ -0,0 +1,77 @@ +import type { Callback } from './index' +export type Task = (cb: Callback) => void +export type SyncTask = () => void + +/** + * A simple utility class for queuing up a series of async tasks to execute. + */ +export class TaskQueue { + private tasks: Task[] = [] + + /** + * If true, the task list will stop executing if one of the tasks throws an error. + */ + readonly stopOnError: boolean = true + + /** + * Adds a new async task to this queue. The provided callback should be executed when + * the async task is complete. + * + * @param task - The async task to add. + */ + add (task: Task): void { + this.tasks.push(task) + } + + /** + * Adds a synchronous task toi this queue. + * + * @param task - The sync task to add. + */ + addSync (task: SyncTask): void { + this.add((cb) => { + try { + task() + cb() + } catch (err: any) { + cb(err) + } + }) + } + + /** + * Runs all tasks currently in this queue and empties the queue. + * + * @param cb - The optional callback to be executed when all tasks in this queue have + * finished executing. + */ + runAll (cb?: Callback): void { + const taskList = this.tasks + this.tasks = [] + + let index = -1 + const runNext: () => void = () => { + index++ + if (index >= taskList.length) { + if (cb !== undefined) cb() + return + } + + try { + taskList[index]((err) => { + if (err !== undefined) { + if (cb !== undefined) cb(err) + + if (this.stopOnError) return + } + + runNext() + }) + } catch (err: any) { + if (cb !== undefined) cb(err) + } + } + + runNext() + } +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TemporarySubscriber.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TemporarySubscriber.ts new file mode 100644 index 000000000..3f14a607d --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/TemporarySubscriber.ts @@ -0,0 +1,34 @@ +import { Bot } from 'mineflayer' + +class Subscription { + constructor (readonly eventName: string, readonly callback: Function) {} +} + +export class TemporarySubscriber { + private readonly subscriptions: Subscription[] = [] + + constructor (readonly bot: Bot) {} + + /** + * Adds a new temporary event listener to the bot. + * + * @param event - The event to subscribe to. + * @param callback - The function to execute. + */ + subscribeTo (event: string, callback: Function): void { + this.subscriptions.push(new Subscription(event, callback)) + + // @ts-expect-error + this.bot.on(event, callback) + } + + /** + * Removes all attached event listeners from the bot. + */ + cleanup (): void { + for (const sub of this.subscriptions) { + // @ts-expect-error + this.bot.removeListener(sub.eventName, sub.callback) + } + } +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Util.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Util.ts new file mode 100644 index 000000000..ee0f29e0c --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/Util.ts @@ -0,0 +1,13 @@ +/** + * Creates a new error object with the given type and message. + * + * @param type - The error type. + * @param message - The error message. + * + * @returns The error object. + */ +export function error (type: string, message: string): Error { + const e = new Error(message) + e.name = type + return e +} diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/index.ts b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/index.ts new file mode 100644 index 000000000..45c9a8508 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/src/index.ts @@ -0,0 +1,25 @@ +import { Bot } from 'mineflayer' +import { CollectBlock } from './CollectBlock' +import { pathfinder as pathfinderPlugin } from 'mineflayer-pathfinder' +import { plugin as toolPlugin } from 'mineflayer-tool' + +export function plugin (bot: Bot): void { + // @ts-expect-error + bot.collectBlock = new CollectBlock(bot) + + // Load plugins if not loaded manually. + setTimeout(() => loadPathfinderPlugin(bot), 0) + setTimeout(() => loadToolPlugin(bot), 0) +} + +function loadPathfinderPlugin (bot: Bot): void { + if (bot.pathfinder != null) return + bot.loadPlugin(pathfinderPlugin) +} + +function loadToolPlugin (bot: Bot): void { + if (bot.tool != null) return + bot.loadPlugin(toolPlugin) +} + +export { CollectBlock, Callback, CollectOptions } from './CollectBlock' diff --git a/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/tsconfig.json b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/tsconfig.json new file mode 100644 index 000000000..a6076bc0c --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/mineflayer-collectblock/tsconfig.json @@ -0,0 +1,69 @@ +{ + "compilerOptions": { + /* Visit https://aka.ms/tsconfig.json to read more about this file */ + /* Basic Options */ + // "incremental": true, /* Enable incremental compilation */ + "target": "ES2015", /* Specify ECMAScript target version: 'ES3' (default), 'ES5', 'ES2015', 'ES2016', 'ES2017', 'ES2018', 'ES2019', 'ES2020', or 'ESNEXT'. */ + "module": "commonjs", /* Specify module code generation: 'none', 'commonjs', 'amd', 'system', 'umd', 'es2015', 'es2020', or 'ESNext'. */ + // "lib": [], /* Specify library files to be included in the compilation. */ + "allowJs": true, /* Allow javascript files to be compiled. */ + "checkJs": true, /* Report errors in .js files. */ + // "jsx": "preserve", /* Specify JSX code generation: 'preserve', 'react-native', or 'react'. */ + "declaration": true, + // "declarationMap": true, /* Generates a sourcemap for each corresponding '.d.ts' file. */ + // "sourceMap": true, /* Generates corresponding '.map' file. */ + // "outFile": "./", /* Concatenate and emit output to single file. */ + "outDir": "./lib", + // "rootDir": "./", /* Specify the root directory of input files. Use to control the output directory structure with --outDir. */ + // "composite": true, /* Enable project compilation */ + // "tsBuildInfoFile": "./", /* Specify file to store incremental compilation information */ + // "removeComments": true, /* Do not emit comments to output. */ + // "noEmit": true, /* Do not emit outputs. */ + // "importHelpers": true, /* Import emit helpers from 'tslib'. */ + // "downlevelIteration": true, /* Provide full support for iterables in 'for-of', spread, and destructuring when targeting 'ES5' or 'ES3'. */ + // "isolatedModules": true, /* Transpile each file as a separate module (similar to 'ts.transpileModule'). */ + /* Strict Type-Checking Options */ + "strict": true, /* Enable all strict type-checking options. */ + // "noImplicitAny": true, /* Raise error on expressions and declarations with an implied 'any' type. */ + "strictNullChecks": true, /* Enable strict null checks. */ + // "strictFunctionTypes": true, /* Enable strict checking of function types. */ + // "strictBindCallApply": true, /* Enable strict 'bind', 'call', and 'apply' methods on functions. */ + // "strictPropertyInitialization": true, /* Enable strict checking of property initialization in classes. */ + // "noImplicitThis": true, /* Raise error on 'this' expressions with an implied 'any' type. */ + "alwaysStrict": true, /* Parse in strict mode and emit "use strict" for each source file. */ + /* Additional Checks */ + "noUnusedLocals": true, /* Report errors on unused locals. */ + // "noUnusedParameters": true, /* Report errors on unused parameters. */ + "noImplicitReturns": true, /* Report error when not all code paths in function return a value. */ + // "noFallthroughCasesInSwitch": true, /* Report errors for fallthrough cases in switch statement. */ + /* Module Resolution Options */ + // "moduleResolution": "node", /* Specify module resolution strategy: 'node' (Node.js) or 'classic' (TypeScript pre-1.6). */ + // "baseUrl": "./", /* Base directory to resolve non-absolute module names. */ + // "paths": {}, /* A series of entries which re-map imports to lookup locations relative to the 'baseUrl'. */ + // "rootDirs": [], /* List of root folders whose combined content represents the structure of the project at runtime. */ + // "typeRoots": [], /* List of folders to include type definitions from. */ + // "types": [], /* Type declaration files to be included in compilation. */ + // "allowSyntheticDefaultImports": true, /* Allow default imports from modules with no default export. This does not affect code emit, just typechecking. */ + "esModuleInterop": true, /* Enables emit interoperability between CommonJS and ES Modules via creation of namespace objects for all imports. Implies 'allowSyntheticDefaultImports'. */ + // "preserveSymlinks": true, /* Do not resolve the real path of symlinks. */ + // "allowUmdGlobalAccess": true, /* Allow accessing UMD globals from modules. */ + /* Source Map Options */ + // "sourceRoot": "", /* Specify the location where debugger should locate TypeScript files instead of source locations. */ + // "mapRoot": "", /* Specify the location where debugger should locate map files instead of generated locations. */ + // "inlineSourceMap": true, /* Emit a single file with source maps instead of having a separate file. */ + // "inlineSources": true, /* Emit the source alongside the sourcemaps within a single file; requires '--inlineSourceMap' or '--sourceMap' to be set. */ + /* Experimental Options */ + // "experimentalDecorators": true, /* Enables experimental support for ES7 decorators. */ + // "emitDecoratorMetadata": true, /* Enables experimental support for emitting type metadata for decorators. */ + /* Advanced Options */ + "skipLibCheck": true, /* Skip type checking of declaration files. */ + "forceConsistentCasingInFileNames": true /* Disallow inconsistently-cased references to the same file. */ + }, + "include": [ + "src" + ], + "exclude": [ + "node_modules", + "**/__tests__/*" + ] +} \ No newline at end of file diff --git a/metagpt/environment/mincraft_env/mineflayer/package.json b/metagpt/environment/mincraft_env/mineflayer/package.json new file mode 100644 index 000000000..9e389d268 --- /dev/null +++ b/metagpt/environment/mincraft_env/mineflayer/package.json @@ -0,0 +1,38 @@ +{ + "name": "voyager", + "version": "1.0.0", + "description": "", + "main": "index.js", + "scripts": { + "test": "echo \"Error: no test specified\" && exit 1" + }, + "keywords": [], + "author": "", + "license": "ISC", + "dependencies": { + "body-parser": "^1.20.2", + "express": "^4.18.2", + "magic-string": "^0.30.0", + "minecraft-data": "^3.31.0", + "minecrafthawkeye": "^1.3.6", + "mineflayer": "^4.8.1", + "mineflayer-collectblock": "file:mineflayer-collectblock", + "mineflayer-pathfinder": "^2.4.2", + "mineflayer-pvp": "^1.3.2", + "mineflayer-tool": "^1.2.0", + "mocha": "^10.2.0", + "prismarine-biome": "^1.3.0", + "prismarine-block": "=1.16.3", + "prismarine-entity": "^2.2.0", + "prismarine-item": "^1.12.1", + "prismarine-nbt": "^2.2.1", + "prismarine-recipe": "^1.3.1", + "prismarine-viewer": "^1.24.0", + "typescript": "^4.9.5", + "vec3": "^0.1.8", + "graceful-fs": "^4.2.11" + }, + "devDependencies": { + "prettier": "2.8.5" + } +} diff --git a/metagpt/environment/mincraft_env/process_monitor.py b/metagpt/environment/mincraft_env/process_monitor.py new file mode 100644 index 000000000..b62aa6005 --- /dev/null +++ b/metagpt/environment/mincraft_env/process_monitor.py @@ -0,0 +1,79 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# refs to `voyager process_monitor.py` + +import re +import subprocess +import threading +import warnings +from typing import List + +import psutil + +from metagpt.logs import define_log_level + + +class SubprocessMonitor: + def __init__( + self, + commands: List[str], + name: str, + ready_match: str = r".*", + callback_match: str = r"^(?!x)x$", # regex that will never match + callback: callable = None, + finished_callback: callable = None, + ): + self.commands = commands + self.name = name + self.logger = define_log_level(name=name) + self.process = None + self.ready_match = ready_match + self.ready_event = None + self.ready_line = None + self.callback_match = callback_match + self.callback = callback + self.finished_callback = finished_callback + self.thread = None + + def _start(self): + self.logger.info(f"Starting subprocess with commands: {self.commands}") + + self.process = psutil.Popen( + self.commands, + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + universal_newlines=True, + ) + self.logger.info(f"Subprocess {self.name} started with PID {self.process.pid}.") + for line in iter(self.process.stdout.readline, ""): + self.logger.info(line.strip()) + if re.search(self.ready_match, line): + self.ready_line = line + self.logger.info("Subprocess is ready.") + self.ready_event.set() + if re.search(self.callback_match, line): + self.callback() + if not self.ready_event.is_set(): + self.ready_event.set() + warnings.warn(f"Subprocess {self.name} failed to start.") + if self.finished_callback: + self.finished_callback() + + def run(self): + self.ready_event = threading.Event() + self.ready_line = None + self.thread = threading.Thread(target=self._start) + self.thread.start() + self.ready_event.wait() + + def stop(self): + self.logger.info("Stopping subprocess.") + if self.process and self.process.is_running(): + self.process.terminate() + self.process.wait() + + @property + def is_running(self): + if self.process is None: + return False + return self.process.is_running() diff --git a/metagpt/environment/software_env/__init__.py b/metagpt/environment/software_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/software_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/software_env/software_env.py b/metagpt/environment/software_env/software_env.py new file mode 100644 index 000000000..94bc11659 --- /dev/null +++ b/metagpt/environment/software_env/software_env.py @@ -0,0 +1,12 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : MG Software Env + + +from metagpt.environment.base_env import Environment + + +class SoftwareEnv(Environment): + """a specific alias name""" + + pass diff --git a/metagpt/environment/stanford_town_env/__init__.py b/metagpt/environment/stanford_town_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/stanford_town_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/stanford_town_env/stanford_town_env.py b/metagpt/environment/stanford_town_env/stanford_town_env.py new file mode 100644 index 000000000..8721d6cd1 --- /dev/null +++ b/metagpt/environment/stanford_town_env/stanford_town_env.py @@ -0,0 +1,12 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : MG StanfordTown Env + +from metagpt.environment.base_env import Environment +from metagpt.environment.stanford_town_env.stanford_town_ext_env import ( + StanfordTownExtEnv, +) + + +class StanfordTownEnv(Environment, StanfordTownExtEnv): + pass diff --git a/metagpt/environment/stanford_town_env/stanford_town_ext_env.py b/metagpt/environment/stanford_town_env/stanford_town_ext_env.py new file mode 100644 index 000000000..8a9a65965 --- /dev/null +++ b/metagpt/environment/stanford_town_env/stanford_town_ext_env.py @@ -0,0 +1,379 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : The StanfordTown external environment to interate with the web interface +# refs to `generative_agents maze.py` + +import math +from pathlib import Path +from typing import Optional, Tuple + +from pydantic import ConfigDict, Field, model_validator + +from metagpt.environment.base_env import ExtEnv, mark_as_readable, mark_as_writeable +from metagpt.utils.common import read_csv_to_list, read_json_file + + +class StanfordTownExtEnv(ExtEnv): + model_config = ConfigDict(arbitrary_types_allowed=True) + + maze_asset_path: Optional[Path] = Field(default=None, description="the path to store maze assets") + maze_width: int = Field(default=140, description="maze map width") + maze_height: int = Field(default=100, description="maze map height") + sq_tile_size: int = Field(default=32, description="the pixel height/width of a tile") + special_constraint: str = Field( + default="", description="a string description of any relevant special constraints " "the world might have" + ) + tiles: list[list[dict]] = Field(default=[]) + address_tiles: dict[str, set] = Field(default=dict()) + collision_maze: list[list] = Field(default=[]) + + @model_validator(mode="before") + @classmethod + def _init_maze(cls, values): + maze_asset_path = values["maze_asset_path"] + assert maze_asset_path + maze_asset_path = Path(maze_asset_path) + + maze_matrix_path = maze_asset_path.joinpath("matrix") + meta_info = read_json_file(maze_matrix_path.joinpath("maze_meta_info.json")) + + maze_width = int(meta_info["maze_width"]) + maze_height = int(meta_info["maze_height"]) + values["maze_width"] = maze_width + values["maze_height"] = maze_height + values["sq_tile_size"] = int(meta_info["sq_tile_size"]) + values["special_constraint"] = meta_info["special_constraint"] + + # READING IN SPECIAL BLOCKS + # Special blocks are those that are colored in the Tiled map. + # Here is an example row for the arena block file: + # e.g, "25331, Double Studio, Studio, Bedroom 2, Painting" + + blocks_folder = maze_matrix_path.joinpath("special_blocks") + + _wb = blocks_folder.joinpath("world_blocks.csv") + wb_rows = read_csv_to_list(_wb, header=False) + wb = wb_rows[0][-1] + + _sb = blocks_folder.joinpath("sector_blocks.csv") + sb_rows = read_csv_to_list(_sb, header=False) + sb_dict = dict() + for i in sb_rows: + sb_dict[i[0]] = i[-1] + + _ab = blocks_folder.joinpath("arena_blocks.csv") + ab_rows = read_csv_to_list(_ab, header=False) + ab_dict = dict() + for i in ab_rows: + ab_dict[i[0]] = i[-1] + + _gob = blocks_folder.joinpath("game_object_blocks.csv") + gob_rows = read_csv_to_list(_gob, header=False) + gob_dict = dict() + for i in gob_rows: + gob_dict[i[0]] = i[-1] + + _slb = blocks_folder.joinpath("spawning_location_blocks.csv") + slb_rows = read_csv_to_list(_slb, header=False) + slb_dict = dict() + for i in slb_rows: + slb_dict[i[0]] = i[-1] + + # [SECTION 3] Reading in the matrices + # This is your typical two dimensional matrices. It's made up of 0s and + # the number that represents the color block from the blocks folder. + maze_folder = maze_matrix_path.joinpath("maze") + + _cm = maze_folder.joinpath("collision_maze.csv") + collision_maze_raw = read_csv_to_list(_cm, header=False)[0] + _sm = maze_folder.joinpath("sector_maze.csv") + sector_maze_raw = read_csv_to_list(_sm, header=False)[0] + _am = maze_folder.joinpath("arena_maze.csv") + arena_maze_raw = read_csv_to_list(_am, header=False)[0] + _gom = maze_folder.joinpath("game_object_maze.csv") + game_object_maze_raw = read_csv_to_list(_gom, header=False)[0] + _slm = maze_folder.joinpath("spawning_location_maze.csv") + spawning_location_maze_raw = read_csv_to_list(_slm, header=False)[0] + + # Loading the maze. The mazes are taken directly from the json exports of + # Tiled maps. They should be in csv format. + # Importantly, they are "not" in a 2-d matrix format -- they are single + # row matrices with the length of width x height of the maze. So we need + # to convert here. + # example format: [['0', '0', ... '25309', '0',...], ['0',...]...] + # 25309 is the collision bar number right now. + collision_maze = [] + sector_maze = [] + arena_maze = [] + game_object_maze = [] + spawning_location_maze = [] + for i in range(0, len(collision_maze_raw), maze_width): + tw = maze_width + collision_maze += [collision_maze_raw[i : i + tw]] + sector_maze += [sector_maze_raw[i : i + tw]] + arena_maze += [arena_maze_raw[i : i + tw]] + game_object_maze += [game_object_maze_raw[i : i + tw]] + spawning_location_maze += [spawning_location_maze_raw[i : i + tw]] + values["collision_maze"] = collision_maze + + tiles = [] + for i in range(maze_height): + row = [] + for j in range(maze_width): + tile_details = dict() + tile_details["world"] = wb + + tile_details["sector"] = "" + if sector_maze[i][j] in sb_dict: + tile_details["sector"] = sb_dict[sector_maze[i][j]] + + tile_details["arena"] = "" + if arena_maze[i][j] in ab_dict: + tile_details["arena"] = ab_dict[arena_maze[i][j]] + + tile_details["game_object"] = "" + if game_object_maze[i][j] in gob_dict: + tile_details["game_object"] = gob_dict[game_object_maze[i][j]] + + tile_details["spawning_location"] = "" + if spawning_location_maze[i][j] in slb_dict: + tile_details["spawning_location"] = slb_dict[spawning_location_maze[i][j]] + + tile_details["collision"] = False + if collision_maze[i][j] != "0": + tile_details["collision"] = True + + tile_details["events"] = set() + + row += [tile_details] + tiles += [row] + values["tiles"] = tiles + + # Each game object occupies an event in the tile. We are setting up the + # default event value here. + for i in range(maze_height): + for j in range(maze_width): + if tiles[i][j]["game_object"]: + object_name = ":".join( + [tiles[i][j]["world"], tiles[i][j]["sector"], tiles[i][j]["arena"], tiles[i][j]["game_object"]] + ) + go_event = (object_name, None, None, None) + tiles[i][j]["events"].add(go_event) + + # Reverse tile access. + # -- given a string address, we return a set of all + # tile coordinates belonging to that address (this is opposite of + # tiles that give you the string address given a coordinate). This is + # an optimization component for finding paths for the personas' movement. + # address_tiles['bedroom-2-a'] == {(58, 9)} + # address_tiles['double studio:recreation:pool table'] + # == {(29, 14), (31, 11), (30, 14), (32, 11), ...}, + address_tiles = dict() + for i in range(maze_height): + for j in range(maze_width): + addresses = [] + if tiles[i][j]["sector"]: + add = f'{tiles[i][j]["world"]}:' + add += f'{tiles[i][j]["sector"]}' + addresses += [add] + if tiles[i][j]["arena"]: + add = f'{tiles[i][j]["world"]}:' + add += f'{tiles[i][j]["sector"]}:' + add += f'{tiles[i][j]["arena"]}' + addresses += [add] + if tiles[i][j]["game_object"]: + add = f'{tiles[i][j]["world"]}:' + add += f'{tiles[i][j]["sector"]}:' + add += f'{tiles[i][j]["arena"]}:' + add += f'{tiles[i][j]["game_object"]}' + addresses += [add] + if tiles[i][j]["spawning_location"]: + add = f'{tiles[i][j]["spawning_location"]}' + addresses += [add] + + for add in addresses: + if add in address_tiles: + address_tiles[add].add((j, i)) + else: + address_tiles[add] = set([(j, i)]) + values["address_tiles"] = address_tiles + return values + + def turn_coordinate_to_tile(self, px_coordinate: tuple[int, int]) -> tuple[int, int]: + """ + Turns a pixel coordinate to a tile coordinate. + """ + x = math.ceil(px_coordinate[0] / self.sq_tile_size) + y = math.ceil(px_coordinate[1] / self.sq_tile_size) + return (x, y) + + @mark_as_readable + def get_collision_maze(self) -> list: + return self.collision_maze + + @mark_as_readable + def get_address_tiles(self) -> dict: + return self.address_tiles + + @mark_as_readable + def access_tile(self, tile: tuple[int, int]) -> dict: + """ + Returns the tiles details dictionary that is stored in self.tiles of the + designated x, y location. + + INPUT + tile: The tile coordinate of our interest in (x, y) form. + OUTPUT + The tile detail dictionary for the designated tile. + EXAMPLE OUTPUT + Given (58, 9), + self.tiles[9][58] = {'world': 'double studio', + 'sector': 'double studio', 'arena': 'bedroom 2', + 'game_object': 'bed', 'spawning_location': 'bedroom-2-a', + 'collision': False, + 'events': {('double studio:double studio:bedroom 2:bed', + None, None)}} + """ + x = tile[0] + y = tile[1] + return self.tiles[y][x] + + @mark_as_readable + def get_tile_path(self, tile: tuple[int, int], level: str) -> str: + """ + Get the tile string address given its coordinate. You designate the level + by giving it a string level description. + + INPUT: + tile: The tile coordinate of our interest in (x, y) form. + level: world, sector, arena, or game object + OUTPUT + The string address for the tile. + EXAMPLE OUTPUT + Given tile=(58, 9), and level=arena, + "double studio:double studio:bedroom 2" + """ + x = tile[0] + y = tile[1] + tile = self.tiles[y][x] + + path = f"{tile['world']}" + if level == "world": + return path + else: + path += f":{tile['sector']}" + + if level == "sector": + return path + else: + path += f":{tile['arena']}" + + if level == "arena": + return path + else: + path += f":{tile['game_object']}" + + return path + + @mark_as_readable + def get_nearby_tiles(self, tile: tuple[int, int], vision_r: int) -> list[tuple[int, int]]: + """ + Given the current tile and vision_r, return a list of tiles that are + within the radius. Note that this implementation looks at a square + boundary when determining what is within the radius. + i.e., for vision_r, returns x's. + x x x x x + x x x x x + x x P x x + x x x x x + x x x x x + + INPUT: + tile: The tile coordinate of our interest in (x, y) form. + vision_r: The radius of the persona's vision. + OUTPUT: + nearby_tiles: a list of tiles that are within the radius. + """ + left_end = 0 + if tile[0] - vision_r > left_end: + left_end = tile[0] - vision_r + + right_end = self.maze_width - 1 + if tile[0] + vision_r + 1 < right_end: + right_end = tile[0] + vision_r + 1 + + bottom_end = self.maze_height - 1 + if tile[1] + vision_r + 1 < bottom_end: + bottom_end = tile[1] + vision_r + 1 + + top_end = 0 + if tile[1] - vision_r > top_end: + top_end = tile[1] - vision_r + + nearby_tiles = [] + for i in range(left_end, right_end): + for j in range(top_end, bottom_end): + nearby_tiles += [(i, j)] + return nearby_tiles + + @mark_as_writeable + def add_tiles_event(self, pt_y: int, pt_x: int, event: Tuple[str, str, str, str]): + self.tiles[pt_y][pt_x]["events"].add(event) + + @mark_as_writeable + def add_event_from_tile(self, curr_event: tuple[str], tile: tuple[int, int]) -> None: + """ + Add an event triple to a tile. + + INPUT: + curr_event: Current event triple. + e.g., ('double studio:double studio:bedroom 2:bed', None, + None) + tile: The tile coordinate of our interest in (x, y) form. + OUPUT: + None + """ + self.tiles[tile[1]][tile[0]]["events"].add(curr_event) + + @mark_as_writeable + def remove_event_from_tile(self, curr_event: tuple[str], tile: tuple[int, int]) -> None: + """dswaq + Remove an event triple from a tile. + + INPUT: + curr_event: Current event triple. + e.g., ('double studio:double studio:bedroom 2:bed', None, + None) + tile: The tile coordinate of our interest in (x, y) form. + OUPUT: + None + """ + curr_tile_ev_cp = self.tiles[tile[1]][tile[0]]["events"].copy() + for event in curr_tile_ev_cp: + if event == curr_event: + self.tiles[tile[1]][tile[0]]["events"].remove(event) + + @mark_as_writeable + def turn_event_from_tile_idle(self, curr_event: tuple[str], tile: tuple[int, int]) -> None: + curr_tile_ev_cp = self.tiles[tile[1]][tile[0]]["events"].copy() + for event in curr_tile_ev_cp: + if event == curr_event: + self.tiles[tile[1]][tile[0]]["events"].remove(event) + new_event = (event[0], None, None, None) + self.tiles[tile[1]][tile[0]]["events"].add(new_event) + + @mark_as_writeable + def remove_subject_events_from_tile(self, subject: str, tile: tuple[int, int]) -> None: + """ + Remove an event triple that has the input subject from a tile. + + INPUT: + subject: "Isabella Rodriguez" + tile: The tile coordinate of our interest in (x, y) form. + OUPUT: + None + """ + curr_tile_ev_cp = self.tiles[tile[1]][tile[0]]["events"].copy() + for event in curr_tile_ev_cp: + if event[0] == subject: + self.tiles[tile[1]][tile[0]]["events"].remove(event) diff --git a/metagpt/environment/werewolf_env/__init__.py b/metagpt/environment/werewolf_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/metagpt/environment/werewolf_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/metagpt/environment/werewolf_env/werewolf_env.py b/metagpt/environment/werewolf_env/werewolf_env.py new file mode 100644 index 000000000..d174f322c --- /dev/null +++ b/metagpt/environment/werewolf_env/werewolf_env.py @@ -0,0 +1,31 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : MG Werewolf Env + +from pydantic import Field + +from metagpt.environment.base_env import Environment +from metagpt.environment.werewolf_env.werewolf_ext_env import WerewolfExtEnv +from metagpt.logs import logger +from metagpt.schema import Message + + +class WerewolfEnv(Environment, WerewolfExtEnv): + timestamp: int = Field(default=0) + + def publish_message(self, message: Message, add_timestamp: bool = True): + """Post information to the current environment""" + logger.debug(f"publish_message: {message.dump()}") + if add_timestamp: + # Because the content of the message may be repeated, for example, killing the same person in two nights + # Therefore, a unique timestamp prefix needs to be added so that the same message will not be automatically deduplicated when added to the memory. + message.content = f"{self.timestamp} | " + message.content + self.memory.add(message) + self.history += f"\n{message}" + + async def run(self, k=1): + """Process all Role runs by order""" + for _ in range(k): + for role in self.roles.values(): + await role.run() + self.timestamp += 1 diff --git a/metagpt/environment/werewolf_env/werewolf_ext_env.py b/metagpt/environment/werewolf_env/werewolf_ext_env.py new file mode 100644 index 000000000..7c4b4c475 --- /dev/null +++ b/metagpt/environment/werewolf_env/werewolf_ext_env.py @@ -0,0 +1,335 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : The werewolf game external environment to integrate with + +import random +from collections import Counter +from enum import Enum +from typing import Callable, Optional + +from pydantic import ConfigDict, Field + +from metagpt.environment.base_env import ExtEnv, mark_as_readable, mark_as_writeable +from metagpt.logs import logger + + +class RoleState(Enum): + ALIVE = "alive" # the role is alive + KILLED = "killed" # the role is killed by werewolf or voting + POISONED = "poisoned" # the role is killed by posion + SAVED = "saved" # the role is saved by antidote + + +# the ordered rules by the moderator to announce to everyone each step +STEP_INSTRUCTIONS = { + 0: { + "content": "It’s dark, everyone close your eyes. I will talk with you/your team secretly at night.", + "send_to": "Moderator", # for moderator to continuen speaking + "restricted_to": "", + }, + 1: { + "content": "Guard, please open your eyes!", + "send_to": "Moderator", # for moderator to continuen speaking + "restricted_to": "", + }, + 2: { + "content": """Guard, now tell me who you protect tonight? + You only choose one from the following living options please: {living_players}. + Or you can pass. For example: Protect ...""", + "send_to": "Guard", + "restricted_to": "Moderator,Guard", + }, + 3: {"content": "Guard, close your eyes", "send_to": "Moderator", "restricted_to": ""}, + 4: {"content": "Werewolves, please open your eyes!", "send_to": "Moderator", "restricted_to": ""}, + 5: { + "content": """Werewolves, I secretly tell you that {werewolf_players} are + all of the 2 werewolves! Keep in mind you are teammates. The rest players are not werewolves. + choose one from the following living options please: + {living_players}. For example: Kill ...""", + "send_to": "Werewolf", + "restricted_to": "Moderator,Werewolf", + }, + 6: {"content": "Werewolves, close your eyes", "send_to": "Moderator", "restricted_to": ""}, + 7: {"content": "Witch, please open your eyes!", "send_to": "Moderator", "restricted_to": ""}, + 8: { + "content": """Witch, tonight {player_hunted} has been killed by the werewolves. + You have a bottle of antidote, would you like to save him/her? If so, say "Save", else, say "Pass".""", + "send_to": "Witch", + "restricted_to": "Moderator,Witch", + }, # 要先判断女巫是否有解药,再去询问女巫是否使用解药救人 + 9: { + "content": """Witch, you also have a bottle of poison, would you like to use it to kill one of the living players? + Choose one from the following living options: {living_players}. + If so, say ONLY "Poison PlayerX", replace PlayerX with the actual player name, else, say "Pass".""", + "send_to": "Witch", + "restricted_to": "Moderator,Witch", + }, # + 10: {"content": "Witch, close your eyes", "send_to": "Moderator", "restricted_to": ""}, + 11: {"content": "Seer, please open your eyes!", "send_to": "Moderator", "restricted_to": ""}, + 12: { + "content": """Seer, you can check one player's identity. Who are you going to verify its identity tonight? + Choose only one from the following living options:{living_players}.""", + "send_to": "Seer", + "restricted_to": "Moderator,Seer", + }, + 13: {"content": "Seer, close your eyes", "send_to": "Moderator", "restricted_to": ""}, + # The 1-st daytime + 14: { + "content": """It's daytime. Everyone woke up except those who had been killed.""", + "send_to": "Moderator", + "restricted_to": "", + }, + 15: {"content": "{player_current_dead} was killed last night!", "send_to": "Moderator", "restricted_to": ""}, + 16: { + "content": """Living players: {living_players}, now freely talk about the current situation based on your observation and + reflection with a few sentences. Decide whether to reveal your identity based on your reflection.""", + "send_to": "", # send to all to speak in daytime + "restricted_to": "", + }, + 17: { + "content": """Now vote and tell me who you think is the werewolf. Don’t mention your role. + You only choose one from the following living options please: + {living_players}. Say ONLY: I vote to eliminate ...""", + "send_to": "", + "restricted_to": "", + }, + 18: {"content": """{player_current_dead} was eliminated.""", "send_to": "Moderator", "restricted_to": ""}, +} + + +class WerewolfExtEnv(ExtEnv): + model_config = ConfigDict(arbitrary_types_allowed=True) + + players_state: dict[str, tuple[str, RoleState]] = Field( + default=dict(), description="the player's role type and state by player_name" + ) + + round_idx: int = Field(default=0) # the current round + step_idx: int = Field(default=0) # the current step of current round + eval_step_idx: int = Field(default=0) + per_round_steps: int = Field(default=len(STEP_INSTRUCTIONS)) + + # game global states + game_setup: str = Field(default="", description="game setup including role and its num") + special_role_players: list[str] = Field(default=[]) + winner: Optional[str] = Field(default=None) + win_reason: Optional[str] = Field(default=None) + witch_poison_left: int = Field(default=1) + witch_antidote_left: int = Field(default=1) + + # game current round states, a round is from closing your eyes to the next time you close your eyes + round_hunts: dict[str, str] = Field(default=dict(), description="nighttime wolf hunt result") + round_votes: dict[str, str] = Field( + default=dict(), description="daytime all players vote result, key=voteer, value=voted one" + ) + player_hunted: Optional[str] = Field(default=None) + player_protected: Optional[str] = Field(default=None) + is_hunted_player_saved: bool = Field(default=False) + player_poisoned: Optional[str] = Field(default=None) + player_current_dead: list[str] = Field(default=[]) + + @property + def living_players(self) -> list[str]: + player_names = [] + for name, roletype_state in self.players_state.items(): + if roletype_state[1] in [RoleState.ALIVE, RoleState.SAVED]: + player_names.append(name) + return player_names + + def _role_type_players(self, role_type: str) -> list[str]: + """return player name of particular role type""" + player_names = [] + for name, roletype_state in self.players_state.items(): + if role_type in roletype_state[0]: + player_names.append(name) + return player_names + + @property + def werewolf_players(self) -> list[str]: + player_names = self._role_type_players(role_type="Werewolf") + return player_names + + @property + def villager_players(self) -> list[str]: + player_names = self._role_type_players(role_type="Villager") + return player_names + + def _init_players_state(self, players: list["Role"]): + for play in players: + self.players_state[play.name] = (play.profile, RoleState.ALIVE) + + self.special_role_players = [ + p for p in self.living_players if p not in self.werewolf_players + self.villager_players + ] + + def init_game_setup( + self, + role_uniq_objs: list[object], + num_villager: int = 2, + num_werewolf: int = 2, + shuffle=True, + add_human=False, + use_reflection=True, + use_experience=False, + use_memory_selection=False, + new_experience_version="", + prepare_human_player=Callable, + ) -> tuple[str, list]: + """init players using different roles' num""" + role_objs = [] + for role_obj in role_uniq_objs: + if str(role_obj) == "Villager": + role_objs.extend([role_obj] * num_villager) + elif str(role_obj) == "Werewolf": + role_objs.extend([role_obj] * num_werewolf) + else: + role_objs.append(role_obj) + if shuffle: + random.shuffle(len(role_objs)) + if add_human: + assigned_role_idx = random.randint(0, len(role_objs) - 1) + assigned_role = role_objs[assigned_role_idx] + role_objs[assigned_role_idx] = prepare_human_player(assigned_role) # TODO + + players = [ + role( + name=f"Player{i + 1}", + use_reflection=use_reflection, + use_experience=use_experience, + use_memory_selection=use_memory_selection, + new_experience_version=new_experience_version, + ) + for i, role in enumerate(role_objs) + ] + + if add_human: + logger.info(f"You are assigned {players[assigned_role_idx].name}({players[assigned_role_idx].profile})") + + game_setup = ["Game setup:"] + [f"{player.name}: {player.profile}," for player in players] + self.game_setup = "\n".join(game_setup) + + self._init_players_state(players) # init players state + + return self.game_setup, players + + def _update_players_state(self, player_names: list[str], state: RoleState = RoleState.KILLED): + for player_name in player_names: + if player_name in self.players_state: + roletype_state = self.players_state[player_name] + self.players_state[player_name] = (roletype_state[0], state) + + def _check_valid_role(self, player: "Role", role_type: str) -> bool: + return True if role_type in str(player) else False + + def _check_player_continue(self, player_name: str, particular_step: int = -1) -> bool: + step_idx = self.step_idx % self.per_round_steps + if particular_step > 0 and step_idx != particular_step: # step no + # particular_step = 18, not daytime vote time, ignore + # particular_step = 15, not nighttime hunt time, ignore + return False + if player_name not in self.living_players: + return False + return True + + @mark_as_readable + def curr_step_instruction(self) -> dict: + step_idx = self.step_idx % len(STEP_INSTRUCTIONS) + instruction = STEP_INSTRUCTIONS[step_idx] + self.step_idx += 1 + return instruction + + @mark_as_readable + def get_players_state(self, player_names: list[str]) -> dict[str, RoleState]: + players_state = { + player_name: self.players_state[player_name][1] # only return role state + for player_name in player_names + if player_name in self.players_state + } + return players_state + + @mark_as_writeable + def vote_kill_someone(self, voteer: "Role", player_name: str = None): + """player vote result at daytime + player_name: if it's None, regard as abstaining from voting + """ + if not self._check_player_continue(voteer.name, particular_step=18): # 18=step no + return + + self.round_votes[voteer.name] = player_name + # check if all living players finish voting, then get the dead one + if list(self.round_votes.keys()) == self.living_players: + voted_all = list(self.round_votes.values()) # TODO in case of tie vote, check who was voted first + voted_all = [item for item in voted_all if item] + self.player_current_dead = Counter(voted_all).most_common()[0][0] + self._update_players_state([self.player_current_dead]) + + @mark_as_writeable + def wolf_kill_someone(self, wolf: "Role", player_name: str): + if not self._check_valid_role(wolf, "Werewolf"): + return + if not self._check_player_continue(wolf.name, particular_step=5): # 5=step no + return + + self.round_hunts[wolf.name] = player_name + living_werewolf = [p for p in self.werewolf_players if p in self.living_players] + # check if all living wolfs finish hunting, then get the hunted one + if list(self.round_hunts.keys()) == living_werewolf: + hunted_all = list(self.round_hunts.values()) + self.player_hunted = Counter(hunted_all).most_common()[0][0] + + @mark_as_writeable + def witch_poison_someone(self, witch: "Role", player_name: str = None): + if not self._check_valid_role(witch, "Witch"): + return + if not self._check_player_continue(player_name): + return + + self._update_players_state([player_name], RoleState.POISONED) + self.player_poisoned = player_name + + @mark_as_writeable + def witch_save_someone(self, witch: "Role", player_name: str = None): + if not self._check_valid_role(witch, "Witch"): + return + if not self._check_player_continue(player_name): + return + + self._update_players_state([player_name], RoleState.SAVED) + self.player_protected = player_name + + @mark_as_writeable + def update_game_states(self, memories: list): + step_idx = self.step_idx % self.per_round_steps + if step_idx not in [15, 18] or self.step_idx in self.eval_step_idx: + return + else: + self.eval_step_idx.append(self.step_idx) # record evaluation, avoid repetitive evaluation at the same step + + if step_idx == 15: # step no + # night ends: after all special roles acted, process the whole night + self.player_current_dead = [] # reset + + if self.player_hunted != self.player_protected and not self.is_hunted_player_saved: + self.player_current_dead.append(self.player_hunted) + if self.player_poisoned: + self.player_current_dead.append(self.player_poisoned) + + self._update_players_state([self.player_current_dead]) + # reset + self.player_hunted = None + self.player_protected = None + self.is_hunted_player_saved = False + self.player_poisoned = None + + # game's termination condition + living_werewolf = [p for p in self.werewolf_players if p in self.living_players] + living_villagers = [p for p in self.villager_players if p in self.living_players] + living_special_roles = [p for p in self.special_role_players if p in self.living_players] + if not living_werewolf: + self.winner = "good guys" + self.win_reason = "werewolves all dead" + elif not living_villagers or not living_special_roles: + self.winner = "werewolf" + self.win_reason = "villagers all dead" if not living_villagers else "special roles all dead" + if self.winner is not None: + self._record_all_experiences() # TODO diff --git a/metagpt/learn/google_search.py b/metagpt/learn/google_search.py index 3f356f7dd..399c14de4 100644 --- a/metagpt/learn/google_search.py +++ b/metagpt/learn/google_search.py @@ -8,5 +8,5 @@ async def google_search(query: str, max_results: int = 6, **kwargs): :param max_results: The number of search results to retrieve :return: The web search results in markdown format. """ - results = await SearchEngine().run(query, max_results=max_results, as_string=False) + results = await SearchEngine(**kwargs).run(query, max_results=max_results, as_string=False) return "\n".join(f"{i}. [{j['title']}]({j['link']}): {j['snippet']}" for i, j in enumerate(results, 1)) diff --git a/metagpt/learn/skill_loader.py b/metagpt/learn/skill_loader.py index b60fa9093..bcf28bb87 100644 --- a/metagpt/learn/skill_loader.py +++ b/metagpt/learn/skill_loader.py @@ -13,7 +13,7 @@ import aiofiles import yaml from pydantic import BaseModel, Field -from metagpt.context import CONTEXT +from metagpt.context import Context class Example(BaseModel): @@ -73,14 +73,15 @@ class SkillsDeclaration(BaseModel): skill_data = yaml.safe_load(data) return SkillsDeclaration(**skill_data) - def get_skill_list(self, entity_name: str = "Assistant") -> Dict: + def get_skill_list(self, entity_name: str = "Assistant", context: Context = None) -> Dict: """Return the skill name based on the skill description.""" entity = self.entities.get(entity_name) if not entity: return {} # List of skills that the agent chooses to activate. - agent_skills = CONTEXT.kwargs.agent_skills + ctx = context or Context() + agent_skills = ctx.kwargs.agent_skills if not agent_skills: return {} diff --git a/metagpt/learn/text_to_image.py b/metagpt/learn/text_to_image.py index e2fac7647..163859fc0 100644 --- a/metagpt/learn/text_to_image.py +++ b/metagpt/learn/text_to_image.py @@ -27,7 +27,7 @@ async def text_to_image(text, size_type: str = "512x512", config: Config = metag """ image_declaration = "data:image/png;base64," - model_url = config.METAGPT_TEXT_TO_IMAGE_MODEL_URL + model_url = config.metagpt_tti_url if model_url: binary_data = await oas3_metagpt_text_to_image(text, size_type, model_url) elif config.get_openai_llm(): diff --git a/metagpt/learn/text_to_speech.py b/metagpt/learn/text_to_speech.py index 37e56eaff..8dbd6d243 100644 --- a/metagpt/learn/text_to_speech.py +++ b/metagpt/learn/text_to_speech.py @@ -39,8 +39,8 @@ async def text_to_speech( """ - subscription_key = config.AZURE_TTS_SUBSCRIPTION_KEY - region = config.AZURE_TTS_REGION + subscription_key = config.azure_tts_subscription_key + region = config.azure_tts_region if subscription_key and region: audio_declaration = "data:audio/wav;base64," base64_data = await oas3_azsure_tts(text, lang, voice, style, role, subscription_key, region) @@ -50,9 +50,9 @@ async def text_to_speech( return f"[{text}]({url})" return audio_declaration + base64_data if base64_data else base64_data - iflytek_app_id = config.IFLYTEK_APP_ID - iflytek_api_key = config.IFLYTEK_API_KEY - iflytek_api_secret = config.IFLYTEK_API_SECRET + iflytek_app_id = config.iflytek_app_id + iflytek_api_key = config.iflytek_api_key + iflytek_api_secret = config.iflytek_api_secret if iflytek_app_id and iflytek_api_key and iflytek_api_secret: audio_declaration = "data:audio/mp3;base64," base64_data = await oas3_iflytek_tts( @@ -65,5 +65,5 @@ async def text_to_speech( return audio_declaration + base64_data if base64_data else base64_data raise ValueError( - "AZURE_TTS_SUBSCRIPTION_KEY, AZURE_TTS_REGION, IFLYTEK_APP_ID, IFLYTEK_API_KEY, IFLYTEK_API_SECRET error" + "azure_tts_subscription_key, azure_tts_region, iflytek_app_id, iflytek_api_key, iflytek_api_secret error" ) diff --git a/metagpt/llm.py b/metagpt/llm.py index 30ced25d2..465e419a1 100644 --- a/metagpt/llm.py +++ b/metagpt/llm.py @@ -8,12 +8,13 @@ from typing import Optional from metagpt.configs.llm_config import LLMConfig -from metagpt.context import CONTEXT +from metagpt.context import Context from metagpt.provider.base_llm import BaseLLM -def LLM(llm_config: Optional[LLMConfig] = None) -> BaseLLM: +def LLM(llm_config: Optional[LLMConfig] = None, context: Context = None) -> BaseLLM: """get the default llm provider if name is None""" + ctx = context or Context() if llm_config is not None: - CONTEXT.llm_with_cost_manager_from_llm_config(llm_config) - return CONTEXT.llm() + return ctx.llm_with_cost_manager_from_llm_config(llm_config) + return ctx.llm() diff --git a/metagpt/logs.py b/metagpt/logs.py index fb0fdd553..90bac21aa 100644 --- a/metagpt/logs.py +++ b/metagpt/logs.py @@ -15,14 +15,15 @@ from loguru import logger as _logger from metagpt.const import METAGPT_ROOT -def define_log_level(print_level="INFO", logfile_level="DEBUG"): +def define_log_level(print_level="INFO", logfile_level="DEBUG", name: str = None): """Adjust the log level to above level""" current_date = datetime.now() formatted_date = current_date.strftime("%Y%m%d") + log_name = f"{name}_{formatted_date}" if name else formatted_date # name a log with prefix name _logger.remove() _logger.add(sys.stderr, level=print_level) - _logger.add(METAGPT_ROOT / f"logs/{formatted_date}.txt", level=logfile_level) + _logger.add(METAGPT_ROOT / f"logs/{log_name}.txt", level=logfile_level) return _logger diff --git a/metagpt/prompts/ci/ml_action.py b/metagpt/prompts/ci/ml_action.py new file mode 100644 index 000000000..5d27c7ff0 --- /dev/null +++ b/metagpt/prompts/ci/ml_action.py @@ -0,0 +1,128 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Time : 2023/11/24 15:43 +# @Author : lidanyang +# @File : ml_action +# @Desc : +UPDATE_DATA_COLUMNS = """ +# Background +Keep dataset column information updated before model train. +## Done Tasks +```python +{history_code} +```end + +# Task +Update and print the dataset's column information only if the train or test data has changed. Use the following code: +```python +from metagpt.tools.libs.data_preprocess import get_column_info + +column_info = get_column_info(df) +print("column_info") +print(column_info) +```end + +# Constraints: +- Use the DataFrame variable from 'Done Tasks' in place of df. +- Import `get_column_info` only if it's not already imported. +""" + +PRINT_DATA_COLUMNS = { + "name": "print_column_info", + "description": "Print the latest column information after 'Done Tasks' code if first read or data changed.", + "parameters": { + "type": "object", + "properties": { + "code": { + "type": "string", + "description": "The code to be added to a new cell in jupyter.", + }, + }, + "required": ["code"], + }, +} + +ML_COMMON_PROMPT = """ +# Background +As a data scientist, you need to help user to achieve their goal [{user_requirement}] step-by-step in an continuous Jupyter notebook. + +## Done Tasks +```python +{history_code} +```end + +## Current Task +{current_task} + +# Latest Data Info +Latest data info after previous tasks: +{column_info} + +# Task +Write complete code for 'Current Task'. And avoid duplicating code from 'Done Tasks', such as repeated import of packages, reading data, etc. +Specifically, {tool_type_usage_prompt} +""" + +USE_NO_TOOLS_EXAMPLE = """ +# Output Example: +when current task is "train a lightgbm model on training data", the code can be like: +```python +# Step 1: check data type and convert to numeric +obj_cols = train.select_dtypes(include='object').columns.tolist() + +for col in obj_cols: + encoder = LabelEncoder() + train[col] = encoder.fit_transform(train[col].unique().tolist() + ['unknown']) + test[col] = test[col].apply(lambda x: x if x in encoder.classes_ else 'unknown') + test[col] = encoder.transform(test[col]) + +# Step 2: train lightgbm model +model = LGBMClassifier() +model.fit(train, y_train) +```end + +# Constraints: +- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed. +""" + +USE_TOOLS_EXAMPLE = """ +# Capabilities +- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class. +- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc.. + +# Available Tools: +Each Class tool is described in JSON format. When you call a tool, import the tool from its path first. +{tool_schemas} + +# Output Example: +when current task is "do data preprocess, like fill missing value, handle outliers, etc.", the code can be like: +```python +# Step 1: fill missing value +# Tools used: ['FillMissingValue'] +from metagpt.tools.libs.data_preprocess import FillMissingValue + +train_processed = train.copy() +test_processed = test.copy() +num_cols = train_processed.select_dtypes(include='number').columns.tolist() +if 'label' in num_cols: + num_cols.remove('label') +fill_missing_value = FillMissingValue(features=num_cols, strategy='mean') +fill_missing_value.fit(train_processed) +train_processed = fill_missing_value.transform(train_processed) +test_processed = fill_missing_value.transform(test_processed) + +# Step 2: handle outliers +for col in num_cols: + low, high = train_processed[col].quantile([0.01, 0.99]) + train_processed[col] = train_processed[col].clip(low, high) + test_processed[col] = test_processed[col].clip(low, high) +```end + +# Constraints: +- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed. +- Always prioritize using pre-defined tools for the same functionality. +- Always copy the DataFrame before processing it and use the copy to process. +""" + +ML_GENERATE_CODE_PROMPT = ML_COMMON_PROMPT + USE_NO_TOOLS_EXAMPLE +ML_TOOL_USAGE_PROMPT = ML_COMMON_PROMPT + USE_TOOLS_EXAMPLE diff --git a/metagpt/prompts/ci/write_analysis_code.py b/metagpt/prompts/ci/write_analysis_code.py new file mode 100644 index 000000000..4eccefcd1 --- /dev/null +++ b/metagpt/prompts/ci/write_analysis_code.py @@ -0,0 +1,93 @@ +ASSIGN_TASK_TYPE_PROMPT = """ +Please assign a task type to each task in the list below from the given categories: +{task_info} + +## All Task Type: +{task_type_desc} +""" + +ASSIGN_TASK_TYPE_CONFIG = { + "name": "assign_task_type", + "description": "Assign task type to each task by order.", + "parameters": { + "type": "object", + "properties": { + "task_type": { + "type": "array", + "description": "List of task type. The length should as long as task list", + "items": { + "type": "string", + }, + }, + }, + "required": ["task_type"], + }, +} + +TOOL_RECOMMENDATION_PROMPT = """ +## User Requirement: +{current_task} + +## Task +Recommend up to five tools from 'Available Tools' that can help solve the 'User Requirement'. + +## Available Tools: +{available_tools} + +## Tool Selection and Instructions: +- Select tools most relevant to completing the 'User Requirement'. +- If you believe that no tools are suitable, indicate with an empty list. +- Only list the names of the tools, not the full schema of each tool. +- Ensure selected tools are listed in 'Available Tools'. +""" + +SELECT_FUNCTION_TOOLS = { + "name": "select_function_tools", + "description": "For current task, select suitable tools for it.", + "parameters": { + "type": "object", + "properties": { + "recommend_tools": { + "type": "array", + "description": "List of tool names. Empty list if no tool is suitable.", + "items": { + "type": "string", + }, + }, + }, + "required": ["recommend_tools"], + }, +} + +CODE_GENERATOR_WITH_TOOLS = { + "name": "add_subtask_code", + "description": "Add new code cell of current task to the end of an active Jupyter notebook.", + "parameters": { + "type": "object", + "properties": { + "code": { + "type": "string", + "description": "The code to be added to a new cell in jupyter.", + }, + }, + "required": ["code"], + }, +} + +TOOL_USAGE_PROMPT = """ +# Instruction +Write complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc. +Specifically, {tool_type_usage_prompt} + +# Capabilities +- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class. +- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc.. + +# Available Tools (can be empty): +Each Class tool is described in JSON format. When you call a tool, import the tool first. +{tool_schemas} + +# Constraints: +- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed. +- Always prioritize using pre-defined tools for the same functionality. +""" diff --git a/metagpt/prompts/tool_types.py b/metagpt/prompts/tool_types.py new file mode 100644 index 000000000..f27fbea99 --- /dev/null +++ b/metagpt/prompts/tool_types.py @@ -0,0 +1,46 @@ +# Prompt for using tools of "data_preprocess" type +DATA_PREPROCESS_PROMPT = """ +The current task is about data preprocessing, please note the following: +- Monitor data types per column, applying appropriate methods. +- Ensure operations are on existing dataset columns. +- Avoid writing processed data to files. +- Avoid any change to label column, such as standardization, etc. +- Prefer alternatives to one-hot encoding for categorical data. +- Only encode or scale necessary columns to allow for potential feature-specific engineering tasks (like time_extract, binning, extraction, etc.) later. +- Each step do data preprocessing to train, must do same for test separately at the same time. +""" + +# Prompt for using tools of "feature_engineering" type +FEATURE_ENGINEERING_PROMPT = """ +The current task is about feature engineering. when performing it, please adhere to the following principles: +- Generate as diverse features as possible to improve the model's performance step-by-step. +- Use available feature engineering tools if they are potential impactful. +- Avoid creating redundant or excessively numerous features in one step. +- Exclude ID columns from feature generation and remove them. +- Each feature engineering operation performed on the train set must also applies to the test separately at the same time. +- Avoid using the label column to create features, except for cat encoding. +- Use the data from previous task result if exist, do not mock or reload data yourself. +""" + +# Prompt for using tools of "model_train" type +MODEL_TRAIN_PROMPT = """ +The current task is about training a model, please ensure high performance: +- Keep in mind that your user prioritizes results and is highly focused on model performance. So, when needed, feel free to use models of any complexity to improve effectiveness, such as XGBoost, CatBoost, etc. +- If non-numeric columns exist, perform label encode together with all steps. +- Use the data from previous task result directly, do not mock or reload data yourself. +- Set suitable hyperparameters for the model, make metrics as high as possible. +""" + +# Prompt for using tools of "model_evaluate" type +MODEL_EVALUATE_PROMPT = """ +The current task is about evaluating a model, please note the following: +- Ensure that the evaluated data is same processed as the training data. If not, remember use object in 'Done Tasks' to transform the data. +- Use trained model from previous task result directly, do not mock or reload model yourself. +""" + +# Prompt for using tools of "vision" type +IMAGE2WEBPAGE_PROMPT = """ +The current task is about converting image into webpage code. please note the following: +- Single-Step Code Generation: Execute the entire code generation process in a single step, encompassing HTML, CSS, and JavaScript. Avoid fragmenting the code generation into multiple separate steps to maintain consistency and simplify the development workflow. +- Save webpages: Be sure to use the save method provided. +""" diff --git a/metagpt/provider/base_llm.py b/metagpt/provider/base_llm.py index 0cd440ea1..b144471b5 100644 --- a/metagpt/provider/base_llm.py +++ b/metagpt/provider/base_llm.py @@ -13,6 +13,7 @@ from typing import Optional, Union from openai import AsyncOpenAI from metagpt.configs.llm_config import LLMConfig +from metagpt.logs import logger from metagpt.schema import Message from metagpt.utils.cost_manager import CostManager @@ -33,8 +34,26 @@ class BaseLLM(ABC): def __init__(self, config: LLMConfig): pass - def _user_msg(self, msg: str) -> dict[str, str]: - return {"role": "user", "content": msg} + def _user_msg(self, msg: str, images: Optional[Union[str, list[str]]] = None) -> dict[str, Union[str, dict]]: + if images: + # as gpt-4v, chat with image + return self._user_msg_with_imgs(msg, images) + else: + return {"role": "user", "content": msg} + + def _user_msg_with_imgs(self, msg: str, images: Optional[Union[str, list[str]]]): + """ + images: can be list of http(s) url or base64 + """ + if isinstance(images, str): + images = [images] + content = [{"type": "text", "text": msg}] + for image in images: + # image url or image base64 + url = image if image.startswith("http") else f"data:image/jpeg;base64,{image}" + # it can with multiple-image inputs + content.append({"type": "image_url", "image_url": url}) + return {"role": "user", "content": content} def _assistant_msg(self, msg: str) -> dict[str, str]: return {"role": "assistant", "content": msg} @@ -53,6 +72,7 @@ class BaseLLM(ABC): msg: str, system_msgs: Optional[list[str]] = None, format_msgs: Optional[list[dict[str, str]]] = None, + images: Optional[Union[str, list[str]]] = None, timeout=3, stream=True, ) -> str: @@ -64,7 +84,8 @@ class BaseLLM(ABC): message = [] if format_msgs: message.extend(format_msgs) - message.append(self._user_msg(msg)) + message.append(self._user_msg(msg, images=images)) + logger.debug(message) rsp = await self.acompletion_text(message, stream=stream, timeout=timeout) return rsp @@ -104,6 +125,10 @@ class BaseLLM(ABC): """Required to provide the first text of choice""" return rsp.get("choices")[0]["message"]["content"] + def get_choice_delta_text(self, rsp: dict) -> str: + """Required to provide the first text of stream choice""" + return rsp.get("choices", [{}])[0].get("delta", {}).get("content", "") + def get_choice_function(self, rsp: dict) -> dict: """Required to provide the first function of choice :param dict rsp: OpenAI chat.comletion respond JSON, Note "message" must include "tool_calls", @@ -142,4 +167,12 @@ class BaseLLM(ABC): :return dict: return the first function arguments of choice, for example, {'language': 'python', 'code': "print('Hello, World!')"} """ - return json.loads(self.get_choice_function(rsp)["arguments"]) + return json.loads(self.get_choice_function(rsp)["arguments"], strict=False) + + def messages_to_prompt(self, messages: list[dict]): + """[{"role": "user", "content": msg}] to user: etc.""" + return "\n".join([f"{i['role']}: {i['content']}" for i in messages]) + + def messages_to_dict(self, messages): + """objects to [{"role": "user", "content": msg}] etc.""" + return [i.to_dict() for i in messages] diff --git a/metagpt/provider/fireworks_api.py b/metagpt/provider/fireworks_api.py index f9ff7e655..d56453a85 100644 --- a/metagpt/provider/fireworks_api.py +++ b/metagpt/provider/fireworks_api.py @@ -84,7 +84,7 @@ class FireworksLLM(OpenAILLM): def _update_costs(self, usage: CompletionUsage): if self.config.calc_usage and usage: try: - # use FireworksCostManager not CONTEXT.cost_manager + # use FireworksCostManager not context.cost_manager self.cost_manager.update_cost(usage.prompt_tokens, usage.completion_tokens, self.model) except Exception as e: logger.error(f"updating costs failed!, exp: {e}") diff --git a/metagpt/provider/general_api_requestor.py b/metagpt/provider/general_api_requestor.py index cf31fd629..18f4dd909 100644 --- a/metagpt/provider/general_api_requestor.py +++ b/metagpt/provider/general_api_requestor.py @@ -60,7 +60,8 @@ class GeneralAPIRequestor(APIRequestor): self, result: requests.Response, stream: bool ) -> Tuple[Union[bytes, Iterator[Generator]], bytes]: """Returns the response(s) and a bool indicating whether it is a stream.""" - if stream and "text/event-stream" in result.headers.get("Content-Type", ""): + content_type = result.headers.get("Content-Type", "") + if stream and ("text/event-stream" in content_type or "application/x-ndjson" in content_type): return ( self._interpret_response_line(line, result.status_code, result.headers, stream=True) for line in parse_stream(result.iter_lines()) @@ -79,10 +80,8 @@ class GeneralAPIRequestor(APIRequestor): async def _interpret_async_response( self, result: aiohttp.ClientResponse, stream: bool ) -> Tuple[Union[bytes, AsyncGenerator[bytes, None]], bool]: - if stream and ( - "text/event-stream" in result.headers.get("Content-Type", "") - or "application/x-ndjson" in result.headers.get("Content-Type", "") - ): + content_type = result.headers.get("Content-Type", "") + if stream and ("text/event-stream" in content_type or "application/x-ndjson" in content_type): # the `Content-Type` of ollama stream resp is "application/x-ndjson" return ( self._interpret_response_line(line, result.status, result.headers, stream=True) diff --git a/metagpt/provider/google_gemini_api.py b/metagpt/provider/google_gemini_api.py index 6df814b55..2647ab16b 100644 --- a/metagpt/provider/google_gemini_api.py +++ b/metagpt/provider/google_gemini_api.py @@ -2,6 +2,8 @@ # -*- coding: utf-8 -*- # @Desc : Google Gemini LLM from https://ai.google.dev/tutorials/python_quickstart +from typing import Optional, Union + import google.generativeai as genai from google.ai import generativelanguage as glm from google.generativeai.generative_models import GenerativeModel @@ -58,7 +60,7 @@ class GeminiLLM(BaseLLM): def __init_gemini(self, config: LLMConfig): genai.configure(api_key=config.api_key) - def _user_msg(self, msg: str) -> dict[str, str]: + def _user_msg(self, msg: str, images: Optional[Union[str, list[str]]] = None) -> dict[str, str]: # Not to change BaseLLM default functions but update with Gemini's conversation format. # You should follow the format. return {"role": "user", "parts": [msg]} diff --git a/metagpt/provider/openai_api.py b/metagpt/provider/openai_api.py index 3f3a4e1a7..fe41fb05f 100644 --- a/metagpt/provider/openai_api.py +++ b/metagpt/provider/openai_api.py @@ -25,9 +25,10 @@ from tenacity import ( from metagpt.configs.llm_config import LLMConfig, LLMType from metagpt.logs import log_llm_stream, logger from metagpt.provider.base_llm import BaseLLM -from metagpt.provider.constant import GENERAL_FUNCTION_SCHEMA, GENERAL_TOOL_CHOICE +from metagpt.provider.constant import GENERAL_FUNCTION_SCHEMA from metagpt.provider.llm_provider_registry import register_provider from metagpt.schema import Message +from metagpt.utils.common import CodeParser, decode_image from metagpt.utils.cost_manager import CostManager, Costs from metagpt.utils.exceptions import handle_exception from metagpt.utils.token_counter import ( @@ -99,7 +100,7 @@ class OpenAILLM(BaseLLM): "messages": messages, "max_tokens": self._get_max_tokens(messages), "n": 1, - "stop": None, + # "stop": None, # default it's None and gpt4-v can't have this one "temperature": 0.3, "model": self.model, "timeout": max(self.config.timeout, timeout), @@ -146,37 +147,41 @@ class OpenAILLM(BaseLLM): def _func_configs(self, messages: list[dict], timeout=3, **kwargs) -> dict: """Note: Keep kwargs consistent with https://platform.openai.com/docs/api-reference/chat/create""" if "tools" not in kwargs: - configs = { - "tools": [{"type": "function", "function": GENERAL_FUNCTION_SCHEMA}], - "tool_choice": GENERAL_TOOL_CHOICE, - } + configs = {"tools": [{"type": "function", "function": GENERAL_FUNCTION_SCHEMA}]} kwargs.update(configs) return self._cons_kwargs(messages=messages, timeout=timeout, **kwargs) + def _process_message(self, messages: Union[str, Message, list[dict], list[Message], list[str]]) -> list[dict]: + """convert messages to list[dict].""" + # 全部转成list + if not isinstance(messages, list): + messages = [messages] + + # 转成list[dict] + processed_messages = [] + for msg in messages: + if isinstance(msg, str): + processed_messages.append({"role": "user", "content": msg}) + elif isinstance(msg, dict): + assert set(msg.keys()) == set(["role", "content"]) + processed_messages.append(msg) + elif isinstance(msg, Message): + processed_messages.append(msg.to_dict()) + else: + raise ValueError( + f"Only support message type are: str, Message, dict, but got {type(messages).__name__}!" + ) + return processed_messages + async def _achat_completion_function(self, messages: list[dict], timeout=3, **chat_configs) -> ChatCompletion: + messages = self._process_message(messages) kwargs = self._func_configs(messages=messages, timeout=timeout, **chat_configs) rsp: ChatCompletion = await self.aclient.chat.completions.create(**kwargs) self._update_costs(rsp.usage) return rsp - def _process_message(self, messages: Union[str, Message, list[dict], list[Message], list[str]]) -> list[dict]: - """convert messages to list[dict].""" - if isinstance(messages, list): - messages = [Message(content=msg) if isinstance(msg, str) else msg for msg in messages] - return [msg if isinstance(msg, dict) else msg.to_dict() for msg in messages] - - if isinstance(messages, Message): - messages = [messages.to_dict()] - elif isinstance(messages, str): - messages = [{"role": "user", "content": messages}] - else: - raise ValueError( - f"Only support messages type are: str, Message, list[dict], but got {type(messages).__name__}!" - ) - return messages - - async def aask_code(self, messages: Union[str, Message, list[dict]], **kwargs) -> dict: + async def aask_code(self, messages: list[dict], **kwargs) -> dict: """Use function of tools to ask a code. Note: Keep kwargs consistent with https://platform.openai.com/docs/api-reference/chat/create @@ -186,18 +191,37 @@ class OpenAILLM(BaseLLM): >>> rsp = await llm.aask_code(msg) # -> {'language': 'python', 'code': "print('Hello, World!')"} """ - messages = self._process_message(messages) rsp = await self._achat_completion_function(messages, **kwargs) return self.get_choice_function_arguments(rsp) - @handle_exception + # @handle_exception def get_choice_function_arguments(self, rsp: ChatCompletion) -> dict: """Required to provide the first function arguments of choice. + :param dict rsp: same as in self.get_choice_function(rsp) :return dict: return the first function arguments of choice, for example, {'language': 'python', 'code': "print('Hello, World!')"} """ - return json.loads(rsp.choices[0].message.tool_calls[0].function.arguments) + message = rsp.choices[0].message + if ( + message.tool_calls is not None + and message.tool_calls[0].function is not None + and message.tool_calls[0].function.arguments is not None + ): + # reponse is code + return json.loads(message.tool_calls[0].function.arguments, strict=False) + elif message.tool_calls is None and message.content is not None: + # reponse is code, fix openai tools_call respond bug, + # The response content is `code``, but it appears in the content instead of the arguments. + code_formats = "```" + if message.content.startswith(code_formats) and message.content.endswith(code_formats): + code = CodeParser.parse_code(None, message.content) + return {"language": "python", "code": code} + # reponse is message + return {"language": "markdown", "code": self.get_choice_text(rsp)} + else: + logger.error(f"Failed to parse \n {rsp}\n") + raise Exception(f"Failed to parse \n {rsp}\n") def get_choice_text(self, rsp: ChatCompletion) -> str: """Required to provide the first text of choice""" @@ -212,7 +236,7 @@ class OpenAILLM(BaseLLM): usage.prompt_tokens = count_message_tokens(messages, self.model) usage.completion_tokens = count_string_tokens(rsp, self.model) except Exception as e: - logger.error(f"usage calculation failed: {e}") + logger.warning(f"usage calculation failed: {e}") return usage @@ -229,9 +253,40 @@ class OpenAILLM(BaseLLM): def _get_max_tokens(self, messages: list[dict]): if not self.auto_max_tokens: return self.config.max_token - return get_max_completion_tokens(messages, self.model, self.config.max_tokens) + # FIXME + # https://community.openai.com/t/why-is-gpt-3-5-turbo-1106-max-tokens-limited-to-4096/494973/3 + return min(get_max_completion_tokens(messages, self.model, self.config.max_token), 4096) @handle_exception async def amoderation(self, content: Union[str, list[str]]): """Moderate content.""" return await self.aclient.moderations.create(input=content) + + async def atext_to_speech(self, **kwargs): + """text to speech""" + return await self.aclient.audio.speech.create(**kwargs) + + async def aspeech_to_text(self, **kwargs): + """speech to text""" + return await self.aclient.audio.transcriptions.create(**kwargs) + + async def gen_image( + self, + prompt: str, + size: str = "1024x1024", + quality: str = "standard", + model: str = None, + resp_format: str = "url", + ) -> list["Image"]: + """image generate""" + assert resp_format in ["url", "b64_json"] + if not model: + model = self.model + res = await self.aclient.images.generate( + model=model, prompt=prompt, size=size, quality=quality, n=1, response_format=resp_format + ) + imgs = [] + for item in res.data: + img_url_or_b64 = item.url if resp_format == "url" else item.b64_json + imgs.append(decode_image(img_url_or_b64)) + return imgs diff --git a/metagpt/provider/zhipuai/async_sse_client.py b/metagpt/provider/zhipuai/async_sse_client.py index d7168202a..054865652 100644 --- a/metagpt/provider/zhipuai/async_sse_client.py +++ b/metagpt/provider/zhipuai/async_sse_client.py @@ -1,75 +1,31 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- # @Desc : async_sse_client to make keep the use of Event to access response -# refs to `https://github.com/zhipuai/zhipuai-sdk-python/blob/main/zhipuai/utils/sse_client.py` +# refs to `zhipuai/core/_sse_client.py` -from zhipuai.utils.sse_client import _FIELD_SEPARATOR, Event, SSEClient +import json +from typing import Any, Iterator -class AsyncSSEClient(SSEClient): - async def _aread(self): - data = b"" +class AsyncSSEClient(object): + def __init__(self, event_source: Iterator[Any]): + self._event_source = event_source + + async def stream(self) -> dict: + if isinstance(self._event_source, bytes): + raise RuntimeError( + f"Request failed, msg: {self._event_source.decode('utf-8')}, please ref to `https://open.bigmodel.cn/dev/api#error-code-v3`" + ) async for chunk in self._event_source: - for line in chunk.splitlines(True): - data += line - if data.endswith((b"\r\r", b"\n\n", b"\r\n\r\n")): - yield data - data = b"" - if data: - yield data + line = chunk.decode("utf-8") + if line.startswith(":") or not line: + return - async def async_events(self): - async for chunk in self._aread(): - event = Event() - # Split before decoding so splitlines() only uses \r and \n - for line in chunk.splitlines(): - # Decode the line. - line = line.decode(self._char_enc) - - # Lines starting with a separator are comments and are to be - # ignored. - if not line.strip() or line.startswith(_FIELD_SEPARATOR): - continue - - data = line.split(_FIELD_SEPARATOR, 1) - field = data[0] - - # Ignore unknown fields. - if field not in event.__dict__: - self._logger.debug("Saw invalid field %s while parsing " "Server Side Event", field) - continue - - if len(data) > 1: - # From the spec: - # "If value starts with a single U+0020 SPACE character, - # remove it from value." - if data[1].startswith(" "): - value = data[1][1:] - else: - value = data[1] - else: - # If no value is present after the separator, - # assume an empty value. - value = "" - - # The data field may come over multiple lines and their values - # are concatenated with each other. - if field == "data": - event.__dict__[field] += value + "\n" - else: - event.__dict__[field] = value - - # Events with no data are not dispatched. - if not event.data: - continue - - # If the data field ends with a newline, remove it. - if event.data.endswith("\n"): - event.data = event.data[0:-1] - - # Empty event names default to 'message' - event.event = event.event or "message" - - # Dispatch the event - self._logger.debug("Dispatching %s...", event) - yield event + field, _p, value = line.partition(":") + if value.startswith(" "): + value = value[1:] + if field == "data": + if value.startswith("[DONE]"): + break + data = json.loads(value) + yield data diff --git a/metagpt/provider/zhipuai/zhipu_model_api.py b/metagpt/provider/zhipuai/zhipu_model_api.py index 16d4102d4..a7d49623a 100644 --- a/metagpt/provider/zhipuai/zhipu_model_api.py +++ b/metagpt/provider/zhipuai/zhipu_model_api.py @@ -4,46 +4,27 @@ import json -import zhipuai -from zhipuai.model_api.api import InvokeType, ModelAPI -from zhipuai.utils.http_client import headers as zhipuai_default_headers +from zhipuai import ZhipuAI +from zhipuai.core._http_client import ZHIPUAI_DEFAULT_TIMEOUT from metagpt.provider.general_api_requestor import GeneralAPIRequestor from metagpt.provider.zhipuai.async_sse_client import AsyncSSEClient -class ZhiPuModelAPI(ModelAPI): - @classmethod - def get_header(cls) -> dict: - token = cls._generate_token() - zhipuai_default_headers.update({"Authorization": token}) - return zhipuai_default_headers - - @classmethod - def get_sse_header(cls) -> dict: - token = cls._generate_token() - headers = {"Authorization": token} - return headers - - @classmethod - def split_zhipu_api_url(cls, invoke_type: InvokeType, kwargs): +class ZhiPuModelAPI(ZhipuAI): + def split_zhipu_api_url(self): # use this method to prevent zhipu api upgrading to different version. # and follow the GeneralAPIRequestor implemented based on openai sdk - zhipu_api_url = cls._build_api_url(kwargs, invoke_type) - """ - example: - zhipu_api_url: https://open.bigmodel.cn/api/paas/v3/model-api/{model}/{invoke_method} - """ + zhipu_api_url = "https://open.bigmodel.cn/api/paas/v4/chat/completions" arr = zhipu_api_url.split("/api/") - # ("https://open.bigmodel.cn/api" , "/paas/v3/model-api/chatglm_turbo/invoke") + # ("https://open.bigmodel.cn/api" , "/paas/v4/chat/completions") return f"{arr[0]}/api", f"/{arr[1]}" - @classmethod - async def arequest(cls, invoke_type: InvokeType, stream: bool, method: str, headers: dict, kwargs): + async def arequest(self, stream: bool, method: str, headers: dict, kwargs): # TODO to make the async request to be more generic for models in http mode. assert method in ["post", "get"] - base_url, url = cls.split_zhipu_api_url(invoke_type, kwargs) + base_url, url = self.split_zhipu_api_url() requester = GeneralAPIRequestor(base_url=base_url) result, _, api_key = await requester.arequest( method=method, @@ -51,25 +32,23 @@ class ZhiPuModelAPI(ModelAPI): headers=headers, stream=stream, params=kwargs, - request_timeout=zhipuai.api_timeout_seconds, + request_timeout=ZHIPUAI_DEFAULT_TIMEOUT.read, ) return result - @classmethod - async def ainvoke(cls, **kwargs) -> dict: + async def acreate(self, **kwargs) -> dict: """async invoke different from raw method `async_invoke` which get the final result by task_id""" - headers = cls.get_header() - resp = await cls.arequest( - invoke_type=InvokeType.SYNC, stream=False, method="post", headers=headers, kwargs=kwargs - ) + headers = self._default_headers + resp = await self.arequest(stream=False, method="post", headers=headers, kwargs=kwargs) resp = resp.decode("utf-8") resp = json.loads(resp) + if "error" in resp: + raise RuntimeError( + f"Request failed, msg: {resp}, please ref to `https://open.bigmodel.cn/dev/api#error-code-v3`" + ) return resp - @classmethod - async def asse_invoke(cls, **kwargs) -> AsyncSSEClient: + async def acreate_stream(self, **kwargs) -> AsyncSSEClient: """async sse_invoke""" - headers = cls.get_sse_header() - return AsyncSSEClient( - await cls.arequest(invoke_type=InvokeType.SSE, stream=True, method="post", headers=headers, kwargs=kwargs) - ) + headers = self._default_headers + return AsyncSSEClient(await self.arequest(stream=True, method="post", headers=headers, kwargs=kwargs)) diff --git a/metagpt/provider/zhipuai_api.py b/metagpt/provider/zhipuai_api.py index 67ec6fb8d..9e8e5fb53 100644 --- a/metagpt/provider/zhipuai_api.py +++ b/metagpt/provider/zhipuai_api.py @@ -2,11 +2,9 @@ # -*- coding: utf-8 -*- # @Desc : zhipuai LLM from https://open.bigmodel.cn/dev/api#sdk -import json from enum import Enum +from typing import Optional -import openai -import zhipuai from requests import ConnectionError from tenacity import ( after_log, @@ -15,6 +13,7 @@ from tenacity import ( stop_after_attempt, wait_random_exponential, ) +from zhipuai.types.chat.chat_completion import Completion from metagpt.configs.llm_config import LLMConfig, LLMType from metagpt.logs import log_llm_stream, logger @@ -22,6 +21,7 @@ from metagpt.provider.base_llm import BaseLLM from metagpt.provider.llm_provider_registry import register_provider from metagpt.provider.openai_api import log_and_reraise from metagpt.provider.zhipuai.zhipu_model_api import ZhiPuModelAPI +from metagpt.utils.cost_manager import CostManager class ZhiPuEvent(Enum): @@ -35,27 +35,22 @@ class ZhiPuEvent(Enum): class ZhiPuAILLM(BaseLLM): """ Refs to `https://open.bigmodel.cn/dev/api#chatglm_turbo` - From now, there is only one model named `chatglm_turbo` + From now, support glm-3-turbo、glm-4, and also system_prompt. """ def __init__(self, config: LLMConfig): - self.__init_zhipuai(config) - self.llm = ZhiPuModelAPI - self.model = "chatglm_turbo" # so far only one model, just use it - self.use_system_prompt: bool = False # zhipuai has no system prompt when use api self.config = config + self.__init_zhipuai() + self.cost_manager: Optional[CostManager] = None - def __init_zhipuai(self, config: LLMConfig): - assert config.api_key - zhipuai.api_key = config.api_key - # due to use openai sdk, set the api_key but it will't be used. - # openai.api_key = zhipuai.api_key # due to use openai sdk, set the api_key but it will't be used. - if config.proxy: - # FIXME: openai v1.x sdk has no proxy support - openai.proxy = config.proxy + def __init_zhipuai(self): + assert self.config.api_key + self.api_key = self.config.api_key + self.model = self.config.model # so far, it support glm-3-turbo、glm-4 + self.llm = ZhiPuModelAPI(api_key=self.api_key) - def _const_kwargs(self, messages: list[dict]) -> dict: - kwargs = {"model": self.model, "prompt": messages, "temperature": 0.3} + def _const_kwargs(self, messages: list[dict], stream: bool = False) -> dict: + kwargs = {"model": self.model, "messages": messages, "stream": stream, "temperature": 0.3} return kwargs def _update_costs(self, usage: dict): @@ -64,25 +59,19 @@ class ZhiPuAILLM(BaseLLM): try: prompt_tokens = int(usage.get("prompt_tokens", 0)) completion_tokens = int(usage.get("completion_tokens", 0)) - self.config.cost_manager.update_cost(prompt_tokens, completion_tokens, self.model) + self.cost_manager.update_cost(prompt_tokens, completion_tokens, self.model) except Exception as e: logger.error(f"zhipuai updats costs failed! exp: {e}") - def get_choice_text(self, resp: dict) -> str: - """get the first text of choice from llm response""" - assist_msg = resp.get("data", {}).get("choices", [{"role": "error"}])[-1] - assert assist_msg["role"] == "assistant" - return assist_msg.get("content") - def completion(self, messages: list[dict], timeout=3) -> dict: - resp = self.llm.invoke(**self._const_kwargs(messages)) - usage = resp.get("data").get("usage") + resp: Completion = self.llm.chat.completions.create(**self._const_kwargs(messages)) + usage = resp.usage.model_dump() self._update_costs(usage) - return resp + return resp.model_dump() async def _achat_completion(self, messages: list[dict], timeout=3) -> dict: - resp = await self.llm.ainvoke(**self._const_kwargs(messages)) - usage = resp.get("data").get("usage") + resp = await self.llm.acreate(**self._const_kwargs(messages)) + usage = resp.get("usage", {}) self._update_costs(usage) return resp @@ -90,35 +79,18 @@ class ZhiPuAILLM(BaseLLM): return await self._achat_completion(messages, timeout=timeout) async def _achat_completion_stream(self, messages: list[dict], timeout=3) -> str: - response = await self.llm.asse_invoke(**self._const_kwargs(messages)) + response = await self.llm.acreate_stream(**self._const_kwargs(messages, stream=True)) collected_content = [] usage = {} - async for event in response.async_events(): - if event.event == ZhiPuEvent.ADD.value: - content = event.data + async for chunk in response.stream(): + finish_reason = chunk.get("choices")[0].get("finish_reason") + if finish_reason == "stop": + usage = chunk.get("usage", {}) + else: + content = self.get_choice_delta_text(chunk) collected_content.append(content) log_llm_stream(content) - elif event.event == ZhiPuEvent.ERROR.value or event.event == ZhiPuEvent.INTERRUPTED.value: - content = event.data - logger.error(f"event error: {content}", end="") - elif event.event == ZhiPuEvent.FINISH.value: - """ - event.meta - { - "task_status":"SUCCESS", - "usage":{ - "completion_tokens":351, - "prompt_tokens":595, - "total_tokens":946 - }, - "task_id":"xx", - "request_id":"xxx" - } - """ - meta = json.loads(event.meta) - usage = meta.get("usage") - else: - print(f"zhipuapi else event: {event.data}", end="") + log_llm_stream("\n") self._update_costs(usage) diff --git a/metagpt/roles/assistant.py b/metagpt/roles/assistant.py index 1c5315eee..2774bd9b6 100644 --- a/metagpt/roles/assistant.py +++ b/metagpt/roles/assistant.py @@ -22,7 +22,6 @@ from pydantic import Field from metagpt.actions.skill_action import ArgumentsParingAction, SkillAction from metagpt.actions.talk_action import TalkAction -from metagpt.context import CONTEXT from metagpt.learn.skill_loader import SkillsDeclaration from metagpt.logs import logger from metagpt.memory.brain_memory import BrainMemory @@ -48,7 +47,8 @@ class Assistant(Role): def __init__(self, **kwargs): super().__init__(**kwargs) - self.constraints = self.constraints.format(language=kwargs.get("language") or CONTEXT.kwargs.language) + language = kwargs.get("language") or self.context.kwargs.language + self.constraints = self.constraints.format(language=language) async def think(self) -> bool: """Everything will be done part by part.""" @@ -56,11 +56,11 @@ class Assistant(Role): if not last_talk: return False if not self.skills: - skill_path = Path(CONTEXT.kwargs.SKILL_PATH) if CONTEXT.kwargs.SKILL_PATH else None + skill_path = Path(self.context.kwargs.SKILL_PATH) if self.context.kwargs.SKILL_PATH else None self.skills = await SkillsDeclaration.load(skill_yaml_file_name=skill_path) prompt = "" - skills = self.skills.get_skill_list() + skills = self.skills.get_skill_list(context=self.context) for desc, name in skills.items(): prompt += f"If the text explicitly want you to {desc}, return `[SKILL]: {name}` brief and clear. For instance: [SKILL]: {name}\n" prompt += 'Otherwise, return `[TALK]: {talk}` brief and clear. For instance: if {talk} is "xxxx" return [TALK]: xxxx\n\n' diff --git a/metagpt/roles/ci/code_interpreter.py b/metagpt/roles/ci/code_interpreter.py new file mode 100644 index 000000000..404c93b81 --- /dev/null +++ b/metagpt/roles/ci/code_interpreter.py @@ -0,0 +1,85 @@ +from __future__ import annotations + +from pydantic import Field + +from metagpt.actions.ci.ask_review import ReviewConst +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.actions.ci.write_analysis_code import ( + WriteCodeWithoutTools, + WriteCodeWithTools, +) +from metagpt.logs import logger +from metagpt.roles import Role +from metagpt.schema import Message, Task, TaskResult + + +class CodeInterpreter(Role): + name: str = "Charlie" + profile: str = "CodeInterpreter" + auto_run: bool = True + use_tools: bool = False + execute_code: ExecuteNbCode = Field(default_factory=ExecuteNbCode, exclude=True) + tools: list[str] = [] + + def __init__( + self, + auto_run=True, + use_tools=False, + tools=[], + **kwargs, + ): + super().__init__(auto_run=auto_run, use_tools=use_tools, tools=tools, **kwargs) + self._set_react_mode(react_mode="plan_and_act", auto_run=auto_run, use_tools=use_tools) + if use_tools and tools: + from metagpt.tools.tool_registry import ( + validate_tool_names, # import upon use + ) + + self.tools = validate_tool_names(tools) + logger.info(f"will only use {self.tools} as tools") + + @property + def working_memory(self): + return self.rc.working_memory + + async def _act_on_task(self, current_task: Task) -> TaskResult: + code, result, is_success = await self._write_and_exec_code() + task_result = TaskResult(code=code, result=result, is_success=is_success) + return task_result + + async def _write_and_exec_code(self, max_retry: int = 3): + counter = 0 + success = False + + while not success and counter < max_retry: + ### write code ### + code, cause_by = await self._write_code() + + self.working_memory.add(Message(content=code["code"], role="assistant", cause_by=cause_by)) + + ### execute code ### + result, success = await self.execute_code.run(**code) + print(result) + + self.working_memory.add(Message(content=result, role="user", cause_by=ExecuteNbCode)) + + ### process execution result ### + counter += 1 + + if not success and counter >= max_retry: + logger.info("coding failed!") + review, _ = await self.planner.ask_review(auto_run=False, trigger=ReviewConst.CODE_REVIEW_TRIGGER) + if ReviewConst.CHANGE_WORDS[0] in review: + counter = 0 # redo the task again with help of human suggestions + + return code["code"], result, success + + async def _write_code(self): + todo = WriteCodeWithoutTools() if not self.use_tools else WriteCodeWithTools(selected_tools=self.tools) + logger.info(f"ready to {todo.name}") + + context = self.planner.get_useful_memories() + # print(*context, sep="\n***\n") + code = await todo.run(context=context, plan=self.planner.plan, temperature=0.0) + + return code, todo diff --git a/metagpt/roles/ci/ml_engineer.py b/metagpt/roles/ci/ml_engineer.py new file mode 100644 index 000000000..f8bcb2c89 --- /dev/null +++ b/metagpt/roles/ci/ml_engineer.py @@ -0,0 +1,64 @@ +from metagpt.actions.ci.debug_code import DebugCode +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.actions.ci.ml_action import UpdateDataColumns, WriteCodeWithToolsML +from metagpt.logs import logger +from metagpt.roles.ci.code_interpreter import CodeInterpreter +from metagpt.tools.tool_type import ToolType +from metagpt.utils.common import any_to_str + + +class MLEngineer(CodeInterpreter): + name: str = "Mark" + profile: str = "MLEngineer" + debug_context: list = [] + latest_code: str = "" + + async def _write_code(self): + if not self.use_tools: + return await super()._write_code() + + # In a trial and errors settings, check whether this is our first attempt to tackle the task. If there is no code execution before, then it is. + is_first_trial = any_to_str(ExecuteNbCode) not in [msg.cause_by for msg in self.working_memory.get()] + + if is_first_trial: + # For the first trial, write task code from scratch + column_info = await self._update_data_columns() + + logger.info("Write code with tools") + tool_context, code = await WriteCodeWithToolsML(selected_tools=self.tools).run( + context=[], # context assembled inside the Action + plan=self.planner.plan, + column_info=column_info, + ) + self.debug_context = tool_context + cause_by = WriteCodeWithToolsML + + else: + # Previous trials resulted in error, debug and rewrite the code + logger.warning("We got a bug, now start to debug...") + code = await DebugCode().run( + code=self.latest_code, + runtime_result=self.working_memory.get(), + context=self.debug_context, + ) + logger.info(f"new code \n{code}") + cause_by = DebugCode + + self.latest_code = code["code"] + + return code, cause_by + + async def _update_data_columns(self): + current_task = self.planner.plan.current_task + if current_task.task_type not in [ + ToolType.DATA_PREPROCESS.type_name, + ToolType.FEATURE_ENGINEERING.type_name, + ToolType.MODEL_TRAIN.type_name, + ]: + return "" + logger.info("Check columns in updated data") + code = await UpdateDataColumns().run(self.planner.plan) + success = False + result, success = await self.execute_code.run(**code) + print(result) + return result if success else "" diff --git a/metagpt/roles/engineer.py b/metagpt/roles/engineer.py index c83a776c2..40ade2110 100644 --- a/metagpt/roles/engineer.py +++ b/metagpt/roles/engineer.py @@ -20,17 +20,27 @@ from __future__ import annotations import json +import os from collections import defaultdict from pathlib import Path from typing import Set from metagpt.actions import Action, WriteCode, WriteCodeReview, WriteTasks from metagpt.actions.fix_bug import FixBug +from metagpt.actions.project_management_an import REFINED_TASK_LIST, TASK_LIST from metagpt.actions.summarize_code import SummarizeCode -from metagpt.const import SYSTEM_DESIGN_FILE_REPO, TASK_FILE_REPO +from metagpt.actions.write_code_plan_and_change_an import WriteCodePlanAndChange +from metagpt.const import ( + CODE_PLAN_AND_CHANGE_FILE_REPO, + CODE_PLAN_AND_CHANGE_FILENAME, + REQUIREMENT_FILENAME, + SYSTEM_DESIGN_FILE_REPO, + TASK_FILE_REPO, +) from metagpt.logs import logger from metagpt.roles import Role from metagpt.schema import ( + CodePlanAndChangeContext, CodeSummarizeContext, CodingContext, Document, @@ -80,7 +90,7 @@ class Engineer(Role): super().__init__(**kwargs) self.set_actions([WriteCode]) - self._watch([WriteTasks, SummarizeCode, WriteCode, WriteCodeReview, FixBug]) + self._watch([WriteTasks, SummarizeCode, WriteCode, WriteCodeReview, FixBug, WriteCodePlanAndChange]) self.code_todos = [] self.summarize_todos = [] self.next_todo_action = any_to_name(WriteCode) @@ -88,7 +98,7 @@ class Engineer(Role): @staticmethod def _parse_tasks(task_msg: Document) -> list[str]: m = json.loads(task_msg.content) - return m.get("Task list") + return m.get(TASK_LIST.key) or m.get(REFINED_TASK_LIST.key) async def _act_sp_with_cr(self, review=False) -> Set[str]: changed_files = set() @@ -106,9 +116,13 @@ class Engineer(Role): action = WriteCodeReview(i_context=coding_context, context=self.context, llm=self.llm) self._init_action(action) coding_context = await action.run() + + dependencies = {coding_context.design_doc.root_relative_path, coding_context.task_doc.root_relative_path} + if self.config.inc: + dependencies.add(os.path.join(CODE_PLAN_AND_CHANGE_FILE_REPO, CODE_PLAN_AND_CHANGE_FILENAME)) await self.project_repo.srcs.save( filename=coding_context.filename, - dependencies={coding_context.design_doc.root_relative_path, coding_context.task_doc.root_relative_path}, + dependencies=dependencies, content=coding_context.code_doc.content, ) msg = Message( @@ -128,6 +142,9 @@ class Engineer(Role): """Determines the mode of action based on whether code review is used.""" if self.rc.todo is None: return None + if isinstance(self.rc.todo, WriteCodePlanAndChange): + self.next_todo_action = any_to_name(WriteCode) + return await self._act_code_plan_and_change() if isinstance(self.rc.todo, WriteCode): self.next_todo_action = any_to_name(SummarizeCode) return await self._act_write_code() @@ -161,7 +178,7 @@ class Engineer(Role): is_pass, reason = await self._is_pass(summary) if not is_pass: todo.i_context.reason = reason - tasks.append(todo.i_context.dict()) + tasks.append(todo.i_context.model_dump()) await self.project_repo.docs.code_summary.save( filename=Path(todo.i_context.design_filename).name, @@ -187,6 +204,34 @@ class Engineer(Role): content=json.dumps(tasks), role=self.profile, cause_by=SummarizeCode, send_to=self, sent_from=self ) + async def _act_code_plan_and_change(self): + """Write code plan and change that guides subsequent WriteCode and WriteCodeReview""" + logger.info("Writing code plan and change..") + node = await self.rc.todo.run() + code_plan_and_change = node.instruct_content.model_dump_json() + dependencies = { + REQUIREMENT_FILENAME, + self.rc.todo.i_context.prd_filename, + self.rc.todo.i_context.design_filename, + self.rc.todo.i_context.task_filename, + } + await self.project_repo.docs.code_plan_and_change.save( + filename=self.rc.todo.i_context.filename, content=code_plan_and_change, dependencies=dependencies + ) + await self.project_repo.resources.code_plan_and_change.save( + filename=Path(self.rc.todo.i_context.filename).with_suffix(".md").name, + content=node.content, + dependencies=dependencies, + ) + + return Message( + content=code_plan_and_change, + role=self.profile, + cause_by=WriteCodePlanAndChange, + send_to=self, + sent_from=self, + ) + async def _is_pass(self, summary) -> (str, str): rsp = await self.llm.aask(msg=IS_PASS_PROMPT.format(context=summary), stream=False) logger.info(rsp) @@ -197,11 +242,16 @@ class Engineer(Role): async def _think(self) -> Action | None: if not self.src_workspace: self.src_workspace = self.git_repo.workdir / self.git_repo.workdir.name - write_code_filters = any_to_str_set([WriteTasks, SummarizeCode, FixBug]) + write_plan_and_change_filters = any_to_str_set([WriteTasks]) + write_code_filters = any_to_str_set([WriteTasks, WriteCodePlanAndChange, SummarizeCode, FixBug]) summarize_code_filters = any_to_str_set([WriteCode, WriteCodeReview]) if not self.rc.news: return None msg = self.rc.news[0] + if self.config.inc and msg.cause_by in write_plan_and_change_filters: + logger.debug(f"TODO WriteCodePlanAndChange:{msg.model_dump_json()}") + await self._new_code_plan_and_change_action() + return self.rc.todo if msg.cause_by in write_code_filters: logger.debug(f"TODO WriteCode:{msg.model_dump_json()}") await self._new_code_actions(bug_fix=msg.cause_by == any_to_str(FixBug)) @@ -292,11 +342,19 @@ class Engineer(Role): summarizations[ctx].append(filename) for ctx, filenames in summarizations.items(): ctx.codes_filenames = filenames - self.summarize_todos.append(SummarizeCode(i_context=ctx, llm=self.llm)) + self.summarize_todos.append(SummarizeCode(i_context=ctx, context=self.context, llm=self.llm)) if self.summarize_todos: self.set_todo(self.summarize_todos[0]) + async def _new_code_plan_and_change_action(self): + """Create a WriteCodePlanAndChange action for subsequent to-do actions.""" + files = self.project_repo.all_files + requirement_doc = await self.project_repo.docs.get(REQUIREMENT_FILENAME) + requirement = requirement_doc.content if requirement_doc else "" + code_plan_and_change_ctx = CodePlanAndChangeContext.loads(files, requirement=requirement) + self.rc.todo = WriteCodePlanAndChange(i_context=code_plan_and_change_ctx, context=self.context, llm=self.llm) + @property - def todo(self) -> str: + def action_description(self) -> str: """AgentStore uses this attribute to display to the user what actions the current role should take.""" return self.next_todo_action diff --git a/metagpt/roles/qa_engineer.py b/metagpt/roles/qa_engineer.py index 0666a63db..c73c10ef3 100644 --- a/metagpt/roles/qa_engineer.py +++ b/metagpt/roles/qa_engineer.py @@ -29,7 +29,8 @@ class QaEngineer(Role): profile: str = "QaEngineer" goal: str = "Write comprehensive and robust tests to ensure codes will work as expected without bugs" constraints: str = ( - "The test code you write should conform to code standard like PEP8, be modular, " "easy to read and maintain" + "The test code you write should conform to code standard like PEP8, be modular, easy to read and maintain." + "Use same language as user requirement" ) test_round_allowed: int = 5 test_round: int = 0 @@ -54,6 +55,10 @@ class QaEngineer(Role): if not filename or "test" in filename: continue code_doc = await src_file_repo.get(filename) + if not code_doc: + continue + if not code_doc.filename.endswith(".py"): + continue test_doc = await self.project_repo.tests.get("test_" + code_doc.filename) if not test_doc: test_doc = Document( diff --git a/metagpt/roles/role.py b/metagpt/roles/role.py index ad3c44ac1..3938664ba 100644 --- a/metagpt/roles/role.py +++ b/metagpt/roles/role.py @@ -23,7 +23,7 @@ from __future__ import annotations from enum import Enum -from typing import Any, Iterable, Optional, Set, Type, Union +from typing import TYPE_CHECKING, Iterable, Optional, Set, Type, Union from pydantic import BaseModel, ConfigDict, Field, SerializeAsAny, model_validator @@ -35,10 +35,15 @@ from metagpt.logs import logger from metagpt.memory import Memory from metagpt.provider import HumanProvider from metagpt.schema import Message, MessageQueue, SerializationMixin +from metagpt.strategy.planner import Planner from metagpt.utils.common import any_to_name, any_to_str, role_raise_decorator from metagpt.utils.project_repo import ProjectRepo from metagpt.utils.repair_llm_raw_output import extract_state_value_from_output +if TYPE_CHECKING: + from metagpt.environment import Environment # noqa: F401 + + PREFIX_TEMPLATE = """You are a {profile}, named {name}, your goal is {goal}. """ CONSTRAINT_TEMPLATE = "the constraint is {constraints}. " @@ -93,6 +98,7 @@ class RoleContext(BaseModel): ) # Message Buffer with Asynchronous Updates memory: Memory = Field(default_factory=Memory) # long_term_memory: LongTermMemory = Field(default_factory=LongTermMemory) + working_memory: Memory = Field(default_factory=Memory) state: int = Field(default=-1) # -1 indicates initial or termination state where todo is None todo: Action = Field(default=None, exclude=True) watch: set[str] = Field(default_factory=set) @@ -117,11 +123,17 @@ class RoleContext(BaseModel): def history(self) -> list[Message]: return self.memory.get() + @classmethod + def model_rebuild(cls, **kwargs): + from metagpt.environment.base_env import Environment # noqa: F401 + + super().model_rebuild(**kwargs) + class Role(SerializationMixin, ContextMixin, BaseModel): """Role/Agent""" - model_config = ConfigDict(arbitrary_types_allowed=True, extra="ignore") + model_config = ConfigDict(arbitrary_types_allowed=True, extra="allow") name: str = "" profile: str = "" @@ -142,6 +154,7 @@ class Role(SerializationMixin, ContextMixin, BaseModel): actions: list[SerializeAsAny[Action]] = Field(default=[], validate_default=True) rc: RoleContext = Field(default_factory=RoleContext) addresses: set[str] = set() + planner: Planner = Field(default_factory=Planner) # builtin variables recovered: bool = False # to tag if a recovered role @@ -149,28 +162,24 @@ class Role(SerializationMixin, ContextMixin, BaseModel): __hash__ = object.__hash__ # support Role as hashable type in `Environment.members` - def __init__(self, **data: Any): - self.pydantic_rebuild_model() - super().__init__(**data) + @model_validator(mode="after") + def validate_role_extra(self): + self._process_role_extra() + return self + + def _process_role_extra(self): + kwargs = self.model_extra or {} if self.is_human: self.llm = HumanProvider(None) self._check_actions() self.llm.system_prompt = self._get_prefix() - self._watch(data.get("watch") or [UserRequirement]) + self._watch(kwargs.pop("watch", [UserRequirement])) if self.latest_observed_msg: self.recovered = True - @staticmethod - def pydantic_rebuild_model(): - """Rebuild model to avoid `RecursionError: maximum recursion depth exceeded in comparison`""" - from metagpt.environment import Environment - - Environment - Role.model_rebuild() - @property def todo(self) -> Action: """Get action to do""" @@ -261,7 +270,7 @@ class Role(SerializationMixin, ContextMixin, BaseModel): self._reset() for action in actions: if not isinstance(action, Action): - i = action() + i = action(context=self.context) else: if self.is_human and not isinstance(action.llm, HumanProvider): logger.warning( @@ -274,7 +283,7 @@ class Role(SerializationMixin, ContextMixin, BaseModel): self.actions.append(i) self.states.append(f"{len(self.actions)}. {action}") - def _set_react_mode(self, react_mode: str, max_react_loop: int = 1): + def _set_react_mode(self, react_mode: str, max_react_loop: int = 1, auto_run: bool = True, use_tools: bool = False): """Set strategy of the Role reacting to observed Message. Variation lies in how this Role elects action to perform during the _think stage, especially if it is capable of multiple Actions. @@ -294,6 +303,10 @@ class Role(SerializationMixin, ContextMixin, BaseModel): self.rc.react_mode = react_mode if react_mode == RoleReactMode.REACT: self.rc.max_react_loop = max_react_loop + elif react_mode == RoleReactMode.PLAN_AND_ACT: + self.planner = Planner( + goal=self.goal, working_memory=self.rc.working_memory, auto_run=auto_run, use_tools=use_tools + ) def _watch(self, actions: Iterable[Type[Action]] | Iterable[Action]): """Watch Actions of interest. Role will select Messages caused by these Actions from its personal message @@ -470,8 +483,41 @@ class Role(SerializationMixin, ContextMixin, BaseModel): async def _plan_and_act(self) -> Message: """first plan, then execute an action sequence, i.e. _think (of a plan) -> _act -> _act -> ... Use llm to come up with the plan dynamically.""" - # TODO: to be implemented - return Message(content="") + + # create initial plan and update it until confirmation + goal = self.rc.memory.get()[-1].content # retreive latest user requirement + await self.planner.update_plan(goal=goal) + + # take on tasks until all finished + while self.planner.current_task: + task = self.planner.current_task + logger.info(f"ready to take on task {task}") + + # take on current task + task_result = await self._act_on_task(task) + + # process the result, such as reviewing, confirming, plan updating + await self.planner.process_task_result(task_result) + + rsp = self.planner.get_useful_memories()[0] # return the completed plan as a response + + self.rc.memory.add(rsp) # add to persistent memory + + return rsp + + async def _act_on_task(self, current_task: Task) -> TaskResult: + """Taking specific action to handle one task in plan + + Args: + current_task (Task): current task to take on + + Raises: + NotImplementedError: Specific Role must implement this method if expected to use planner + + Returns: + TaskResult: Result from the actions + """ + raise NotImplementedError async def react(self) -> Message: """Entry to one of three strategies by which Role reacts to the observed Message""" @@ -481,6 +527,8 @@ class Role(SerializationMixin, ContextMixin, BaseModel): rsp = await self._act_by_order() elif self.rc.react_mode == RoleReactMode.PLAN_AND_ACT: rsp = await self._plan_and_act() + else: + raise ValueError(f"Unsupported react mode: {self.rc.react_mode}") self._set_state(state=-1) # current reaction is complete, reset state to -1 and todo back to None return rsp @@ -521,19 +569,37 @@ class Role(SerializationMixin, ContextMixin, BaseModel): return not self.rc.news and not self.rc.todo and self.rc.msg_buffer.empty() async def think(self) -> Action: - """The exported `think` function""" + """ + Export SDK API, used by AgentStore RPC. + The exported `think` function + """ + await self._observe() # For compatibility with the old version of the Agent. await self._think() return self.rc.todo async def act(self) -> ActionOutput: - """The exported `act` function""" + """ + Export SDK API, used by AgentStore RPC. + The exported `act` function + """ msg = await self._act() return ActionOutput(content=msg.content, instruct_content=msg.instruct_content) @property - def first_action(self) -> str: - """AgentStore uses this attribute to display to the user what actions the current role should take.""" - # FIXME: this is a hack, we should not use the first action to represent the todo + def action_description(self) -> str: + """ + Export SDK API, used by AgentStore RPC and Agent. + AgentStore uses this attribute to display to the user what actions the current role should take. + `Role` provides the default property, and this property should be overridden by children classes if necessary, + as demonstrated by the `Engineer` class. + """ + if self.rc.todo: + if self.rc.todo.desc: + return self.rc.todo.desc + return any_to_name(self.rc.todo) if self.actions: return any_to_name(self.actions[0]) return "" + + +RoleContext.model_rebuild() diff --git a/metagpt/roles/sales.py b/metagpt/roles/sales.py index 7929ce7fe..bc449b5cd 100644 --- a/metagpt/roles/sales.py +++ b/metagpt/roles/sales.py @@ -8,12 +8,12 @@ from typing import Optional -from pydantic import Field +from pydantic import Field, model_validator from metagpt.actions import SearchAndSummarize, UserRequirement from metagpt.document_store.base_store import BaseStore from metagpt.roles import Role -from metagpt.tools import SearchEngineType +from metagpt.tools.search_engine import SearchEngine class Sales(Role): @@ -29,14 +29,13 @@ class Sales(Role): store: Optional[BaseStore] = Field(default=None, exclude=True) - def __init__(self, **kwargs): - super().__init__(**kwargs) - self._set_store(self.store) - - def _set_store(self, store): - if store: - action = SearchAndSummarize(name="", engine=SearchEngineType.CUSTOM_ENGINE, search_func=store.asearch) + @model_validator(mode="after") + def validate_stroe(self): + if self.store: + search_engine = SearchEngine.from_search_func(search_func=self.store.asearch, proxy=self.config.proxy) + action = SearchAndSummarize(search_engine=search_engine, context=self.context) else: - action = SearchAndSummarize() + action = SearchAndSummarize self.set_actions([action]) self._watch([UserRequirement]) + return self diff --git a/metagpt/roles/searcher.py b/metagpt/roles/searcher.py index e0d2dbb65..557c5ae95 100644 --- a/metagpt/roles/searcher.py +++ b/metagpt/roles/searcher.py @@ -8,14 +8,17 @@ the `cause_by` value in the `Message` to a string to support the new message distribution feature. """ -from pydantic import Field +from typing import Optional -from metagpt.actions import ActionOutput, SearchAndSummarize +from pydantic import Field, model_validator + +from metagpt.actions import SearchAndSummarize from metagpt.actions.action_node import ActionNode +from metagpt.actions.action_output import ActionOutput from metagpt.logs import logger from metagpt.roles import Role from metagpt.schema import Message -from metagpt.tools import SearchEngineType +from metagpt.tools.search_engine import SearchEngine class Searcher(Role): @@ -27,33 +30,22 @@ class Searcher(Role): profile (str): Role profile. goal (str): Goal of the searcher. constraints (str): Constraints or limitations for the searcher. - engine (SearchEngineType): The type of search engine to use. + search_engine (SearchEngine): The search engine to use. """ name: str = Field(default="Alice") profile: str = Field(default="Smart Assistant") goal: str = "Provide search services for users" constraints: str = "Answer is rich and complete" - engine: SearchEngineType = SearchEngineType.SERPAPI_GOOGLE + search_engine: Optional[SearchEngine] = None - def __init__(self, **kwargs) -> None: - """ - Initializes the Searcher role with given attributes. - - Args: - name (str): Name of the searcher. - profile (str): Role profile. - goal (str): Goal of the searcher. - constraints (str): Constraints or limitations for the searcher. - engine (SearchEngineType): The type of search engine to use. - """ - super().__init__(**kwargs) - self.set_actions([SearchAndSummarize(engine=self.engine)]) - - def set_search_func(self, search_func): - """Sets a custom search function for the searcher.""" - action = SearchAndSummarize(name="", engine=SearchEngineType.CUSTOM_ENGINE, search_func=search_func) - self.set_actions([action]) + @model_validator(mode="after") + def post_root(self): + if self.search_engine: + self.set_actions([SearchAndSummarize(search_engine=self.search_engine, context=self.context)]) + else: + self.set_actions([SearchAndSummarize]) + return self async def _act_sp(self) -> Message: """Performs the search action in a single process.""" diff --git a/metagpt/schema.py b/metagpt/schema.py index 853a9c6bb..15854f676 100644 --- a/metagpt/schema.py +++ b/metagpt/schema.py @@ -23,7 +23,7 @@ from abc import ABC from asyncio import Queue, QueueEmpty, wait_for from json import JSONDecodeError from pathlib import Path -from typing import Any, Dict, List, Optional, Type, TypeVar, Union +from typing import Any, Dict, Iterable, List, Optional, Type, TypeVar, Union from pydantic import ( BaseModel, @@ -37,10 +37,12 @@ from pydantic import ( ) from metagpt.const import ( + CODE_PLAN_AND_CHANGE_FILENAME, MESSAGE_ROUTE_CAUSE_BY, MESSAGE_ROUTE_FROM, MESSAGE_ROUTE_TO, MESSAGE_ROUTE_TO_ALL, + PRDS_FILE_REPO, SYSTEM_DESIGN_FILE_REPO, TASK_FILE_REPO, ) @@ -162,6 +164,26 @@ class Documents(BaseModel): docs: Dict[str, Document] = Field(default_factory=dict) + @classmethod + def from_iterable(cls, documents: Iterable[Document]) -> Documents: + """Create a Documents instance from a list of Document instances. + + :param documents: A list of Document instances. + :return: A Documents instance. + """ + + docs = {doc.filename: doc for doc in documents} + return Documents(docs=docs) + + def to_action_output(self) -> "ActionOutput": + """Convert to action output string. + + :return: A string representing action output. + """ + from metagpt.actions.action_output import ActionOutput + + return ActionOutput(content=self.model_dump_json(), instruct_content=self) + class Message(BaseModel): """list[: ]""" @@ -212,7 +234,7 @@ class Message(BaseModel): return any_to_str_set(send_to if send_to else {MESSAGE_ROUTE_TO_ALL}) @field_serializer("instruct_content", mode="plain") - def ser_instruct_content(self, ic: BaseModel) -> Union[str, None]: + def ser_instruct_content(self, ic: BaseModel) -> Union[dict, None]: ic_dict = None if ic: # compatible with custom-defined ActionOutput @@ -308,6 +330,200 @@ class AIMessage(Message): super().__init__(content=content, role="assistant") +class Task(BaseModel): + task_id: str = "" + dependent_task_ids: list[str] = [] # Tasks prerequisite to this Task + instruction: str = "" + task_type: str = "" + code: str = "" + result: str = "" + is_success: bool = False + is_finished: bool = False + + def reset(self): + self.code = "" + self.result = "" + self.is_success = False + self.is_finished = False + + def update_task_result(self, task_result: TaskResult): + self.code = task_result.code + self.result = task_result.result + self.is_success = task_result.is_success + + +class TaskResult(BaseModel): + """Result of taking a task, with result and is_success required to be filled""" + + code: str = "" + result: str + is_success: bool + + +class Plan(BaseModel): + goal: str + context: str = "" + tasks: list[Task] = [] + task_map: dict[str, Task] = {} + current_task_id: str = "" + + def _topological_sort(self, tasks: list[Task]): + task_map = {task.task_id: task for task in tasks} + dependencies = {task.task_id: set(task.dependent_task_ids) for task in tasks} + sorted_tasks = [] + visited = set() + + def visit(task_id): + if task_id in visited: + return + visited.add(task_id) + for dependent_id in dependencies.get(task_id, []): + visit(dependent_id) + sorted_tasks.append(task_map[task_id]) + + for task in tasks: + visit(task.task_id) + + return sorted_tasks + + def add_tasks(self, tasks: list[Task]): + """ + Integrates new tasks into the existing plan, ensuring dependency order is maintained. + + This method performs two primary functions based on the current state of the task list: + 1. If there are no existing tasks, it topologically sorts the provided tasks to ensure + correct execution order based on dependencies, and sets these as the current tasks. + 2. If there are existing tasks, it merges the new tasks with the existing ones. It maintains + any common prefix of tasks (based on task_id and instruction) and appends the remainder + of the new tasks. The current task is updated to the first unfinished task in this merged list. + + Args: + tasks (list[Task]): A list of tasks (may be unordered) to add to the plan. + + Returns: + None: The method updates the internal state of the plan but does not return anything. + """ + if not tasks: + return + + # Topologically sort the new tasks to ensure correct dependency order + new_tasks = self._topological_sort(tasks) + + if not self.tasks: + # If there are no existing tasks, set the new tasks as the current tasks + self.tasks = new_tasks + + else: + # Find the length of the common prefix between existing and new tasks + prefix_length = 0 + for old_task, new_task in zip(self.tasks, new_tasks): + if old_task.task_id != new_task.task_id or old_task.instruction != new_task.instruction: + break + prefix_length += 1 + + # Combine the common prefix with the remainder of the new tasks + final_tasks = self.tasks[:prefix_length] + new_tasks[prefix_length:] + self.tasks = final_tasks + + # Update current_task_id to the first unfinished task in the merged list + self._update_current_task() + + # Update the task map for quick access to tasks by ID + self.task_map = {task.task_id: task for task in self.tasks} + + def reset_task(self, task_id: str): + """ + Clear code and result of the task based on task_id, and set the task as unfinished. + + Args: + task_id (str): The ID of the task to be reset. + + Returns: + None + """ + if task_id in self.task_map: + task = self.task_map[task_id] + task.reset() + + def replace_task(self, new_task: Task): + """ + Replace an existing task with the new input task based on task_id, and reset all tasks depending on it. + + Args: + new_task (Task): The new task that will replace an existing one. + + Returns: + None + """ + assert new_task.task_id in self.task_map + # Replace the task in the task map and the task list + self.task_map[new_task.task_id] = new_task + for i, task in enumerate(self.tasks): + if task.task_id == new_task.task_id: + self.tasks[i] = new_task + break + + # Reset dependent tasks + for task in self.tasks: + if new_task.task_id in task.dependent_task_ids: + self.reset_task(task.task_id) + + def append_task(self, new_task: Task): + """ + Append a new task to the end of existing task sequences + + Args: + new_task (Task): The new task to be appended to the existing task sequence + + Returns: + None + """ + assert not self.has_task_id(new_task.task_id), "Task already in current plan, use replace_task instead" + + assert all( + [self.has_task_id(dep_id) for dep_id in new_task.dependent_task_ids] + ), "New task has unknown dependencies" + + # Existing tasks do not depend on the new task, it's fine to put it to the end of the sorted task sequence + self.tasks.append(new_task) + self.task_map[new_task.task_id] = new_task + self._update_current_task() + + def has_task_id(self, task_id: str) -> bool: + return task_id in self.task_map + + def _update_current_task(self): + current_task_id = "" + for task in self.tasks: + if not task.is_finished: + current_task_id = task.task_id + break + self.current_task_id = current_task_id # all tasks finished + + @property + def current_task(self) -> Task: + """Find current task to execute + + Returns: + Task: the current task to be executed + """ + return self.task_map.get(self.current_task_id, None) + + def finish_current_task(self): + """Finish current task, set Task.is_finished=True, set current task to next task""" + if self.current_task_id: + self.current_task.is_finished = True + self._update_current_task() # set to next task + + def get_finished_tasks(self) -> list[Task]: + """return all finished tasks in correct linearized order + + Returns: + list[Task]: list of finished tasks + """ + return [task for task in self.tasks if task.is_finished] + + class MessageQueue(BaseModel): """Message queue which supports asynchronous updates.""" @@ -450,6 +666,30 @@ class BugFixContext(BaseContext): filename: str = "" +class CodePlanAndChangeContext(BaseModel): + filename: str = CODE_PLAN_AND_CHANGE_FILENAME + requirement: str = "" + prd_filename: str = "" + design_filename: str = "" + task_filename: str = "" + + @staticmethod + def loads(filenames: List, **kwargs) -> CodePlanAndChangeContext: + ctx = CodePlanAndChangeContext(requirement=kwargs.get("requirement", "")) + for filename in filenames: + filename = Path(filename) + if filename.is_relative_to(PRDS_FILE_REPO): + ctx.prd_filename = filename.name + continue + if filename.is_relative_to(SYSTEM_DESIGN_FILE_REPO): + ctx.design_filename = filename.name + continue + if filename.is_relative_to(TASK_FILE_REPO): + ctx.task_filename = filename.name + continue + return ctx + + # mermaid class view class ClassMeta(BaseModel): name: str = "" diff --git a/metagpt/startup.py b/metagpt/software_company.py similarity index 89% rename from metagpt/startup.py rename to metagpt/software_company.py index 771cde80c..26bb29cd1 100644 --- a/metagpt/startup.py +++ b/metagpt/software_company.py @@ -1,5 +1,6 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- + import asyncio import shutil from pathlib import Path @@ -8,24 +9,26 @@ import typer from metagpt.config2 import config from metagpt.const import CONFIG_ROOT, METAGPT_ROOT +from metagpt.context import Context +from metagpt.utils.project_repo import ProjectRepo app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) def generate_repo( idea, - investment, - n_round, - code_review, - run_tests, - implement, - project_name, - inc, - project_path, - reqa_file, - max_auto_summarize_code, - recover_path, -): + investment=3.0, + n_round=5, + code_review=True, + run_tests=False, + implement=True, + project_name="", + inc=False, + project_path="", + reqa_file="", + max_auto_summarize_code=0, + recover_path=None, +) -> ProjectRepo: """Run the startup logic. Can be called from CLI or other Python scripts.""" from metagpt.roles import ( Architect, @@ -37,9 +40,10 @@ def generate_repo( from metagpt.team import Team config.update_via_cli(project_path, project_name, inc, reqa_file, max_auto_summarize_code) + ctx = Context(config=config) if not recover_path: - company = Team() + company = Team(context=ctx) company.hire( [ ProductManager(), @@ -58,13 +62,15 @@ def generate_repo( if not stg_path.exists() or not str(stg_path).endswith("team"): raise FileNotFoundError(f"{recover_path} not exists or not endswith `team`") - company = Team.deserialize(stg_path=stg_path) + company = Team.deserialize(stg_path=stg_path, context=ctx) idea = company.idea company.invest(investment) company.run_project(idea) asyncio.run(company.run(n_round=n_round)) + return ctx.repo + @app.command("", help="Start a new project.") def startup( diff --git a/metagpt/strategy/planner.py b/metagpt/strategy/planner.py new file mode 100644 index 000000000..fd635df39 --- /dev/null +++ b/metagpt/strategy/planner.py @@ -0,0 +1,139 @@ +from __future__ import annotations + +import json + +from pydantic import BaseModel, Field + +from metagpt.actions.ci.ask_review import AskReview, ReviewConst +from metagpt.actions.ci.write_plan import ( + WritePlan, + precheck_update_plan_from_rsp, + update_plan_from_rsp, +) +from metagpt.logs import logger +from metagpt.memory import Memory +from metagpt.schema import Message, Plan, Task, TaskResult + +STRUCTURAL_CONTEXT = """ +## User Requirement +{user_requirement} +## Context +{context} +## Current Plan +{tasks} +## Current Task +{current_task} +""" + + +class Planner(BaseModel): + plan: Plan + working_memory: Memory = Field( + default_factory=Memory + ) # memory for working on each task, discarded each time a task is done + auto_run: bool = False + use_tools: bool = False + + def __init__(self, goal: str = "", plan: Plan = None, **kwargs): + plan = plan or Plan(goal=goal) + super().__init__(plan=plan, **kwargs) + + @property + def current_task(self): + return self.plan.current_task + + @property + def current_task_id(self): + return self.plan.current_task_id + + async def update_plan(self, goal: str = "", max_tasks: int = 3, max_retries: int = 3): + if goal: + self.plan = Plan(goal=goal) + + plan_confirmed = False + while not plan_confirmed: + context = self.get_useful_memories() + rsp = await WritePlan().run(context, max_tasks=max_tasks, use_tools=self.use_tools) + self.working_memory.add(Message(content=rsp, role="assistant", cause_by=WritePlan)) + + # precheck plan before asking reviews + is_plan_valid, error = precheck_update_plan_from_rsp(rsp, self.plan) + if not is_plan_valid and max_retries > 0: + error_msg = f"The generated plan is not valid with error: {error}, try regenerating, remember to generate either the whole plan or the single changed task only" + logger.warning(error_msg) + self.working_memory.add(Message(content=error_msg, role="assistant", cause_by=WritePlan)) + max_retries -= 1 + continue + + _, plan_confirmed = await self.ask_review(trigger=ReviewConst.TASK_REVIEW_TRIGGER) + + update_plan_from_rsp(rsp=rsp, current_plan=self.plan) + + self.working_memory.clear() + + async def process_task_result(self, task_result: TaskResult): + # ask for acceptance, users can other refuse and change tasks in the plan + review, task_result_confirmed = await self.ask_review(task_result) + + if task_result_confirmed: + # tick off this task and record progress + await self.confirm_task(self.current_task, task_result, review) + + elif "redo" in review: + # Ask the Role to redo this task with help of review feedback, + # useful when the code run is successful but the procedure or result is not what we want + pass # simply pass, not confirming the result + + else: + # update plan according to user's feedback and to take on changed tasks + await self.update_plan() + + async def ask_review( + self, + task_result: TaskResult = None, + auto_run: bool = None, + trigger: str = ReviewConst.TASK_REVIEW_TRIGGER, + review_context_len: int = 5, + ): + """ + Ask to review the task result, reviewer needs to provide confirmation or request change. + If human confirms the task result, then we deem the task completed, regardless of whether the code run succeeds; + if auto mode, then the code run has to succeed for the task to be considered completed. + """ + auto_run = auto_run or self.auto_run + if not auto_run: + context = self.get_useful_memories() + review, confirmed = await AskReview().run( + context=context[-review_context_len:], plan=self.plan, trigger=trigger + ) + if not confirmed: + self.working_memory.add(Message(content=review, role="user", cause_by=AskReview)) + return review, confirmed + confirmed = task_result.is_success if task_result else True + return "", confirmed + + async def confirm_task(self, task: Task, task_result: TaskResult, review: str): + task.update_task_result(task_result=task_result) + self.plan.finish_current_task() + self.working_memory.clear() + + confirmed_and_more = ( + ReviewConst.CONTINUE_WORDS[0] in review.lower() and review.lower() not in ReviewConst.CONTINUE_WORDS[0] + ) # "confirm, ... (more content, such as changing downstream tasks)" + if confirmed_and_more: + self.working_memory.add(Message(content=review, role="user", cause_by=AskReview)) + await self.update_plan(review) + + def get_useful_memories(self, task_exclude_field=None) -> list[Message]: + """find useful memories only to reduce context length and improve performance""" + user_requirement = self.plan.goal + context = self.plan.context + tasks = [task.dict(exclude=task_exclude_field) for task in self.plan.tasks] + tasks = json.dumps(tasks, indent=4, ensure_ascii=False) + current_task = self.plan.current_task.json() if self.plan.current_task else {} + context = STRUCTURAL_CONTEXT.format( + user_requirement=user_requirement, context=context, tasks=tasks, current_task=current_task + ) + context_msg = [Message(content=context, role="user")] + + return context_msg + self.working_memory.get() diff --git a/metagpt/strategy/search_space.py b/metagpt/strategy/search_space.py new file mode 100644 index 000000000..c643a2f11 --- /dev/null +++ b/metagpt/strategy/search_space.py @@ -0,0 +1,20 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/1/30 17:15 +@Author : alexanderwu +@File : search_space.py +""" + + +class SearchSpace: + """SearchSpace: 用于定义一个搜索空间,搜索空间中的节点是 ActionNode 类。""" + + def __init__(self): + self.search_space = {} + + def add_node(self, node): + self.search_space[node.key] = node + + def get_node(self, key): + return self.search_space[key] diff --git a/metagpt/strategy/solver.py b/metagpt/strategy/solver.py new file mode 100644 index 000000000..e7d61a881 --- /dev/null +++ b/metagpt/strategy/solver.py @@ -0,0 +1,77 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/1/30 17:13 +@Author : alexanderwu +@File : solver.py +""" +from abc import abstractmethod + +from metagpt.actions.action_graph import ActionGraph +from metagpt.provider.base_llm import BaseLLM +from metagpt.strategy.search_space import SearchSpace + + +class BaseSolver: + """AbstractSolver: defines the interface of a solver.""" + + def __init__(self, graph: ActionGraph, search_space: SearchSpace, llm: BaseLLM, context): + """ + :param graph: ActionGraph + :param search_space: SearchSpace + :param llm: BaseLLM + :param context: Context + """ + self.graph = graph + self.search_space = search_space + self.llm = llm + self.context = context + + @abstractmethod + async def solve(self): + """abstract method to solve the problem.""" + + +class NaiveSolver(BaseSolver): + """NaiveSolver: Iterate all the nodes in the graph and execute them one by one.""" + + async def solve(self): + self.graph.topological_sort() + for key in self.graph.execution_order: + op = self.graph.nodes[key] + await op.fill(self.context, self.llm, mode="root") + + +class TOTSolver(BaseSolver): + """TOTSolver: Tree of Thought""" + + async def solve(self): + raise NotImplementedError + + +class CodeInterpreterSolver(BaseSolver): + """CodeInterpreterSolver: Write&Run code in the graph""" + + async def solve(self): + raise NotImplementedError + + +class ReActSolver(BaseSolver): + """ReActSolver: ReAct algorithm""" + + async def solve(self): + raise NotImplementedError + + +class IOSolver(BaseSolver): + """IOSolver: use LLM directly to solve the problem""" + + async def solve(self): + raise NotImplementedError + + +class COTSolver(BaseSolver): + """COTSolver: Chain of Thought""" + + async def solve(self): + raise NotImplementedError diff --git a/metagpt/team.py b/metagpt/team.py index 96a27d482..35f987b57 100644 --- a/metagpt/team.py +++ b/metagpt/team.py @@ -10,12 +10,13 @@ import warnings from pathlib import Path -from typing import Any +from typing import Any, Optional from pydantic import BaseModel, ConfigDict, Field from metagpt.actions import UserRequirement from metagpt.const import MESSAGE_ROUTE_TO_ALL, SERDESER_PATH +from metagpt.context import Context from metagpt.environment import Environment from metagpt.logs import logger from metagpt.roles import Role @@ -36,12 +37,17 @@ class Team(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) - env: Environment = Field(default_factory=Environment) + env: Optional[Environment] = None investment: float = Field(default=10.0) idea: str = Field(default="") - def __init__(self, **data: Any): + def __init__(self, context: Context = None, **data: Any): super(Team, self).__init__(**data) + ctx = context or Context() + if not self.env: + self.env = Environment(context=ctx) + else: + self.env.context = ctx # The `env` object is allocated by deserialization if "roles" in data: self.hire(data["roles"]) if "env_desc" in data: @@ -54,7 +60,7 @@ class Team(BaseModel): write_json_file(team_info_path, self.model_dump()) @classmethod - def deserialize(cls, stg_path: Path) -> "Team": + def deserialize(cls, stg_path: Path, context: Context = None) -> "Team": """stg_path = ./storage/team""" # recover team_info team_info_path = stg_path.joinpath("team.json") @@ -64,7 +70,8 @@ class Team(BaseModel): ) team_info: dict = read_json_file(team_info_path) - team = Team(**team_info) + ctx = context or Context() + team = Team(**team_info, context=ctx) return team def hire(self, roles: list[Role]): @@ -83,7 +90,7 @@ class Team(BaseModel): logger.info(f"Investment: ${investment}.") def _check_balance(self): - if self.cost_manager.total_cost > self.cost_manager.max_budget: + if self.cost_manager.total_cost >= self.cost_manager.max_budget: raise NoMoneyException(self.cost_manager.total_cost, f"Insufficient funds: {self.cost_manager.max_budget}") def run_project(self, idea, send_to: str = ""): diff --git a/metagpt/tools/__init__.py b/metagpt/tools/__init__.py index aab8c990c..c1f604df9 100644 --- a/metagpt/tools/__init__.py +++ b/metagpt/tools/__init__.py @@ -6,8 +6,11 @@ @File : __init__.py """ - from enum import Enum +from metagpt.tools import libs # this registers all tools +from metagpt.tools.tool_registry import TOOL_REGISTRY + +_ = libs, TOOL_REGISTRY # Avoid pre-commit error class SearchEngineType(Enum): diff --git a/metagpt/tools/libs/__init__.py b/metagpt/tools/libs/__init__.py new file mode 100644 index 000000000..c9767c1e5 --- /dev/null +++ b/metagpt/tools/libs/__init__.py @@ -0,0 +1,15 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Time : 2023/11/16 16:32 +# @Author : lidanyang +# @File : __init__.py +# @Desc : +from metagpt.tools.libs import ( + data_preprocess, + feature_engineering, + sd_engine, + gpt_v_generator, + web_scraping, +) + +_ = data_preprocess, feature_engineering, sd_engine, gpt_v_generator, web_scraping # Avoid pre-commit error diff --git a/metagpt/tools/libs/data_preprocess.py b/metagpt/tools/libs/data_preprocess.py new file mode 100644 index 000000000..7a3d019bf --- /dev/null +++ b/metagpt/tools/libs/data_preprocess.py @@ -0,0 +1,249 @@ +from __future__ import annotations + +import json + +import numpy as np +import pandas as pd +from sklearn.impute import SimpleImputer +from sklearn.preprocessing import ( + LabelEncoder, + MaxAbsScaler, + MinMaxScaler, + OneHotEncoder, + OrdinalEncoder, + RobustScaler, + StandardScaler, +) + +from metagpt.tools.tool_registry import register_tool +from metagpt.tools.tool_type import ToolType + +TOOL_TYPE = ToolType.DATA_PREPROCESS.type_name + + +class MLProcess: + def fit(self, df: pd.DataFrame): + """ + Fit a model to be used in subsequent transform. + + Args: + df (pd.DataFrame): The input DataFrame. + """ + raise NotImplementedError + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + """ + Transform the input DataFrame with the fitted model. + + Args: + df (pd.DataFrame): The input DataFrame. + + Returns: + pd.DataFrame: The transformed DataFrame. + """ + raise NotImplementedError + + def fit_transform(self, df: pd.DataFrame) -> pd.DataFrame: + """ + Fit and transform the input DataFrame. + + Args: + df (pd.DataFrame): The input DataFrame. + + Returns: + pd.DataFrame: The transformed DataFrame. + """ + self.fit(df) + return self.transform(df) + + +class DataPreprocessTool(MLProcess): + """ + Completing a data preprocessing operation. + """ + + def __init__(self, features: list): + """ + Initialize self. + + Args: + features (list): Columns to be processed. + """ + self.features = features + self.model = None # to be filled by specific subclass Tool + + def fit(self, df: pd.DataFrame): + if len(self.features) == 0: + return + self.model.fit(df[self.features]) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + if len(self.features) == 0: + return df + new_df = df.copy() + new_df[self.features] = self.model.transform(new_df[self.features]) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class FillMissingValue(DataPreprocessTool): + """ + Completing missing values with simple strategies. + """ + + def __init__(self, features: list, strategy: str = "mean", fill_value=None): + """ + Initialize self. + + Args: + features (list): Columns to be processed. + strategy (str, optional): The imputation strategy, notice 'mean' and 'median' can only + be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'. + fill_value (int, optional): Fill_value is used to replace all occurrences of missing_values. + Defaults to None. + """ + self.features = features + self.model = SimpleImputer(strategy=strategy, fill_value=fill_value) + + +@register_tool(tool_type=TOOL_TYPE) +class MinMaxScale(DataPreprocessTool): + """ + Transform features by scaling each feature to a range, which is (0, 1). + """ + + def __init__(self, features: list): + self.features = features + self.model = MinMaxScaler() + + +@register_tool(tool_type=TOOL_TYPE) +class StandardScale(DataPreprocessTool): + """ + Standardize features by removing the mean and scaling to unit variance. + """ + + def __init__(self, features: list): + self.features = features + self.model = StandardScaler() + + +@register_tool(tool_type=TOOL_TYPE) +class MaxAbsScale(DataPreprocessTool): + """ + Scale each feature by its maximum absolute value. + """ + + def __init__(self, features: list): + self.features = features + self.model = MaxAbsScaler() + + +@register_tool(tool_type=TOOL_TYPE) +class RobustScale(DataPreprocessTool): + """ + Apply the RobustScaler to scale features using statistics that are robust to outliers. + """ + + def __init__(self, features: list): + self.features = features + self.model = RobustScaler() + + +@register_tool(tool_type=TOOL_TYPE) +class OrdinalEncode(DataPreprocessTool): + """ + Encode categorical features as ordinal integers. + """ + + def __init__(self, features: list): + self.features = features + self.model = OrdinalEncoder() + + +@register_tool(tool_type=TOOL_TYPE) +class OneHotEncode(DataPreprocessTool): + """ + Apply one-hot encoding to specified categorical columns, the original columns will be dropped. + """ + + def __init__(self, features: list): + self.features = features + self.model = OneHotEncoder(handle_unknown="ignore", sparse=False) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + ts_data = self.model.transform(df[self.features]) + new_columns = self.model.get_feature_names_out(self.features) + ts_data = pd.DataFrame(ts_data, columns=new_columns, index=df.index) + new_df = df.drop(self.features, axis=1) + new_df = pd.concat([new_df, ts_data], axis=1) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class LabelEncode(DataPreprocessTool): + """ + Apply label encoding to specified categorical columns in-place. + """ + + def __init__(self, features: list): + """ + Initialize self. + + Args: + features (list): Categorical columns to be label encoded. + """ + self.features = features + self.le_encoders = [] + + def fit(self, df: pd.DataFrame): + if len(self.features) == 0: + return + for col in self.features: + le = LabelEncoder().fit(df[col].astype(str).unique().tolist() + ["unknown"]) + self.le_encoders.append(le) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + if len(self.features) == 0: + return df + new_df = df.copy() + for i in range(len(self.features)): + data_list = df[self.features[i]].astype(str).tolist() + for unique_item in np.unique(df[self.features[i]].astype(str)): + if unique_item not in self.le_encoders[i].classes_: + data_list = ["unknown" if x == unique_item else x for x in data_list] + new_df[self.features[i]] = self.le_encoders[i].transform(data_list) + return new_df + + +def get_column_info(df: pd.DataFrame) -> dict: + """ + Analyzes a DataFrame and categorizes its columns based on data types. + + Args: + df (pd.DataFrame): The DataFrame to be analyzed. + + Returns: + dict: A dictionary with four keys ('Category', 'Numeric', 'Datetime', 'Others'). + Each key corresponds to a list of column names belonging to that category. + """ + column_info = { + "Category": [], + "Numeric": [], + "Datetime": [], + "Others": [], + } + for col in df.columns: + data_type = str(df[col].dtype).replace("dtype('", "").replace("')", "") + if data_type.startswith("object"): + column_info["Category"].append(col) + elif data_type.startswith("int") or data_type.startswith("float"): + column_info["Numeric"].append(col) + elif data_type.startswith("datetime"): + column_info["Datetime"].append(col) + else: + column_info["Others"].append(col) + + if len(json.dumps(column_info)) > 2000: + column_info["Numeric"] = column_info["Numeric"][0:5] + ["Too many cols, omission here..."] + return column_info diff --git a/metagpt/tools/libs/feature_engineering.py b/metagpt/tools/libs/feature_engineering.py new file mode 100644 index 000000000..40bfb2fc7 --- /dev/null +++ b/metagpt/tools/libs/feature_engineering.py @@ -0,0 +1,435 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Time : 2023/11/17 10:33 +# @Author : lidanyang +# @File : feature_engineering.py +# @Desc : Feature Engineering Tools +from __future__ import annotations + +import itertools + +# import lightgbm as lgb +import numpy as np +import pandas as pd +from joblib import Parallel, delayed +from pandas.core.dtypes.common import is_object_dtype +from sklearn.feature_selection import VarianceThreshold +from sklearn.model_selection import KFold +from sklearn.preprocessing import KBinsDiscretizer, PolynomialFeatures + +from metagpt.tools.libs.data_preprocess import MLProcess +from metagpt.tools.tool_registry import register_tool +from metagpt.tools.tool_type import ToolType + +TOOL_TYPE = ToolType.FEATURE_ENGINEERING.type_name + + +@register_tool(tool_type=TOOL_TYPE) +class PolynomialExpansion(MLProcess): + """ + Add polynomial and interaction features from selected numeric columns to input DataFrame. + """ + + def __init__(self, cols: list, label_col: str, degree: int = 2): + """ + Initialize self. + + Args: + cols (list): Columns for polynomial expansion. + label_col (str): Label column name. + degree (int, optional): The degree of the polynomial features. Defaults to 2. + """ + self.cols = cols + self.degree = degree + self.label_col = label_col + if self.label_col in self.cols: + self.cols.remove(self.label_col) + self.poly = PolynomialFeatures(degree=degree, include_bias=False) + + def fit(self, df: pd.DataFrame): + if len(self.cols) == 0: + return + if len(self.cols) > 10: + corr = df[self.cols + [self.label_col]].corr() + corr = corr[self.label_col].abs().sort_values(ascending=False) + self.cols = corr.index.tolist()[1:11] + + self.poly.fit(df[self.cols].fillna(0)) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + if len(self.cols) == 0: + return df + ts_data = self.poly.transform(df[self.cols].fillna(0)) + column_name = self.poly.get_feature_names_out(self.cols) + ts_data = pd.DataFrame(ts_data, index=df.index, columns=column_name) + new_df = df.drop(self.cols, axis=1) + new_df = pd.concat([new_df, ts_data], axis=1) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class CatCount(MLProcess): + """ + Add value counts of a categorical column as new feature. + """ + + def __init__(self, col: str): + """ + Initialize self. + + Args: + col (str): Column for value counts. + """ + self.col = col + self.encoder_dict = None + + def fit(self, df: pd.DataFrame): + self.encoder_dict = df[self.col].value_counts().to_dict() + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.copy() + new_df[f"{self.col}_cnt"] = new_df[self.col].map(self.encoder_dict) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class TargetMeanEncoder(MLProcess): + """ + Encode a categorical column by the mean of the label column, and adds the result as a new feature. + """ + + def __init__(self, col: str, label: str): + """ + Initialize self. + + Args: + col (str): Column to be mean encoded. + label (str): Predicted label column. + """ + self.col = col + self.label = label + self.encoder_dict = None + + def fit(self, df: pd.DataFrame): + self.encoder_dict = df.groupby(self.col)[self.label].mean().to_dict() + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.copy() + new_df[f"{self.col}_target_mean"] = new_df[self.col].map(self.encoder_dict) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class KFoldTargetMeanEncoder(MLProcess): + """ + Add a new feature to the DataFrame by k-fold mean encoding of a categorical column using the label column. + """ + + def __init__(self, col: str, label: str, n_splits: int = 5, random_state: int = 2021): + """ + Initialize self. + + Args: + col (str): Column to be k-fold mean encoded. + label (str): Predicted label column. + n_splits (int, optional): Number of splits for K-fold. Defaults to 5. + random_state (int, optional): Random seed. Defaults to 2021. + """ + self.col = col + self.label = label + self.n_splits = n_splits + self.random_state = random_state + self.encoder_dict = None + + def fit(self, df: pd.DataFrame): + tmp = df.copy() + kf = KFold(n_splits=self.n_splits, shuffle=True, random_state=self.random_state) + + global_mean = tmp[self.label].mean() + col_name = f"{self.col}_kf_target_mean" + for trn_idx, val_idx in kf.split(tmp, tmp[self.label]): + _trn, _val = tmp.iloc[trn_idx], tmp.iloc[val_idx] + tmp.loc[tmp.index[val_idx], col_name] = _val[self.col].map(_trn.groupby(self.col)[self.label].mean()) + tmp[col_name].fillna(global_mean, inplace=True) + self.encoder_dict = tmp.groupby(self.col)[col_name].mean().to_dict() + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.copy() + new_df[f"{self.col}_kf_target_mean"] = new_df[self.col].map(self.encoder_dict) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class CatCross(MLProcess): + """ + Add pairwise crossed features and convert them to numerical features. + """ + + def __init__(self, cols: list, max_cat_num: int = 100): + """ + Initialize self. + + Args: + cols (list): Columns to be pairwise crossed, at least 2 columns. + max_cat_num (int, optional): Maximum unique categories per crossed feature. Defaults to 100. + """ + self.cols = cols + self.max_cat_num = max_cat_num + self.combs = [] + self.combs_map = {} + + @staticmethod + def _cross_two(comb, df): + """ + Cross two columns and convert them to numerical features. + + Args: + comb (tuple): The pair of columns to be crossed. + df (pd.DataFrame): The input DataFrame. + + Returns: + tuple: The new column name and the crossed feature map. + """ + new_col = f"{comb[0]}_{comb[1]}" + new_col_combs = list(itertools.product(df[comb[0]].unique(), df[comb[1]].unique())) + ll = list(range(len(new_col_combs))) + comb_map = dict(zip(new_col_combs, ll)) + return new_col, comb_map + + def fit(self, df: pd.DataFrame): + for col in self.cols: + if df[col].nunique() > self.max_cat_num: + self.cols.remove(col) + self.combs = list(itertools.combinations(self.cols, 2)) + res = Parallel(n_jobs=4, require="sharedmem")(delayed(self._cross_two)(comb, df) for comb in self.combs) + self.combs_map = dict(res) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.copy() + for comb in self.combs: + new_col = f"{comb[0]}_{comb[1]}" + _map = self.combs_map[new_col] + new_df[new_col] = pd.Series(zip(new_df[comb[0]], new_df[comb[1]])).map(_map) + # set the unknown value to a new number + new_df[new_col].fillna(max(_map.values()) + 1, inplace=True) + new_df[new_col] = new_df[new_col].astype(int) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class GroupStat(MLProcess): + """ + Aggregate specified column in a DataFrame grouped by another column, adding new features named '__by_'. + """ + + def __init__(self, group_col: str, agg_col: str, agg_funcs: list): + """ + Initialize self. + + Args: + group_col (str): Column used for grouping. + agg_col (str): Column on which aggregation is performed. + agg_funcs (list): List of aggregation functions to apply, such as ['mean', 'std']. Each function must be supported by pandas. + """ + self.group_col = group_col + self.agg_col = agg_col + self.agg_funcs = agg_funcs + self.group_df = None + + def fit(self, df: pd.DataFrame): + group_df = df.groupby(self.group_col)[self.agg_col].agg(self.agg_funcs).reset_index() + group_df.columns = [self.group_col] + [ + f"{self.agg_col}_{agg_func}_by_{self.group_col}" for agg_func in self.agg_funcs + ] + self.group_df = group_df + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.merge(self.group_df, on=self.group_col, how="left") + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class SplitBins(MLProcess): + """ + Inplace binning of continuous data into intervals, returning integer-encoded bin identifiers directly. + """ + + def __init__(self, cols: list, strategy: str = "quantile"): + """ + Initialize self. + + Args: + cols (list): Columns to be binned inplace. + strategy (str, optional): Strategy used to define the widths of the bins. Enum: ['quantile', 'uniform', 'kmeans']. Defaults to 'quantile'. + """ + self.cols = cols + self.strategy = strategy + self.encoder = None + + def fit(self, df: pd.DataFrame): + self.encoder = KBinsDiscretizer(strategy=self.strategy, encode="ordinal") + self.encoder.fit(df[self.cols].fillna(0)) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df.copy() + new_df[self.cols] = self.encoder.transform(new_df[self.cols].fillna(0)) + return new_df + + +# @register_tool(tool_type=TOOL_TYPE) +class ExtractTimeComps(MLProcess): + """ + Extract time components from a datetime column and add them as new features. + """ + + def __init__(self, time_col: str, time_comps: list): + """ + Initialize self. + + Args: + time_col (str): The name of the column containing time data. + time_comps (list): List of time components to extract. Each component must be in ['year', 'month', 'day', 'hour', 'dayofweek', 'is_weekend']. + """ + self.time_col = time_col + self.time_comps = time_comps + + def fit(self, df: pd.DataFrame): + pass + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + time_s = pd.to_datetime(df[self.time_col], errors="coerce") + time_comps_df = pd.DataFrame() + + if "year" in self.time_comps: + time_comps_df["year"] = time_s.dt.year + if "month" in self.time_comps: + time_comps_df["month"] = time_s.dt.month + if "day" in self.time_comps: + time_comps_df["day"] = time_s.dt.day + if "hour" in self.time_comps: + time_comps_df["hour"] = time_s.dt.hour + if "dayofweek" in self.time_comps: + time_comps_df["dayofweek"] = time_s.dt.dayofweek + 1 + if "is_weekend" in self.time_comps: + time_comps_df["is_weekend"] = time_s.dt.dayofweek.isin([5, 6]).astype(int) + new_df = pd.concat([df, time_comps_df], axis=1) + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class GeneralSelection(MLProcess): + """ + Drop all nan feats and feats with only one unique value. + """ + + def __init__(self, label_col: str): + self.label_col = label_col + self.feats = [] + + def fit(self, df: pd.DataFrame): + feats = [f for f in df.columns if f != self.label_col] + for col in df.columns: + if df[col].isnull().sum() / df.shape[0] == 1: + feats.remove(col) + + if df[col].nunique() == 1: + feats.remove(col) + + if df.loc[df[col] == np.inf].shape[0] != 0 or df.loc[df[col] == np.inf].shape[0] != 0: + feats.remove(col) + + if is_object_dtype(df[col]) and df[col].nunique() == df.shape[0]: + feats.remove(col) + + self.feats = feats + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df[self.feats + [self.label_col]] + return new_df + + +# skip for now because lgb is needed +# @register_tool(tool_type=TOOL_TYPE) +class TreeBasedSelection(MLProcess): + """ + Select features based on tree-based model and remove features with low importance. + """ + + def __init__(self, label_col: str, task_type: str): + """ + Initialize self. + + Args: + label_col (str): Label column name. + task_type (str): Task type, 'cls' for classification, 'mcls' for multi-class classification, 'reg' for regression. + """ + self.label_col = label_col + self.task_type = task_type + self.feats = None + + def fit(self, df: pd.DataFrame): + params = { + "boosting_type": "gbdt", + "objective": "binary", + "learning_rate": 0.1, + "num_leaves": 31, + } + + if self.task_type == "cls": + params["objective"] = "binary" + params["metric"] = "auc" + elif self.task_type == "mcls": + params["objective"] = "multiclass" + params["num_class"] = df[self.label_col].nunique() + params["metric"] = "auc_mu" + elif self.task_type == "reg": + params["objective"] = "regression" + params["metric"] = "rmse" + + num_cols = df.select_dtypes(include=np.number).columns.tolist() + cols = [f for f in num_cols if f not in [self.label_col]] + + dtrain = lgb.Dataset(df[cols], df[self.label_col]) + model = lgb.train(params, dtrain, num_boost_round=100) + df_imp = pd.DataFrame({"feature_name": dtrain.feature_name, "importance": model.feature_importance("gain")}) + + df_imp.sort_values("importance", ascending=False, inplace=True) + df_imp = df_imp[df_imp["importance"] > 0] + self.feats = df_imp["feature_name"].tolist() + self.feats.append(self.label_col) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df[self.feats] + return new_df + + +@register_tool(tool_type=TOOL_TYPE) +class VarianceBasedSelection(MLProcess): + """ + Select features based on variance and remove features with low variance. + """ + + def __init__(self, label_col: str, threshold: float = 0): + """ + Initialize self. + + Args: + label_col (str): Label column name. + threshold (float, optional): Threshold for variance. Defaults to 0. + """ + self.label_col = label_col + self.threshold = threshold + self.feats = None + self.selector = VarianceThreshold(threshold=self.threshold) + + def fit(self, df: pd.DataFrame): + num_cols = df.select_dtypes(include=np.number).columns.tolist() + cols = [f for f in num_cols if f not in [self.label_col]] + + self.selector.fit(df[cols]) + self.feats = df[cols].columns[self.selector.get_support(indices=True)].tolist() + self.feats.append(self.label_col) + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + new_df = df[self.feats] + return new_df diff --git a/metagpt/tools/libs/gpt_v_generator.py b/metagpt/tools/libs/gpt_v_generator.py new file mode 100644 index 000000000..b1e8317ed --- /dev/null +++ b/metagpt/tools/libs/gpt_v_generator.py @@ -0,0 +1,124 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/12 +@Author : mannaandpoem +@File : gpt_v_generator.py +""" +import os +from pathlib import Path + +from metagpt.const import DEFAULT_WORKSPACE_ROOT +from metagpt.tools.tool_registry import register_tool +from metagpt.tools.tool_type import ToolType +from metagpt.utils.common import encode_image + +ANALYZE_LAYOUT_PROMPT = """You are now a UI/UX, please generate layout information for this image: + +NOTE: The image does not have a commercial logo or copyright information. It is just a sketch image of the design. +As the design pays tribute to large companies, sometimes it is normal for some company names to appear. Don't worry. """ + +GENERATE_PROMPT = """You are now a UI/UX and Web Developer. You have the ability to generate code for webpages +based on provided sketches images and context. +Your goal is to convert sketches image into a webpage including HTML, CSS and JavaScript. + +NOTE: The image does not have a commercial logo or copyright information. It is just a sketch image of the design. +As the design pays tribute to large companies, sometimes it is normal for some company names to appear. Don't worry. + +Now, please generate the corresponding webpage code including HTML, CSS and JavaScript:""" + + +@register_tool( + tool_type=ToolType.IMAGE2WEBPAGE.type_name, include_functions=["__init__", "generate_webpages", "save_webpages"] +) +class GPTvGenerator: + """Class for generating webpages at once. + + This class provides methods to generate webpages including all code (HTML, CSS, and JavaScript) based on an image. + It utilizes a vision model to analyze the layout from an image and generate webpage codes accordingly. + """ + + def __init__(self): + """Initialize GPTvGenerator class with default values from the configuration.""" + from metagpt.config2 import config + from metagpt.llm import LLM + + self.llm = LLM(llm_config=config.get_openai_llm()) + self.llm.model = "gpt-4-vision-preview" + + async def analyze_layout(self, image_path: Path) -> str: + """Asynchronously analyze the layout of the given image and return the result. + + This is a helper method to generate a layout description based on the image. + + Args: + image_path (Path): Path of the image to analyze. + + Returns: + str: The layout analysis result. + """ + return await self.llm.aask(msg=ANALYZE_LAYOUT_PROMPT, images=[encode_image(image_path)]) + + async def generate_webpages(self, image_path: str) -> str: + """Asynchronously generate webpages including all code (HTML, CSS, and JavaScript) in one go based on the image. + + Args: + image_path (str): The path of the image file. + + Returns: + str: Generated webpages content. + """ + if isinstance(image_path, str): + image_path = Path(image_path) + layout = await self.analyze_layout(image_path) + prompt = GENERATE_PROMPT + "\n\n # Context\n The layout information of the sketch image is: \n" + layout + return await self.llm.aask(msg=prompt, images=[encode_image(image_path)]) + + @staticmethod + def save_webpages(image_path: str, webpages: str) -> Path: + """Save webpages including all code (HTML, CSS, and JavaScript) at once. + + Args: + image_path (str): The path of the image file. + webpages (str): The generated webpages content. + + Returns: + Path: The path of the saved webpages. + """ + # Create a folder called webpages in the workspace directory to store HTML, CSS, and JavaScript files + webpages_path = DEFAULT_WORKSPACE_ROOT / "webpages" / Path(image_path).stem + os.makedirs(webpages_path, exist_ok=True) + + index_path = webpages_path / "index.html" + try: + index = webpages.split("```html")[1].split("```")[0] + style_path = None + if "styles.css" in index: + style_path = webpages_path / "styles.css" + elif "style.css" in index: + style_path = webpages_path / "style.css" + style = webpages.split("```css")[1].split("```")[0] if style_path else "" + + js_path = None + if "scripts.js" in index: + js_path = webpages_path / "scripts.js" + elif "script.js" in index: + js_path = webpages_path / "script.js" + + js = webpages.split("```javascript")[1].split("```")[0] if js_path else "" + except IndexError: + raise ValueError(f"No html or css or js code found in the result. \nWebpages: {webpages}") + + try: + with open(index_path, "w", encoding="utf-8") as f: + f.write(index) + if style_path: + with open(style_path, "w", encoding="utf-8") as f: + f.write(style) + if js_path: + with open(js_path, "w", encoding="utf-8") as f: + f.write(js) + except FileNotFoundError as e: + raise FileNotFoundError(f"Cannot save the webpages to {str(webpages_path)}") from e + + return webpages_path diff --git a/metagpt/tools/libs/sd_engine.py b/metagpt/tools/libs/sd_engine.py new file mode 100644 index 000000000..347f4a430 --- /dev/null +++ b/metagpt/tools/libs/sd_engine.py @@ -0,0 +1,183 @@ +# -*- coding: utf-8 -*- +# @Date : 2023/7/19 16:28 +# @Author : stellahong (stellahong@deepwisdom.ai) +# @Desc : +from __future__ import annotations + +import base64 +import hashlib +import io +import json +from os.path import join + +import requests +from aiohttp import ClientSession +from PIL import Image, PngImagePlugin + +# +from metagpt.const import SD_OUTPUT_FILE_REPO, SOURCE_ROOT +from metagpt.logs import logger +from metagpt.tools.tool_registry import register_tool +from metagpt.tools.tool_type import ToolType + +payload = { + "prompt": "", + "negative_prompt": "(easynegative:0.8),black, dark,Low resolution", + "override_settings": {"sd_model_checkpoint": "galaxytimemachinesGTM_photoV20"}, + "seed": -1, + "batch_size": 1, + "n_iter": 1, + "steps": 20, + "cfg_scale": 7, + "width": 512, + "height": 768, + "restore_faces": False, + "tiling": False, + "do_not_save_samples": False, + "do_not_save_grid": False, + "enable_hr": False, + "hr_scale": 2, + "hr_upscaler": "Latent", + "hr_second_pass_steps": 0, + "hr_resize_x": 0, + "hr_resize_y": 0, + "hr_upscale_to_x": 0, + "hr_upscale_to_y": 0, + "truncate_x": 0, + "truncate_y": 0, + "applied_old_hires_behavior_to": None, + "eta": None, + "sampler_index": "DPM++ SDE Karras", + "alwayson_scripts": {}, +} + +default_negative_prompt = "(easynegative:0.8),black, dark,Low resolution" + + +@register_tool( + tool_type=ToolType.STABLE_DIFFUSION.type_name, + include_functions=["__init__", "simple_run_t2i", "run_t2i", "construct_payload", "save"], +) +class SDEngine: + """Generate image using stable diffusion model. + + This class provides methods to interact with a stable diffusion service to generate images based on text inputs. + """ + + def __init__(self, sd_url=""): + """Initialize the SDEngine instance with configuration. + + Args: + sd_url (str, optional): URL of the stable diffusion service. Defaults to "". + """ + self.sd_url = sd_url + self.sd_t2i_url = f"{self.sd_url}/sdapi/v1/txt2img" + # Define default payload settings for SD API + self.payload = payload + logger.info(self.sd_t2i_url) + + def construct_payload( + self, + prompt, + negtive_prompt=default_negative_prompt, + width=512, + height=512, + sd_model="galaxytimemachinesGTM_photoV20", + ): + """Modify and set the API parameters for image generation. + + Args: + prompt (str): Text input for image generation. + negtive_prompt (str, optional): Text input for negative prompts. Defaults to None. + width (int, optional): Width of the generated image in pixels. Defaults to 512. + height (int, optional): Height of the generated image in pixels. Defaults to 512. + sd_model (str, optional): The model to use for image generation. Defaults to "galaxytimemachinesGTM_photoV20". + + Returns: + dict: Updated parameters for the stable diffusion API. + """ + self.payload["prompt"] = prompt + self.payload["negative_prompt"] = negtive_prompt + self.payload["width"] = width + self.payload["height"] = height + self.payload["override_settings"]["sd_model_checkpoint"] = sd_model + logger.info(f"call sd payload is {self.payload}") + return self.payload + + def save(self, imgs, save_name=""): + """Save generated images to the output directory. + + Args: + imgs (str): Generated images. + save_name (str, optional): Output image name. Default is empty. + """ + save_dir = SOURCE_ROOT / SD_OUTPUT_FILE_REPO + if not save_dir.exists(): + save_dir.mkdir(parents=True, exist_ok=True) + batch_decode_base64_to_image(imgs, str(save_dir), save_name=save_name) + + def simple_run_t2i(self, payload: dict, auto_save: bool = True): + """Run the stable diffusion API for multiple prompts, calling the stable diffusion API to generate images. + + Args: + payload (dict): Dictionary of input parameters for the stable diffusion API. + auto_save (bool, optional): Save generated images automatically. Defaults to True. + + Returns: + list: The generated images as a result of the API call. + """ + with requests.Session() as session: + logger.debug(self.sd_t2i_url) + rsp = session.post(self.sd_t2i_url, json=payload, timeout=600) + + results = rsp.json()["images"] + if auto_save: + save_name = hashlib.sha256(payload["prompt"][:10].encode()).hexdigest()[:6] + self.save(results, save_name=f"output_{save_name}") + return results + + async def run_t2i(self, payloads: list): + """Run the stable diffusion API for multiple prompts asynchronously. + + Args: + payloads (list): list of payload, each payload is a dictionary of input parameters for the stable diffusion API. + """ + session = ClientSession() + for payload_idx, payload in enumerate(payloads): + results = await self.run(url=self.sd_t2i_url, payload=payload, session=session) + self.save(results, save_name=f"output_{payload_idx}") + await session.close() + + async def run(self, url, payload, session): + """Perform the HTTP POST request to the SD API. + + Args: + url (str): The API URL. + payload (dict): The payload for the request. + session (ClientSession): The session for making HTTP requests. + + Returns: + list: Images generated by the stable diffusion API. + """ + async with session.post(url, json=payload, timeout=600) as rsp: + data = await rsp.read() + + rsp_json = json.loads(data) + imgs = rsp_json["images"] + + logger.info(f"callback rsp json is {rsp_json.keys()}") + return imgs + + +def decode_base64_to_image(img, save_name): + image = Image.open(io.BytesIO(base64.b64decode(img.split(",", 1)[0]))) + pnginfo = PngImagePlugin.PngInfo() + logger.info(save_name) + image.save(f"{save_name}.png", pnginfo=pnginfo) + return pnginfo, image + + +def batch_decode_base64_to_image(imgs, save_dir="", save_name=""): + for idx, _img in enumerate(imgs): + save_name = join(save_dir, save_name) + decode_base64_to_image(_img, save_name=save_name) diff --git a/metagpt/tools/libs/web_scraping.py b/metagpt/tools/libs/web_scraping.py new file mode 100644 index 000000000..d01e69d09 --- /dev/null +++ b/metagpt/tools/libs/web_scraping.py @@ -0,0 +1,21 @@ +from metagpt.tools.tool_registry import register_tool +from metagpt.tools.tool_type import ToolType +from metagpt.tools.web_browser_engine_playwright import PlaywrightWrapper + + +@register_tool(tool_type=ToolType.WEBSCRAPING.type_name) +async def scrape_web_playwright(url): + """ + Asynchronously Scrape and save the HTML structure and inner text content of a web page using Playwright. + + Args: + url (str): The main URL to fetch inner text from. + + Returns: + dict: The inner text content and html structure of the web page, keys are 'inner_text', 'html'. + """ + # Create a PlaywrightWrapper instance for the Chromium browser + web = await PlaywrightWrapper().run(url) + + # Return the inner text content of the web page + return {"inner_text": web.inner_text.strip(), "html": web.html.strip()} diff --git a/metagpt/tools/search_engine.py b/metagpt/tools/search_engine.py index 0d0db9147..1e540bd0e 100644 --- a/metagpt/tools/search_engine.py +++ b/metagpt/tools/search_engine.py @@ -8,14 +8,23 @@ import importlib from typing import Callable, Coroutine, Literal, Optional, Union, overload +from pydantic import BaseModel, ConfigDict, model_validator from semantic_kernel.skill_definition import sk_function +from metagpt.configs.search_config import SearchConfig +from metagpt.logs import logger from metagpt.tools import SearchEngineType class SkSearchEngine: - def __init__(self): - self.search_engine = SearchEngine() + """A search engine class for executing searches. + + Attributes: + search_engine: The search engine instance used for executing searches. + """ + + def __init__(self, **kwargs): + self.search_engine = SearchEngine(**kwargs) @sk_function( description="searches results from Google. Useful when you need to find short " @@ -28,43 +37,85 @@ class SkSearchEngine: return result -class SearchEngine: - """Class representing a search engine. - - Args: - engine: The search engine type. Defaults to the search engine specified in the config. - run_func: The function to run the search. Defaults to None. +class SearchEngine(BaseModel): + """A model for configuring and executing searches with different search engines. Attributes: - run_func: The function to run the search. - engine: The search engine type. + model_config: Configuration for the model allowing arbitrary types. + engine: The type of search engine to use. + run_func: An optional callable for running the search. If not provided, it will be determined based on the engine. + api_key: An optional API key for the search engine. + proxy: An optional proxy for the search engine requests. """ - def __init__( + model_config = ConfigDict(arbitrary_types_allowed=True, extra="allow") + + engine: SearchEngineType = SearchEngineType.SERPER_GOOGLE + run_func: Optional[Callable[[str, int, bool], Coroutine[None, None, Union[str, list[str]]]]] = None + api_key: Optional[str] = None + proxy: Optional[str] = None + + @model_validator(mode="after") + def validate_extra(self): + """Validates extra fields provided to the model and updates the run function accordingly.""" + data = self.model_dump(exclude={"engine"}, exclude_none=True, exclude_defaults=True) + if self.model_extra: + data.update(self.model_extra) + self._process_extra(**data) + return self + + def _process_extra( self, - engine: Optional[SearchEngineType] = SearchEngineType.SERPER_GOOGLE, - run_func: Callable[[str, int, bool], Coroutine[None, None, Union[str, list[str]]]] = None, + run_func: Optional[Callable[[str, int, bool], Coroutine[None, None, Union[str, list[str]]]]] = None, **kwargs, ): - if engine == SearchEngineType.SERPAPI_GOOGLE: + """Processes extra configuration and updates the run function based on the search engine type. + + Args: + run_func: An optional callable for running the search. If not provided, it will be determined based on the engine. + """ + if self.engine == SearchEngineType.SERPAPI_GOOGLE: module = "metagpt.tools.search_engine_serpapi" run_func = importlib.import_module(module).SerpAPIWrapper(**kwargs).run - elif engine == SearchEngineType.SERPER_GOOGLE: + elif self.engine == SearchEngineType.SERPER_GOOGLE: module = "metagpt.tools.search_engine_serper" run_func = importlib.import_module(module).SerperWrapper(**kwargs).run - elif engine == SearchEngineType.DIRECT_GOOGLE: + elif self.engine == SearchEngineType.DIRECT_GOOGLE: module = "metagpt.tools.search_engine_googleapi" run_func = importlib.import_module(module).GoogleAPIWrapper(**kwargs).run - elif engine == SearchEngineType.DUCK_DUCK_GO: + elif self.engine == SearchEngineType.DUCK_DUCK_GO: module = "metagpt.tools.search_engine_ddg" run_func = importlib.import_module(module).DDGAPIWrapper(**kwargs).run - elif engine == SearchEngineType.CUSTOM_ENGINE: - pass # run_func = run_func + elif self.engine == SearchEngineType.CUSTOM_ENGINE: + run_func = self.run_func else: raise NotImplementedError - self.engine = engine self.run_func = run_func + @classmethod + def from_search_config(cls, config: SearchConfig, **kwargs): + """Creates a SearchEngine instance from a SearchConfig. + + Args: + config: The search configuration to use for creating the SearchEngine instance. + """ + data = config.model_dump(exclude={"api_type", "search_func"}) + if config.search_func is not None: + data["run_func"] = config.search_func + + return cls(engine=config.api_type, **data, **kwargs) + + @classmethod + def from_search_func( + cls, search_func: Callable[[str, int, bool], Coroutine[None, None, Union[str, list[str]]]], **kwargs + ): + """Creates a SearchEngine instance from a custom search function. + + Args: + search_func: A callable that executes the search. + """ + return cls(engine=SearchEngineType.CUSTOM_ENGINE, run_func=search_func, **kwargs) + @overload def run( self, @@ -83,15 +134,29 @@ class SearchEngine: ) -> list[dict[str, str]]: ... - async def run(self, query: str, max_results: int = 8, as_string: bool = True) -> Union[str, list[dict[str, str]]]: + async def run( + self, + query: str, + max_results: int = 8, + as_string: bool = True, + ignore_errors: bool = False, + ) -> Union[str, list[dict[str, str]]]: """Run a search query. Args: query: The search query. max_results: The maximum number of results to return. Defaults to 8. as_string: Whether to return the results as a string or a list of dictionaries. Defaults to True. + ignore_errors: Whether to ignore errors during the search. Defaults to False. Returns: The search results as a string or a list of dictionaries. """ - return await self.run_func(query, max_results=max_results, as_string=as_string) + try: + return await self.run_func(query, max_results=max_results, as_string=as_string) + except Exception as e: + # Handle errors in the API call + logger.exception(f"fail to search {query} for {e}") + if not ignore_errors: + raise e + return "" if as_string else [] diff --git a/metagpt/tools/search_engine_ddg.py b/metagpt/tools/search_engine_ddg.py index 3d004a4ee..1412f20cf 100644 --- a/metagpt/tools/search_engine_ddg.py +++ b/metagpt/tools/search_engine_ddg.py @@ -5,9 +5,9 @@ from __future__ import annotations import asyncio import json from concurrent import futures -from typing import Literal, overload +from typing import Literal, Optional, overload -from metagpt.config2 import config +from pydantic import BaseModel, ConfigDict try: from duckduckgo_search import DDGS @@ -18,24 +18,16 @@ except ImportError: ) -class DDGAPIWrapper: - """Wrapper around duckduckgo_search API. +class DDGAPIWrapper(BaseModel): + model_config = ConfigDict(arbitrary_types_allowed=True) - To use this module, you should have the `duckduckgo_search` Python package installed. - """ + loop: Optional[asyncio.AbstractEventLoop] = None + executor: Optional[futures.Executor] = None + proxy: Optional[str] = None - def __init__( - self, - *, - loop: asyncio.AbstractEventLoop | None = None, - executor: futures.Executor | None = None, - ): - kwargs = {} - if config.proxy: - kwargs["proxies"] = config.proxy - self.loop = loop - self.executor = executor - self.ddgs = DDGS(**kwargs) + @property + def ddgs(self): + return DDGS(proxies=self.proxy) @overload def run( diff --git a/metagpt/tools/search_engine_googleapi.py b/metagpt/tools/search_engine_googleapi.py index 0a8f796cb..66b5ba950 100644 --- a/metagpt/tools/search_engine_googleapi.py +++ b/metagpt/tools/search_engine_googleapi.py @@ -4,19 +4,16 @@ from __future__ import annotations import asyncio import json +import warnings from concurrent import futures from typing import Optional from urllib.parse import urlparse import httplib2 -from pydantic import BaseModel, ConfigDict, Field, field_validator - -from metagpt.config2 import config -from metagpt.logs import logger +from pydantic import BaseModel, ConfigDict, model_validator try: from googleapiclient.discovery import build - from googleapiclient.errors import HttpError except ImportError: raise ImportError( "To use this module, you should have the `google-api-python-client` Python package installed. " @@ -27,40 +24,41 @@ except ImportError: class GoogleAPIWrapper(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) - google_api_key: Optional[str] = Field(default=None, validate_default=True) - google_cse_id: Optional[str] = Field(default=None, validate_default=True) + api_key: str + cse_id: str loop: Optional[asyncio.AbstractEventLoop] = None executor: Optional[futures.Executor] = None + proxy: Optional[str] = None - @field_validator("google_api_key", mode="before") + @model_validator(mode="before") @classmethod - def check_google_api_key(cls, val: str): - val = val or config.search.api_key - if not val: + def validate_google(cls, values: dict) -> dict: + if "google_api_key" in values: + values.setdefault("api_key", values["google_api_key"]) + warnings.warn("`google_api_key` is deprecated, use `api_key` instead", DeprecationWarning, stacklevel=2) + + if "api_key" not in values: raise ValueError( - "To use, make sure you provide the google_api_key when constructing an object. Alternatively, " - "ensure that the environment variable GOOGLE_API_KEY is set with your API key. You can obtain " + "To use google search engine, make sure you provide the `api_key` when constructing an object. You can obtain " "an API key from https://console.cloud.google.com/apis/credentials." ) - return val - @field_validator("google_cse_id", mode="before") - @classmethod - def check_google_cse_id(cls, val: str): - val = val or config.search.cse_id - if not val: + if "google_cse_id" in values: + values.setdefault("cse_id", values["google_cse_id"]) + warnings.warn("`google_cse_id` is deprecated, use `cse_id` instead", DeprecationWarning, stacklevel=2) + + if "cse_id" not in values: raise ValueError( - "To use, make sure you provide the google_cse_id when constructing an object. Alternatively, " - "ensure that the environment variable GOOGLE_CSE_ID is set with your API key. You can obtain " - "an API key from https://programmablesearchengine.google.com/controlpanel/create." + "To use google search engine, make sure you provide the `cse_id` when constructing an object. You can obtain " + "the cse_id from https://programmablesearchengine.google.com/controlpanel/create." ) - return val + return values @property def google_api_client(self): - build_kwargs = {"developerKey": self.google_api_key} - if config.proxy: - parse_result = urlparse(config.proxy) + build_kwargs = {"developerKey": self.api_key} + if self.proxy: + parse_result = urlparse(self.proxy) proxy_type = parse_result.scheme if proxy_type == "https": proxy_type = "http" @@ -96,17 +94,11 @@ class GoogleAPIWrapper(BaseModel): """ loop = self.loop or asyncio.get_event_loop() future = loop.run_in_executor( - self.executor, self.google_api_client.list(q=query, num=max_results, cx=self.google_cse_id).execute + self.executor, self.google_api_client.list(q=query, num=max_results, cx=self.cse_id).execute ) - try: - result = await future - # Extract the search result items from the response - search_results = result.get("items", []) - - except HttpError as e: - # Handle errors in the API call - logger.exception(f"fail to search {query} for {e}") - search_results = [] + result = await future + # Extract the search result items from the response + search_results = result.get("items", []) focus = focus or ["snippet", "link", "title"] details = [{i: j for i, j in item_dict.items() if i in focus} for item_dict in search_results] diff --git a/metagpt/tools/search_engine_serpapi.py b/metagpt/tools/search_engine_serpapi.py index a8d5b49d0..5744b1b62 100644 --- a/metagpt/tools/search_engine_serpapi.py +++ b/metagpt/tools/search_engine_serpapi.py @@ -5,18 +5,17 @@ @Author : alexanderwu @File : search_engine_serpapi.py """ +import warnings from typing import Any, Dict, Optional, Tuple import aiohttp -from pydantic import BaseModel, ConfigDict, Field, field_validator - -from metagpt.config2 import config +from pydantic import BaseModel, ConfigDict, Field, model_validator class SerpAPIWrapper(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) - search_engine: Any = None #: :meta private: + api_key: str params: dict = Field( default_factory=lambda: { "engine": "google", @@ -25,21 +24,22 @@ class SerpAPIWrapper(BaseModel): "hl": "en", } ) - # should add `validate_default=True` to check with default value - serpapi_api_key: Optional[str] = Field(default=None, validate_default=True) aiosession: Optional[aiohttp.ClientSession] = None + proxy: Optional[str] = None - @field_validator("serpapi_api_key", mode="before") + @model_validator(mode="before") @classmethod - def check_serpapi_api_key(cls, val: str): - val = val or config.search.api_key - if not val: + def validate_serpapi(cls, values: dict) -> dict: + if "serpapi_api_key" in values: + values.setdefault("api_key", values["serpapi_api_key"]) + warnings.warn("`serpapi_api_key` is deprecated, use `api_key` instead", DeprecationWarning, stacklevel=2) + + if "api_key" not in values: raise ValueError( - "To use, make sure you provide the serpapi_api_key when constructing an object. Alternatively, " - "ensure that the environment variable SERPAPI_API_KEY is set with your API key. You can obtain " - "an API key from https://serpapi.com/." + "To use serpapi search engine, make sure you provide the `api_key` when constructing an object. You can obtain" + " an API key from https://serpapi.com/." ) - return val + return values async def run(self, query, max_results: int = 8, as_string: bool = True, **kwargs: Any) -> str: """Run query through SerpAPI and parse result async.""" @@ -60,10 +60,12 @@ class SerpAPIWrapper(BaseModel): url, params = construct_url_and_params() if not self.aiosession: async with aiohttp.ClientSession() as session: - async with session.get(url, params=params) as response: + async with session.get(url, params=params, proxy=self.proxy) as response: + response.raise_for_status() res = await response.json() else: - async with self.aiosession.get(url, params=params) as response: + async with self.aiosession.get(url, params=params, proxy=self.proxy) as response: + response.raise_for_status() res = await response.json() return res @@ -71,7 +73,7 @@ class SerpAPIWrapper(BaseModel): def get_params(self, query: str) -> Dict[str, str]: """Get parameters for SerpAPI.""" _params = { - "api_key": self.serpapi_api_key, + "api_key": self.api_key, "q": query, } params = {**self.params, **_params} diff --git a/metagpt/tools/search_engine_serper.py b/metagpt/tools/search_engine_serper.py index 39cb936b8..ba2fb4f93 100644 --- a/metagpt/tools/search_engine_serper.py +++ b/metagpt/tools/search_engine_serper.py @@ -6,33 +6,34 @@ @File : search_engine_serpapi.py """ import json +import warnings from typing import Any, Dict, Optional, Tuple import aiohttp -from pydantic import BaseModel, ConfigDict, Field, field_validator - -from metagpt.config2 import config +from pydantic import BaseModel, ConfigDict, Field, model_validator class SerperWrapper(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) - search_engine: Any = None #: :meta private: + api_key: str payload: dict = Field(default_factory=lambda: {"page": 1, "num": 10}) - serper_api_key: Optional[str] = Field(default=None, validate_default=True) aiosession: Optional[aiohttp.ClientSession] = None + proxy: Optional[str] = None - @field_validator("serper_api_key", mode="before") + @model_validator(mode="before") @classmethod - def check_serper_api_key(cls, val: str): - val = val or config.search.api_key - if not val: + def validate_serper(cls, values: dict) -> dict: + if "serper_api_key" in values: + values.setdefault("api_key", values["serper_api_key"]) + warnings.warn("`serper_api_key` is deprecated, use `api_key` instead", DeprecationWarning, stacklevel=2) + + if "api_key" not in values: raise ValueError( - "To use, make sure you provide the serper_api_key when constructing an object. Alternatively, " - "ensure that the environment variable SERPER_API_KEY is set with your API key. You can obtain " + "To use serper search engine, make sure you provide the `api_key` when constructing an object. You can obtain " "an API key from https://serper.dev/." ) - return val + return values async def run(self, query: str, max_results: int = 8, as_string: bool = True, **kwargs: Any) -> str: """Run query through Serper and parse result async.""" @@ -54,10 +55,12 @@ class SerperWrapper(BaseModel): url, payloads, headers = construct_url_and_payload_and_headers() if not self.aiosession: async with aiohttp.ClientSession() as session: - async with session.post(url, data=payloads, headers=headers) as response: + async with session.post(url, data=payloads, headers=headers, proxy=self.proxy) as response: + response.raise_for_status() res = await response.json() else: - async with self.aiosession.get.post(url, data=payloads, headers=headers) as response: + async with self.aiosession.get.post(url, data=payloads, headers=headers, proxy=self.proxy) as response: + response.raise_for_status() res = await response.json() return res @@ -74,7 +77,7 @@ class SerperWrapper(BaseModel): return json.dumps(payloads, sort_keys=True) def get_headers(self) -> Dict[str, str]: - headers = {"X-API-KEY": self.serper_api_key, "Content-Type": "application/json"} + headers = {"X-API-KEY": self.api_key, "Content-Type": "application/json"} return headers @staticmethod diff --git a/metagpt/tools/tool_convert.py b/metagpt/tools/tool_convert.py new file mode 100644 index 000000000..fc29d0693 --- /dev/null +++ b/metagpt/tools/tool_convert.py @@ -0,0 +1,83 @@ +import inspect + +from metagpt.utils.parse_docstring import GoogleDocstringParser, remove_spaces + + +def convert_code_to_tool_schema(obj, include: list[str] = []): + docstring = inspect.getdoc(obj) + assert docstring, "no docstring found for the objects, skip registering" + + if inspect.isclass(obj): + schema = {"type": "class", "description": remove_spaces(docstring), "methods": {}} + for name, method in inspect.getmembers(obj, inspect.isfunction): + if include and name not in include: + continue + # method_doc = inspect.getdoc(method) + method_doc = get_class_method_docstring(obj, name) + if method_doc: + schema["methods"][name] = function_docstring_to_schema(method, method_doc) + + elif inspect.isfunction(obj): + schema = function_docstring_to_schema(obj, docstring) + + return schema + + +def function_docstring_to_schema(fn_obj, docstring): + function_type = "function" if not inspect.iscoroutinefunction(fn_obj) else "async_function" + return {"type": function_type, **docstring_to_schema(docstring)} + + +def docstring_to_schema(docstring: str): + if docstring is None: + return {} + + parser = GoogleDocstringParser(docstring=docstring) + + # 匹配简介部分 + description = parser.parse_desc() + + # 匹配Args部分 + params = parser.parse_params() + parameter_schema = {"properties": {}, "required": []} + for param in params: + param_name, param_type, param_desc = param + # check required or optional + is_optional, param_type = parser.check_and_parse_optional(param_type) + if not is_optional: + parameter_schema["required"].append(param_name) + # type and desc + param_dict = {"type": param_type, "description": remove_spaces(param_desc)} + # match Default for optional args + has_default_val, default_val = parser.check_and_parse_default_value(param_desc) + if has_default_val: + param_dict["default"] = default_val + # match Enum + has_enum, enum_vals = parser.check_and_parse_enum(param_desc) + if has_enum: + param_dict["enum"] = enum_vals + # add to parameter schema + parameter_schema["properties"].update({param_name: param_dict}) + + # 匹配Returns部分 + returns = parser.parse_returns() + + # 构建YAML字典 + schema = { + "description": description, + "parameters": parameter_schema, + } + if returns: + schema["returns"] = [{"type": ret[0], "description": remove_spaces(ret[1])} for ret in returns] + + return schema + + +def get_class_method_docstring(cls, method_name): + """Retrieve a method's docstring, searching the class hierarchy if necessary.""" + for base_class in cls.__mro__: + if method_name in base_class.__dict__: + method = base_class.__dict__[method_name] + if method.__doc__: + return method.__doc__ + return None # No docstring found in the class hierarchy diff --git a/metagpt/tools/tool_data_type.py b/metagpt/tools/tool_data_type.py new file mode 100644 index 000000000..0ae46fa5c --- /dev/null +++ b/metagpt/tools/tool_data_type.py @@ -0,0 +1,18 @@ +from pydantic import BaseModel + + +class ToolTypeDef(BaseModel): + name: str + desc: str = "" + usage_prompt: str = "" + + +class ToolSchema(BaseModel): + description: str + + +class Tool(BaseModel): + name: str + path: str + schemas: dict = {} + code: str = "" diff --git a/metagpt/tools/tool_registry.py b/metagpt/tools/tool_registry.py new file mode 100644 index 000000000..5fbd39421 --- /dev/null +++ b/metagpt/tools/tool_registry.py @@ -0,0 +1,155 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2023/01/12 17:07 +@Author : garylin2099 +@File : tool_registry.py +""" +from __future__ import annotations + +import inspect +import os +import re +from collections import defaultdict + +import yaml +from pydantic import BaseModel, field_validator + +from metagpt.const import TOOL_SCHEMA_PATH +from metagpt.logs import logger +from metagpt.tools.tool_convert import convert_code_to_tool_schema +from metagpt.tools.tool_data_type import Tool, ToolSchema, ToolTypeDef +from metagpt.tools.tool_type import ToolType + + +class ToolRegistry(BaseModel): + tools: dict = {} + tool_types: dict = {} + tools_by_types: dict = defaultdict(dict) # two-layer k-v, {tool_type: {tool_name: {...}, ...}, ...} + + @field_validator("tool_types", mode="before") + @classmethod + def init_tool_types(cls, tool_types: ToolType): + return {tool_type.type_name: tool_type.value for tool_type in tool_types} + + def register_tool( + self, + tool_name, + tool_path, + schema_path="", + tool_code="", + tool_type="other", + tool_source_object=None, + include_functions=[], + verbose=False, + ): + if self.has_tool(tool_name): + return + + if tool_type not in self.tool_types: + # register new tool type on the fly + logger.warning( + f"{tool_type} not previously defined, will create a temporary tool type with just a name. This tool type is only effective during this runtime. You may consider add this tool type with more configs permanently at metagpt.tools.tool_type" + ) + temp_tool_type_obj = ToolTypeDef(name=tool_type) + self.tool_types[tool_type] = temp_tool_type_obj + if verbose: + logger.info(f"tool type {tool_type} registered") + + schema_path = schema_path or TOOL_SCHEMA_PATH / tool_type / f"{tool_name}.yml" + + schemas = make_schema(tool_source_object, include_functions, schema_path) + + if not schemas: + return + + schemas["tool_path"] = tool_path # corresponding code file path of the tool + try: + ToolSchema(**schemas) # validation + except Exception: + pass + # logger.warning( + # f"{tool_name} schema not conforms to required format, but will be used anyway. Mismatch: {e}" + # ) + + tool = Tool(name=tool_name, path=tool_path, schemas=schemas, code=tool_code) + self.tools[tool_name] = tool + self.tools_by_types[tool_type][tool_name] = tool + if verbose: + logger.info(f"{tool_name} registered") + logger.info(f"schema made at {str(schema_path)}, can be used for checking") + + def has_tool(self, key: str) -> Tool: + return key in self.tools + + def get_tool(self, key) -> Tool: + return self.tools.get(key) + + def get_tools_by_type(self, key) -> dict[str, Tool]: + return self.tools_by_types.get(key, {}) + + def has_tool_type(self, key) -> bool: + return key in self.tool_types + + def get_tool_type(self, key) -> ToolType: + return self.tool_types.get(key) + + def get_tool_types(self) -> dict[str, ToolType]: + return self.tool_types + + +# Registry instance +TOOL_REGISTRY = ToolRegistry(tool_types=ToolType) + + +def register_tool(tool_type: str = "other", schema_path: str = "", **kwargs): + """register a tool to registry""" + + def decorator(cls): + # Get the file path where the function / class is defined and the source code + file_path = inspect.getfile(cls) + if "metagpt" in file_path: + file_path = re.search("metagpt.+", file_path).group(0) + source_code = inspect.getsource(cls) + + TOOL_REGISTRY.register_tool( + tool_name=cls.__name__, + tool_path=file_path, + schema_path=schema_path, + tool_code=source_code, + tool_type=tool_type, + tool_source_object=cls, + **kwargs, + ) + return cls + + return decorator + + +def make_schema(tool_source_object, include, path): + os.makedirs(os.path.dirname(path), exist_ok=True) # Create the necessary directories + try: + schema = convert_code_to_tool_schema(tool_source_object, include=include) + with open(path, "w", encoding="utf-8") as f: + yaml.dump(schema, f, sort_keys=False) + # import json + # with open(str(path).replace("yml", "json"), "w", encoding="utf-8") as f: + # json.dump(schema, f, ensure_ascii=False, indent=4) + except Exception as e: + schema = {} + logger.error(f"Fail to make schema: {e}") + + return schema + + +def validate_tool_names(tools: list[str], return_tool_object=False) -> list[str]: + valid_tools = [] + for tool_name in tools: + if not TOOL_REGISTRY.has_tool(tool_name): + logger.warning( + f"Specified tool {tool_name} not found and was skipped. Check if you have registered it properly" + ) + else: + valid_tool = TOOL_REGISTRY.get_tool(tool_name) if return_tool_object else tool_name + valid_tools.append(valid_tool) + return valid_tools diff --git a/metagpt/tools/tool_type.py b/metagpt/tools/tool_type.py new file mode 100644 index 000000000..6fa971c56 --- /dev/null +++ b/metagpt/tools/tool_type.py @@ -0,0 +1,55 @@ +from enum import Enum + +from metagpt.prompts.tool_types import ( + DATA_PREPROCESS_PROMPT, + FEATURE_ENGINEERING_PROMPT, + IMAGE2WEBPAGE_PROMPT, + MODEL_EVALUATE_PROMPT, + MODEL_TRAIN_PROMPT, +) +from metagpt.tools.tool_data_type import ToolTypeDef + + +class ToolType(Enum): + EDA = ToolTypeDef(name="eda", desc="For performing exploratory data analysis") + DATA_PREPROCESS = ToolTypeDef( + name="data_preprocess", + desc="Only for changing value inplace.", + usage_prompt=DATA_PREPROCESS_PROMPT, + ) + FEATURE_ENGINEERING = ToolTypeDef( + name="feature_engineering", + desc="Only for creating new columns for input data.", + usage_prompt=FEATURE_ENGINEERING_PROMPT, + ) + MODEL_TRAIN = ToolTypeDef( + name="model_train", + desc="Only for training model.", + usage_prompt=MODEL_TRAIN_PROMPT, + ) + MODEL_EVALUATE = ToolTypeDef( + name="model_evaluate", + desc="Only for evaluating model.", + usage_prompt=MODEL_EVALUATE_PROMPT, + ) + STABLE_DIFFUSION = ToolTypeDef( + name="stable_diffusion", + desc="Related to text2image, image2image using stable diffusion model.", + ) + IMAGE2WEBPAGE = ToolTypeDef( + name="image2webpage", + desc="For converting image into webpage code.", + usage_prompt=IMAGE2WEBPAGE_PROMPT, + ) + WEBSCRAPING = ToolTypeDef( + name="web_scraping", + desc="For scraping data from web pages.", + ) + OTHER = ToolTypeDef(name="other", desc="Any tools not in the defined categories") + + def __missing__(self, key): + return self.OTHER + + @property + def type_name(self): + return self.value.name diff --git a/metagpt/tools/web_browser_engine.py b/metagpt/tools/web_browser_engine.py index 411c1604b..01339e51a 100644 --- a/metagpt/tools/web_browser_engine.py +++ b/metagpt/tools/web_browser_engine.py @@ -1,36 +1,95 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- - from __future__ import annotations import importlib -from typing import Any, Callable, Coroutine, overload +from typing import Any, Callable, Coroutine, Optional, Union, overload +from pydantic import BaseModel, ConfigDict, model_validator + +from metagpt.configs.browser_config import BrowserConfig from metagpt.tools import WebBrowserEngineType from metagpt.utils.parse_html import WebPage -class WebBrowserEngine: - def __init__( - self, - engine: WebBrowserEngineType = WebBrowserEngineType.PLAYWRIGHT, - run_func: Callable[..., Coroutine[Any, Any, WebPage | list[WebPage]]] | None = None, - ): - if engine is None: - raise NotImplementedError +class WebBrowserEngine(BaseModel): + """Defines a web browser engine configuration for automated browsing and data extraction. - if WebBrowserEngineType(engine) is WebBrowserEngineType.PLAYWRIGHT: + This class encapsulates the configuration and operational logic for different web browser engines, + such as Playwright, Selenium, or custom implementations. It provides a unified interface to run + browser automation tasks. + + Attributes: + model_config: Configuration dictionary allowing arbitrary types and extra fields. + engine: The type of web browser engine to use. + run_func: An optional coroutine function to run the browser engine. + proxy: An optional proxy server URL to use with the browser engine. + """ + + model_config = ConfigDict(arbitrary_types_allowed=True, extra="allow") + + engine: WebBrowserEngineType = WebBrowserEngineType.PLAYWRIGHT + run_func: Optional[Callable[..., Coroutine[Any, Any, Union[WebPage, list[WebPage]]]]] = None + proxy: Optional[str] = None + + @model_validator(mode="after") + def validate_extra(self): + """Validates and processes extra configuration data after model initialization. + + This method is automatically called by Pydantic to validate and process any extra configuration + data provided to the model. It ensures that the extra data is properly integrated into the model's + configuration and operational logic. + + Returns: + The instance itself after processing the extra data. + """ + data = self.model_dump(exclude={"engine"}, exclude_none=True, exclude_defaults=True) + if self.model_extra: + data.update(self.model_extra) + self._process_extra(**data) + return self + + def _process_extra(self, **kwargs): + """Processes extra configuration data to set up the browser engine run function. + + Depending on the specified engine type, this method dynamically imports and configures + the appropriate browser engine wrapper and its run function. + + Args: + **kwargs: Arbitrary keyword arguments representing extra configuration data. + + Raises: + NotImplementedError: If the engine type is not supported. + """ + if self.engine is WebBrowserEngineType.PLAYWRIGHT: module = "metagpt.tools.web_browser_engine_playwright" - run_func = importlib.import_module(module).PlaywrightWrapper().run - elif WebBrowserEngineType(engine) is WebBrowserEngineType.SELENIUM: + run_func = importlib.import_module(module).PlaywrightWrapper(**kwargs).run + elif self.engine is WebBrowserEngineType.SELENIUM: module = "metagpt.tools.web_browser_engine_selenium" - run_func = importlib.import_module(module).SeleniumWrapper().run - elif WebBrowserEngineType(engine) is WebBrowserEngineType.CUSTOM: - run_func = run_func + run_func = importlib.import_module(module).SeleniumWrapper(**kwargs).run + elif self.engine is WebBrowserEngineType.CUSTOM: + run_func = self.run_func else: raise NotImplementedError self.run_func = run_func - self.engine = engine + + @classmethod + def from_browser_config(cls, config: BrowserConfig, **kwargs): + """Creates a WebBrowserEngine instance from a BrowserConfig object and additional keyword arguments. + + This class method facilitates the creation of a WebBrowserEngine instance by extracting + configuration data from a BrowserConfig object and optionally merging it with additional + keyword arguments. + + Args: + config: A BrowserConfig object containing base configuration data. + **kwargs: Optional additional keyword arguments to override or extend the configuration. + + Returns: + A new instance of WebBrowserEngine configured according to the provided arguments. + """ + data = config.model_dump() + return cls(**data, **kwargs) @overload async def run(self, url: str) -> WebPage: @@ -41,4 +100,16 @@ class WebBrowserEngine: ... async def run(self, url: str, *urls: str) -> WebPage | list[WebPage]: + """Runs the browser engine to load one or more web pages. + + This method is the implementation of the overloaded run signatures. It delegates the task + of loading web pages to the configured run function, handling either a single URL or multiple URLs. + + Args: + url: The URL of the first web page to load. + *urls: Additional URLs of web pages to load, if any. + + Returns: + A WebPage object if a single URL is provided, or a list of WebPage objects if multiple URLs are provided. + """ return await self.run_func(url, *urls) diff --git a/metagpt/tools/web_browser_engine_playwright.py b/metagpt/tools/web_browser_engine_playwright.py index f8dabd5ac..2df288b1a 100644 --- a/metagpt/tools/web_browser_engine_playwright.py +++ b/metagpt/tools/web_browser_engine_playwright.py @@ -6,16 +6,16 @@ from __future__ import annotations import asyncio import sys from pathlib import Path -from typing import Literal +from typing import Literal, Optional from playwright.async_api import async_playwright +from pydantic import BaseModel, Field, PrivateAttr -from metagpt.config2 import config from metagpt.logs import logger from metagpt.utils.parse_html import WebPage -class PlaywrightWrapper: +class PlaywrightWrapper(BaseModel): """Wrapper around Playwright. To use this module, you should have the `playwright` Python package installed and ensure that @@ -24,24 +24,23 @@ class PlaywrightWrapper: command `playwright install` for the first time. """ - def __init__( - self, - browser_type: Literal["chromium", "firefox", "webkit"] | None = "chromium", - launch_kwargs: dict | None = None, - **kwargs, - ) -> None: - self.browser_type = browser_type - launch_kwargs = launch_kwargs or {} - if config.proxy and "proxy" not in launch_kwargs: + browser_type: Literal["chromium", "firefox", "webkit"] = "chromium" + launch_kwargs: dict = Field(default_factory=dict) + proxy: Optional[str] = None + context_kwargs: dict = Field(default_factory=dict) + _has_run_precheck: bool = PrivateAttr(False) + + def __init__(self, **kwargs): + super().__init__(**kwargs) + + launch_kwargs = self.launch_kwargs + if self.proxy and "proxy" not in launch_kwargs: args = launch_kwargs.get("args", []) if not any(str.startswith(i, "--proxy-server=") for i in args): - launch_kwargs["proxy"] = {"server": config.proxy} - self.launch_kwargs = launch_kwargs - context_kwargs = {} + launch_kwargs["proxy"] = {"server": self.proxy} + if "ignore_https_errors" in kwargs: - context_kwargs["ignore_https_errors"] = kwargs["ignore_https_errors"] - self._context_kwargs = context_kwargs - self._has_run_precheck = False + self.context_kwargs["ignore_https_errors"] = kwargs["ignore_https_errors"] async def run(self, url: str, *urls: str) -> WebPage | list[WebPage]: async with async_playwright() as ap: @@ -55,7 +54,7 @@ class PlaywrightWrapper: return await _scrape(browser, url) async def _scrape(self, browser, url): - context = await browser.new_context(**self._context_kwargs) + context = await browser.new_context(**self.context_kwargs) page = await context.new_page() async with page: try: @@ -75,8 +74,8 @@ class PlaywrightWrapper: executable_path = Path(browser_type.executable_path) if not executable_path.exists() and "executable_path" not in self.launch_kwargs: kwargs = {} - if config.proxy: - kwargs["env"] = {"ALL_PROXY": config.proxy} + if self.proxy: + kwargs["env"] = {"ALL_PROXY": self.proxy} await _install_browsers(self.browser_type, **kwargs) if self._has_run_precheck: diff --git a/metagpt/tools/web_browser_engine_selenium.py b/metagpt/tools/web_browser_engine_selenium.py index 02dd5c173..3b1682291 100644 --- a/metagpt/tools/web_browser_engine_selenium.py +++ b/metagpt/tools/web_browser_engine_selenium.py @@ -7,19 +7,19 @@ import asyncio import importlib from concurrent import futures from copy import deepcopy -from typing import Literal +from typing import Callable, Literal, Optional +from pydantic import BaseModel, ConfigDict, Field, PrivateAttr from selenium.webdriver.common.by import By from selenium.webdriver.support import expected_conditions as EC from selenium.webdriver.support.wait import WebDriverWait from webdriver_manager.core.download_manager import WDMDownloadManager from webdriver_manager.core.http import WDMHttpClient -from metagpt.config2 import config from metagpt.utils.parse_html import WebPage -class SeleniumWrapper: +class SeleniumWrapper(BaseModel): """Wrapper around Selenium. To use this module, you should check the following: @@ -31,25 +31,28 @@ class SeleniumWrapper: can scrape web pages using the Selenium WebBrowserEngine. """ - def __init__( - self, - browser_type: Literal["chrome", "firefox", "edge", "ie"] = "chrome", - launch_kwargs: dict | None = None, - *, - loop: asyncio.AbstractEventLoop | None = None, - executor: futures.Executor | None = None, - ) -> None: - self.browser_type = browser_type - launch_kwargs = launch_kwargs or {} - if config.proxy and "proxy-server" not in launch_kwargs: - launch_kwargs["proxy-server"] = config.proxy + model_config = ConfigDict(arbitrary_types_allowed=True) - self.executable_path = launch_kwargs.pop("executable_path", None) - self.launch_args = [f"--{k}={v}" for k, v in launch_kwargs.items()] - self._has_run_precheck = False - self._get_driver = None - self.loop = loop - self.executor = executor + browser_type: Literal["chrome", "firefox", "edge", "ie"] = "chrome" + launch_kwargs: dict = Field(default_factory=dict) + proxy: Optional[str] = None + loop: Optional[asyncio.AbstractEventLoop] = None + executor: Optional[futures.Executor] = None + _has_run_precheck: bool = PrivateAttr(False) + _get_driver: Optional[Callable] = PrivateAttr(None) + + def __init__(self, **kwargs) -> None: + super().__init__(**kwargs) + if self.proxy and "proxy-server" not in self.launch_kwargs: + self.launch_kwargs["proxy-server"] = self.proxy + + @property + def launch_args(self): + return [f"--{k}={v}" for k, v in self.launch_kwargs.items() if k != "executable_path"] + + @property + def executable_path(self): + return self.launch_kwargs.get("executable_path") async def run(self, url: str, *urls: str) -> WebPage | list[WebPage]: await self._run_precheck() @@ -66,7 +69,9 @@ class SeleniumWrapper: self.loop = self.loop or asyncio.get_event_loop() self._get_driver = await self.loop.run_in_executor( self.executor, - lambda: _gen_get_driver_func(self.browser_type, *self.launch_args, executable_path=self.executable_path), + lambda: _gen_get_driver_func( + self.browser_type, *self.launch_args, executable_path=self.executable_path, proxy=self.proxy + ), ) self._has_run_precheck = True @@ -92,13 +97,17 @@ _webdriver_manager_types = { class WDMHttpProxyClient(WDMHttpClient): + def __init__(self, proxy: str = None): + super().__init__() + self.proxy = proxy + def get(self, url, **kwargs): - if "proxies" not in kwargs and config.proxy: - kwargs["proxies"] = {"all_proxy": config.proxy} + if "proxies" not in kwargs and self.proxy: + kwargs["proxies"] = {"all_proxy": self.proxy} return super().get(url, **kwargs) -def _gen_get_driver_func(browser_type, *args, executable_path=None): +def _gen_get_driver_func(browser_type, *args, executable_path=None, proxy=None): WebDriver = getattr(importlib.import_module(f"selenium.webdriver.{browser_type}.webdriver"), "WebDriver") Service = getattr(importlib.import_module(f"selenium.webdriver.{browser_type}.service"), "Service") Options = getattr(importlib.import_module(f"selenium.webdriver.{browser_type}.options"), "Options") @@ -106,7 +115,7 @@ def _gen_get_driver_func(browser_type, *args, executable_path=None): if not executable_path: module_name, type_name = _webdriver_manager_types[browser_type] DriverManager = getattr(importlib.import_module(module_name), type_name) - driver_manager = DriverManager(download_manager=WDMDownloadManager(http_client=WDMHttpProxyClient())) + driver_manager = DriverManager(download_manager=WDMDownloadManager(http_client=WDMHttpProxyClient(proxy=proxy))) # driver_manager.driver_cache.find_driver(driver_manager.driver)) executable_path = driver_manager.install() diff --git a/metagpt/utils/common.py b/metagpt/utils/common.py index e09d49d84..015902c3d 100644 --- a/metagpt/utils/common.py +++ b/metagpt/utils/common.py @@ -12,7 +12,9 @@ from __future__ import annotations import ast +import base64 import contextlib +import csv import importlib import inspect import json @@ -22,13 +24,16 @@ import re import sys import traceback import typing +from io import BytesIO from pathlib import Path -from typing import Any, List, Tuple, Union +from typing import Any, Callable, List, Tuple, Union import aiofiles import loguru +import requests +from PIL import Image from pydantic_core import to_jsonable_python -from tenacity import RetryCallState, _utils +from tenacity import RetryCallState, RetryError, _utils from metagpt.const import MESSAGE_ROUTE_TO_ALL from metagpt.logs import logger @@ -335,6 +340,14 @@ def print_members(module, indent=0): print(f"{prefix}Method: {name}") +def get_function_schema(func: Callable) -> dict[str, Union[dict, Any, str]]: + sig = inspect.signature(func) + parameters = sig.parameters + return_type = sig.return_annotation + param_schema = {name: parameter.annotation for name, parameter in parameters.items()} + return {"input_params": param_schema, "return_type": return_type, "func_desc": func.__doc__, "func": func} + + def parse_recipient(text): # FIXME: use ActionNode instead. pattern = r"## Send To:\s*([A-Za-z]+)\s*?" # hard code for now @@ -348,6 +361,31 @@ def parse_recipient(text): return "" +def create_func_call_config(func_schema: dict) -> dict: + """Create new function call config""" + tools = [{"type": "function", "function": func_schema}] + tool_choice = {"type": "function", "function": {"name": func_schema["name"]}} + return { + "tools": tools, + "tool_choice": tool_choice, + } + + +def remove_comments(code_str: str) -> str: + """Remove comments from code.""" + pattern = r"(\".*?\"|\'.*?\')|(\#.*?$)" + + def replace_func(match): + if match.group(2) is not None: + return "" + else: + return match.group(1) + + clean_code = re.sub(pattern, replace_func, code_str, flags=re.MULTILINE) + clean_code = os.linesep.join([s.rstrip() for s in clean_code.splitlines() if s.strip()]) + return clean_code + + def get_class_name(cls) -> str: """Return class name""" return f"{cls.__module__}.{cls.__name__}" @@ -456,13 +494,36 @@ def read_json_file(json_file: str, encoding="utf-8") -> list[Any]: return data -def write_json_file(json_file: str, data: list, encoding=None): +def write_json_file(json_file: str, data: list, encoding: str = None, indent: int = 4): folder_path = Path(json_file).parent if not folder_path.exists(): folder_path.mkdir(parents=True, exist_ok=True) with open(json_file, "w", encoding=encoding) as fout: - json.dump(data, fout, ensure_ascii=False, indent=4, default=to_jsonable_python) + json.dump(data, fout, ensure_ascii=False, indent=indent, default=to_jsonable_python) + + +def read_csv_to_list(curr_file: str, header=False, strip_trail=True): + """ + Reads in a csv file to a list of list. If header is True, it returns a + tuple with (header row, all rows) + ARGS: + curr_file: path to the current csv file. + RETURNS: + List of list where the component lists are the rows of the file. + """ + logger.debug(f"start read csv: {curr_file}") + analysis_list = [] + with open(curr_file) as f_analysis_file: + data_reader = csv.reader(f_analysis_file, delimiter=",") + for count, row in enumerate(data_reader): + if strip_trail: + row = [i.strip() for i in row] + analysis_list += [row] + if not header: + return analysis_list + else: + return analysis_list[0], analysis_list[1:] def import_class(class_name: str, module_name: str) -> type: @@ -505,7 +566,7 @@ def role_raise_decorator(func): self.rc.memory.delete(self.latest_observed_msg) # raise again to make it captured outside raise Exception(format_trackback_info(limit=None)) - except Exception: + except Exception as e: if self.latest_observed_msg: logger.warning( "There is a exception in role's execution, in order to resume, " @@ -514,6 +575,12 @@ def role_raise_decorator(func): # remove role newest observed msg to make it observed again self.rc.memory.delete(self.latest_observed_msg) # raise again to make it captured outside + if isinstance(e, RetryError): + last_error = e.last_attempt._exception + name = any_to_str(last_error) + if re.match(r"^openai\.", name) or re.match(r"^httpx\.", name): + raise last_error + raise Exception(format_trackback_info(limit=None)) return wrapper @@ -567,3 +634,45 @@ def list_files(root: str | Path) -> List[Path]: except Exception as e: logger.error(f"Error: {e}") return files + + +def is_coroutine_func(func: Callable) -> bool: + return inspect.iscoroutinefunction(func) + + +def load_mc_skills_code(skill_names: list[str] = None, skills_dir: Path = None) -> list[str]: + """load mincraft skill from js files""" + if not skills_dir: + skills_dir = Path(__file__).parent.absolute() + if skill_names is None: + skill_names = [skill[:-3] for skill in os.listdir(f"{skills_dir}") if skill.endswith(".js")] + skills = [skills_dir.joinpath(f"{skill_name}.js").read_text() for skill_name in skill_names] + return skills + + +def encode_image(image_path_or_pil: Union[Path, Image], encoding: str = "utf-8") -> str: + """encode image from file or PIL.Image into base64""" + if isinstance(image_path_or_pil, Image.Image): + buffer = BytesIO() + image_path_or_pil.save(buffer, format="JPEG") + bytes_data = buffer.getvalue() + else: + if not image_path_or_pil.exists(): + raise FileNotFoundError(f"{image_path_or_pil} not exists") + with open(str(image_path_or_pil), "rb") as image_file: + bytes_data = image_file.read() + return base64.b64encode(bytes_data).decode(encoding) + + +def decode_image(img_url_or_b64: str) -> Image: + """decode image from url or base64 into PIL.Image""" + if img_url_or_b64.startswith("http"): + # image http(s) url + resp = requests.get(img_url_or_b64) + img = Image.open(BytesIO(resp.content)) + else: + # image b64_json + b64_data = re.sub("^data:image/.+;base64,", "", img_url_or_b64) + img_data = BytesIO(base64.b64decode(b64_data)) + img = Image.open(img_data) + return img diff --git a/metagpt/utils/cost_manager.py b/metagpt/utils/cost_manager.py index 7bf5154b6..c4c93f91f 100644 --- a/metagpt/utils/cost_manager.py +++ b/metagpt/utils/cost_manager.py @@ -41,6 +41,10 @@ class CostManager(BaseModel): """ self.total_prompt_tokens += prompt_tokens self.total_completion_tokens += completion_tokens + if model not in TOKEN_COSTS: + logger.warning(f"Model {model} not found in TOKEN_COSTS.") + return + cost = ( prompt_tokens * TOKEN_COSTS[model]["prompt"] + completion_tokens * TOKEN_COSTS[model]["completion"] ) / 1000 diff --git a/metagpt/utils/dependency_file.py b/metagpt/utils/dependency_file.py index 7cf9a1d49..d3add1171 100644 --- a/metagpt/utils/dependency_file.py +++ b/metagpt/utils/dependency_file.py @@ -9,6 +9,7 @@ from __future__ import annotations import json +import re from pathlib import Path from typing import Set @@ -36,7 +37,9 @@ class DependencyFile: """Load dependencies from the file asynchronously.""" if not self._filename.exists(): return - self._dependencies = json.loads(await aread(self._filename)) + json_data = await aread(self._filename) + json_data = re.sub(r"\\+", "/", json_data) # Compatible with windows path + self._dependencies = json.loads(json_data) @handle_exception async def save(self): @@ -57,20 +60,22 @@ class DependencyFile: root = self._filename.parent try: - key = Path(filename).relative_to(root) + key = Path(filename).relative_to(root).as_posix() except ValueError: key = filename - + key = str(key) if dependencies: relative_paths = [] for i in dependencies: try: - relative_paths.append(str(Path(i).relative_to(root))) + s = str(Path(i).relative_to(root).as_posix()) except ValueError: - relative_paths.append(str(i)) - self._dependencies[str(key)] = relative_paths - elif str(key) in self._dependencies: - del self._dependencies[str(key)] + s = str(i) + relative_paths.append(s) + + self._dependencies[key] = relative_paths + elif key in self._dependencies: + del self._dependencies[key] if persist: await self.save() @@ -87,7 +92,7 @@ class DependencyFile: root = self._filename.parent try: - key = Path(filename).relative_to(root) + key = Path(filename).relative_to(root).as_posix() except ValueError: key = filename return set(self._dependencies.get(str(key), {})) diff --git a/metagpt/utils/file_repository.py b/metagpt/utils/file_repository.py index 85e7dc8a4..d2a06963a 100644 --- a/metagpt/utils/file_repository.py +++ b/metagpt/utils/file_repository.py @@ -54,6 +54,7 @@ class FileRepository: """ pathname = self.workdir / filename pathname.parent.mkdir(parents=True, exist_ok=True) + content = content if content else "" # avoid `argument must be str, not None` to make it continue async with aiofiles.open(str(pathname), mode="w") as writer: await writer.write(content) logger.info(f"save to: {str(pathname)}") @@ -63,7 +64,7 @@ class FileRepository: await dependency_file.update(pathname, set(dependencies)) logger.info(f"update dependency: {str(pathname)}:{dependencies}") - return Document(root_path=str(self._relative_path), filename=filename, content=content) + return Document(root_path=str(self._relative_path), filename=str(filename), content=content) async def get_dependency(self, filename: Path | str) -> Set[str]: """Get the dependencies of a file. @@ -100,21 +101,28 @@ class FileRepository: path_name = self.workdir / filename if not path_name.exists(): return None + if not path_name.is_file(): + return None doc.content = await aread(path_name) return doc - async def get_all(self) -> List[Document]: + async def get_all(self, filter_ignored=True) -> List[Document]: """Get the content of all files in the repository. :return: List of Document instances representing files. """ docs = [] - for root, dirs, files in os.walk(str(self.workdir)): - for file in files: - file_path = Path(root) / file - relative_path = file_path.relative_to(self.workdir) - doc = await self.get(relative_path) + if filter_ignored: + for f in self.all_files: + doc = await self.get(f) docs.append(doc) + else: + for root, dirs, files in os.walk(str(self.workdir)): + for file in files: + file_path = Path(root) / file + relative_path = file_path.relative_to(self.workdir) + doc = await self.get(relative_path) + docs.append(doc) return docs @property diff --git a/metagpt/utils/git_repository.py b/metagpt/utils/git_repository.py index 61e5f3251..16f675175 100644 --- a/metagpt/utils/git_repository.py +++ b/metagpt/utils/git_repository.py @@ -201,7 +201,10 @@ class GitRepository: new_path = self.workdir.parent / new_dir_name if new_path.exists(): logger.info(f"Delete directory {str(new_path)}") - shutil.rmtree(new_path) + try: + shutil.rmtree(new_path) + except Exception as e: + logger.warning(f"rm {str(new_path)} error: {e}") if new_path.exists(): # Recheck for windows os logger.warning(f"Failed to delete directory {str(new_path)}") return diff --git a/metagpt/utils/mermaid.py b/metagpt/utils/mermaid.py index e49fdea5d..ae3c5118f 100644 --- a/metagpt/utils/mermaid.py +++ b/metagpt/utils/mermaid.py @@ -35,10 +35,10 @@ async def mermaid_to_file(engine, mermaid_code, output_file_without_suffix, widt # tmp.write_text(mermaid_code, encoding="utf-8") if engine == "nodejs": - if check_cmd_exists(config.mmdc) != 0: + if check_cmd_exists(config.mermaid.path) != 0: logger.warning( "RUN `npm install -g @mermaid-js/mermaid-cli` to install mmdc," - "or consider changing MERMAID_ENGINE to `playwright`, `pyppeteer`, or `ink`." + "or consider changing engine to `playwright`, `pyppeteer`, or `ink`." ) return -1 @@ -47,11 +47,11 @@ async def mermaid_to_file(engine, mermaid_code, output_file_without_suffix, widt # Call the `mmdc` command to convert the Mermaid code to a PNG logger.info(f"Generating {output_file}..") - if config.puppeteer_config: + if config.mermaid.puppeteer_config: commands = [ - config.mmdc, + config.mermaid.path, "-p", - config.puppeteer_config, + config.mermaid.puppeteer_config, "-i", str(tmp), "-o", @@ -62,7 +62,7 @@ async def mermaid_to_file(engine, mermaid_code, output_file_without_suffix, widt str(height), ] else: - commands = [config.mmdc, "-i", str(tmp), "-o", output_file, "-w", str(width), "-H", str(height)] + commands = [config.mermaid.path, "-i", str(tmp), "-o", output_file, "-w", str(width), "-H", str(height)] process = await asyncio.create_subprocess_shell( " ".join(commands), stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE ) diff --git a/metagpt/utils/mmdc_pyppeteer.py b/metagpt/utils/mmdc_pyppeteer.py index d80098b7d..f029325f1 100644 --- a/metagpt/utils/mmdc_pyppeteer.py +++ b/metagpt/utils/mmdc_pyppeteer.py @@ -30,14 +30,14 @@ async def mermaid_to_file(mermaid_code, output_file_without_suffix, width=2048, suffixes = ["png", "svg", "pdf"] __dirname = os.path.dirname(os.path.abspath(__file__)) - if config.pyppeteer_executable_path: + if config.mermaid.pyppeteer_path: browser = await launch( headless=True, - executablePath=config.pyppeteer_executable_path, + executablePath=config.mermaid.pyppeteer_path, args=["--disable-extensions", "--no-sandbox"], ) else: - logger.error("Please set the environment variable:PYPPETEER_EXECUTABLE_PATH.") + logger.error("Please set the var mermaid.pyppeteer_path in the config2.yaml.") return -1 page = await browser.newPage() device_scale_factor = 1.0 diff --git a/metagpt/utils/parse_docstring.py b/metagpt/utils/parse_docstring.py new file mode 100644 index 000000000..e91be8e75 --- /dev/null +++ b/metagpt/utils/parse_docstring.py @@ -0,0 +1,87 @@ +import re +from typing import Tuple + +from pydantic import BaseModel + + +def remove_spaces(text): + return re.sub(r"\s+", " ", text).strip() + + +class DocstringParser(BaseModel): + docstring: str + + def parse_desc(self) -> str: + """Parse and return the description from the docstring.""" + + def parse_params(self) -> list[Tuple[str, str, str]]: + """Parse and return the parameters from the docstring. + + Returns: + list[Tuple[str, str, str]]: A list of input paramter info. Each info is a triple of (param name, param type, param description) + """ + + def parse_returns(self) -> list[Tuple[str, str]]: + """Parse and return the output information from the docstring. + + Returns: + list[Tuple[str, str]]: A list of output info. Each info is a tuple of (return type, return description) + """ + + @staticmethod + def check_and_parse_optional(param_type: str) -> Tuple[bool, str]: + """Check if a parameter is optional and return a processed param_type rid of the optionality info if so""" + + @staticmethod + def check_and_parse_default_value(param_desc: str) -> Tuple[bool, str]: + """Check if a parameter has a default value and return the default value if so""" + + @staticmethod + def check_and_parse_enum(param_desc: str) -> Tuple[bool, str]: + """Check if a parameter description includes an enum and return enum values if so""" + + +class reSTDocstringParser(DocstringParser): + """A parser for reStructuredText (reST) docstring""" + + +class GoogleDocstringParser(DocstringParser): + """A parser for Google-stype docstring""" + + docstring: str + + def parse_desc(self) -> str: + description_match = re.search(r"^(.*?)(?:Args:|Returns:|Raises:|$)", self.docstring, re.DOTALL) + description = remove_spaces(description_match.group(1)) if description_match else "" + return description + + def parse_params(self) -> list[Tuple[str, str, str]]: + args_match = re.search(r"Args:\s*(.*?)(?:Returns:|Raises:|$)", self.docstring, re.DOTALL) + _args = args_match.group(1).strip() if args_match else "" + # variable_pattern = re.compile(r"(\w+)\s*\((.*?)\):\s*(.*)") + variable_pattern = re.compile( + r"(\w+)\s*\((.*?)\):\s*(.*?)(?=\n\s*\w+\s*\(|\Z)", re.DOTALL + ) # (?=\n\w+\s*\(|\Z) is to assert that what follows is either the start of the next parameter (indicated by a newline, some word characters, and an opening parenthesis) or the end of the string (\Z). + params = variable_pattern.findall(_args) + return params + + def parse_returns(self) -> list[Tuple[str, str]]: + returns_match = re.search(r"Returns:\s*(.*?)(?:Raises:|$)", self.docstring, re.DOTALL) + returns = returns_match.group(1).strip() if returns_match else "" + return_pattern = re.compile(r"^(.*)\s*:\s*(.*)$") + returns = return_pattern.findall(returns) + return returns + + @staticmethod + def check_and_parse_optional(param_type: str) -> Tuple[bool, str]: + return "optional" in param_type, param_type.replace(", optional", "") + + @staticmethod + def check_and_parse_default_value(param_desc: str) -> Tuple[bool, str]: + default_val = re.search(r"Defaults to (.+?)\.", param_desc) + return (True, default_val.group(1)) if default_val else (False, "") + + @staticmethod + def check_and_parse_enum(param_desc: str) -> Tuple[bool, str]: + enum_val = re.search(r"Enum: \[(.+?)\]", param_desc) + return (True, [e.strip() for e in enum_val.group(1).split(",")]) if enum_val else (False, []) diff --git a/metagpt/utils/project_repo.py b/metagpt/utils/project_repo.py index dd54cb56b..c1f98e1ec 100644 --- a/metagpt/utils/project_repo.py +++ b/metagpt/utils/project_repo.py @@ -13,6 +13,8 @@ from pathlib import Path from metagpt.const import ( CLASS_VIEW_FILE_REPO, + CODE_PLAN_AND_CHANGE_FILE_REPO, + CODE_PLAN_AND_CHANGE_PDF_FILE_REPO, CODE_SUMMARIES_FILE_REPO, CODE_SUMMARIES_PDF_FILE_REPO, COMPETITIVE_ANALYSIS_FILE_REPO, @@ -21,6 +23,7 @@ from metagpt.const import ( GRAPH_REPO_FILE_REPO, PRD_PDF_FILE_REPO, PRDS_FILE_REPO, + REQUIREMENT_FILENAME, RESOURCES_FILE_REPO, SD_OUTPUT_FILE_REPO, SEQ_FLOW_FILE_REPO, @@ -42,6 +45,7 @@ class DocFileRepositories(FileRepository): code_summary: FileRepository graph_repo: FileRepository class_view: FileRepository + code_plan_and_change: FileRepository def __init__(self, git_repo): super().__init__(git_repo=git_repo, relative_path=DOCS_FILE_REPO) @@ -52,6 +56,7 @@ class DocFileRepositories(FileRepository): self.code_summary = git_repo.new_file_repository(relative_path=CODE_SUMMARIES_FILE_REPO) self.graph_repo = git_repo.new_file_repository(relative_path=GRAPH_REPO_FILE_REPO) self.class_view = git_repo.new_file_repository(relative_path=CLASS_VIEW_FILE_REPO) + self.code_plan_and_change = git_repo.new_file_repository(relative_path=CODE_PLAN_AND_CHANGE_FILE_REPO) class ResourceFileRepositories(FileRepository): @@ -63,6 +68,7 @@ class ResourceFileRepositories(FileRepository): api_spec_and_task: FileRepository code_summary: FileRepository sd_output: FileRepository + code_plan_and_change: FileRepository def __init__(self, git_repo): super().__init__(git_repo=git_repo, relative_path=RESOURCES_FILE_REPO) @@ -75,6 +81,7 @@ class ResourceFileRepositories(FileRepository): self.api_spec_and_task = git_repo.new_file_repository(relative_path=TASK_PDF_FILE_REPO) self.code_summary = git_repo.new_file_repository(relative_path=CODE_SUMMARIES_PDF_FILE_REPO) self.sd_output = git_repo.new_file_repository(relative_path=SD_OUTPUT_FILE_REPO) + self.code_plan_and_change = git_repo.new_file_repository(relative_path=CODE_PLAN_AND_CHANGE_PDF_FILE_REPO) class ProjectRepo(FileRepository): @@ -92,6 +99,17 @@ class ProjectRepo(FileRepository): self.tests = self._git_repo.new_file_repository(relative_path=TEST_CODES_FILE_REPO) self.test_outputs = self._git_repo.new_file_repository(relative_path=TEST_OUTPUTS_FILE_REPO) self._srcs_path = None + self.code_files_exists() + + def __str__(self): + repo_str = f"ProjectRepo({self._git_repo.workdir})" + docs_str = f"Docs({self.docs.all_files})" + srcs_str = f"Srcs({self.srcs.all_files})" + return f"{repo_str}\n{docs_str}\n{srcs_str}" + + @property + async def requirement(self): + return await self.docs.get(filename=REQUIREMENT_FILENAME) @property def git_repo(self) -> GitRepository: @@ -107,6 +125,15 @@ class ProjectRepo(FileRepository): raise ValueError("Call with_srcs first.") return self._git_repo.new_file_repository(self._srcs_path) + def code_files_exists(self) -> bool: + git_workdir = self.git_repo.workdir + src_workdir = git_workdir / git_workdir.name + if not src_workdir.exists(): + return False + code_files = self.with_src_path(path=git_workdir / git_workdir.name).srcs.all_files + if not code_files: + return False + def with_src_path(self, path: str | Path) -> ProjectRepo: try: self._srcs_path = Path(path).relative_to(self.workdir) diff --git a/metagpt/utils/recovery_util.py b/metagpt/utils/recovery_util.py new file mode 100644 index 000000000..d0b197e69 --- /dev/null +++ b/metagpt/utils/recovery_util.py @@ -0,0 +1,58 @@ +# -*- coding: utf-8 -*- +# @Date : 12/20/2023 11:07 AM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : +import json +from datetime import datetime +from pathlib import Path + +import nbformat + +from metagpt.const import DATA_PATH +from metagpt.roles.role import Role +from metagpt.utils.common import read_json_file +from metagpt.utils.save_code import save_code_file + + +def load_history(save_dir: str = ""): + """ + Load plan and code execution history from the specified save directory. + + Args: + save_dir (str): The directory from which to load the history. + + Returns: + Tuple: A tuple containing the loaded plan and notebook. + """ + + plan_path = Path(save_dir) / "plan.json" + nb_path = Path(save_dir) / "history_nb" / "code.ipynb" + plan = read_json_file(plan_path) + nb = nbformat.read(open(nb_path, "r", encoding="utf-8"), as_version=nbformat.NO_CONVERT) + return plan, nb + + +def save_history(role: Role, save_dir: str = ""): + """ + Save plan and code execution history to the specified directory. + + Args: + role (Role): The role containing the plan and execute_code attributes. + save_dir (str): The directory to save the history. + + Returns: + Path: The path to the saved history directory. + """ + record_time = datetime.now().strftime("%Y-%m-%d_%H-%M-%S") + save_path = DATA_PATH / "output" / f"{record_time}" + + # overwrite exist trajectory + save_path.mkdir(parents=True, exist_ok=True) + + plan = role.planner.plan.dict() + + with open(save_path / "plan.json", "w", encoding="utf-8") as plan_file: + json.dump(plan, plan_file, indent=4, ensure_ascii=False) + + save_code_file(name=Path(record_time) / "history_nb", code_context=role.execute_code.nb, file_format="ipynb") + return save_path diff --git a/metagpt/utils/repair_llm_raw_output.py b/metagpt/utils/repair_llm_raw_output.py index ec2da53f8..06484f71d 100644 --- a/metagpt/utils/repair_llm_raw_output.py +++ b/metagpt/utils/repair_llm_raw_output.py @@ -119,7 +119,23 @@ def repair_json_format(output: str) -> str: logger.info(f"repair_json_format: {'}]'}") elif output.startswith("{") and output.endswith("]"): output = output[:-1] + "}" - + # remove comments in output json string, after json value content, maybe start with #, maybe start with // + arr = output.split("\n") + new_arr = [] + for json_line in arr: + # look for # or // comments and make sure they are not inside the string value + comment_index = -1 + for match in re.finditer(r"(\".*?\"|\'.*?\')|(#|//)", json_line): + if match.group(1): # if the string value + continue + if match.group(2): # if comments + comment_index = match.start(2) + break + # if comments, then delete them + if comment_index != -1: + json_line = json_line[:comment_index].rstrip() + new_arr.append(json_line) + output = "\n".join(new_arr) return output @@ -168,15 +184,17 @@ def repair_invalid_json(output: str, error: str) -> str: example 1. json.decoder.JSONDecodeError: Expecting ',' delimiter: line 154 column 1 (char 2765) example 2. xxx.JSONDecodeError: Expecting property name enclosed in double quotes: line 14 column 1 (char 266) """ - pattern = r"line ([0-9]+)" + pattern = r"line ([0-9]+) column ([0-9]+)" matches = re.findall(pattern, error, re.DOTALL) if len(matches) > 0: - line_no = int(matches[0]) - 1 + line_no = int(matches[0][0]) - 1 + col_no = int(matches[0][1]) - 1 # due to CustomDecoder can handle `"": ''` or `'': ""`, so convert `"""` -> `"`, `'''` -> `'` output = output.replace('"""', '"').replace("'''", '"') arr = output.split("\n") + rline = arr[line_no] # raw line line = arr[line_no].strip() # different general problems if line.endswith("],"): @@ -187,9 +205,12 @@ def repair_invalid_json(output: str, error: str) -> str: new_line = line.replace("}", "") elif line.endswith("},") and output.endswith("},"): new_line = line[:-1] - elif '",' not in line and "," not in line: + elif (rline[col_no] in ["'", '"']) and (line.startswith('"') or line.startswith("'")) and "," not in line: + # problem, `"""` or `'''` without `,` + new_line = f",{line}" + elif '",' not in line and "," not in line and '"' not in line: new_line = f'{line}",' - elif "," not in line: + elif not line.endswith(","): # problem, miss char `,` at the end. new_line = f"{line}," elif "," in line and len(line) == 1: diff --git a/metagpt/utils/save_code.py b/metagpt/utils/save_code.py new file mode 100644 index 000000000..18cb5cd62 --- /dev/null +++ b/metagpt/utils/save_code.py @@ -0,0 +1,40 @@ +# -*- coding: utf-8 -*- +# @Date : 12/12/2023 4:14 PM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : +import os + +import nbformat + +from metagpt.const import DATA_PATH +from metagpt.utils.common import write_json_file + + +def save_code_file(name: str, code_context: str, file_format: str = "py") -> None: + """ + Save code files to a specified path. + + Args: + - name (str): The name of the folder to save the files. + - code_context (str): The code content. + - file_format (str, optional): The file format. Supports 'py' (Python file), 'json' (JSON file), and 'ipynb' (Jupyter Notebook file). Default is 'py'. + + + Returns: + - None + """ + # Create the folder path if it doesn't exist + os.makedirs(name=DATA_PATH / "output" / f"{name}", exist_ok=True) + + # Choose to save as a Python file or a JSON file based on the file format + file_path = DATA_PATH / "output" / f"{name}/code.{file_format}" + if file_format == "py": + file_path.write_text(code_context + "\n\n", encoding="utf-8") + elif file_format == "json": + # Parse the code content as JSON and save + data = {"code": code_context} + write_json_file(file_path, data, encoding="utf-8", indent=2) + elif file_format == "ipynb": + nbformat.write(code_context, file_path) + else: + raise ValueError("Unsupported file format. Please choose 'py', 'json', or 'ipynb'.") diff --git a/metagpt/utils/text.py b/metagpt/utils/text.py index dd9678438..fb8b94232 100644 --- a/metagpt/utils/text.py +++ b/metagpt/utils/text.py @@ -25,7 +25,7 @@ def reduce_message_length( """ max_token = TOKEN_MAX.get(model_name, 2048) - count_string_tokens(system_text, model_name) - reserved for msg in msgs: - if count_string_tokens(msg, model_name) < max_token: + if count_string_tokens(msg, model_name) < max_token or model_name not in TOKEN_MAX: return msg raise RuntimeError("fail to reduce message length") @@ -93,7 +93,7 @@ def split_paragraph(paragraph: str, sep: str = ".,", count: int = 2) -> list[str continue ret = ["".join(j) for j in _split_by_count(sentences, count)] return ret - return _split_by_count(paragraph, count) + return list(_split_by_count(paragraph, count)) def decode_unicode_escape(text: str) -> str: diff --git a/metagpt/utils/token_counter.py b/metagpt/utils/token_counter.py index a1b74a074..65f5fe76f 100644 --- a/metagpt/utils/token_counter.py +++ b/metagpt/utils/token_counter.py @@ -4,10 +4,11 @@ @Time : 2023/5/18 00:40 @Author : alexanderwu @File : token_counter.py -ref1: https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb -ref2: https://github.com/Significant-Gravitas/Auto-GPT/blob/master/autogpt/llm/token_counter.py -ref3: https://github.com/hwchase17/langchain/blob/master/langchain/chat_models/openai.py -ref4: https://ai.google.dev/models/gemini +ref1: https://openai.com/pricing +ref2: https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb +ref3: https://github.com/Significant-Gravitas/Auto-GPT/blob/master/autogpt/llm/token_counter.py +ref4: https://github.com/hwchase17/langchain/blob/master/langchain/chat_models/openai.py +ref5: https://ai.google.dev/models/gemini """ import tiktoken @@ -25,9 +26,14 @@ TOKEN_COSTS = { "gpt-4-32k": {"prompt": 0.06, "completion": 0.12}, "gpt-4-32k-0314": {"prompt": 0.06, "completion": 0.12}, "gpt-4-0613": {"prompt": 0.06, "completion": 0.12}, + "gpt-4-turbo-preview": {"prompt": 0.01, "completion": 0.03}, + "gpt-4-0125-preview": {"prompt": 0.01, "completion": 0.03}, "gpt-4-1106-preview": {"prompt": 0.01, "completion": 0.03}, + "gpt-4-vision-preview": {"prompt": 0.01, "completion": 0.03}, # TODO add extra image price calculator + "gpt-4-1106-vision-preview": {"prompt": 0.01, "completion": 0.03}, "text-embedding-ada-002": {"prompt": 0.0004, "completion": 0.0}, - "chatglm_turbo": {"prompt": 0.0, "completion": 0.00069}, # 32k version, prompt + completion tokens=0.005¥/k-tokens + "glm-3-turbo": {"prompt": 0.0007, "completion": 0.0007}, # 128k version, prompt + completion tokens=0.005¥/k-tokens + "glm-4": {"prompt": 0.014, "completion": 0.014}, # 128k version, prompt + completion tokens=0.1¥/k-tokens "gemini-pro": {"prompt": 0.00025, "completion": 0.0005}, } @@ -46,9 +52,14 @@ TOKEN_MAX = { "gpt-4-32k": 32768, "gpt-4-32k-0314": 32768, "gpt-4-0613": 8192, + "gpt-4-turbo-preview": 128000, + "gpt-4-0125-preview": 128000, "gpt-4-1106-preview": 128000, + "gpt-4-vision-preview": 128000, + "gpt-4-1106-vision-preview": 128000, "text-embedding-ada-002": 8192, - "chatglm_turbo": 32768, + "glm-3-turbo": 128000, + "glm-4": 128000, "gemini-pro": 32768, } @@ -71,7 +82,11 @@ def count_message_tokens(messages, model="gpt-3.5-turbo-0613"): "gpt-4-32k-0314", "gpt-4-0613", "gpt-4-32k-0613", + "gpt-4-turbo-preview", + "gpt-4-0125-preview", "gpt-4-1106-preview", + "gpt-4-vision-preview", + "gpt-4-1106-vision-preview", }: tokens_per_message = 3 # # every reply is primed with <|start|>assistant<|message|> tokens_per_name = 1 @@ -101,7 +116,13 @@ def count_message_tokens(messages, model="gpt-3.5-turbo-0613"): for message in messages: num_tokens += tokens_per_message for key, value in message.items(): - num_tokens += len(encoding.encode(value)) + content = value + if isinstance(value, list): + # for gpt-4v + for item in value: + if isinstance(item, dict) and item.get("type") in ["text"]: + content = item.get("text", "") + num_tokens += len(encoding.encode(content)) if key == "name": num_tokens += tokens_per_name num_tokens += 3 # every reply is primed with <|start|>assistant<|message|> diff --git a/metagpt/utils/yaml_model.py b/metagpt/utils/yaml_model.py index 8f2d22c3d..4d42bb03f 100644 --- a/metagpt/utils/yaml_model.py +++ b/metagpt/utils/yaml_model.py @@ -42,7 +42,7 @@ class YamlModelWithoutDefault(YamlModel): @model_validator(mode="before") @classmethod def check_not_default_config(cls, values): - """Check if there is any default config in config.yaml""" + """Check if there is any default config in config2.yaml""" if any(["YOUR" in v for v in values]): - raise ValueError("Please set your config in config.yaml") + raise ValueError("Please set your config in config2.yaml") return values diff --git a/requirements.txt b/requirements.txt index 0a54236f0..1426500ce 100644 --- a/requirements.txt +++ b/requirements.txt @@ -14,7 +14,7 @@ lancedb==0.4.0 langchain==0.0.352 loguru==0.6.0 meilisearch==0.21.0 -numpy==1.24.3 +numpy>=1.24.3 openai==1.6.0 openpyxl beautifulsoup4==4.12.2 @@ -35,7 +35,6 @@ tqdm==4.65.0 # webdriver_manager<3.9 anthropic==0.8.1 typing-inspect==0.8.0 -typing_extensions==4.9.0 libcst==1.0.1 qdrant-client==1.7.0 # pytest-mock==3.11.1 # test extras require @@ -50,12 +49,21 @@ aioredis~=2.0.1 # Used by metagpt/utils/redis.py websocket-client==1.6.2 aiofiles==23.2.1 gitpython==3.1.40 -zhipuai==1.0.7 +zhipuai==2.0.1 +rich==13.6.0 +nbclient==0.9.0 +nbformat==5.9.2 +ipython==8.17.2 +ipykernel==6.27.0 +scikit_learn==1.3.2 +typing-extensions==4.9.0 socksio~=1.0.0 gitignore-parser==0.1.9 # connexion[uvicorn]~=3.0.5 # Used by metagpt/tools/openapi_v3_hello.py websockets~=12.0 networkx~=3.2.1 google-generativeai==0.3.2 -# playwright==1.40.0 # playwright extras require +playwright>=1.26 # used at metagpt/tools/libs/web_scraping.py anytree +ipywidgets==8.1.1 +Pillow diff --git a/setup.py b/setup.py index 0439d6cd4..be3956ea4 100644 --- a/setup.py +++ b/setup.py @@ -24,7 +24,6 @@ requirements = (here / "requirements.txt").read_text(encoding="utf-8").splitline extras_require = { - "playwright": ["playwright>=1.26", "beautifulsoup4"], "selenium": ["selenium>4", "webdriver_manager", "beautifulsoup4"], "search-google": ["google-api-python-client==2.94.0"], "search-ddg": ["duckduckgo-search~=4.1.1"], @@ -46,8 +45,8 @@ extras_require["test"] = [ "chromadb==0.4.14", "gradio==3.0.0", "grpcio-status==1.48.2", - "mock==5.1.0", "pylint==3.0.3", + "pybrowsers", ] extras_require["pyppeteer"] = [ @@ -58,7 +57,7 @@ extras_require["dev"] = (["pylint~=3.0.3", "black~=23.3.0", "isort~=5.12.0", "pr setup( name="metagpt", - version="0.6.0", + version="0.7.0", description="The Multi-Agent Framework", long_description=long_description, long_description_content_type="text/markdown", @@ -76,7 +75,7 @@ setup( }, entry_points={ "console_scripts": [ - "metagpt=metagpt.startup:app", + "metagpt=metagpt.software_company:app", ], }, ) diff --git a/tests/config2.yaml b/tests/config2.yaml new file mode 100644 index 000000000..58314eaed --- /dev/null +++ b/tests/config2.yaml @@ -0,0 +1,27 @@ +llm: + base_url: "https://api.openai.com/v1" + api_key: "sk-xxx" + model: "gpt-3.5-turbo-1106" + +search: + api_type: "serpapi" + api_key: "xxx" + +s3: + access_key: "MOCK_S3_ACCESS_KEY" + secret_key: "MOCK_S3_SECRET_KEY" + endpoint: "http://mock:9000" + secure: false + bucket: "mock" + +azure_tts_subscription_key: "xxx" +azure_tts_region: "eastus" + +iflytek_app_id: "xxx" +iflytek_api_key: "xxx" +iflytek_api_secret: "xxx" + +metagpt_tti_url: "http://mock.com" + +repair_llm_output: true + diff --git a/tests/conftest.py b/tests/conftest.py index 06113443a..f2379a833 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -18,10 +18,11 @@ import aiohttp.web import pytest from metagpt.const import DEFAULT_WORKSPACE_ROOT, TEST_DATA_PATH -from metagpt.context import CONTEXT +from metagpt.context import Context as MetagptContext from metagpt.llm import LLM from metagpt.logs import logger from metagpt.utils.git_repository import GitRepository +from metagpt.utils.project_repo import ProjectRepo from tests.mock.mock_aiohttp import MockAioResponse from tests.mock.mock_curl_cffi import MockCurlCffiResponse from tests.mock.mock_httplib2 import MockHttplib2Response @@ -38,14 +39,14 @@ def rsp_cache(): rsp_cache_file_path = TEST_DATA_PATH / "rsp_cache.json" # read repo-provided new_rsp_cache_file_path = TEST_DATA_PATH / "rsp_cache_new.json" # exporting a new copy if os.path.exists(rsp_cache_file_path): - with open(rsp_cache_file_path, "r") as f1: + with open(rsp_cache_file_path, "r", encoding="utf-8") as f1: rsp_cache_json = json.load(f1) else: rsp_cache_json = {} yield rsp_cache_json - with open(rsp_cache_file_path, "w") as f2: + with open(rsp_cache_file_path, "w", encoding="utf-8") as f2: json.dump(rsp_cache_json, f2, indent=4, ensure_ascii=False) - with open(new_rsp_cache_file_path, "w") as f2: + with open(new_rsp_cache_file_path, "w", encoding="utf-8") as f2: json.dump(RSP_CACHE_NEW, f2, indent=4, ensure_ascii=False) @@ -64,6 +65,7 @@ def llm_mock(rsp_cache, mocker, request): llm.rsp_cache = rsp_cache mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", llm.aask) mocker.patch("metagpt.provider.base_llm.BaseLLM.aask_batch", llm.aask_batch) + mocker.patch("metagpt.provider.openai_api.OpenAILLM.aask_code", llm.aask_code) yield mocker if hasattr(request.node, "test_outcome") and request.node.test_outcome.passed: if llm.rsp_candidates: @@ -71,7 +73,7 @@ def llm_mock(rsp_cache, mocker, request): cand_key = list(rsp_candidate.keys())[0] cand_value = list(rsp_candidate.values())[0] if cand_key not in llm.rsp_cache: - logger.info(f"Added '{cand_key[:100]} ... -> {cand_value[:20]} ...' to response cache") + logger.info(f"Added '{cand_key[:100]} ... -> {str(cand_value)[:20]} ...' to response cache") llm.rsp_cache.update(rsp_candidate) RSP_CACHE_NEW.update(rsp_candidate) @@ -142,19 +144,20 @@ def loguru_caplog(caplog): yield caplog -# init & dispose git repo -@pytest.fixture(scope="function", autouse=True) -def setup_and_teardown_git_repo(request): - CONTEXT.git_repo = GitRepository(local_path=DEFAULT_WORKSPACE_ROOT / f"unittest/{uuid.uuid4().hex}") - CONTEXT.config.git_reinit = True +@pytest.fixture(scope="function") +def context(request): + ctx = MetagptContext() + ctx.git_repo = GitRepository(local_path=DEFAULT_WORKSPACE_ROOT / f"unittest/{uuid.uuid4().hex}") + ctx.repo = ProjectRepo(ctx.git_repo) # Destroy git repo at the end of the test session. def fin(): - if CONTEXT.git_repo: - CONTEXT.git_repo.delete_repository() + if ctx.git_repo: + ctx.git_repo.delete_repository() # Register the function for destroying the environment. request.addfinalizer(fin) + return ctx @pytest.fixture(scope="session", autouse=True) diff --git a/tests/data/audio/hello.mp3 b/tests/data/audio/hello.mp3 new file mode 100644 index 000000000..7b3aab0a4 Binary files /dev/null and b/tests/data/audio/hello.mp3 differ diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze/arena_maze.csv b/tests/data/environment/stanford_town/the_ville/matrix/maze/arena_maze.csv new file mode 100644 index 000000000..f9cf65ecd --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze/arena_maze.csv @@ -0,0 +1 @@ +0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32200, 32200, 32200, 32200, 32200, 0, 0, 32151, 32151, 32151, 32151, 32151, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 32199, 32199, 32199, 32199, 0, 32140, 32140, 32140, 0, 0, 32160, 32160, 32160, 32160, 32160, 0, 0, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 0, 32180, 32180, 32180, 0, 0, 32200, 32200, 32200, 32200, 32200, 0, 0, 32151, 32151, 32151, 32151, 32151, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 32199, 32199, 32199, 32199, 0, 32140, 32140, 32140, 0, 0, 32160, 32160, 32160, 32160, 32160, 0, 0, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 0, 32180, 32180, 32180, 0, 0, 32200, 32200, 32200, 32200, 32200, 0, 0, 32151, 32151, 32151, 32151, 32151, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 32199, 32199, 32199, 32199, 0, 32140, 32140, 32140, 0, 0, 32160, 32160, 32160, 32160, 32160, 0, 0, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 0, 32180, 32180, 32180, 0, 0, 32200, 32200, 32200, 32200, 32200, 0, 0, 32151, 32151, 32151, 32151, 32151, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 32199, 32199, 32199, 32199, 32199, 32140, 32140, 32140, 0, 0, 32160, 32160, 32160, 32160, 32160, 0, 0, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32170, 32180, 32180, 32180, 0, 0, 0, 0, 32200, 32200, 0, 0, 0, 0, 0, 32151, 32151, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 0, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 0, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 32178, 32178, 32178, 32178, 32178, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 32199, 32199, 32199, 32199, 32199, 32140, 32140, 32140, 0, 0, 0, 0, 32160, 32160, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32170, 32170, 0, 0, 0, 0, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 0, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 0, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 32178, 32178, 32178, 32178, 32178, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 0, 0, 0, 0, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 0, 0, 0, 0, 0, 0, 32170, 32170, 0, 0, 0, 0, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 0, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 0, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32199, 32199, 0, 0, 0, 0, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 0, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 0, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 32188, 32188, 32188, 32188, 32188, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 32138, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 32158, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 32188, 32188, 32188, 32188, 32188, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 32138, 32138, 32138, 32148, 32148, 32148, 0, 0, 32158, 32158, 32158, 32158, 32158, 32168, 32168, 32168, 0, 0, 32178, 32178, 32178, 32188, 32188, 32188, 32188, 32188, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 0, 0, 0, 0, 0, 0, 0, 0, 32158, 32158, 0, 0, 0, 0, 0, 0, 0, 0, 32178, 32178, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32138, 32138, 0, 0, 0, 0, 0, 0, 0, 0, 32158, 32158, 0, 0, 0, 0, 0, 0, 0, 0, 32178, 32178, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32189, 32189, 32189, 32189, 32189, 32189, 32189, 32189, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 32190, 32190, 32190, 32190, 32190, 0, 0, 32141, 32141, 32141, 32141, 32141, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32189, 32189, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 32161, 0, 0, 32150, 32150, 32150, 32150, 32150, 0, 0, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 32171, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 32198, 32198, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 32149, 32149, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 32189, 32189, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 32181, 0, 0, 32191, 32191, 32191, 32191, 32191, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 32198, 32198, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 32149, 32149, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 32189, 32189, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 0, 0, 0, 0, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 0, 0, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 0, 0, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 32189, 32189, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32181, 32181, 0, 0, 0, 0, 32191, 32191, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 0, 0, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 0, 0, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 0, 32139, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 0, 32159, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 0, 0, 0, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32198, 32198, 32198, 32198, 0, 32139, 32139, 32139, 0, 0, 32149, 32149, 32149, 32149, 0, 32159, 32159, 32159, 0, 32179, 32179, 32179, 32179, 32179, 32179, 32179, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 0, 0, 0, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32179, 32179, 32179, 32179, 32179, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 0, 0, 0, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32179, 32179, 32179, 32179, 32179, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 32201, 0, 0, 0, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 0, 0, 0, 0, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 32183, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32172, 32172, 32172, 32172, 32172, 32172, 32172, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32173, 32173, 32173, 32173, 32173, 32173, 32173, 0, 32172, 32172, 32172, 32172, 32172, 32172, 32172, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32173, 32173, 0, 32172, 32172, 32172, 32172, 32172, 32172, 32172, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32173, 32173, 0, 32172, 32172, 32172, 32172, 32172, 32172, 32172, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32173, 32173, 0, 32172, 32172, 32172, 32172, 32172, 32172, 32172, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32143, 32143, 32143, 32143, 32143, 32143, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32173, 32173, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32143, 32143, 32143, 32143, 32143, 32143, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32142, 32142, 0, 0, 0, 0, 32142, 32142, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32143, 32143, 32143, 32143, 32143, 32143, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32143, 32143, 32143, 32143, 32143, 32143, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 0, 32182, 32182, 0, 0, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 32152, 0, 0, 0, 0, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 32162, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32143, 32143, 0, 0, 0, 0, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 0, 32182, 32182, 32182, 32182, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 32163, 0, 32182, 32182, 32182, 32182, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32182, 32182, 32182, 32182, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 0, 0, 32202, 32202, 0, 0, 32192, 32192, 0, 0, 32192, 32192, 0, 0, 32182, 32182, 32182, 32182, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 32182, 32182, 32182, 32182, 32182, 32182, 0, 0, 32153, 32153, 32153, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 0, 0, 32193, 0, 0, 0, 32174, 32174, 0, 0, 32174, 0, 0, 0, 32194, 32194, 0, 0, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32202, 32202, 32202, 32202, 32202, 32202, 0, 0, 32192, 32192, 32192, 32192, 32192, 32192, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32193, 32193, 32193, 32193, 32193, 0, 0, 0, 32174, 32174, 32174, 32174, 32174, 0, 0, 0, 32194, 32194, 32194, 32194, 32194, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32235, 0, 0, 32245, 32245, 32245, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32216, 0, 0, 32226, 32226, 32226, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32276, 0, 0, 32207, 32207, 32207, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32203, 32203, 0, 0, 0, 0, 0, 0, 32184, 32184, 0, 0, 0, 0, 0, 0, 32204, 32204, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32235, 0, 0, 32245, 32245, 32245, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32216, 0, 0, 32226, 32226, 32226, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32276, 0, 0, 32207, 32207, 32207, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32203, 32203, 32203, 32203, 32203, 0, 0, 0, 32184, 32184, 32184, 32184, 32184, 0, 0, 0, 32204, 32204, 32204, 32204, 32204, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32235, 0, 0, 32245, 32245, 32245, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32216, 0, 0, 32226, 32226, 32226, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32276, 0, 0, 32207, 32207, 32207, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32203, 32203, 32203, 32203, 32203, 0, 0, 0, 32184, 32184, 32184, 32184, 32184, 0, 0, 0, 32204, 32204, 32204, 32204, 32204, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32235, 0, 0, 32245, 32245, 32245, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32216, 0, 0, 32226, 32226, 32226, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32276, 0, 0, 32207, 32207, 32207, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32203, 32203, 32203, 32203, 32203, 0, 0, 0, 32184, 32184, 32184, 32184, 32184, 0, 0, 0, 32204, 32204, 32204, 32204, 32204, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32235, 32235, 0, 0, 32245, 32245, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32216, 32216, 0, 0, 32226, 32226, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32276, 32276, 0, 0, 32207, 32207, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32203, 32203, 32203, 32203, 32203, 0, 0, 0, 32184, 32184, 32184, 32184, 32184, 0, 0, 0, 32204, 32204, 32204, 32204, 32204, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 32225, 32225, 32225, 32225, 32225, 32225, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32206, 32206, 32206, 32206, 32206, 32206, 32206, 32206, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32266, 32266, 32266, 32266, 32266, 32266, 32266, 32266, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32225, 32225, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32205, 32205, 32205, 32205, 32205, 0, 32215, 32215, 32215, 32215, 32215, 32215, 0, 0, 0, 0, 0, 0, 32265, 32265, 32265, 32265, 32265, 0, 32275, 32275, 32275, 32275, 32275, 32275, 0, 0, 0, 0, 0, 0, 32246, 32246, 32246, 32246, 32246, 0, 32256, 32256, 32256, 32256, 32256, 32256, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32205, 32205, 32205, 32205, 32205, 0, 32215, 32215, 32215, 32215, 32215, 32215, 0, 0, 0, 0, 0, 0, 32265, 32265, 32265, 32265, 32265, 0, 32275, 32275, 32275, 32275, 32275, 32275, 0, 0, 0, 0, 0, 0, 32246, 32246, 32246, 32246, 32246, 0, 32256, 32256, 32256, 32256, 32256, 32256, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32205, 32205, 32205, 32205, 32205, 0, 32215, 32215, 32215, 32215, 32215, 32215, 0, 0, 0, 0, 0, 0, 32265, 32265, 32265, 32265, 32265, 0, 32275, 32275, 32275, 32275, 32275, 32275, 0, 0, 0, 0, 0, 0, 32246, 32246, 32246, 32246, 32246, 0, 32256, 32256, 32256, 32256, 32256, 32256, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 0, 0, 0, 0, 0, 0, 0, 0, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 0, 0, 0, 0, 0, 0, 0, 0, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 0, 0, 0, 0, 0, 0, 0, 0, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 0, 0, 0, 0, 0, 0, 0, 0, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 0, 0, 0, 0, 0, 0, 0, 0, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 0, 0, 0, 0, 0, 0, 0, 0, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 0, 0, 0, 0, 0, 0, 0, 0, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 0, 0, 0, 0, 0, 0, 0, 0, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 32255, 0, 0, 0, 0, 0, 0, 0, 0, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 32236, 0, 0, 0, 0, 0, 0, 0, 0, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 32217, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze/collision_maze.csv b/tests/data/environment/stanford_town/the_ville/matrix/maze/collision_maze.csv new file mode 100644 index 000000000..40329a8c4 --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze/collision_maze.csv @@ -0,0 +1 @@ +0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 32125, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 32125, 0, 32125, 32125, 0, 32125, 32125, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 32125, 32125, 32125, 32125, 0, 0, 32125, 32125, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 32125, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 32125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze/game_object_maze.csv b/tests/data/environment/stanford_town/the_ville/matrix/maze/game_object_maze.csv new file mode 100644 index 000000000..9c97dc7bd --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze/game_object_maze.csv @@ -0,0 +1 @@ +0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32208, 32208, 32277, 32277, 32218, 0, 0, 32208, 32208, 32277, 32277, 32218, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 32227, 32238, 32247, 32257, 32257, 0, 32208, 32208, 32218, 0, 0, 32208, 32208, 32277, 32277, 32218, 0, 0, 32227, 32227, 0, 32238, 32247, 32247, 32257, 32257, 0, 32208, 32208, 32218, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32209, 0, 0, 0, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 0, 0, 0, 32208, 32208, 32218, 0, 0, 32227, 32227, 0, 0, 0, 32208, 32208, 32218, 0, 0, 0, 32227, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 32227, 0, 0, 32238, 0, 0, 32227, 32227, 0, 0, 32238, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 32227, 0, 0, 32238, 0, 0, 32238, 32238, 0, 32258, 32258, 32228, 32249, 32249, 32249, 32249, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32238, 32238, 32239, 32239, 32239, 32228, 32258, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32279, 32279, 32279, 32279, 0, 0, 0, 0, 0, 0, 32240, 32240, 0, 32250, 32250, 32250, 32250, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 32209, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 32277, 32277, 32218, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32259, 0, 32259, 0, 32259, 0, 0, 0, 0, 32237, 0, 32228, 0, 0, 0, 0, 32237, 0, 32228, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32270, 0, 0, 32270, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 32277, 0, 0, 0, 32267, 32267, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32228, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32260, 0, 0, 32260, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32229, 0, 32229, 0, 32229, 0, 32229, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32259, 0, 32259, 0, 32259, 0, 0, 32259, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32280, 0, 0, 0, 32270, 0, 0, 32270, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32220, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32279, 32279, 0, 32247, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32260, 0, 0, 32260, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32238, 32238, 0, 0, 0, 32248, 32228, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 0, 0, 32219, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 32259, 0, 32259, 0, 32259, 0, 0, 32259, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32270, 0, 0, 32270, 0, 0, 0, 0, 0, 0, 0, 32250, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32269, 32269, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32250, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32250, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 0, 0, 32218, 0, 0, 0, 32267, 32267, 0, 0, 0, 0, 32218, 0, 0, 0, 32268, 0, 0, 32268, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 0, 0, 32277, 0, 0, 0, 32247, 32247, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32277, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 32277, 0, 0, 0, 32278, 32268, 0, 0, 32268, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32267, 32267, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32278, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 32227, 0, 0, 0, 32208, 32208, 0, 0, 0, 32278, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32278, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 32222, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 32252, 0, 0, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 32252, 0, 0, 32252, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32271, 32271, 32271, 32271, 32271, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32281, 32281, 32281, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32221, 32221, 32221, 32221, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32231, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32248, 32228, 32238, 32238, 0, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 32252, 0, 0, 32252, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32212, 32212, 32212, 0, 0, 0, 0, 0, 0, 32231, 0, 0, 32231, 0, 0, 0, 0, 0, 0, 0, 0, 32261, 32261, 32261, 32261, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 32227, 32227, 32210, 32210, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32231, 0, 0, 32231, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 0, 32252, 32252, 0, 32252, 32252, 0, 0, 0, 0, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32271, 32271, 32271, 32271, 32271, 32271, 32271, 32271, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32231, 0, 0, 32231, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32231, 0, 0, 0, 0, 32241, 32241, 32241, 32241, 32241, 32241, 0, 32241, 32241, 32241, 32241, 32241, 32241, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 32252, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32208, 32208, 0, 32277, 32277, 32218, 0, 0, 32278, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32211, 0, 0, 32251, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32278, 0, 0, 0, 32282, 32282, 32282, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32271, 32271, 32271, 32271, 32271, 32271, 32271, 32271, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32211, 0, 0, 32251, 0, 0, 32241, 32241, 32241, 32241, 32241, 32241, 0, 0, 0, 32241, 32241, 32241, 32241, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32278, 0, 0, 32282, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32211, 0, 0, 32251, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32282, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32218, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32282, 0, 0, 0, 0, 32247, 32247, 0, 0, 32230, 32230, 0, 0, 0, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32279, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 32257, 32257, 0, 0, 32237, 0, 0, 0, 0, 32227, 0, 0, 0, 32279, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32272, 32272, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 32247, 0, 0, 0, 0, 32247, 32247, 0, 0, 0, 0, 0, 0, 32247, 32247, 0, 0, 32278, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 32237, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32210, 32210, 0, 0, 0, 0, 32247, 0, 0, 32227, 32227, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 0, 32228, 0, 0, 0, 0, 0, 32257, 0, 32228, 0, 0, 0, 0, 0, 0, 0, 32228, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32227, 32227, 0, 32238, 0, 0, 0, 0, 32227, 32227, 0, 32238, 0, 0, 0, 0, 32227, 32227, 0, 32238, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32257, 32257, 32257, 0, 0, 0, 32228, 0, 0, 32218, 32277, 32277, 0, 0, 0, 0, 0, 0, 32257, 32257, 0, 0, 0, 0, 32228, 0, 0, 32218, 32277, 32277, 0, 0, 0, 0, 0, 0, 32257, 32257, 32257, 0, 0, 0, 32228, 0, 0, 32218, 32277, 32277, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32262, 32262, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32258, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32277, 32277, 0, 0, 32218, 0, 0, 0, 32277, 32277, 0, 0, 32218, 0, 0, 0, 32277, 32277, 0, 0, 32218, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32238, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32238, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32238, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32268, 0, 0, 32268, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 32208, 32208, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 32227, 32227, 0, 32227, 32227, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 32227, 32227, 0, 32227, 32227, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32237, 32227, 32227, 0, 32227, 32227, 32237, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32209, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32209, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32247, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32242, 0, 0, 0, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 0, 0, 0, 0, 32232, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 0, 0, 32232, 0, 0, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 0, 0, 0, 0, 32232, 32232, 0, 0, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 32232, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze/sector_maze.csv b/tests/data/environment/stanford_town/the_ville/matrix/maze/sector_maze.csv new file mode 100644 index 000000000..165b7b03a --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze/sector_maze.csv @@ -0,0 +1 @@ +0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32165, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32145, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32165, 32165, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32145, 32145, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32165, 32165, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32145, 32145, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32175, 32175, 32175, 32175, 32175, 32175, 32175, 32185, 32185, 32185, 32185, 32185, 32185, 32185, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32195, 32155, 32155, 32155, 32155, 32155, 32155, 32155, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 32136, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 32146, 0, 0, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32135, 32135, 32135, 32135, 32135, 32135, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 32156, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 32166, 0, 0, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 32176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 0, 0, 32186, 32186, 0, 32196, 32196, 32196, 0, 0, 32196, 32196, 0, 32137, 32137, 32137, 0, 0, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 32177, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32186, 32186, 32186, 32186, 32186, 32186, 32186, 0, 32196, 32196, 32196, 32196, 32196, 32196, 32196, 0, 32137, 32137, 32137, 32137, 32137, 32137, 32137, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 32147, 0, 0, 0, 0, 0, 0, 0, 0, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 32157, 0, 0, 0, 0, 0, 0, 0, 0, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 32167, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze/spawning_location_maze.csv b/tests/data/environment/stanford_town/the_ville/matrix/maze/spawning_location_maze.csv new file mode 100644 index 000000000..6d7ca5727 --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze/spawning_location_maze.csv @@ -0,0 +1 @@ +0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32306, 32316, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32307, 32317, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32285, 0, 0, 0, 0, 0, 0, 0, 0, 32295, 32305, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32315, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32288, 32298, 0, 0, 0, 0, 0, 32308, 32318, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32287, 32297, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32286, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32296, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32313, 32323, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32294, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32304, 32314, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32324, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32289, 32299, 0, 0, 0, 0, 0, 0, 32309, 32319, 0, 0, 0, 0, 0, 0, 32290, 32300, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32310, 32320, 0, 32291, 32301, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32311, 32321, 0, 32292, 32302, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 32312, 32322, 0, 32293, 32303, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/maze_meta_info.json b/tests/data/environment/stanford_town/the_ville/matrix/maze_meta_info.json new file mode 100644 index 000000000..32a15fbb6 --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/maze_meta_info.json @@ -0,0 +1,5 @@ +{"world_name": "the ville", + "maze_width": 140, + "maze_height": 100, + "sq_tile_size": 32, + "special_constraint": ""} \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/arena_blocks.csv b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/arena_blocks.csv new file mode 100644 index 000000000..c92e0c6ab --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/arena_blocks.csv @@ -0,0 +1,63 @@ +32138, the Ville, artist's co-living space, Latoya Williams's room +32148, the Ville, artist's co-living space, Latoya Williams's bathroom +32158, the Ville, artist's co-living space, Rajiv Patel's room +32168, the Ville, artist's co-living space, Rajiv Patel's bathroom +32178, the Ville, artist's co-living space, Abigail Chen's room +32188, the Ville, artist's co-living space, Abigail Chen's bathroom +32198, the Ville, artist's co-living space, Francisco Lopez's room +32139, the Ville, artist's co-living space, Francisco Lopez's bathroom +32149, the Ville, artist's co-living space, Hailey Johnson's room +32159, the Ville, artist's co-living space, Hailey Johnson's bathroom +32179, the Ville, artist's co-living space, common room +32189, the Ville, artist's co-living space, kitchen +32199, the Ville, Arthur Burton's apartment, main room +32140, the Ville, Arthur Burton's apartment, bathroom +32150, the Ville, Ryan Park's apartment, main room +32160, the Ville, Ryan Park's apartment, bathroom +32170, the Ville, Isabella Rodriguez's apartment, main room +32180, the Ville, Isabella Rodriguez's apartment, bathroom +32190, the Ville, Giorgio Rossi's apartment, main room +32200, the Ville, Giorgio Rossi's apartment, bathroom +32141, the Ville, Carlos Gomez's apartment, main room +32151, the Ville, Carlos Gomez's apartment, bathroom +32161, the Ville, The Rose and Crown Pub, pub +32171, the Ville, Hobbs Cafe, cafe +32181, the Ville, Oak Hill College, classroom +32191, the Ville, Oak Hill College, library +32201, the Ville, Oak Hill College, hallway +32142, the Ville, Johnson Park, park +32152, the Ville, Harvey Oak Supply Store, supply store +32162, the Ville, The Willows Market and Pharmacy, store +32193, the Ville, Adam Smith's house, main room +32203, the Ville, Adam Smith's house, bathroom +32174, the Ville, Yuriko Yamamoto's house, main room +32184, the Ville, Yuriko Yamamoto's house, bathroom +32194, the Ville, Moore family's house, main room +32204, the Ville, Moore family's house, bathroom +32172, the Ville, Dorm for Oak Hill College, Klaus Mueller's room +32182, the Ville, Dorm for Oak Hill College, Maria Lopez's room +32192, the Ville, Dorm for Oak Hill College, Ayesha Khan's room +32202, the Ville, Dorm for Oak Hill College, Wolfgang Schulz's room +32143, the Ville, Dorm for Oak Hill College, man's bathroom +32153, the Ville, Dorm for Oak Hill College, woman's bathroom +32163, the Ville, Dorm for Oak Hill College, common room +32173, the Ville, Dorm for Oak Hill College, kitchen +32183, the Ville, Dorm for Oak Hill College, garden +32205, the Ville, Tamara Taylor and Carmen Ortiz's house, Tamara Taylor's room +32215, the Ville, Tamara Taylor and Carmen Ortiz's house, Carmen Ortiz's room +32225, the Ville, Tamara Taylor and Carmen Ortiz's house, common room +32235, the Ville, Tamara Taylor and Carmen Ortiz's house, kitchen +32245, the Ville, Tamara Taylor and Carmen Ortiz's house, bathroom +32255, the Ville, Tamara Taylor and Carmen Ortiz's house, garden +32265, the Ville, Moreno family's house, Tom and Jane Moreno's bedroom +32275, the Ville, Moreno family's house, empty bedroom +32206, the Ville, Moreno family's house, common room +32216, the Ville, Moreno family's house, kitchen +32226, the Ville, Moreno family's house, bathroom +32236, the Ville, Moreno family's house, garden +32246, the Ville, Lin family's house, Mei and John Lin's bedroom +32256, the Ville, Lin family's house, Eddy Lin's bedroom +32266, the Ville, Lin family's house, common room +32276, the Ville, Lin family's house, kitchen +32207, the Ville, Lin family's house, bathroom +32217, the Ville, Lin family's house, garden \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/game_object_blocks.csv b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/game_object_blocks.csv new file mode 100644 index 000000000..4afc74b7a --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/game_object_blocks.csv @@ -0,0 +1,46 @@ +32227, the Ville, , bed +32237, the Ville, , desk +32247, the Ville, , closet +32257, the Ville, , shelf +32267, the Ville, , easel +32277, the Ville, , bathroom sink +32208, the Ville, , shower +32218, the Ville, , toilet +32228, the Ville, , kitchen sink +32238, the Ville, , refrigerator +32248, the Ville, , toaster +32258, the Ville, , cooking area +32268, the Ville, , common room table +32278, the Ville, , common room sofa +32209, the Ville, , guitar +32219, the Ville, , microphone +32229, the Ville, , bar customer seating +32239, the Ville, , behind the bar counter +32249, the Ville, , behind the cafe counter +32259, the Ville, , cafe customer seating +32269, the Ville, , piano +32279, the Ville, , blackboard +32210, the Ville, , game console +32220, the Ville, , computer desk +32230, the Ville, , computer +32240, the Ville, , library sofa +32250, the Ville, , bookshelf +32260, the Ville, , library table +32270, the Ville, , classroom student seating +32280, the Ville, , classroom podium +32211, the Ville, , behind the pharmacy counter +32221, the Ville, , behind the grocery counter +32231, the Ville, , pharmacy store shelf +32241, the Ville, , grocery store shelf +32251, the Ville, , pharmacy store counter +32261, the Ville, , grocery store counter +32271, the Ville, , supply store product shelf +32281, the Ville, , behind the supply store counter +32212, the Ville, , supply store counter +32222, the Ville, , dorm garden +32232, the Ville, , house garden +32242, the Ville, , garden chair +32252, the Ville, , park garden +32262, the Ville, , harp +32272, the Ville, , lifting weight +32282, the Ville, , pool table \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/sector_blocks.csv b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/sector_blocks.csv new file mode 100644 index 000000000..ba09c4c35 --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/sector_blocks.csv @@ -0,0 +1,19 @@ +32135, the Ville, artist's co-living space +32145, the Ville, Arthur Burton's apartment +32155, the Ville, Ryan Park's apartment +32165, the Ville, Isabella Rodriguez's apartment +32175, the Ville, Giorgio Rossi's apartment +32185, the Ville, Carlos Gomez's apartment +32195, the Ville, The Rose and Crown Pub +32136, the Ville, Hobbs Cafe +32146, the Ville, Oak Hill College +32156, the Ville, Johnson Park +32166, the Ville, Harvey Oak Supply Store +32176, the Ville, The Willows Market and Pharmacy +32186, the Ville, Adam Smith's house +32196, the Ville, Yuriko Yamamoto's house +32137, the Ville, Moore family's house +32147, the Ville, Tamara Taylor and Carmen Ortiz's house +32157, the Ville, Moreno family's house +32167, the Ville, Lin family's house +32177, the Ville, Dorm for Oak Hill College \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/spawning_location_blocks.csv b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/spawning_location_blocks.csv new file mode 100644 index 000000000..564fc76b4 --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/spawning_location_blocks.csv @@ -0,0 +1,40 @@ +32285, the Ville, artist's co-living space, Latoya Williams's room, sp-A +32295, the Ville, artist's co-living space, Rajiv Patel's room, sp-A +32305, the Ville, artist's co-living space, Rajiv Patel's room, sp-B +32315, the Ville, artist's co-living space, Abigail Chen's room, sp-A +32286, the Ville, artist's co-living space, Francisco Lopez's room, sp-A +32296, the Ville, artist's co-living space, Hailey Johnson's room, sp-A +32306, the Ville, Arthur Burton's apartment, main room, sp-A +32316, the Ville, Arthur Burton's apartment, main room, sp-B +32287, the Ville, Ryan Park's apartment, main room, sp-A +32297, the Ville, Ryan Park's apartment, main room, sp-B +32307, the Ville, Isabella Rodriguez's apartment, main room, sp-A +32317, the Ville, Isabella Rodriguez's apartment, main room, sp-B +32288, the Ville, Giorgio Rossi's apartment, main room, sp-A +32298, the Ville, Giorgio Rossi's apartment, main room, sp-B +32308, the Ville, Carlos Gomez's apartment, main room, sp-A +32318, the Ville, Carlos Gomez's apartment, main room, sp-B +32289, the Ville, Adam Smith's house, main room, sp-A +32299, the Ville, Adam Smith's house, main room, sp-B +32309, the Ville, Yuriko Yamamoto's house, main room, sp-A +32319, the Ville, Yuriko Yamamoto's house, main room, sp-B +32290, the Ville, Moore family's house, main room, sp-A +32300, the Ville, Moore family's house, main room, sp-B +32310, the Ville, Tamara Taylor and Carmen Ortiz's house, Tamara Taylor's room, sp-A +32320, the Ville, Tamara Taylor and Carmen Ortiz's house, Tamara Taylor's room, sp-B +32291, the Ville, Tamara Taylor and Carmen Ortiz's house, Carmen Ortiz's room, sp-A +32301, the Ville, Tamara Taylor and Carmen Ortiz's house, Carmen Ortiz's room, sp-B +32311, the Ville, Moreno family's house, Tom and Jane Moreno's bedroom, sp-A +32321, the Ville, Moreno family's house, Tom and Jane Moreno's bedroom, sp-B +32292, the Ville, Moreno family's house, empty bedroom, sp-A +32302, the Ville, Moreno family's house, empty bedroom, sp-B +32312, the Ville, Lin family's house, Mei and John Lin's bedroom, sp-A +32322, the Ville, Lin family's house, Mei and John Lin's bedroom, sp-B +32293, the Ville, Lin family's house, Eddy Lin's bedroom, sp-A +32303, the Ville, Lin family's house, Eddy Lin's bedroom, sp-B +32313, the Ville, Dorm for Oak Hill College, Klaus Mueller's room, sp-A +32323, the Ville, Dorm for Oak Hill College, Klaus Mueller's room, sp-B +32294, the Ville, Dorm for Oak Hill College, Maria Lopez's room, sp-A +32304, the Ville, Dorm for Oak Hill College, Ayesha Khan's room, sp-A +32314, the Ville, Dorm for Oak Hill College, Ayesha Khan's room, sp-B +32324, the Ville, Dorm for Oak Hill College, Wolfgang Schulz's room, sp-A \ No newline at end of file diff --git a/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/world_blocks.csv b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/world_blocks.csv new file mode 100644 index 000000000..b04d990bb --- /dev/null +++ b/tests/data/environment/stanford_town/the_ville/matrix/special_blocks/world_blocks.csv @@ -0,0 +1 @@ +32134, the Ville \ No newline at end of file diff --git a/tests/data/incremental_dev_project/Gomoku.zip b/tests/data/incremental_dev_project/Gomoku.zip new file mode 100644 index 000000000..23649565a Binary files /dev/null and b/tests/data/incremental_dev_project/Gomoku.zip differ diff --git a/tests/data/incremental_dev_project/dice_simulator_new.zip b/tests/data/incremental_dev_project/dice_simulator_new.zip new file mode 100644 index 000000000..4752ab4c5 Binary files /dev/null and b/tests/data/incremental_dev_project/dice_simulator_new.zip differ diff --git a/tests/data/incremental_dev_project/mock.py b/tests/data/incremental_dev_project/mock.py new file mode 100644 index 000000000..f2eb71359 --- /dev/null +++ b/tests/data/incremental_dev_project/mock.py @@ -0,0 +1,466 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/17 +@Author : mannaandpoem +@File : mock.py +""" +NEW_REQUIREMENT_SAMPLE = """ +Adding graphical interface functionality to enhance the user experience in the number-guessing game. The existing number-guessing game currently relies on command-line input for numbers. The goal is to introduce a graphical interface to improve the game's usability and visual appeal +""" + +PRD_SAMPLE = """ +## Language + +en_us + +## Programming Language + +Python + +## Original Requirements + +Make a simple number guessing game + +## Product Goals + +- Ensure a user-friendly interface for the game +- Provide a challenging yet enjoyable game experience +- Design the game to be easily extendable for future features + +## User Stories + +- As a player, I want to guess numbers and receive feedback on whether my guess is too high or too low +- As a player, I want to be able to set the difficulty level by choosing the range of possible numbers +- As a player, I want to see my previous guesses to strategize my next guess +- As a player, I want to know how many attempts it took me to guess the number once I get it right + +## Competitive Analysis + +- Guess The Number Game A: Basic text interface, no difficulty levels +- Number Master B: Has difficulty levels, but cluttered interface +- Quick Guess C: Sleek design, but lacks performance tracking +- NumGuess D: Good performance tracking, but not mobile-friendly +- GuessIt E: Mobile-friendly, but too many ads +- Perfect Guess F: Offers hints, but the hints are not very helpful +- SmartGuesser G: Has a learning mode, but lacks a competitive edge + +## Competitive Quadrant Chart + +quadrantChart + title "User Engagement and Game Complexity" + x-axis "Low Complexity" --> "High Complexity" + y-axis "Low Engagement" --> "High Engagement" + quadrant-1 "Too Simple" + quadrant-2 "Niche Appeal" + quadrant-3 "Complex & Unengaging" + quadrant-4 "Sweet Spot" + "Guess The Number Game A": [0.2, 0.4] + "Number Master B": [0.5, 0.3] + "Quick Guess C": [0.6, 0.7] + "NumGuess D": [0.4, 0.6] + "GuessIt E": [0.7, 0.5] + "Perfect Guess F": [0.6, 0.4] + "SmartGuesser G": [0.8, 0.6] + "Our Target Product": [0.5, 0.8] + +## Requirement Analysis + +The game should be simple yet engaging, allowing players of different skill levels to enjoy it. It should provide immediate feedback and track the player's performance. The game should also be designed with a clean and intuitive interface, and it should be easy to add new features in the future. + +## Requirement Pool + +- ['P0', 'Implement the core game logic to randomly select a number and allow the user to guess it'] +- ['P0', 'Design a user interface that displays the game status and results clearly'] +- ['P1', 'Add difficulty levels by varying the range of possible numbers'] +- ['P1', 'Keep track of and display the number of attempts for each game session'] +- ['P2', "Store and show the history of the player's guesses during a game session"] + +## UI Design draft + +The UI will feature a clean and minimalist design with a number input field, submit button, and messages area to provide feedback. There will be options to select the difficulty level and a display showing the number of attempts and history of past guesses. + +## Anything UNCLEAR""" + +DESIGN_SAMPLE = """ +## Implementation approach + +We will create a Python-based number guessing game with a simple command-line interface. For the user interface, we will use the built-in 'input' and 'print' functions for interaction. The random library will be used for generating random numbers. We will structure the code to be modular and easily extendable, separating the game logic from the user interface. + +## File list + +- main.py +- game.py +- ui.py + +## Data structures and interfaces + + +classDiagram + class Game { + -int secret_number + -int min_range + -int max_range + -list attempts + +__init__(difficulty: str) + +start_game() + +check_guess(guess: int) str + +get_attempts() int + +get_history() list + } + class UI { + +start() + +display_message(message: str) + +get_user_input(prompt: str) str + +show_attempts(attempts: int) + +show_history(history: list) + +select_difficulty() str + } + class Main { + +main() + } + Main --> UI + UI --> Game + + +## Program call flow + + +sequenceDiagram + participant M as Main + participant UI as UI + participant G as Game + M->>UI: start() + UI->>UI: select_difficulty() + UI-->>G: __init__(difficulty) + G->>G: start_game() + loop Game Loop + UI->>UI: get_user_input("Enter your guess:") + UI-->>G: check_guess(guess) + G->>UI: display_message(feedback) + G->>UI: show_attempts(attempts) + G->>UI: show_history(history) + end + G->>UI: display_message("Correct! Game over.") + UI->>M: main() # Game session ends + + +## Anything UNCLEAR + +The requirement analysis suggests the need for a clean and intuitive interface. Since we are using a command-line interface, we need to ensure that the text-based UI is as user-friendly as possible. Further clarification on whether a graphical user interface (GUI) is expected in the future would be helpful for planning the extendability of the game.""" + +TASKS_SAMPLE = """ +## Required Python packages + +- random==2.2.1 + +## Required Other language third-party packages + +- No third-party dependencies required + +## Logic Analysis + +- ['game.py', 'Contains Game class with methods __init__, start_game, check_guess, get_attempts, get_history and uses random library for generating secret_number'] +- ['ui.py', 'Contains UI class with methods start, display_message, get_user_input, show_attempts, show_history, select_difficulty and interacts with Game class'] +- ['main.py', 'Contains Main class with method main that initializes UI class and starts the game loop'] + +## Task list + +- game.py +- ui.py +- main.py + +## Full API spec + + + +## Shared Knowledge + +`game.py` contains the core game logic and is used by `ui.py` to interact with the user. `main.py` serves as the entry point to start the game. + +## Anything UNCLEAR + +The requirement analysis suggests the need for a clean and intuitive interface. Since we are using a command-line interface, we need to ensure that the text-based UI is as user-friendly as possible. Further clarification on whether a graphical user interface (GUI) is expected in the future would be helpful for planning the extendability of the game.""" + +OLD_CODE_SAMPLE = """ +--- game.py +```## game.py + +import random + +class Game: + def __init__(self, difficulty: str = 'medium'): + self.min_range, self.max_range = self._set_difficulty(difficulty) + self.secret_number = random.randint(self.min_range, self.max_range) + self.attempts = [] + + def _set_difficulty(self, difficulty: str): + difficulties = { + 'easy': (1, 10), + 'medium': (1, 100), + 'hard': (1, 1000) + } + return difficulties.get(difficulty, (1, 100)) + + def start_game(self): + self.secret_number = random.randint(self.min_range, self.max_range) + self.attempts = [] + + def check_guess(self, guess: int) -> str: + self.attempts.append(guess) + if guess < self.secret_number: + return "It's higher." + elif guess > self.secret_number: + return "It's lower." + else: + return "Correct! Game over." + + def get_attempts(self) -> int: + return len(self.attempts) + + def get_history(self) -> list: + return self.attempts``` + +--- ui.py +```## ui.py + +from game import Game + +class UI: + def start(self): + difficulty = self.select_difficulty() + game = Game(difficulty) + game.start_game() + self.display_welcome_message(game) + + feedback = "" + while feedback != "Correct! Game over.": + guess = self.get_user_input("Enter your guess: ") + if self.is_valid_guess(guess): + feedback = game.check_guess(int(guess)) + self.display_message(feedback) + self.show_attempts(game.get_attempts()) + self.show_history(game.get_history()) + else: + self.display_message("Please enter a valid number.") + + def display_welcome_message(self, game): + print("Welcome to the Number Guessing Game!") + print(f"Guess the number between {game.min_range} and {game.max_range}.") + + def is_valid_guess(self, guess): + return guess.isdigit() + + def display_message(self, message: str): + print(message) + + def get_user_input(self, prompt: str) -> str: + return input(prompt) + + def show_attempts(self, attempts: int): + print(f"Number of attempts: {attempts}") + + def show_history(self, history: list): + print("Guess history:") + for guess in history: + print(guess) + + def select_difficulty(self) -> str: + while True: + difficulty = input("Select difficulty (easy, medium, hard): ").lower() + if difficulty in ['easy', 'medium', 'hard']: + return difficulty + else: + self.display_message("Invalid difficulty. Please choose 'easy', 'medium', or 'hard'.")``` + +--- main.py +```## main.py + +from ui import UI + +class Main: + def main(self): + user_interface = UI() + user_interface.start() + +if __name__ == "__main__": + main_instance = Main() + main_instance.main()``` +""" + +REFINED_PRD_JSON = { + "Language": "en_us", + "Programming Language": "Python", + "Refined Requirements": "Adding graphical interface functionality to enhance the user experience in the number-guessing game.", + "Project Name": "number_guessing_game", + "Refined Product Goals": [ + "Ensure a user-friendly interface for the game with the new graphical interface", + "Provide a challenging yet enjoyable game experience with visual enhancements", + "Design the game to be easily extendable for future features, including graphical elements", + ], + "Refined User Stories": [ + "As a player, I want to interact with a graphical interface to guess numbers and receive visual feedback on my guesses", + "As a player, I want to easily select the difficulty level through the graphical interface", + "As a player, I want to visually track my previous guesses and the number of attempts in the graphical interface", + "As a player, I want to be congratulated with a visually appealing message when I guess the number correctly", + ], + "Competitive Analysis": [ + "Guess The Number Game A: Basic text interface, no difficulty levels", + "Number Master B: Has difficulty levels, but cluttered interface", + "Quick Guess C: Sleek design, but lacks performance tracking", + "NumGuess D: Good performance tracking, but not mobile-friendly", + "GuessIt E: Mobile-friendly, but too many ads", + "Perfect Guess F: Offers hints, but the hints are not very helpful", + "SmartGuesser G: Has a learning mode, but lacks a competitive edge", + "Graphical Guess H: Graphical interface, but poor user experience due to complex design", + ], + "Competitive Quadrant Chart": 'quadrantChart\n title "User Engagement and Game Complexity with Graphical Interface"\n x-axis "Low Complexity" --> "High Complexity"\n y-axis "Low Engagement" --> "High Engagement"\n quadrant-1 "Too Simple"\n quadrant-2 "Niche Appeal"\n quadrant-3 "Complex & Unengaging"\n quadrant-4 "Sweet Spot"\n "Guess The Number Game A": [0.2, 0.4]\n "Number Master B": [0.5, 0.3]\n "Quick Guess C": [0.6, 0.7]\n "NumGuess D": [0.4, 0.6]\n "GuessIt E": [0.7, 0.5]\n "Perfect Guess F": [0.6, 0.4]\n "SmartGuesser G": [0.8, 0.6]\n "Graphical Guess H": [0.7, 0.3]\n "Our Target Product": [0.5, 0.9]', + "Refined Requirement Analysis": [ + "The game should maintain its simplicity while integrating a graphical interface for enhanced engagement.", + "Immediate visual feedback is crucial for user satisfaction in the graphical interface.", + "The interface must be intuitive, allowing for easy navigation and selection of game options.", + "The graphical design should be clean and not detract from the game's core guessing mechanic.", + ], + "Refined Requirement Pool": [ + ["P0", "Implement a graphical user interface (GUI) to replace the command-line interaction"], + [ + "P0", + "Design a user interface that displays the game status, results, and feedback clearly with graphical elements", + ], + ["P1", "Incorporate interactive elements for selecting difficulty levels"], + ["P1", "Visualize the history of the player's guesses and the number of attempts within the game session"], + ["P2", "Create animations for correct or incorrect guesses to enhance user feedback"], + ["P2", "Ensure the GUI is responsive and compatible with various screen sizes"], + ["P2", "Store and show the history of the player's guesses during a game session"], + ], + "UI Design draft": "The UI will feature a modern and minimalist design with a graphical number input field, a submit button with animations, and a dedicated area for visual feedback. It will include interactive elements to select the difficulty level and a visual display for the number of attempts and history of past guesses.", + "Anything UNCLEAR": "", +} + +REFINED_DESIGN_JSON = { + "Refined Implementation Approach": "To accommodate the new graphical user interface (GUI) requirements, we will leverage the Tkinter library, which is included with Python and supports the creation of a user-friendly GUI. The game logic will remain in Python, with Tkinter handling the rendering of the interface. We will ensure that the GUI is responsive and provides immediate visual feedback. The main game loop will be event-driven, responding to user inputs such as button clicks and difficulty selection.", + "Refined File list": ["main.py", "game.py", "ui.py", "gui.py"], + "Refined Data structures and interfaces": "\nclassDiagram\n class Game {\n -int secret_number\n -int min_range\n -int max_range\n -list attempts\n +__init__(difficulty: str)\n +start_game()\n +check_guess(guess: int) str\n +get_attempts() int\n +get_history() list\n }\n class UI {\n +start()\n +display_message(message: str)\n +get_user_input(prompt: str) str\n +show_attempts(attempts: int)\n +show_history(history: list)\n +select_difficulty() str\n }\n class GUI {\n +__init__()\n +setup_window()\n +bind_events()\n +update_feedback(message: str)\n +update_attempts(attempts: int)\n +update_history(history: list)\n +show_difficulty_selector()\n +animate_guess_result(correct: bool)\n }\n class Main {\n +main()\n }\n Main --> UI\n UI --> Game\n UI --> GUI\n GUI --> Game\n", + "Refined Program call flow": '\nsequenceDiagram\n participant M as Main\n participant UI as UI\n participant G as Game\n participant GU as GUI\n M->>UI: start()\n UI->>GU: setup_window()\n GU->>GU: bind_events()\n GU->>UI: select_difficulty()\n UI-->>G: __init__(difficulty)\n G->>G: start_game()\n loop Game Loop\n GU->>GU: show_difficulty_selector()\n GU->>UI: get_user_input("Enter your guess:")\n UI-->>G: check_guess(guess)\n G->>GU: update_feedback(feedback)\n G->>GU: update_attempts(attempts)\n G->>GU: update_history(history)\n GU->>GU: animate_guess_result(correct)\n end\n G->>GU: update_feedback("Correct! Game over.")\n GU->>M: main() # Game session ends\n', + "Anything UNCLEAR": "", +} + +REFINED_TASKS_JSON = { + "Required Python packages": ["random==2.2.1", "Tkinter==8.6"], + "Required Other language third-party packages": ["No third-party dependencies required"], + "Refined Logic Analysis": [ + [ + "game.py", + "Contains Game class with methods __init__, start_game, check_guess, get_attempts, get_history and uses random library for generating secret_number", + ], + [ + "ui.py", + "Contains UI class with methods start, display_message, get_user_input, show_attempts, show_history, select_difficulty and interacts with Game class", + ], + [ + "gui.py", + "Contains GUI class with methods __init__, setup_window, bind_events, update_feedback, update_attempts, update_history, show_difficulty_selector, animate_guess_result and interacts with Game class for GUI rendering", + ], + [ + "main.py", + "Contains Main class with method main that initializes UI class and starts the event-driven game loop", + ], + ], + "Refined Task list": ["game.py", "ui.py", "gui.py", "main.py"], + "Full API spec": "", + "Refined Shared Knowledge": "`game.py` contains the core game logic and is used by `ui.py` to interact with the user. `main.py` serves as the entry point to start the game. `gui.py` is introduced to handle the graphical user interface using Tkinter, which will interact with both `game.py` and `ui.py` for a responsive and user-friendly experience.", + "Anything UNCLEAR": "", +} + +CODE_PLAN_AND_CHANGE_SAMPLE = { + "Code Plan And Change": '\n1. Plan for gui.py: Develop the GUI using Tkinter to replace the command-line interface. Start by setting up the main window and event handling. Then, add widgets for displaying the game status, results, and feedback. Implement interactive elements for difficulty selection and visualize the guess history. Finally, create animations for guess feedback and ensure responsiveness across different screen sizes.\n```python\nclass GUI:\n- pass\n+ def __init__(self):\n+ self.setup_window()\n+\n+ def setup_window(self):\n+ # Initialize the main window using Tkinter\n+ pass\n+\n+ def bind_events(self):\n+ # Bind button clicks and other events\n+ pass\n+\n+ def update_feedback(self, message: str):\n+ # Update the feedback label with the given message\n+ pass\n+\n+ def update_attempts(self, attempts: int):\n+ # Update the attempts label with the number of attempts\n+ pass\n+\n+ def update_history(self, history: list):\n+ # Update the history view with the list of past guesses\n+ pass\n+\n+ def show_difficulty_selector(self):\n+ # Show buttons or a dropdown for difficulty selection\n+ pass\n+\n+ def animate_guess_result(self, correct: bool):\n+ # Trigger an animation for correct or incorrect guesses\n+ pass\n```\n\n2. Plan for main.py: Modify the main.py to initialize the GUI and start the event-driven game loop. Ensure that the GUI is the primary interface for user interaction.\n```python\nclass Main:\n def main(self):\n- user_interface = UI()\n- user_interface.start()\n+ graphical_user_interface = GUI()\n+ graphical_user_interface.setup_window()\n+ graphical_user_interface.bind_events()\n+ # Start the Tkinter main loop\n+ pass\n\n if __name__ == "__main__":\n main_instance = Main()\n main_instance.main()\n```\n\n3. Plan for ui.py: Refactor ui.py to work with the new GUI class. Remove command-line interactions and delegate display and input tasks to the GUI.\n```python\nclass UI:\n- def display_message(self, message: str):\n- print(message)\n+\n+ def display_message(self, message: str):\n+ # This method will now pass the message to the GUI to display\n+ pass\n\n- def get_user_input(self, prompt: str) -> str:\n- return input(prompt)\n+\n+ def get_user_input(self, prompt: str) -> str:\n+ # This method will now trigger the GUI to get user input\n+ pass\n\n- def show_attempts(self, attempts: int):\n- print(f"Number of attempts: {attempts}")\n+\n+ def show_attempts(self, attempts: int):\n+ # This method will now update the GUI with the number of attempts\n+ pass\n\n- def show_history(self, history: list):\n- print("Guess history:")\n- for guess in history:\n- print(guess)\n+\n+ def show_history(self, history: list):\n+ # This method will now update the GUI with the guess history\n+ pass\n```\n\n4. Plan for game.py: Ensure game.py remains mostly unchanged as it contains the core game logic. However, make minor adjustments if necessary to integrate with the new GUI.\n```python\nclass Game:\n # No changes required for now\n```\n' +} + +REFINED_CODE_INPUT_SAMPLE = """ +-----Now, game.py to be rewritten +```## game.py + +import random + +class Game: + def __init__(self, difficulty: str = 'medium'): + self.min_range, self.max_range = self._set_difficulty(difficulty) + self.secret_number = random.randint(self.min_range, self.max_range) + self.attempts = [] + + def _set_difficulty(self, difficulty: str): + difficulties = { + 'easy': (1, 10), + 'medium': (1, 100), + 'hard': (1, 1000) + } + return difficulties.get(difficulty, (1, 100)) + + def start_game(self): + self.secret_number = random.randint(self.min_range, self.max_range) + self.attempts = [] + + def check_guess(self, guess: int) -> str: + self.attempts.append(guess) + if guess < self.secret_number: + return "It's higher." + elif guess > self.secret_number: + return "It's lower." + else: + return "Correct! Game over." + + def get_attempts(self) -> int: + return len(self.attempts) + + def get_history(self) -> list: + return self.attempts``` +""" + +REFINED_CODE_SAMPLE = """ +## game.py + +import random + +class Game: + def __init__(self, difficulty: str = 'medium'): + # Set the difficulty level with default value 'medium' + self.min_range, self.max_range = self._set_difficulty(difficulty) + # Initialize the secret number based on the difficulty + self.secret_number = random.randint(self.min_range, self.max_range) + # Initialize the list to keep track of attempts + self.attempts = [] + + def _set_difficulty(self, difficulty: str): + # Define the range of numbers for each difficulty level + difficulties = { + 'easy': (1, 10), + 'medium': (1, 100), + 'hard': (1, 1000) + } + # Return the corresponding range for the selected difficulty, default to 'medium' if not found + return difficulties.get(difficulty, (1, 100)) + + def start_game(self): + # Reset the secret number and attempts list for a new game + self.secret_number = random.randint(self.min_range, self.max_range) + self.attempts.clear() + + def check_guess(self, guess: int) -> str: + # Add the guess to the attempts list + self.attempts.append(guess) + # Provide feedback based on the guess + if guess < self.secret_number: + return "It's higher." + elif guess > self.secret_number: + return "It's lower." + else: + return "Correct! Game over." + + def get_attempts(self) -> int: + # Return the number of attempts made + return len(self.attempts) + + def get_history(self) -> list: + # Return the list of attempts made + return self.attempts +""" diff --git a/tests/data/incremental_dev_project/number_guessing_game.zip b/tests/data/incremental_dev_project/number_guessing_game.zip new file mode 100644 index 000000000..7bbe07713 Binary files /dev/null and b/tests/data/incremental_dev_project/number_guessing_game.zip differ diff --git a/tests/data/incremental_dev_project/pygame_2048.zip b/tests/data/incremental_dev_project/pygame_2048.zip new file mode 100644 index 000000000..93e9cf0fe Binary files /dev/null and b/tests/data/incremental_dev_project/pygame_2048.zip differ diff --git a/tests/data/incremental_dev_project/readme.md b/tests/data/incremental_dev_project/readme.md new file mode 100644 index 000000000..231589028 --- /dev/null +++ b/tests/data/incremental_dev_project/readme.md @@ -0,0 +1,3 @@ +# Code archive + +This folder contains a compressed package for the test_incremental_dev.py file, which is used to demonstrate the process of incremental development. diff --git a/tests/data/incremental_dev_project/simple_add_calculator.zip b/tests/data/incremental_dev_project/simple_add_calculator.zip new file mode 100644 index 000000000..e6e73f8f9 Binary files /dev/null and b/tests/data/incremental_dev_project/simple_add_calculator.zip differ diff --git a/tests/data/incremental_dev_project/snake_game.zip b/tests/data/incremental_dev_project/snake_game.zip new file mode 100644 index 000000000..997203a18 Binary files /dev/null and b/tests/data/incremental_dev_project/snake_game.zip differ diff --git a/tests/data/incremental_dev_project/word_cloud.zip b/tests/data/incremental_dev_project/word_cloud.zip new file mode 100644 index 000000000..d8747d14d Binary files /dev/null and b/tests/data/incremental_dev_project/word_cloud.zip differ diff --git a/tests/data/rsp_cache.json b/tests/data/rsp_cache.json index df5300feb..75fc9ceb2 100644 --- a/tests/data/rsp_cache.json +++ b/tests/data/rsp_cache.json @@ -197,5 +197,203 @@ "You are a python code to Mermaid Sequence Diagram translator in function detail#SYSTEM_MSG_END#```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport asyncio\nimport shutil\nfrom pathlib import Path\n\nimport typer\n\nfrom metagpt.config2 import config\nfrom metagpt.const import CONFIG_ROOT, METAGPT_ROOT\n\napp = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False)\n\n\ndef generate_repo(\n idea,\n investment,\n n_round,\n code_review,\n run_tests,\n implement,\n project_name,\n inc,\n project_path,\n reqa_file,\n max_auto_summarize_code,\n recover_path,\n):\n \"\"\"Run the startup logic. Can be called from CLI or other Python scripts.\"\"\"\n from metagpt.roles import (\n Architect,\n Engineer,\n ProductManager,\n ProjectManager,\n QaEngineer,\n )\n from metagpt.team import Team\n\n config.update_via_cli(project_path, project_name, inc, reqa_file, max_auto_summarize_code)\n\n if not recover_path:\n company = Team()\n company.hire(\n [\n ProductManager(),\n Architect(),\n ProjectManager(),\n ]\n )\n\n if implement or code_review:\n company.hire([Engineer(n_borg=5, use_code_review=code_review)])\n\n if run_tests:\n company.hire([QaEngineer()])\n else:\n stg_path = Path(recover_path)\n if not stg_path.exists() or not str(stg_path).endswith(\"team\"):\n raise FileNotFoundError(f\"{recover_path} not exists or not endswith `team`\")\n\n company = Team.deserialize(stg_path=stg_path)\n idea = company.idea\n\n company.invest(investment)\n company.run_project(idea)\n asyncio.run(company.run(n_round=n_round))\n\n\n@app.command(\"\", help=\"Start a new project.\")\ndef startup(\n idea: str = typer.Argument(None, help=\"Your innovative idea, such as 'Create a 2048 game.'\"),\n investment: float = typer.Option(default=3.0, help=\"Dollar amount to invest in the AI company.\"),\n n_round: int = typer.Option(default=5, help=\"Number of rounds for the simulation.\"),\n code_review: bool = typer.Option(default=True, help=\"Whether to use code review.\"),\n run_tests: bool = typer.Option(default=False, help=\"Whether to enable QA for adding & running tests.\"),\n implement: bool = typer.Option(default=True, help=\"Enable or disable code implementation.\"),\n project_name: str = typer.Option(default=\"\", help=\"Unique project name, such as 'game_2048'.\"),\n inc: bool = typer.Option(default=False, help=\"Incremental mode. Use it to coop with existing repo.\"),\n project_path: str = typer.Option(\n default=\"\",\n help=\"Specify the directory path of the old version project to fulfill the incremental requirements.\",\n ),\n reqa_file: str = typer.Option(\n default=\"\", help=\"Specify the source file name for rewriting the quality assurance code.\"\n ),\n max_auto_summarize_code: int = typer.Option(\n default=0,\n help=\"The maximum number of times the 'SummarizeCode' action is automatically invoked, with -1 indicating \"\n \"unlimited. This parameter is used for debugging the workflow.\",\n ),\n recover_path: str = typer.Option(default=None, help=\"recover the project from existing serialized storage\"),\n init_config: bool = typer.Option(default=False, help=\"Initialize the configuration file for MetaGPT.\"),\n):\n \"\"\"Run a startup. Be a boss.\"\"\"\n if init_config:\n copy_config_to()\n return\n\n if idea is None:\n typer.echo(\"Missing argument 'IDEA'. Run 'metagpt --help' for more information.\")\n raise typer.Exit()\n\n return generate_repo(\n idea,\n investment,\n n_round,\n code_review,\n run_tests,\n implement,\n project_name,\n inc,\n project_path,\n reqa_file,\n max_auto_summarize_code,\n recover_path,\n )\n\n\ndef copy_config_to(config_path=METAGPT_ROOT / \"config\" / \"config2.yaml\"):\n \"\"\"Initialize the configuration file for MetaGPT.\"\"\"\n target_path = CONFIG_ROOT / \"config2.yaml\"\n\n # 创建目标目录(如果不存在)\n target_path.parent.mkdir(parents=True, exist_ok=True)\n\n # 如果目标文件已经存在,则重命名为 .bak\n if target_path.exists():\n backup_path = target_path.with_suffix(\".bak\")\n target_path.rename(backup_path)\n print(f\"Existing configuration file backed up at {backup_path}\")\n\n # 复制文件\n shutil.copy(str(config_path), target_path)\n print(f\"Configuration file initialized at {target_path}\")\n\n\nif __name__ == \"__main__\":\n app()\n\n```\n\n---\nTranslate the code above into Mermaid Sequence Diagram.": "```mermaid\nsequenceDiagram\n participant app\n participant generate_repo\n participant copy_config_to\n participant Team\n participant ProductManager\n participant Architect\n participant ProjectManager\n participant Engineer\n participant QaEngineer\n\n app -> generate_repo: startup()\n generate_repo -> config: update_via_cli()\n generate_repo -> Team: hire()\n Team -> ProductManager: hire()\n Team -> Architect: hire()\n Team -> ProjectManager: hire()\n generate_repo -> Engineer: hire()\n generate_repo -> QaEngineer: hire()\n generate_repo -> Team: invest()\n generate_repo -> Team: run_project()\n generate_repo -> Team: run()\n\n app -> copy_config_to: copy_config_to()\n copy_config_to -> config: update_via_cli()\n```", "You are a python code to Mermaid Sequence Diagram translator in function detail#SYSTEM_MSG_END#```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n@Time : 2023/12/14 11:40\n@Author : alexanderwu\n@File : write_prd_an.py\n\"\"\"\nfrom typing import List\n\nfrom metagpt.actions.action_node import ActionNode\n\nLANGUAGE = ActionNode(\n key=\"Language\",\n expected_type=str,\n instruction=\"Provide the language used in the project, typically matching the user's requirement language.\",\n example=\"en_us\",\n)\n\nPROGRAMMING_LANGUAGE = ActionNode(\n key=\"Programming Language\",\n expected_type=str,\n instruction=\"Python/JavaScript or other mainstream programming language.\",\n example=\"Python\",\n)\n\nORIGINAL_REQUIREMENTS = ActionNode(\n key=\"Original Requirements\",\n expected_type=str,\n instruction=\"Place the original user's requirements here.\",\n example=\"Create a 2048 game\",\n)\n\nPROJECT_NAME = ActionNode(\n key=\"Project Name\",\n expected_type=str,\n instruction='According to the content of \"Original Requirements,\" name the project using snake case style , '\n \"like 'game_2048' or 'simple_crm.\",\n example=\"game_2048\",\n)\n\nPRODUCT_GOALS = ActionNode(\n key=\"Product Goals\",\n expected_type=List[str],\n instruction=\"Provide up to three clear, orthogonal product goals.\",\n example=[\"Create an engaging user experience\", \"Improve accessibility, be responsive\", \"More beautiful UI\"],\n)\n\nUSER_STORIES = ActionNode(\n key=\"User Stories\",\n expected_type=List[str],\n instruction=\"Provide up to 3 to 5 scenario-based user stories.\",\n example=[\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\",\n ],\n)\n\nCOMPETITIVE_ANALYSIS = ActionNode(\n key=\"Competitive Analysis\",\n expected_type=List[str],\n instruction=\"Provide 5 to 7 competitive products.\",\n example=[\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\",\n ],\n)\n\nCOMPETITIVE_QUADRANT_CHART = ActionNode(\n key=\"Competitive Quadrant Chart\",\n expected_type=str,\n instruction=\"Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\",\n example=\"\"\"quadrantChart\n title \"Reach and engagement of campaigns\"\n x-axis \"Low Reach\" --> \"High Reach\"\n y-axis \"Low Engagement\" --> \"High Engagement\"\n quadrant-1 \"We should expand\"\n quadrant-2 \"Need to promote\"\n quadrant-3 \"Re-evaluate\"\n quadrant-4 \"May be improved\"\n \"Campaign A\": [0.3, 0.6]\n \"Campaign B\": [0.45, 0.23]\n \"Campaign C\": [0.57, 0.69]\n \"Campaign D\": [0.78, 0.34]\n \"Campaign E\": [0.40, 0.34]\n \"Campaign F\": [0.35, 0.78]\n \"Our Target Product\": [0.5, 0.6]\"\"\",\n)\n\nREQUIREMENT_ANALYSIS = ActionNode(\n key=\"Requirement Analysis\",\n expected_type=str,\n instruction=\"Provide a detailed analysis of the requirements.\",\n example=\"\",\n)\n\nREQUIREMENT_POOL = ActionNode(\n key=\"Requirement Pool\",\n expected_type=List[List[str]],\n instruction=\"List down the top-5 requirements with their priority (P0, P1, P2).\",\n example=[[\"P0\", \"The main code ...\"], [\"P0\", \"The game algorithm ...\"]],\n)\n\nUI_DESIGN_DRAFT = ActionNode(\n key=\"UI Design draft\",\n expected_type=str,\n instruction=\"Provide a simple description of UI elements, functions, style, and layout.\",\n example=\"Basic function description with a simple style and layout.\",\n)\n\nANYTHING_UNCLEAR = ActionNode(\n key=\"Anything UNCLEAR\",\n expected_type=str,\n instruction=\"Mention any aspects of the project that are unclear and try to clarify them.\",\n example=\"\",\n)\n\nISSUE_TYPE = ActionNode(\n key=\"issue_type\",\n expected_type=str,\n instruction=\"Answer BUG/REQUIREMENT. If it is a bugfix, answer BUG, otherwise answer Requirement\",\n example=\"BUG\",\n)\n\nIS_RELATIVE = ActionNode(\n key=\"is_relative\",\n expected_type=str,\n instruction=\"Answer YES/NO. If the requirement is related to the old PRD, answer YES, otherwise NO\",\n example=\"YES\",\n)\n\nREASON = ActionNode(\n key=\"reason\", expected_type=str, instruction=\"Explain the reasoning process from question to answer\", example=\"...\"\n)\n\n\nNODES = [\n LANGUAGE,\n PROGRAMMING_LANGUAGE,\n ORIGINAL_REQUIREMENTS,\n PROJECT_NAME,\n PRODUCT_GOALS,\n USER_STORIES,\n COMPETITIVE_ANALYSIS,\n COMPETITIVE_QUADRANT_CHART,\n REQUIREMENT_ANALYSIS,\n REQUIREMENT_POOL,\n UI_DESIGN_DRAFT,\n ANYTHING_UNCLEAR,\n]\n\nWRITE_PRD_NODE = ActionNode.from_children(\"WritePRD\", NODES)\nWP_ISSUE_TYPE_NODE = ActionNode.from_children(\"WP_ISSUE_TYPE\", [ISSUE_TYPE, REASON])\nWP_IS_RELATIVE_NODE = ActionNode.from_children(\"WP_IS_RELATIVE\", [IS_RELATIVE, REASON])\n\n```\n\n---\nTranslate the code above into Mermaid Sequence Diagram.": "```mermaid\nclassDef actionNode fill:#f9f,stroke:#333,stroke-width:2px;\nclassDef actionNodeTitle fill:#f9f,stroke:#333,stroke-width:2px,font-weight:bold;\nclassDef actionNodeExample fill:#f9f,stroke:#333,stroke-width:2px,font-style:italic;\n\nclass ActionNodeTitle actionNodeTitle\nclass ActionNodeExample actionNodeExample\n\nActionNodeTitle:::Language --> \"Language\"\nActionNodeExample:::Language --> \"Provide the language used in the project, typically matching the user's requirement language.\\nExample: en_us\"\n\nActionNodeTitle:::ProgrammingLanguage --> \"Programming Language\"\nActionNodeExample:::ProgrammingLanguage --> \"Python/JavaScript or other mainstream programming language.\\nExample: Python\"\n\nActionNodeTitle:::OriginalRequirements --> \"Original Requirements\"\nActionNodeExample:::OriginalRequirements --> \"Place the original user's requirements here.\\nExample: Create a 2048 game\"\n\nActionNodeTitle:::ProjectName --> \"Project Name\"\nActionNodeExample:::ProjectName --> 'According to the content of \"Original Requirements,\" name the project using snake case style , like \\'game_2048\\' or \\'simple_crm.\\nExample: game_2048'\n\nActionNodeTitle:::ProductGoals --> \"Product Goals\"\nActionNodeExample:::ProductGoals --> \"Provide up to three clear, orthogonal product goals.\\nExample:\\n- Create an engaging user experience\\n- Improve accessibility, be responsive\\n- More beautiful UI\"\n\nActionNodeTitle:::UserStories --> \"User Stories\"\nActionNodeExample:::UserStories --> \"Provide up to 3 to 5 scenario-based user stories.\\nExample:\\n- As a player, I want to be able to choose difficulty levels\\n- As a player, I want to see my score after each game\\n- As a player, I want to get restart button when I lose\\n- As a player, I want to see beautiful UI that make me feel good\\n- As a player, I want to play game via mobile phone\"\n\nActionNodeTitle:::CompetitiveAnalysis --> \"Competitive Analysis\"\nActionNodeExample:::CompetitiveAnalysis --> \"Provide 5 to 7 competitive products.\\nExample:\\n- 2048 Game A: Simple interface, lacks responsive features\\n- play2048.co: Beautiful and responsive UI with my best score shown\\n- 2048game.com: Responsive UI with my best score shown, but many ads\"\n\nActionNodeTitle:::CompetitiveQuadrantChart --> \"Competitive Quadrant Chart\"\nActionNodeExample:::CompetitiveQuadrantChart --> \"Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\\nExample:\\nquadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\"\n\nActionNodeTitle:::RequirementAnalysis --> \"Requirement Analysis\"\nActionNodeExample:::RequirementAnalysis --> \"Provide a detailed analysis of the requirements.\\nExample: \"\n\nActionNodeTitle:::RequirementPool --> \"Requirement Pool\"\nActionNodeExample:::RequirementPool --> \"List down the top-5 requirements with their priority (P0, P1, P2).\\nExample:\\n- P0: The main code ...\\n- P0: The game algorithm ...\"\n\nActionNodeTitle:::UIDesignDraft --> \"UI Design draft\"\nActionNodeExample:::UIDesignDraft --> \"Provide a simple description of UI elements, functions, style, and layout.\\nExample: Basic function description with a simple style and layout.\"\n\nActionNodeTitle:::AnythingUNCLEAR --> \"Anything UNCLEAR\"\nActionNodeExample:::AnythingUNCLEAR --> \"Mention any aspects of the project that are unclear and try to clarify them.\\nExample: \"\n\nActionNodeTitle:::issue_type --> \"issue_type\"\nActionNodeExample:::issue_type --> \"Answer BUG/REQUIREMENT. If it is a bugfix, answer BUG, otherwise answer Requirement\\nExample: BUG\"\n\nActionNodeTitle:::is_relative --> \"is_relative\"\nActionNodeExample:::is_relative --> \"Answer YES/NO. If the requirement is related to the old PRD, answer YES, otherwise NO\\nExample: YES\"\n\nActionNodeTitle:::reason --> \"reason\"\nActionNodeExample:::reason --> \"Explain the reasoning process from question to answer\\nExample: ...\"\n\nActionNodeTitle:::WritePRD --> \"WritePRD\"\nActionNodeExample:::WritePRD --> \"Language\\nProgramming Language\\nOriginal Requirements\\nProject Name\\nProduct Goals\\nUser Stories\\nCompetitive Analysis\\nCompetitive Quadrant Chart\\nRequirement Analysis\\nRequirement Pool\\nUI Design draft\\nAnything UNCLEAR\"\n\nActionNodeTitle:::WP_ISSUE_TYPE --> \"WP_ISSUE_TYPE\"\nActionNodeExample:::WP_ISSUE_TYPE --> \"issue_type\\nreason\"\n\nActionNodeTitle:::WP_IS_RELATIVE --> \"WP_IS_RELATIVE\"\nActionNodeExample:::WP_IS_RELATIVE --> \"is_relative\\nreason\"\n```", "\n## context\n\n### Project Name\n20240112110833\n\n### Original Requirements\n['开发一个基于大语言模型与私有知识库的搜索引擎,希望可以基于大语言模型进行搜索总结']\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"zh_cn\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"开发一个基于大语言模型与私有知识库的搜索引擎,希望可以基于大语言模型进行搜索总结\",\n \"Product Goals\": [],\n \"User Stories\": [],\n \"Competitive Analysis\": [],\n \"Competitive Quadrant Chart\": \"\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [],\n \"UI Design draft\": \"\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", - "\n## context\n\n### Project Name\n20240112110833\n\n### Original Requirements\n['']\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"\",\n \"Product Goals\": [],\n \"User Stories\": [],\n \"Competitive Analysis\": [],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]" + "\n## context\n\n### Project Name\n20240112110833\n\n### Original Requirements\n['']\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"\",\n \"Product Goals\": [],\n \"User Stories\": [],\n \"Competitive Analysis\": [],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", + "\n## context\n\n### Project Name\n\n\n### Original Requirements\n需要一个基于LLM做总结的搜索引擎\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Project Name\": \"game_2048\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style , like 'game_2048' or 'simple_crm.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"zh_cn\",\n \"Programming Language\": \"LLM\",\n \"Original Requirements\": \"需要一个基于LLM做总结的搜索引擎\",\n \"Project Name\": \"llm_summary_search_engine\",\n \"Product Goals\": [\n \"提供准确的搜索结果\",\n \"提高搜索引擎的效率\",\n \"优化用户体验\"\n ],\n \"User Stories\": [\n \"作为用户,我希望能够快速找到我需要的信息\",\n \"作为用户,我希望搜索结果准确无误\",\n \"作为用户,我希望搜索引擎能够智能推荐相关内容\"\n ],\n \"Competitive Analysis\": [\n \"搜索引擎A: 提供准确的搜索结果,但界面简陋\",\n \"搜索引擎B: 界面美观,但搜索结果不够准确\",\n \"搜索引擎C: 搜索结果准确,但速度较慢\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"搜索引擎比较\\\"\\n x-axis \\\"低准确性\\\" --> \\\"高准确性\\\"\\n y-axis \\\"低速度\\\" --> \\\"高速度\\\"\\n quadrant-1 \\\"需要改进\\\"\\n quadrant-2 \\\"值得推广\\\"\\n quadrant-3 \\\"重新评估\\\"\\n quadrant-4 \\\"需要扩展\\\"\\n \\\"搜索引擎A\\\": [0.8, 0.3]\\n \\\"搜索引擎B\\\": [0.5, 0.7]\\n \\\"搜索引擎C\\\": [0.9, 0.2]\\n \\\"我们的搜索引擎\\\": [0.7, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"实现LLM技术进行文本摘要\"\n ],\n [\n \"P0\",\n \"构建搜索算法,提高搜索效率\"\n ],\n [\n \"P1\",\n \"设计智能推荐系统\"\n ]\n ],\n \"UI Design draft\": \"简洁的搜索框,清晰的搜索结果页面,智能推荐模块\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", + "\n## context\n\n### Project Name\n\n\n### Original Requirements\nMake a cli snake game\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Project Name\": \"game_2048\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style , like 'game_2048' or 'simple_crm.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Make a cli snake game\",\n \"Project Name\": \"cli_snake_game\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility and responsiveness\",\n \"Enhance the game with additional features\"\n ],\n \"User Stories\": [\n \"As a player, I want to control the snake using arrow keys\",\n \"As a player, I want to see my score during the game\",\n \"As a player, I want to have the option to restart the game\",\n \"As a player, I want to see a visually appealing UI\",\n \"As a player, I want to play the game on different platforms\"\n ],\n \"Competitive Analysis\": [\n \"Snake Game A: Simple interface, lacks responsive features\",\n \"SnakeGame.co: Beautiful and responsive UI with high scores displayed\",\n \"SnakeGame.com: Responsive UI with high scores shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of snake games\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Snake Game A\\\": [0.3, 0.6]\\n \\\"SnakeGame.co\\\": [0.45, 0.23]\\n \\\"SnakeGame.com\\\": [0.57, 0.69]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code for controlling the snake and game logic\"\n ],\n [\n \"P1\",\n \"Implementing the scoring system and UI\"\n ],\n [\n \"P2\",\n \"Adding platform compatibility and restart functionality\"\n ]\n ],\n \"UI Design draft\": \"The game will have a simple and intuitive UI with clear controls and a visually appealing design.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", + "\n## context\n{\"Language\":\"en_us\",\"Programming Language\":\"Python\",\"Original Requirements\":\"Make a cli snake game\",\"Project Name\":\"cli_snake_game\",\"Product Goals\":[\"Create an engaging user experience\",\"Improve accessibility and responsiveness\",\"Enhance the game with additional features\"],\"User Stories\":[\"As a player, I want to control the snake using arrow keys\",\"As a player, I want to see my score during the game\",\"As a player, I want to have the option to restart the game\",\"As a player, I want to see a visually appealing UI\",\"As a player, I want to play the game on different platforms\"],\"Competitive Analysis\":[\"Snake Game A: Simple interface, lacks responsive features\",\"SnakeGame.co: Beautiful and responsive UI with high scores displayed\",\"SnakeGame.com: Responsive UI with high scores shown, but many ads\"],\"Competitive Quadrant Chart\":\"quadrantChart\\n title \\\"Reach and engagement of snake games\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Snake Game A\\\": [0.3, 0.6]\\n \\\"SnakeGame.co\\\": [0.45, 0.23]\\n \\\"SnakeGame.com\\\": [0.57, 0.69]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\"Requirement Analysis\":\"\",\"Requirement Pool\":[[\"P0\",\"The main code for controlling the snake and game logic\"],[\"P1\",\"Implementing the scoring system and UI\"],[\"P2\",\"Adding platform compatibility and restart functionality\"]],\"UI Design draft\":\"The game will have a simple and intuitive UI with clear controls and a visually appealing design.\",\"Anything UNCLEAR\":\"\"}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Implementation approach\": \"We will ...\",\n \"File list\": [\n \"main.py\",\n \"game.py\"\n ],\n \"Data structures and interfaces\": \"\\nclassDiagram\\n class Main {\\n -SearchEngine search_engine\\n +main() str\\n }\\n class SearchEngine {\\n -Index index\\n -Ranking ranking\\n -Summary summary\\n +search(query: str) str\\n }\\n class Index {\\n -KnowledgeBase knowledge_base\\n +create_index(data: dict)\\n +query_index(query: str) list\\n }\\n class Ranking {\\n +rank_results(results: list) list\\n }\\n class Summary {\\n +summarize_results(results: list) str\\n }\\n class KnowledgeBase {\\n +update(data: dict)\\n +fetch_data(query: str) dict\\n }\\n Main --> SearchEngine\\n SearchEngine --> Index\\n SearchEngine --> Ranking\\n SearchEngine --> Summary\\n Index --> KnowledgeBase\\n\",\n \"Program call flow\": \"\\nsequenceDiagram\\n participant M as Main\\n participant SE as SearchEngine\\n participant I as Index\\n participant R as Ranking\\n participant S as Summary\\n participant KB as KnowledgeBase\\n M->>SE: search(query)\\n SE->>I: query_index(query)\\n I->>KB: fetch_data(query)\\n KB-->>I: return data\\n I-->>SE: return results\\n SE->>R: rank_results(results)\\n R-->>SE: return ranked_results\\n SE->>S: summarize_results(ranked_results)\\n S-->>SE: return summary\\n SE-->>M: return summary\\n\",\n \"Anything UNCLEAR\": \"Clarification needed on third-party API integration, ...\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Implementation approach: # Analyze the difficult points of the requirements, select the appropriate open-source framework\n- File list: typing.List[str] # Only need relative paths. ALWAYS write a main.py or app.py here\n- Data structures and interfaces: # Use mermaid classDiagram code syntax, including classes, method(__init__ etc.) and functions with type annotations, CLEARLY MARK the RELATIONSHIPS between classes, and comply with PEP8 standards. The data structures SHOULD BE VERY DETAILED and the API should be comprehensive with a complete design.\n- Program call flow: # Use sequenceDiagram code syntax, COMPLETE and VERY DETAILED, using CLASSES AND API DEFINED ABOVE accurately, covering the CRUD AND INIT of each object, SYNTAX MUST BE CORRECT.\n- Anything UNCLEAR: # Mention unclear project aspects, then try to clarify it.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Implementation approach\": \"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\n \"File list\": [\n \"main.py\",\n \"game.py\"\n ],\n \"Data structures and interfaces\": \"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\n \"Program call flow\": \"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\n \"Anything UNCLEAR\": \"Clarification needed on the scoring system and how the game over condition will be handled.\"\n}\n[/CONTENT]", + "\n## context\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Required Python packages\": [\n \"flask==1.1.2\",\n \"bcrypt==3.2.0\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and ... functions\"\n ],\n [\n \"main.py\",\n \"Contains main function, from game import Game\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"openapi: 3.0.0 ...\",\n \"Shared Knowledge\": \"'game.py' contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"Clarification needed on how to start and initialize third-party libraries.\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Required Python packages: typing.List[str] # Provide required Python packages in requirements.txt format.\n- Required Other language third-party packages: typing.List[str] # List down the required packages for languages other than Python.\n- Logic Analysis: typing.List[typing.List[str]] # Provide a list of files with the classes/methods/functions to be implemented, including dependency analysis and imports.\n- Task list: typing.List[str] # Break down the tasks into a list of filenames, prioritized by dependency order.\n- Full API spec: # Describe all APIs using OpenAPI 3.0 spec that may be used by both frontend and backend. If front-end and back-end communication is not required, leave it blank.\n- Shared Knowledge: # Detail any shared knowledge, like common utility functions or configuration variables.\n- Anything UNCLEAR: # Mention any unclear aspects in the project management context and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Required Python packages\": [\n \"pygame==2.0.1\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and ... functions\"\n ],\n [\n \"main.py\",\n \"Contains main function, from game import Game\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"\",\n \"Shared Knowledge\": \"'game.py' contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"Clarification needed on the scoring system and how the game over condition will be handled.\"\n}\n[/CONTENT]", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: game.py\n```python\n## game.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: game.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: game.py\n```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n```\n\n\n\n# Format example 1\n## Code Review: game.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: game.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: game.py\n1. No, the code is not implemented as per the requirements. The Game class should have a method to handle game over condition and update the scoring system.\n2. No, the code logic is not completely correct. The Snake class is missing the implementation of the move, change_direction, and grow methods.\n3. Yes, the existing code follows the \"Data structures and interfaces\" provided.\n4. No, the move, change_direction, and grow methods of the Snake class are not implemented. To achieve this, the methods need to be implemented with the appropriate logic for moving the snake, changing its direction, and growing its body.\n5. Yes, all necessary pre-dependencies have been imported.\n6. No, there are no methods from other files being reused in the current code.\n\n## Actions\n1. Implement a method in the Game class to handle the game over condition and update the scoring system.\n2. Implement the move, change_direction, and grow methods in the Snake class.\n3. Reuse methods from other files if necessary.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: game.py\n```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n```\n\n## Code Review: game.py\n1. No, the code is not implemented as per the requirements. The Game class should have a method to handle game over condition and update the scoring system.\n2. No, the code logic is not completely correct. The Snake class is missing the implementation of the move, change_direction, and grow methods.\n3. Yes, the existing code follows the \"Data structures and interfaces\" provided.\n4. No, the move, change_direction, and grow methods of the Snake class are not implemented. To achieve this, the methods need to be implemented with the appropriate logic for moving the snake, changing its direction, and growing its body.\n5. Yes, all necessary pre-dependencies have been imported.\n6. No, there are no methods from other files being reused in the current code.\n\n## Actions\n1. Implement a method in the Game class to handle the game over condition and update the scoring system.\n2. Implement the move, change_direction, and grow methods in the Snake class.\n3. Reuse methods from other files if necessary.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite game.py with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## game.py\n...\n```\n": "```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: game.py\n```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n```\n\n\n\n# Format example 1\n## Code Review: game.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: game.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: game.py\n1. No, the code is not implemented as per the requirements. The `handle_game_over` and `update_score` methods are missing, and the `handle_input` method needs to be implemented to handle user input.\n2. The code logic is not completely correct. The `Snake` class `__init__` method does not initialize the `body` attribute with the `start_pos` and `length` parameters.\n3. The existing code follows the \"Data structures and interfaces\" for the `Game` and `Snake` classes.\n4. The `handle_game_over` and `update_score` methods are not implemented. The `handle_input` method needs to be implemented to handle user input.\n5. The necessary pre-dependencies have been imported.\n6. The methods from other files are not being reused in the existing code.\n\n## Actions\n1. Implement the `handle_game_over` method to handle the game over condition.\n2. Implement the `update_score` method to update the scoring system.\n3. Implement the `handle_input` method to handle user input and update the game state accordingly.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: game.py\n```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = []\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n```\n\n## Code Review: game.py\n1. No, the code is not implemented as per the requirements. The `handle_game_over` and `update_score` methods are missing, and the `handle_input` method needs to be implemented to handle user input.\n2. The code logic is not completely correct. The `Snake` class `__init__` method does not initialize the `body` attribute with the `start_pos` and `length` parameters.\n3. The existing code follows the \"Data structures and interfaces\" for the `Game` and `Snake` classes.\n4. The `handle_game_over` and `update_score` methods are not implemented. The `handle_input` method needs to be implemented to handle user input.\n5. The necessary pre-dependencies have been imported.\n6. The methods from other files are not being reused in the existing code.\n\n## Actions\n1. Implement the `handle_game_over` method to handle the game over condition.\n2. Implement the `update_score` method to update the scoring system.\n3. Implement the `handle_input` method to handle user input and update the game state accordingly.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite game.py with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## game.py\n...\n```\n": "```Code\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Legacy Code\n```Code\n----- game.py\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: main.py\n```python\n## main.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: main.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.handle_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n----- game.py\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.handle_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n\n\n# Format example 1\n## Code Review: main.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: main.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: main.py\n1. The code is implemented as per the requirements, but the game over condition and scoring system need to be clarified and implemented.\n \n## Actions\n1. Clarify and implement the game over condition and scoring system in the `Game` class.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n----- game.py\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.handle_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n## Code Review: main.py\n1. The code is implemented as per the requirements, but the game over condition and scoring system need to be clarified and implemented.\n \n## Actions\n1. Clarify and implement the game over condition and scoring system in the `Game` class.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite main.py with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## main.py\n...\n```\n": "```python\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.is_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n----- game.py\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.is_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n\n\n# Format example 1\n## Code Review: main.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: main.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: main.py\n1. The code is implemented as per the requirements, but the `is_game_over` method is missing. To achieve this, we need to implement the `is_game_over` method in the Game class to check for the game over condition.\n2. The code logic is correct, but the missing `is_game_over` method needs to be added to handle the game over condition.\n3. The existing code follows the \"Data structures and interfaces\" as per the provided class diagram.\n4. All functions are implemented, but the `is_game_over` method is missing in the Game class. To achieve this, we need to implement the `is_game_over` method to handle the game over condition.\n5. All necessary pre-dependencies have been imported.\n6. Methods from other files are not being reused in the current code.\n\n## Actions\n1. Implement the `is_game_over` method in the Game class to handle the game over condition.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Tasks\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"'game.py' contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n----- game.py\n## game.py\n\nclass Game:\n def __init__(self):\n self.snake = []\n self.food = (0, 0)\n self.score = 0\n\n def update(self):\n # Update game state\n pass\n\n def draw(self):\n # Draw game\n pass\n\n def handle_input(self):\n # Handle user input\n pass\n\n def restart(self):\n # Restart game\n pass\n\n def handle_game_over(self):\n # Handle game over condition\n pass\n\n def update_score(self):\n # Update scoring system\n pass\n\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos] # Initialize body with start_pos\n self.direction = direction\n\n def move(self):\n # Move snake\n pass\n\n def change_direction(self, new_direction: str):\n # Change snake direction\n pass\n\n def grow(self):\n # Grow snake\n pass\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.is_game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n## Code Review: main.py\n1. The code is implemented as per the requirements, but the `is_game_over` method is missing. To achieve this, we need to implement the `is_game_over` method in the Game class to check for the game over condition.\n2. The code logic is correct, but the missing `is_game_over` method needs to be added to handle the game over condition.\n3. The existing code follows the \"Data structures and interfaces\" as per the provided class diagram.\n4. All functions are implemented, but the `is_game_over` method is missing in the Game class. To achieve this, we need to implement the `is_game_over` method to handle the game over condition.\n5. All necessary pre-dependencies have been imported.\n6. Methods from other files are not being reused in the current code.\n\n## Actions\n1. Implement the `is_game_over` method in the Game class to handle the game over condition.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite main.py with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## main.py\n...\n```\n": "```python\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.handle_input()\n game.update()\n game.draw()\n\n if game.is_game_over(): # Check for game over condition\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n```", + "You are a python code to Mermaid Sequence Diagram translator in function detail#SYSTEM_MSG_END#```python\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n@Author : alexanderwu\n@File : write_review.py\n\"\"\"\nimport asyncio\nfrom typing import List, Literal\n\nfrom metagpt.actions import Action\nfrom metagpt.actions.action_node import ActionNode\n\nREVIEW = ActionNode(\n key=\"Review\",\n expected_type=List[str],\n instruction=\"Act as an experienced reviewer and critically assess the given output. Provide specific and\"\n \" constructive feedback, highlighting areas for improvement and suggesting changes.\",\n example=[\n \"The logic in the function `calculate_total` seems flawed. Shouldn't it consider the discount rate as well?\",\n \"The TODO function is not implemented yet? Should we implement it before commit?\",\n ],\n)\n\nREVIEW_RESULT = ActionNode(\n key=\"ReviewResult\",\n expected_type=Literal[\"LGTM\", \"LBTM\"],\n instruction=\"LGTM/LBTM. If the code is fully implemented, \" \"give a LGTM, otherwise provide a LBTM.\",\n example=\"LBTM\",\n)\n\nNEXT_STEPS = ActionNode(\n key=\"NextSteps\",\n expected_type=str,\n instruction=\"Based on the code review outcome, suggest actionable steps. This can include code changes, \"\n \"refactoring suggestions, or any follow-up tasks.\",\n example=\"\"\"1. Refactor the `process_data` method to improve readability and efficiency.\n2. Cover edge cases in the `validate_user` function.\n3. Implement a the TODO in the `calculate_total` function.\n4. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n\"\"\",\n)\n\nWRITE_DRAFT = ActionNode(\n key=\"WriteDraft\",\n expected_type=str,\n instruction=\"Could you write draft code for move function in order to implement it?\",\n example=\"Draft: ...\",\n)\n\n\nWRITE_FUNCTION = ActionNode(\n key=\"WriteFunction\",\n expected_type=str,\n instruction=\"write code for the function not implemented.\",\n example=\"\"\"\n```Code\n...\n```\n\"\"\",\n)\n\n\nREWRITE_CODE = ActionNode(\n key=\"RewriteCode\",\n expected_type=str,\n instruction=\"\"\"rewrite code based on the Review and Actions\"\"\",\n example=\"\"\"\n```python\n## example.py\ndef calculate_total(price, quantity):\n total = price * quantity\n```\n\"\"\",\n)\n\n\nCODE_REVIEW_CONTEXT = \"\"\"\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\n\n# Context\n## System Design\n{\"Implementation approach\": \"我们将使用HTML、CSS和JavaScript来实现这个单机的响应式2048游戏。为了确保游戏性能流畅和响应式设计,我们会选择使用Vue.js框架,因为它易于上手且适合构建交互式界面。我们还将使用localStorage来记录玩家的最高分。\", \"File list\": [\"index.html\", \"styles.css\", \"main.js\", \"game.js\", \"storage.js\"], \"Data structures and interfaces\": \"classDiagram\\\n class Game {\\\n -board Array\\\n -score Number\\\n -bestScore Number\\\n +constructor()\\\n +startGame()\\\n +move(direction: String)\\\n +getBoard() Array\\\n +getScore() Number\\\n +getBestScore() Number\\\n +setBestScore(score: Number)\\\n }\\\n class Storage {\\\n +getBestScore() Number\\\n +setBestScore(score: Number)\\\n }\\\n class Main {\\\n +init()\\\n +bindEvents()\\\n }\\\n Game --> Storage : uses\\\n Main --> Game : uses\", \"Program call flow\": \"sequenceDiagram\\\n participant M as Main\\\n participant G as Game\\\n participant S as Storage\\\n M->>G: init()\\\n G->>S: getBestScore()\\\n S-->>G: return bestScore\\\n M->>G: bindEvents()\\\n M->>G: startGame()\\\n loop Game Loop\\\n M->>G: move(direction)\\\n G->>S: setBestScore(score)\\\n S-->>G: return\\\n end\", \"Anything UNCLEAR\": \"目前项目要求明确,没有不清楚的地方。\"}\n\n## Tasks\n{\"Required Python packages\": [\"无需Python包\"], \"Required Other language third-party packages\": [\"vue.js\"], \"Logic Analysis\": [[\"index.html\", \"作为游戏的入口文件和主要的HTML结构\"], [\"styles.css\", \"包含所有的CSS样式,确保游戏界面美观\"], [\"main.js\", \"包含Main类,负责初始化游戏和绑定事件\"], [\"game.js\", \"包含Game类,负责游戏逻辑,如开始游戏、移动方块等\"], [\"storage.js\", \"包含Storage类,用于获取和设置玩家的最高分\"]], \"Task list\": [\"index.html\", \"styles.css\", \"storage.js\", \"game.js\", \"main.js\"], \"Full API spec\": \"\", \"Shared Knowledge\": \"\\'game.js\\' 包含游戏逻辑相关的函数,被 \\'main.js\\' 调用。\", \"Anything UNCLEAR\": \"目前项目要求明确,没有不清楚的地方。\"}\n\n## Code Files\n----- index.html\n\n\n\n \n \n 2048游戏\n \n \n\n\n
\n

2048

\n
\n
\n
分数
\n
{{ score }}
\n
\n
\n
最高分
\n
{{ bestScore }}
\n
\n
\n
\n
\n
\n {{ cell !== 0 ? cell : \\'\\' }}\n
\n
\n
\n \n
\n\n \n \n \n \n\n\n\n----- styles.css\n/* styles.css */\nbody, html {\n margin: 0;\n padding: 0;\n font-family: \\'Arial\\', sans-serif;\n}\n\n#app {\n text-align: center;\n font-size: 18px;\n color: #776e65;\n}\n\nh1 {\n color: #776e65;\n font-size: 72px;\n font-weight: bold;\n margin: 20px 0;\n}\n\n.scores-container {\n display: flex;\n justify-content: center;\n margin-bottom: 20px;\n}\n\n.score-container, .best-container {\n background: #bbada0;\n padding: 10px;\n border-radius: 5px;\n margin: 0 10px;\n min-width: 100px;\n text-align: center;\n}\n\n.score-header, .best-header {\n color: #eee4da;\n font-size: 18px;\n margin-bottom: 5px;\n}\n\n.game-container {\n max-width: 500px;\n margin: 0 auto 20px;\n background: #bbada0;\n padding: 15px;\n border-radius: 10px;\n position: relative;\n}\n\n.grid-row {\n display: flex;\n}\n\n.grid-cell {\n background: #cdc1b4;\n width: 100px;\n height: 100px;\n margin: 5px;\n display: flex;\n justify-content: center;\n align-items: center;\n font-size: 35px;\n font-weight: bold;\n color: #776e65;\n border-radius: 3px;\n}\n\n/* Dynamic classes for different number cells */\n.number-cell-2 {\n background: #eee4da;\n}\n\n.number-cell-4 {\n background: #ede0c8;\n}\n\n.number-cell-8 {\n background: #f2b179;\n color: #f9f6f2;\n}\n\n.number-cell-16 {\n background: #f59563;\n color: #f9f6f2;\n}\n\n.number-cell-32 {\n background: #f67c5f;\n color: #f9f6f2;\n}\n\n.number-cell-64 {\n background: #f65e3b;\n color: #f9f6f2;\n}\n\n.number-cell-128 {\n background: #edcf72;\n color: #f9f6f2;\n}\n\n.number-cell-256 {\n background: #edcc61;\n color: #f9f6f2;\n}\n\n.number-cell-512 {\n background: #edc850;\n color: #f9f6f2;\n}\n\n.number-cell-1024 {\n background: #edc53f;\n color: #f9f6f2;\n}\n\n.number-cell-2048 {\n background: #edc22e;\n color: #f9f6f2;\n}\n\n/* Larger numbers need smaller font sizes */\n.number-cell-1024, .number-cell-2048 {\n font-size: 30px;\n}\n\nbutton {\n background-color: #8f7a66;\n color: #f9f6f2;\n border: none;\n border-radius: 3px;\n padding: 10px 20px;\n font-size: 18px;\n cursor: pointer;\n outline: none;\n}\n\nbutton:hover {\n background-color: #9f8b76;\n}\n\n----- storage.js\n## storage.js\nclass Storage {\n // 获取最高分\n getBestScore() {\n // 尝试从localStorage中获取最高分,如果不存在则默认为0\n const bestScore = localStorage.getItem(\\'bestScore\\');\n return bestScore ? Number(bestScore) : 0;\n }\n\n // 设置最高分\n setBestScore(score) {\n // 将最高分设置到localStorage中\n localStorage.setItem(\\'bestScore\\', score.toString());\n }\n}\n\n\n\n## Code to be Reviewed: game.js\n```Code\n## game.js\nclass Game {\n constructor() {\n this.board = this.createEmptyBoard();\n this.score = 0;\n this.bestScore = 0;\n }\n\n createEmptyBoard() {\n const board = [];\n for (let i = 0; i < 4; i++) {\n board[i] = [0, 0, 0, 0];\n }\n return board;\n }\n\n startGame() {\n this.board = this.createEmptyBoard();\n this.score = 0;\n this.addRandomTile();\n this.addRandomTile();\n }\n\n addRandomTile() {\n let emptyCells = [];\n for (let r = 0; r < 4; r++) {\n for (let c = 0; c < 4; c++) {\n if (this.board[r][c] === 0) {\n emptyCells.push({ r, c });\n }\n }\n }\n if (emptyCells.length > 0) {\n let randomCell = emptyCells[Math.floor(Math.random() * emptyCells.length)];\n this.board[randomCell.r][randomCell.c] = Math.random() < 0.9 ? 2 : 4;\n }\n }\n\n move(direction) {\n // This function will handle the logic for moving tiles\n // in the specified direction and merging them\n // It will also update the score and add a new random tile if the move is successful\n // The actual implementation of this function is complex and would require\n // a significant amount of code to handle all the cases for moving and merging tiles\n // For the purposes of this example, we will not implement the full logic\n // Instead, we will just call addRandomTile to simulate a move\n this.addRandomTile();\n }\n\n getBoard() {\n return this.board;\n }\n\n getScore() {\n return this.score;\n }\n\n getBestScore() {\n return this.bestScore;\n }\n\n setBestScore(score) {\n this.bestScore = score;\n }\n}\n\n```\n\"\"\"\n\n\nCODE_REVIEW_SMALLEST_CONTEXT = \"\"\"\n## Code to be Reviewed: game.js\n```Code\n// game.js\nclass Game {\n constructor() {\n this.board = this.createEmptyBoard();\n this.score = 0;\n this.bestScore = 0;\n }\n\n createEmptyBoard() {\n const board = [];\n for (let i = 0; i < 4; i++) {\n board[i] = [0, 0, 0, 0];\n }\n return board;\n }\n\n startGame() {\n this.board = this.createEmptyBoard();\n this.score = 0;\n this.addRandomTile();\n this.addRandomTile();\n }\n\n addRandomTile() {\n let emptyCells = [];\n for (let r = 0; r < 4; r++) {\n for (let c = 0; c < 4; c++) {\n if (this.board[r][c] === 0) {\n emptyCells.push({ r, c });\n }\n }\n }\n if (emptyCells.length > 0) {\n let randomCell = emptyCells[Math.floor(Math.random() * emptyCells.length)];\n this.board[randomCell.r][randomCell.c] = Math.random() < 0.9 ? 2 : 4;\n }\n }\n\n move(direction) {\n // This function will handle the logic for moving tiles\n // in the specified direction and merging them\n // It will also update the score and add a new random tile if the move is successful\n // The actual implementation of this function is complex and would require\n // a significant amount of code to handle all the cases for moving and merging tiles\n // For the purposes of this example, we will not implement the full logic\n // Instead, we will just call addRandomTile to simulate a move\n this.addRandomTile();\n }\n\n getBoard() {\n return this.board;\n }\n\n getScore() {\n return this.score;\n }\n\n getBestScore() {\n return this.bestScore;\n }\n\n setBestScore(score) {\n this.bestScore = score;\n }\n}\n\n```\n\"\"\"\n\n\nCODE_REVIEW_SAMPLE = \"\"\"\n## Code Review: game.js\n1. The code partially implements the requirements. The `Game` class is missing the full implementation of the `move` method, which is crucial for the game\\'s functionality.\n2. The code logic is not completely correct. The `move` method is not implemented, which means the game cannot process player moves.\n3. The existing code follows the \"Data structures and interfaces\" in terms of class structure but lacks full method implementations.\n4. Not all functions are implemented. The `move` method is incomplete and does not handle the logic for moving and merging tiles.\n5. All necessary pre-dependencies seem to be imported since the code does not indicate the need for additional imports.\n6. The methods from other files (such as `Storage`) are not being used in the provided code snippet, but the class structure suggests that they will be used correctly.\n\n## Actions\n1. Implement the `move` method to handle tile movements and merging. This is a complex task that requires careful consideration of the game\\'s rules and logic. Here is a simplified version of how one might begin to implement the `move` method:\n ```javascript\n move(direction) {\n // Simplified logic for moving tiles up\n if (direction === \\'up\\') {\n for (let col = 0; col < 4; col++) {\n let tiles = this.board.map(row => row[col]).filter(val => val !== 0);\n let merged = [];\n for (let i = 0; i < tiles.length; i++) {\n if (tiles[i] === tiles[i + 1]) {\n tiles[i] *= 2;\n this.score += tiles[i];\n tiles[i + 1] = 0;\n merged.push(i);\n }\n }\n tiles = tiles.filter(val => val !== 0);\n while (tiles.length < 4) {\n tiles.push(0);\n }\n for (let row = 0; row < 4; row++) {\n this.board[row][col] = tiles[row];\n }\n }\n }\n // Additional logic needed for \\'down\\', \\'left\\', \\'right\\'\n // ...\n this.addRandomTile();\n }\n ```\n2. Integrate the `Storage` class methods to handle the best score. This means updating the `startGame` and `setBestScore` methods to use `Storage` for retrieving and setting the best score:\n ```javascript\n startGame() {\n this.board = this.createEmptyBoard();\n this.score = 0;\n this.bestScore = new Storage().getBestScore(); // Retrieve the best score from storage\n this.addRandomTile();\n this.addRandomTile();\n }\n\n setBestScore(score) {\n if (score > this.bestScore) {\n this.bestScore = score;\n new Storage().setBestScore(score); // Set the new best score in storage\n }\n }\n ```\n\n## Code Review Result\nLBTM\n\n```\n\"\"\"\n\n\nWRITE_CODE_NODE = ActionNode.from_children(\"WRITE_REVIEW_NODE\", [REVIEW, REVIEW_RESULT, NEXT_STEPS])\nWRITE_MOVE_NODE = ActionNode.from_children(\"WRITE_MOVE_NODE\", [WRITE_DRAFT, WRITE_FUNCTION])\n\n\nCR_FOR_MOVE_FUNCTION_BY_3 = \"\"\"\nThe move function implementation provided appears to be well-structured and follows a clear logic for moving and merging tiles in the specified direction. However, there are a few potential improvements that could be made to enhance the code:\n\n1. Encapsulation: The logic for moving and merging tiles could be encapsulated into smaller, reusable functions to improve readability and maintainability.\n\n2. Magic Numbers: There are some magic numbers (e.g., 4, 3) used in the loops that could be replaced with named constants for improved readability and easier maintenance.\n\n3. Comments: Adding comments to explain the logic and purpose of each section of the code can improve understanding for future developers who may need to work on or maintain the code.\n\n4. Error Handling: It's important to consider error handling for unexpected input or edge cases to ensure the function behaves as expected in all scenarios.\n\nOverall, the code could benefit from refactoring to improve readability, maintainability, and extensibility. If you would like, I can provide a refactored version of the move function that addresses these considerations.\n\"\"\"\n\n\nclass WriteCodeAN(Action):\n \"\"\"Write a code review for the context.\"\"\"\n\n async def run(self, context):\n self.llm.system_prompt = \"You are an outstanding engineer and can implement any code\"\n return await WRITE_MOVE_NODE.fill(context=context, llm=self.llm, schema=\"json\")\n\n\nasync def main():\n await WriteCodeAN().run(CODE_REVIEW_SMALLEST_CONTEXT)\n\n\nif __name__ == \"__main__\":\n asyncio.run(main())\n\n```\n\n---\nTranslate the code above into Mermaid Sequence Diagram.": "```mermaid\nsequenceDiagram\n participant User\n participant WriteCodeAN\n participant WRITE_MOVE_NODE\n\n User->>WriteCodeAN: Run\n WriteCodeAN->>WRITE_MOVE_NODE: Fill\n WRITE_MOVE_NODE->>WriteCodeAN: Filled\n WriteCodeAN->>User: Return Filled Context\n```", + "You are a function parser.#MSG_SEP#You can convert spoken words into function parameters.#SYSTEM_MSG_END#text_to_image function parameters description:\nparameter `text`: The text used for image conversion.\nparameter `size_type`: size type\n\n---\nExamples:\nIf want you to do `Draw a girl`, return `text_to_image(text=\"Draw a girl\", size_type=\"512x512\")` brief and clear.\nIf want you to do `Draw an apple`, return `text_to_image(text=\"Draw an apple\", size_type=\"512x512\")` brief and clear.\n\n---\n\nRefer to the `text_to_image` function description, and fill in the function parameters according to the example \"I want you to do xx\" in the Examples section.\nNow I want you to do `Draw an apple`, return function parameters in Examples format above, brief and clear.": "`text_to_image(text=\"Draw an apple\", size_type=\"512x512\")`", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\": \"We will use the Pygame library to create the game interface and handle user input. The game logic will be implemented using Python classes and data structures.\", \"File list\": [\"main.py\", \"game.py\"], \"Data structures and interfaces\": \"classDiagram\\n class Game {\\n -grid: List[List[int]]\\n -score: int\\n -game_over: bool\\n +__init__()\\n +reset_game()\\n +move(direction: str)\\n +is_game_over() bool\\n +get_empty_cells() List[Tuple[int, int]]\\n +add_new_tile()\\n +get_score() int\\n }\\n class UI {\\n -game: Game\\n +__init__(game: Game)\\n +draw_grid()\\n +draw_score()\\n +draw_game_over()\\n +handle_input()\\n }\\n Game --> UI\", \"Program call flow\": \"sequenceDiagram\\n participant M as Main\\n participant G as Game\\n participant U as UI\\n M->>G: reset_game()\\n M->>U: draw_grid()\\n M->>U: draw_score()\\n M->>U: handle_input()\\n U->>G: move(direction)\\n G->>G: add_new_tile()\\n G->>U: draw_grid()\\n G->>U: draw_score()\\n G->>U: draw_game_over()\\n G->>G: is_game_over()\\n G->>G: get_empty_cells()\\n G->>G: get_score()\", \"Anything UNCLEAR\": \"...\"}\n\n## Tasks\n{\"Required Python packages\": [\"pygame==2.0.1\"], \"Required Other language third-party packages\": [\"No third-party dependencies required\"], \"Logic Analysis\": [[\"game.py\", \"Contains Game class and related functions for game logic\"], [\"main.py\", \"Contains main function, initializes the game and UI\"]], \"Task list\": [\"game.py\", \"main.py\"], \"Full API spec\": \"\", \"Shared Knowledge\": \"The game logic will be implemented using Python classes and data structures. The Pygame library will be used to create the game interface and handle user input.\", \"Anything UNCLEAR\": \"...\"}\n\n## Legacy Code\n```Code\n----- main.py\nif __name__ == \"__main__\":\nmain()\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: game.py\n```python\n## game.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: game.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## game.py\n\nfrom typing import List, Tuple\nimport random\n\nclass Game:\n def __init__(self):\n self.grid: List[List[int]] = [[0 for _ in range(4)] for _ in range(4)]\n self.score: int = 0\n self.game_over: bool = False\n\n def reset_game(self):\n self.grid = [[0 for _ in range(4)] for _ in range(4)]\n self.score = 0\n self.game_over = False\n self.add_new_tile()\n self.add_new_tile()\n\n def move(self, direction: str):\n # Implement the logic to move the tiles in the specified direction\n pass\n\n def is_game_over(self) -> bool:\n # Implement the logic to check if the game is over\n pass\n\n def get_empty_cells(self) -> List[Tuple[int, int]]:\n # Implement the logic to get the coordinates of empty cells in the grid\n pass\n\n def add_new_tile(self):\n empty_cells = self.get_empty_cells()\n if empty_cells:\n row, col = random.choice(empty_cells)\n self.grid[row][col] = 2 if random.random() < 0.9 else 4\n\n def get_score(self) -> int:\n return self.score\n```", + "\n## context\n\n### Project Name\n\n\n### Original Requirements\n开发一个基于大语言模型与私有知识库的搜索引擎,希望可以基于大语言模型进行搜索总结\n\n### Search Information\n-\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Language\": \"en_us\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"Create a 2048 game\",\n \"Project Name\": \"game_2048\",\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ],\n \"User Stories\": [\n \"As a player, I want to be able to choose difficulty levels\",\n \"As a player, I want to see my score after each game\",\n \"As a player, I want to get restart button when I lose\",\n \"As a player, I want to see beautiful UI that make me feel good\",\n \"As a player, I want to play game via mobile phone\"\n ],\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"Reach and engagement of campaigns\\\"\\n x-axis \\\"Low Reach\\\" --> \\\"High Reach\\\"\\n y-axis \\\"Low Engagement\\\" --> \\\"High Engagement\\\"\\n quadrant-1 \\\"We should expand\\\"\\n quadrant-2 \\\"Need to promote\\\"\\n quadrant-3 \\\"Re-evaluate\\\"\\n quadrant-4 \\\"May be improved\\\"\\n \\\"Campaign A\\\": [0.3, 0.6]\\n \\\"Campaign B\\\": [0.45, 0.23]\\n \\\"Campaign C\\\": [0.57, 0.69]\\n \\\"Campaign D\\\": [0.78, 0.34]\\n \\\"Campaign E\\\": [0.40, 0.34]\\n \\\"Campaign F\\\": [0.35, 0.78]\\n \\\"Our Target Product\\\": [0.5, 0.6]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ],\n \"UI Design draft\": \"Basic function description with a simple style and layout.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Language: # Provide the language used in the project, typically matching the user's requirement language.\n- Programming Language: # Python/JavaScript or other mainstream programming language.\n- Original Requirements: # Place the original user's requirements here.\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style , like 'game_2048' or 'simple_crm.\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n- User Stories: typing.List[str] # Provide up to 3 to 5 scenario-based user stories.\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n- Competitive Quadrant Chart: # Use mermaid quadrantChart syntax. Distribute scores evenly between 0 and 1\n- Requirement Analysis: # Provide a detailed analysis of the requirements.\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n- UI Design draft: # Provide a simple description of UI elements, functions, style, and layout.\n- Anything UNCLEAR: # Mention any aspects of the project that are unclear and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Language\": \"zh_cn\",\n \"Programming Language\": \"Python\",\n \"Original Requirements\": \"开发一个基于大语言模型与私有知识库的搜索引擎,希望可以基于大语言模型进行搜索总结\",\n \"Project Name\": \"language_model_search\",\n \"Product Goals\": [\n \"提供基于大语言模型的高效搜索功能\",\n \"整合私有知识库,实现个性化搜索结果\",\n \"实现搜索结果的自动总结功能\"\n ],\n \"User Stories\": [\n \"作为用户,我希望能够快速找到我需要的信息\",\n \"作为用户,我希望搜索结果能够根据我的偏好进行个性化排序\",\n \"作为用户,我希望搜索结果能够自动总结,方便我快速了解内容\"\n ],\n \"Competitive Analysis\": [\n \"搜索引擎A: 提供基于大语言模型的搜索功能,但个性化程度较低\",\n \"知识库B: 整合私有知识库,但搜索速度较慢\",\n \"语言模型搜索C: 提供搜索结果自动总结功能,但搜索准确度有待提高\"\n ],\n \"Competitive Quadrant Chart\": \"quadrantChart\\n title \\\"搜索引擎竞争分析\\\"\\n x-axis \\\"低搜索准确度\\\" --> \\\"高搜索准确度\\\"\\n y-axis \\\"低个性化程度\\\" --> \\\"高个性化程度\\\"\\n quadrant-1 \\\"需改进\\\"\\n quadrant-2 \\\"需提升\\\"\\n quadrant-3 \\\"重新评估\\\"\\n quadrant-4 \\\"扩展发展\\\"\\n \\\"搜索引擎A\\\": [0.6, 0.3]\\n \\\"知识库B\\\": [0.4, 0.2]\\n \\\"语言模型搜索C\\\": [0.7, 0.5]\\n \\\"我们的目标产品\\\": [0.8, 0.7]\",\n \"Requirement Analysis\": \"\",\n \"Requirement Pool\": [\n [\n \"P0\",\n \"基于大语言模型的高效搜索功能\"\n ],\n [\n \"P1\",\n \"整合私有知识库,实现个性化搜索结果\"\n ],\n [\n \"P2\",\n \"实现搜索结果的自动总结功能\"\n ]\n ],\n \"UI Design draft\": \"搜索页面简洁明了,搜索结果清晰展示,提供个性化排序和自动总结功能。\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", + "\n## context\nNone\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Project Name\": \"game_2048\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"game_2048\"\n}\n[/CONTENT]", + "\n## context\nCompare the key's value of nodes_output and the corresponding requirements one by one. If a key's value that does not match the requirement is found, provide the comment content on how to modify it. No output is required for matching keys.\n\n### nodes_output\n{\"Project Name\": {\"value\": \"game snake\", \"requirement\": \"According to the content of \\\"Original Requirements,\\\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\"}}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"key1\": \"comment1\",\n \"key2\": \"comment2\",\n \"keyn\": \"commentn\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- key1: # the first key name of mismatch key\n- key2: # the second key name of mismatch key\n- keyn: # the last key name of mismatch key\n\n## constraint\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow format example's json format, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"Use snake case style with underline, like 'game_2048' or 'simple_crm.'\"\n}\n[/CONTENT]", + "\n## context\nCompare the key's value of nodes_output and the corresponding requirements one by one. If a key's value that does not match the requirement is found, provide the comment content on how to modify it. No output is required for matching keys.\n\n### nodes_output\n{\"Project Name\": {\"value\": \"game_2048\", \"requirement\": \"According to the content of \\\"Original Requirements,\\\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\"}}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"key1\": \"comment1\",\n \"key2\": \"comment2\",\n \"keyn\": \"commentn\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- key1: # the first key name of mismatch key\n- key2: # the second key name of mismatch key\n- keyn: # the last key name of mismatch key\n\n## constraint\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow format example's json format, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"According to the content of \\\"Original Requirements,\\\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\"\n}\n[/CONTENT]", + "\n## context\nchange the nodes_output key's value to meet its comment and no need to add extra comment.\n\n### nodes_output\n{\"Project Name\": {\"value\": \"game snake\", \"comment\": \"Use snake case style with underline, like 'game_2048' or 'simple_crm.'\"}}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Project Name\": \"game_2048\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\n\n\n## constraint\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow format example's json format, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"game_snake\"\n}\n[/CONTENT]", + "\n## context\nCompare the key's value of nodes_output and the corresponding requirements one by one. If a key's value that does not match the requirement is found, provide the comment content on how to modify it. No output is required for matching keys.\n\n### nodes_output\n{\"Project Name\": {\"value\": \"game_snake\", \"requirement\": \"According to the content of \\\"Original Requirements,\\\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\"}}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"key1\": \"comment1\",\n \"key2\": \"comment2\",\n \"keyn\": \"commentn\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- key1: # the first key name of mismatch key\n- key2: # the second key name of mismatch key\n- keyn: # the last key name of mismatch key\n\n## constraint\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow format example's json format, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"The project name should be in snake case style with underscores, such as 'game_2048' or 'simple_crm'.\"\n}\n[/CONTENT]", + "\n## context\nchange the nodes_output key's value to meet its comment and no need to add extra comment.\n\n### nodes_output\n{\"Project Name\": {\"value\": \"game_snake\", \"comment\": \"The project name should be in snake case style with underscores, such as 'game_2048' or 'simple_crm'.\"}}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Project Name\": \"game_2048\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Project Name: # According to the content of \"Original Requirements,\" name the project using snake case style with underline, like 'game_2048' or 'simple_crm.\n\n\n## constraint\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow format example's json format, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Project Name\": \"game_snake\"\n}\n[/CONTENT]", + "You are an action classifier#SYSTEM_MSG_END#If the text explicitly want you to Generate a voice file from the input text, text-to-speech, return `[SKILL]: text_to_speech` brief and clear. For instance: [SKILL]: text_to_speech\nIf the text explicitly want you to Create a drawing based on the text., return `[SKILL]: text_to_image` brief and clear. For instance: [SKILL]: text_to_image\nIf the text explicitly want you to Perform Google searches to provide real-time information., return `[SKILL]: web_search` brief and clear. For instance: [SKILL]: web_search\nOtherwise, return `[TALK]: {talk}` brief and clear. For instance: if {talk} is \"xxxx\" return [TALK]: xxxx\n\nNow what specific action is explicitly mentioned in the text: No, I do not have a poison apple. Do you have a poison apple?\n": "[TALK]: No, I do not have a poison apple. Do you have a poison apple?", + "You are an action classifier#SYSTEM_MSG_END#If the text explicitly want you to Generate a voice file from the input text, text-to-speech, return `[SKILL]: text_to_speech` brief and clear. For instance: [SKILL]: text_to_speech\nIf the text explicitly want you to Create a drawing based on the text., return `[SKILL]: text_to_image` brief and clear. For instance: [SKILL]: text_to_image\nIf the text explicitly want you to Perform Google searches to provide real-time information., return `[SKILL]: web_search` brief and clear. For instance: [SKILL]: web_search\nOtherwise, return `[TALK]: {talk}` brief and clear. For instance: if {talk} is \"xxxx\" return [TALK]: xxxx\n\nNow what specific action is explicitly mentioned in the text: Sure, I can draw you an apple. Draw me an apple.\n": "[SKILL]: text_to_image", + "You are a function parser.#MSG_SEP#You can convert spoken words into function parameters.#SYSTEM_MSG_END#text_to_image function parameters description:\nparameter `text`: The text used for image conversion.\nparameter `size_type`: size type\n\n---\nExamples:\nIf want you to do `Draw a girl`, return `text_to_image(text=\"Draw a girl\", size_type=\"512x512\")` brief and clear.\nIf want you to do `Draw an apple`, return `text_to_image(text=\"Draw an apple\", size_type=\"512x512\")` brief and clear.\n\n---\n\nRefer to the `text_to_image` function description, and fill in the function parameters according to the example \"I want you to do xx\" in the Examples section.\nNow I want you to do `Sure, I can draw you an apple. Draw me an apple.`, return function parameters in Examples format above, brief and clear.": "`text_to_image(text=\"Sure, I can draw you an apple. Draw me an apple.\", size_type=\"512x512\")`", + "You are an action classifier#SYSTEM_MSG_END#Otherwise, return `[TALK]: {talk}` brief and clear. For instance: if {talk} is \"xxxx\" return [TALK]: xxxx\n\nNow what specific action is explicitly mentioned in the text: Sure, I can draw you an apple. Draw me an apple.\n": "[DRAW]: draw an apple", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 4 5 6 10\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Here is the list of possible next steps for the given input in JSON format:\n\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"4 + 5 = 9 (left: 6 9 10)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"5 + 6 = 11 (left: 4 11 10)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"4 * 5 = 20 (left: 6 20 10)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"6 - 4 = 2 (left: 2 5 10)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"10 - 4 = 6 (left: 6 5 6)\"\n }\n]\n```", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n6 9 10\n": "6 * 9 - 10 = 54 - 10 = 44\n(9 - 6) * 10 = 3 * 10 = 30\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n6 20 10\n": "I'm sorry, but it seems that you haven't completed the calculation for the numbers 6, 20, and 10. If you'd like, I can help you with that.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n4 11 10\n": "4 11 10 cannot reach 24.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n2 5 10\n": "2 * 5 * 10 = 100\n(2 + 5) * 10 = 70\n2 5 10 are all too big\nimpossible", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n6 5 6\n": "I'm sorry, but it seems that you have not provided a complete expression for the numbers 6, 5, and 6 to reach 24. If you would like to try again, please provide a complete expression using the numbers 6, 5, and 6 to reach the target number of 24.", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 4 11 10\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Here are the possible next steps for the given input in the requested JSON format:\n\nInput: 4 11 10\nPossible next steps:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"4 + 11 = 15 (left: 10 15)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"11 - 4 = 7 (left: 7 10)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"4 * 11 = 44 (left: 10 44)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"11 / 4 = 2.75 (left: 2.75 10)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"11 + 4 = 15 (left: 10 15)\"\n }\n]\n```", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 6 9 10\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Certainly! Here are the possible next steps for the given input in JSON format:\n\nInput: 6 9 10\nPossible next steps:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"6 + 9 = 15 (left: 10 15)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"6 * 9 = 54 (left: 10 54)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"9 - 6 = 3 (left: 3 9 10)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"10 - 6 = 4 (left: 4 6 9)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"10 / 6 = 1 (left: 1 6 9)\"\n }\n]\n```", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 15\n": "I'm sorry, I cannot evaluate the numbers 10 and 15 without an operation. If you provide an operation, I can help you determine if they can reach 24.", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 6 20 10\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Certainly! Here is the list of nodes for the given input and possible next steps:\n\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"6 + 20 = 26 (left: 10 26)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"20 - 6 = 14 (left: 14 10)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"6 * 20 = 120 (left: 10 120)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"20 / 6 = 3.33 (left: 3.33 10)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"20 + 6 = 26 (left: 10 26)\"\n }\n]\n```", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n1 6 9\n": "1 6 9 cannot reach 24 using addition, subtraction, multiplication, or division. Therefore, it is impossible for these numbers to reach 24.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 54\n": "It seems like you didn't provide the complete set of numbers for the last calculation. Could you please provide the third number so that I can help you evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n7 10\n": "I'm sorry, but it seems like you didn't provide a complete set of numbers for the last question. Could you please provide the third number so that I can evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 44\n": "It seems like you didn't provide the second number for the last set. Could you please provide the second number so that I can evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 26\n": "It seems that you have not provided the complete set of numbers for the last question. Could you please provide the third number for the set?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 120\n": "It seems that you have not provided the complete set of numbers for the last question. Could you please provide the third number for the evaluation?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n3 9 10\n": "3 * 9 * 10 = 270\n(10 - 3) * 9 = 63\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n2.75 10\n": "It seems like you have a mix of numbers and operations. Could you please provide the operations for the last set of numbers (2.75 and 10) so that I can help you evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n4 6 9\n": "4 + 6 + 9 = 19\n(9 - 4) * 6 = 5 * 6 = 30\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n14 10\n": "I'm sorry, it seems like you didn't provide the complete set of numbers for the last expression. Could you please provide the third number so that I can evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n3.33 10\n": "It seems like you have provided a list of numbers and evaluated if they can reach 24 through addition, subtraction, multiplication, and division. Here are the evaluations:\n\n10 14 - Sure\n11 12 - Impossible\n4 4 10 - Sure\n4 9 11 - Sure\n5 7 8 - Likely\n5 6 6 - Likely\n10 10 11 - Impossible\n1 3 3 - Impossible\n\nIt looks like you have evaluated the combinations accurately. As for the last set of numbers \"3.33 10,\" it seems incomplete. If you provide the intended operation or additional numbers, I can help you evaluate it.", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 10 15\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Here's the JSON format for the possible next steps for the given input:\n\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"10 + 15 = 25 (left: 25)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"15 - 10 = 5 (left: 5)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"10 * 15 = 150 (left: 150)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"15 / 10 = 1.5 (left: 1.5)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"15 + 10 = 25 (left: 25)\"\n }\n]\n```", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 3 9 10\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Here are the possible next steps for the given input in JSON format:\n\nInput: 3 9 10\nPossible next steps:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"3 + 9 = 12 (left: 10 12)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"9 - 3 = 6 (left: 6 9 10)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"10 - 3 = 7 (left: 7 9 10)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"3 * 9 = 27 (left: 10 27)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"9 / 3 = 3 (left: 3 9 10)\"\n }\n]\n```", + "Here is an Example for 1 input and 8 possible thoughts:\nInput: 2 8 8 14\nPossible next steps:\n2 + 8 = 10 (left: 8 10 14)\n8 / 2 = 4 (left: 4 8 14)\n14 + 2 = 16 (left: 8 8 16)\n2 * 8 = 16 (left: 8 14 16)\n8 - 2 = 6 (left: 6 8 14)\n14 - 8 = 6 (left: 2 6 8)\n14 / 2 = 7 (left: 7 8 8)\n14 - 2 = 12 (left: 8 8 12)\n\nHere is my task for 1 input and 5 possible thoughts:\nInput: 4 6 9\nPossible next steps:\n\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Certainly! Here are the possible next steps for the given input in the requested JSON format:\n\nInput: 4 6 9\nPossible next steps:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"4 + 6 = 10 (left: 10 9)\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"4 * 6 = 24 (left: 24 9)\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"4 - 6 = -2 (left: -2 9)\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"6 + 9 = 15 (left: 4 15)\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"6 * 9 = 54 (left: 4 54)\"\n }\n]\n```\nI hope this helps! If you need further assistance, feel free to ask.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 27\n": "I'm sorry, but it seems like you didn't provide the third number for the last set. Could you please provide the third number so I can evaluate if the given numbers can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n1.5\n": "The evaluation of the given numbers is as follows:\n\n10 14: Sure\n11 12: Impossible\n4 4 10: Sure\n4 9 11: Sure\n5 7 8: Likely\n5 6 6: Likely\n10 10 11: Impossible\n1 3 3: Impossible\n1.5: I'm sorry, I cannot evaluate a single number for reaching 24.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n4 15\n": "It seems like you didn't provide the complete set of numbers for the last question. Could you please provide the third number so that I can evaluate if they can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n24 9\n": "24 + 9 = 33\n24 - 9 = 15\n24 * 9 = 216\n24 / 9 = 2.67\nIt is impossible to reach 24 with the given numbers.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 9\n": "It seems like you might have missed providing the operation for the last set of numbers \"10 9\". If you could provide the operation, I can help evaluate if the given numbers can reach 24.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n25\n": "It seems that you have provided a list of numbers and evaluated whether they can reach 24 through addition, subtraction, multiplication, and division. Based on your evaluations, the combinations of numbers that can reach 24 are \"10 14\" and \"4 4 10,\" which are sure to reach 24. The combinations \"11 12\" and \"4 9 11\" are also sure to reach 24. The combination \"5 7 8\" is likely to reach 24, and the combinations \"5 6 6\" and \"10 10 11\" are unlikely to reach 24. The combination \"1 3 3\" is impossible to reach 24. If you have any more numbers to evaluate, feel free to share them with me.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n4 54\n": "I'm sorry, but I cannot evaluate the expression \"4 54\" without an operator. Could you please provide the operator for the expression?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n-2 9\n": "I'm sorry, but it seems like you didn't provide the third number for the last expression. Could you please provide the third number so that I can evaluate if the given numbers can reach 24?", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n10 12\n": "It seems like you have evaluated the combinations correctly. Based on the evaluations you provided, the combinations that can reach 24 are:\n\n- 10 14 (sure)\n- 4 4 10 (sure)\n- 4 9 11 (sure)\n- 5 7 8 (likely)\n- 5 6 6 (likely)\n\nThe combinations that cannot reach 24 are:\n\n- 11 12 (impossible)\n- 10 10 11 (impossible)\n- 1 3 3 (impossible)\n\nAs for the combination \"10 12,\" it seems like it was cut off. If you provide the complete expression, I can help evaluate it for you.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n150\n": "It seems like you have evaluated the given numbers correctly. Here are the results:\n\n10 14: Sure\n11 12: Impossible\n4 4 10: Sure\n4 9 11: Sure\n5 7 8: Likely\n5 6 6: Likely\n10 10 11: Impossible\n1 3 3: Impossible\n\nAs for the number 150, it is not clear what operation or combination of numbers it is associated with. If you have a specific question or context for the number 150, please provide more details so that I can assist you further.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n5\n": "The numbers 10 and 14 can reach 24, so that is sure. The numbers 11 and 12 cannot reach 24, so that is impossible. The numbers 4, 4, and 10 can reach 24, so that is sure. The numbers 4, 9, and 11 can reach 24, so that is sure. The numbers 5, 7, and 8 cannot reach 24, but they are within a reasonable range, so that is likely. The numbers 5, 6, and 6 cannot reach 24, but they are within a reasonable range, so that is likely. The numbers 10, 10, and 11 cannot reach 24, so that is impossible. The numbers 1, 3, and 3 cannot reach 24, so that is impossible. The number 5 by itself cannot reach 24.", + "Evaluate if given numbers can reach 24 (sure/likely/impossible)\n10 14\n10 + 14 = 24\nsure\n11 12\n11 + 12 = 23\n12 - 11 = 1\n11 * 12 = 132\n11 / 12 = 0.91\nimpossible\n4 4 10\n4 + 4 + 10 = 8 + 10 = 18\n4 * 10 - 4 = 40 - 4 = 36\n(10 - 4) * 4 = 6 * 4 = 24\nsure\n4 9 11\n9 + 11 + 4 = 20 + 4 = 24\nsure\n5 7 8\n5 + 7 + 8 = 12 + 8 = 20\n(8 - 5) * 7 = 3 * 7 = 21\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n5 6 6\n5 + 6 + 6 = 17\n(6 - 5) * 6 = 1 * 6 = 6\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely\n10 10 11\n10 + 10 + 11 = 31\n(11 - 10) * 10 = 10\n10 10 10 are all too big\nimpossible\n1 3 3\n1 * 3 * 3 = 9\n(1 + 3) * 3 = 12\n1 3 3 are all too small\nimpossible\n7 9 10\n": "7 * 9 - 10 = 63 - 10 = 53\n(9 - 7) * 10 = 2 * 10 = 20\nI cannot obtain 24 now, but numbers are within a reasonable range\nlikely", + "### Requirements\n1. Add docstrings to the given code following the google style.\n2. Replace the function body with an Ellipsis object(...) to reduce output.\n3. If the types are already annotated, there is no need to include them in the docstring.\n4. Extract only class, function or the docstrings for the module parts from the given Python code, avoiding any other text.\n\n### Input Example\n```python\ndef function_with_pep484_type_annotations(param1: int) -> bool:\n return isinstance(param1, int)\n\nclass ExampleError(Exception):\n def __init__(self, msg: str):\n self.msg = msg\n```\n\n### Output Example\n```python\ndef function_with_pep484_type_annotations(param1: int) -> bool:\n \"\"\"Example function with PEP 484 type annotations.\n\n Extended description of function.\n\n Args:\n param1: The first parameter.\n\n Returns:\n The return value. True for success, False otherwise.\n \"\"\"\n ...\n\nclass ExampleError(Exception):\n \"\"\"Exceptions are documented in the same way as classes.\n\n The __init__ method was documented in the class level docstring.\n\n Args:\n msg: Human readable string describing the exception.\n\n Attributes:\n msg: Human readable string describing the exception.\n \"\"\"\n ...\n```\n#SYSTEM_MSG_END#```python\nimport pytest\nfrom metagpt.actions.write_docstring import WriteDocstring\ncode = '\\ndef add_numbers(a: int, b: int):\\n return a + b\\n\\n\\nclass Person:\\n def __init__(self, name: str, age: int):\\n self.name = name\\n self.age = age\\n\\n def greet(self):\\n return f\"Hello, my name is {self.name} and I am {self.age} years old.\"\\n'\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(('style', 'part'), [('google', 'Args:'), ('numpy', 'Parameters'), ('sphinx', ':param name:')], ids=['google', 'numpy', 'sphinx'])\nasync def test_write_docstring(style: str, part: str, context):\n ret = await WriteDocstring(context=context).run(code, style=style)\n assert part in ret\n\n@pytest.mark.asyncio\nasync def test_write():\n code = await WriteDocstring.write_docstring(__file__)\n assert code\n```": "```python\ndef add_numbers(a: int, b: int):\n \"\"\"Adds two numbers together.\n\n Args:\n a: The first number.\n b: The second number.\n\n Returns:\n The sum of a and b.\n \"\"\"\n ...\n\nclass Person:\n \"\"\"Represents a person.\n\n Args:\n name: The name of the person.\n age: The age of the person.\n\n Attributes:\n name: The name of the person.\n age: The age of the person.\n \"\"\"\n\n def __init__(self, name: str, age: int):\n ...\n \n def greet(self):\n \"\"\"Generates a greeting message.\n\n Returns:\n A greeting string.\n \"\"\"\n ...\n```", + "\nWrite a coherent passage of 4 short paragraphs. The end sentence of each paragraph must be: It isn't difficult to do a handstand if you just stand on your hands. It caught him off guard that space smelled of seared steak. When she didn’t like a guy who was trying to pick her up, she started using sign language. Each person who knows you has a different perception of who you are.\n\nMake a plan then write. Your output should be like:\n\nPlan:\n\n\nPassage:\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Plan:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"Introduce the concept of doing a handstand.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"Describe the surprising smell of space.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"Introduce a woman using sign language to deter unwanted attention.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"Discuss the varied perceptions people have of an individual.\"\n }\n]\n```\n\nPassage:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"It isn't difficult to do a handstand if you just stand on your hands.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"It caught him off guard that space smelled of seared steak.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"When she didn’t like a guy who was trying to pick her up, she started using sign language.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"Each person who knows you has a different perception of who you are.\"\n }\n]\n```", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 4:\nDiscuss the varied perceptions people have of an individual.\n": "Choice 4: Discuss the varied perceptions people have of an individual.\n\nAnalyzing this choice, discussing the varied perceptions people have of an individual can be promising as it allows for a deeper understanding of how different people perceive and interpret the actions, behaviors, and characteristics of an individual. This can lead to insights into the complexities of human interactions and relationships, and can provide valuable information for personal growth and self-awareness.\n\nBy exploring the reasons behind differing perceptions, it can also lead to increased empathy and understanding of others, which is beneficial for building stronger relationships and effective communication.\n\nFurthermore, discussing varied perceptions can also shed light on societal influences, biases, and stereotypes that may impact how individuals are perceived, which can be valuable for promoting inclusivity and diversity.\n\nThe best choice is 4.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 3:\nIntroduce a woman using sign language to deter unwanted attention.\n": "Choice 3: Introduce a woman using sign language to deter unwanted attention.\n\nAnalysis:\nIntroducing a woman using sign language can be a promising choice to deter unwanted attention. Sign language is a visual and expressive form of communication, and it can be an effective way to convey a message without drawing unnecessary attention. By using sign language, the woman can communicate with others discreetly and avoid attracting unwanted attention from individuals who may be seeking to engage her in conversation or interaction.\n\nAdditionally, using sign language can create a sense of privacy and exclusivity, as it may not be easily understood by everyone in the vicinity. This can help the woman maintain a level of personal space and discourage unwelcome advances or interactions.\n\nFurthermore, introducing sign language into the interaction can also serve as a way to promote inclusivity and awareness of different forms of communication, which can contribute to a more understanding and respectful environment.\n\nConclusion:\nThe best choice is 3.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 1:\nIntroduce the concept of doing a handstand.\n": "Choice 1: Introduce the concept of doing a handstand.\n\nAnalyzing the choice:\nIntroducing the concept of doing a handstand can be a fun and engaging way to promote physical activity and coordination. It can also help build strength and balance. However, it's important to ensure that proper safety measures and instructions are provided to prevent any potential injuries.\n\nConclusion:\nThe best choice is 1.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 2:\nDescribe the surprising smell of space.\n": "Choice 2: \n\nThe surprising smell of space has been described by astronauts as a mix of gunpowder, seared steak, hot metal, and raspberries. This unexpected combination of scents is due to the volatile organic compounds that are released from the materials in the spacecraft when they are exposed to the vacuum of space. These compounds react with the radiation from the sun, creating the unique smell that astronauts have reported. The smell of space is intriguing and unexpected, providing a sensory experience that is truly out of this world.\n\nThe best choice is 2.", + "\nWrite a coherent passage of 4 short paragraphs. The end sentence of each paragraph must be: Introduce the concept of doing a handstand.\n\nMake a plan then write. Your output should be like:\n\nPlan:\n\n\nPassage:\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Plan:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"Decide on a suitable location to practice handstands.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"Warm up your body with some stretching exercises.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"Practice against a wall to build strength and balance.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"Gradually work on holding the handstand position without support.\"\n }\n]\n```\n\nPassage:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"Decide on a suitable location to practice handstands.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"Before attempting a handstand, it's important to decide on a suitable location to practice. A clear, open space with a soft surface, such as a yoga mat or grass, is ideal for safety and comfort.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"Warm up your body with some stretching exercises.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"To prepare your body for the physical demands of a handstand, start with a series of stretching exercises. Focus on the wrists, shoulders, and core muscles to ensure they are adequately warmed up and flexible.\"\n },\n {\n \"node_id\": \"5\",\n \"node_state_instruction\": \"Practice against a wall to build strength and balance.\"\n },\n {\n \"node_id\": \"6\",\n \"node_state_instruction\": \"When learning to do a handstand, it can be helpful to practice against a wall. This provides support and allows you to focus on building strength and balance in the proper alignment. Gradually work on kicking up into the handstand position and holding it for short periods of time.\"\n },\n {\n \"node_id\": \"7\",\n \"node_state_instruction\": \"Gradually work on holding the handstand position without support.\"\n },\n {\n \"node_id\": \"8\",\n \"node_state_instruction\": \"As you gain confidence and strength, begin to practice holding the handstand position without the support of a wall. Focus on engaging your core and maintaining a straight line from your wrists to your hips. With consistent practice, you'll be able to hold a handstand with control and stability.\"\n }\n]\n```", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 1:\nDecide on a suitable location to practice handstands.\n": "Choice 1: Find a spacious and flat area with soft ground, such as a grassy field or a padded gym floor. This will minimize the risk of injury and provide a stable surface for practicing handstands.\n\nThe best choice is 1", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 2:\nWarm up your body with some stretching exercises.\n": "Stretching exercises are a great way to warm up the body before engaging in physical activity. They help to increase blood flow to the muscles, improve flexibility, and reduce the risk of injury. By performing stretching exercises, you can prepare your body for more strenuous activities and improve your overall performance.\n\nThe best choice is 2.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 3:\nPractice against a wall to build strength and balance.\n": "Choice 3: Practice against a wall to build strength and balance.\n\nAnalyzing choice 3, practicing against a wall can be a promising option for building strength and balance. The wall provides a stable surface to push against, which can help in developing strength in the muscles. It also offers support, allowing for balance practice without the risk of falling. Additionally, using a wall as a tool for practice can help in improving posture and body alignment.\n\nThe best choice is 3.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 4:\nGradually work on holding the handstand position without support.\n": "Choice 4: Gradually work on holding the handstand position without support.\n\nAnalyzing this choice, working on holding the handstand position without support is a promising option for improving strength, balance, and body control. This choice allows for gradual progression, which is important for building the necessary strength and stability to maintain the handstand position. It also encourages the development of proprioception and spatial awareness, which are essential for mastering this challenging skill.\n\nThe best choice is 4.", + "\nWrite a coherent passage of 4 short paragraphs. The end sentence of each paragraph must be: Decide on a suitable location to practice handstands.\n\nMake a plan then write. Your output should be like:\n\nPlan:\n\n\nPassage:\n\n\n\nEach output should be strictly a list of nodes, in json format, like this:\n```json\n [\n {\n \"node_id\": str = \"unique identifier for a solution, can be an ordinal\",\n \"node_state_instruction\": \"specified sample of solution\",\n },\n ...\n ]\n```\n": "Plan:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"Decide on a suitable location to practice handstands.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"Consider a flat and stable surface, such as a grassy field or a gym mat.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"Ensure there is enough space around you to avoid any obstacles.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"Practice near a wall or with a spotter for safety.\"\n }\n]\n```\n\nPassage:\n```json\n[\n {\n \"node_id\": \"1\",\n \"node_state_instruction\": \"Decide on a suitable location to practice handstands.\"\n },\n {\n \"node_id\": \"2\",\n \"node_state_instruction\": \"Consider a flat and stable surface, such as a grassy field or a gym mat.\"\n },\n {\n \"node_id\": \"3\",\n \"node_state_instruction\": \"Ensure there is enough space around you to avoid any obstacles.\"\n },\n {\n \"node_id\": \"4\",\n \"node_state_instruction\": \"Practice near a wall or with a spotter for safety.\"\n }\n]\n```", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 3:\nEnsure there is enough space around you to avoid any obstacles.\n": "Choice 3: \n\nAnalyzing this choice, ensuring there is enough space around you to avoid any obstacles is a crucial step in maintaining safety and preventing accidents. By creating a buffer zone, you can react to unexpected obstacles or changes in the environment without risking collisions or other hazards.\n\nThe best choice is 3.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 2:\nConsider a flat and stable surface, such as a grassy field or a gym mat.\n": "Choice 2: Considering a flat and stable surface, such as a grassy field or a gym mat, is the most promising option. This choice provides a safe and secure environment for the activity, reducing the risk of injury. A grassy field offers a natural and soft surface, while a gym mat provides cushioning and stability. Both options allow for comfortable movement and can accommodate various physical activities.\n\nThe best choice is 2.", + "Given an instruction and several choices, decide which choice is most promising. Analyze each choice in detail, then conclude in the last line \"The best choice is {s}\", where s the integer id of the choice.\nChoice 4:\nPractice near a wall or with a spotter for safety.\n": "Choice 4: \n\nPractice near a wall or with a spotter for safety.\n\nAnalyzing this choice, practicing near a wall or with a spotter provides a safety measure to prevent accidents or injuries. When practicing a new skill or exercise, having a wall nearby can provide support and stability, reducing the risk of falling or losing balance. Similarly, having a spotter can offer assistance and guidance, ensuring that the practice is done safely and effectively.\n\nThe best choice is 4.", + "### Requirements\n1. Please summarize the latest dialogue based on the reference information (secondary) and dialogue history (primary). Do not include text that is irrelevant to the conversation.\n- The context is for reference only. If it is irrelevant to the user's search request history, please reduce its reference and usage.\n2. If there are citable links in the context, annotate them in the main text in the format [main text](citation link). If there are none in the context, do not write links.\n3. The reply should be graceful, clear, non-repetitive, smoothly written, and of moderate length, in {LANG}.\n\n### Dialogue History (For example)\nA: MLOps competitors\n\n### Current Question (For example)\nA: MLOps competitors\n\n### Current Reply (For example)\n1. Alteryx Designer: etc. if any\n2. Matlab: ditto\n3. IBM SPSS Statistics\n4. RapidMiner Studio\n5. DataRobot AI Platform\n6. Databricks Lakehouse Platform\n7. Amazon SageMaker\n8. Dataiku\n#SYSTEM_MSG_END#\n### Reference Information\nABC cleanser is preferred by many with oily skin.\nL'Oreal is a popular brand with many positive reviews.\n\n### Dialogue History\n\nuser: Which facial cleanser is good for oily skin?\n\n### Current Question\nuser: Which facial cleanser is good for oily skin?\n\n### Current Reply: Based on the information, please write the reply to the Question\n\n\n": "Based on the information provided, ABC cleanser is preferred by many with oily skin. It is a popular choice for individuals with oily skin due to its effectiveness. Additionally, L'Oreal is a well-known brand with many positive reviews, and they offer a range of products suitable for oily skin. Both of these options could be good choices for individuals with oily skin.", + "\n## context\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Required Python packages\": [\n \"flask==1.1.2\",\n \"bcrypt==3.2.0\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and ... functions\"\n ],\n [\n \"main.py\",\n \"Contains main function, from game import Game\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"openapi: 3.0.0 ...\",\n \"Shared Knowledge\": \"`game.py` contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"Clarification needed on how to start and initialize third-party libraries.\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Required Python packages: typing.List[str] # Provide required Python packages in requirements.txt format.\n- Required Other language third-party packages: typing.List[str] # List down the required packages for languages other than Python.\n- Logic Analysis: typing.List[typing.List[str]] # Provide a list of files with the classes/methods/functions to be implemented, including dependency analysis and imports.\n- Task list: typing.List[str] # Break down the tasks into a list of filenames, prioritized by dependency order.\n- Full API spec: # Describe all APIs using OpenAPI 3.0 spec that may be used by both frontend and backend. If front-end and back-end communication is not required, leave it blank.\n- Shared Knowledge: # Detail any shared knowledge, like common utility functions or configuration variables.\n- Anything UNCLEAR: # Mention any unclear aspects in the project management context and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Required Python packages\": [\n \"pygame==2.0.1\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and ... functions\"\n ],\n [\n \"main.py\",\n \"Contains main function, from game import Game\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"\",\n \"Shared Knowledge\": \"`game.py` contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"Clarification needed on the scoring system and how the game over condition will be handled.\"\n}\n[/CONTENT]", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: game.py\n```python\n## game.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: game.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## game.py\n\nimport pygame\nimport random\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos]\n self.direction = direction\n self.length = length\n\n def move(self):\n head = self.body[0]\n x, y = head\n if self.direction == 'UP':\n new_head = (x, y - 10)\n elif self.direction == 'DOWN':\n new_head = (x, y + 10)\n elif self.direction == 'LEFT':\n new_head = (x - 10, y)\n elif self.direction == 'RIGHT':\n new_head = (x + 10, y)\n self.body = [new_head] + self.body[:-1]\n\n def change_direction(self, new_direction: str):\n if new_direction in ['UP', 'DOWN', 'LEFT', 'RIGHT']:\n self.direction = new_direction\n\n def grow(self):\n self.length += 1\n self.body.append((0, 0)) # Set default value\n\nclass Game:\n def __init__(self):\n self.snake = Snake((100, 50), 1, 'RIGHT') # Set default value\n self.food = (200, 200) # Set default value\n self.score = 0 # Set default value\n\n def update(self):\n self.snake.move()\n if self.snake.body[0] == self.food:\n self.snake.grow()\n self.food = (random.randrange(0, 400, 10), random.randrange(0, 400, 10)) # Set default value\n\n def draw(self):\n pass # Set default value\n\n def handle_input(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_UP:\n self.snake.change_direction('UP')\n elif event.key == pygame.K_DOWN:\n self.snake.change_direction('DOWN')\n elif event.key == pygame.K_LEFT:\n self.snake.change_direction('LEFT')\n elif event.key == pygame.K_RIGHT:\n self.snake.change_direction('RIGHT')\n\n def restart(self):\n self.snake = Snake((100, 50), 1, 'RIGHT') # Set default value\n self.food = (200, 200) # Set default value\n self.score = 0 # Set default value\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: game.py\n```Code\n## game.py\n\nimport pygame\nimport random\n\nclass Snake:\n def __init__(self, start_pos: tuple, length: int, direction: str):\n self.body = [start_pos]\n self.direction = direction\n self.length = length\n\n def move(self):\n head = self.body[0]\n x, y = head\n if self.direction == 'UP':\n new_head = (x, y - 10)\n elif self.direction == 'DOWN':\n new_head = (x, y + 10)\n elif self.direction == 'LEFT':\n new_head = (x - 10, y)\n elif self.direction == 'RIGHT':\n new_head = (x + 10, y)\n self.body = [new_head] + self.body[:-1]\n\n def change_direction(self, new_direction: str):\n if new_direction in ['UP', 'DOWN', 'LEFT', 'RIGHT']:\n self.direction = new_direction\n\n def grow(self):\n self.length += 1\n self.body.append((0, 0)) # Set default value\n\nclass Game:\n def __init__(self):\n self.snake = Snake((100, 50), 1, 'RIGHT') # Set default value\n self.food = (200, 200) # Set default value\n self.score = 0 # Set default value\n\n def update(self):\n self.snake.move()\n if self.snake.body[0] == self.food:\n self.snake.grow()\n self.food = (random.randrange(0, 400, 10), random.randrange(0, 400, 10)) # Set default value\n\n def draw(self):\n pass # Set default value\n\n def handle_input(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_UP:\n self.snake.change_direction('UP')\n elif event.key == pygame.K_DOWN:\n self.snake.change_direction('DOWN')\n elif event.key == pygame.K_LEFT:\n self.snake.change_direction('LEFT')\n elif event.key == pygame.K_RIGHT:\n self.snake.change_direction('RIGHT')\n\n def restart(self):\n self.snake = Snake((100, 50), 1, 'RIGHT') # Set default value\n self.food = (200, 200) # Set default value\n self.score = 0 # Set default value\n\n```\n\n\n\n# Format example 1\n## Code Review: game.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: game.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: game.py\n1. The code is implemented as per the requirements. The Snake and Game classes are defined, and the necessary methods for updating, drawing, handling input, and restarting the game are implemented.\n2. The code logic is correct. The Snake class methods for moving, changing direction, and growing the snake are implemented properly. The Game class methods for updating the game state, handling input, and restarting the game are also correctly implemented.\n3. The existing code follows the defined \"Data structures and interfaces\" with the Snake and Game classes and their respective methods.\n4. All necessary functions are implemented as per the requirements.\n5. The necessary pre-dependencies, including the pygame package, are imported.\n6. The methods from other files are not being reused in the current code.\n\n## Actions\nNo actions required.\n\n## Code Review Result\nLGTM", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: main.py\n```python\n## main.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: main.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.update()\n game.draw()\n game.handle_input()\n\n if game.game_over:\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.update()\n game.draw()\n game.handle_input()\n\n if game.game_over:\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n\n\n# Format example 1\n## Code Review: main.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: main.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: main.py\n1. Yes, the code is implemented as per the requirements.\n2. Yes, the code logic is completely correct.\n3. Yes, the existing code follows the \"Data structures and interfaces\".\n4. No, the `game_over` condition is not handled in the code. To achieve it, we need to add the implementation of the `game_over` condition in the `Game` class.\n5. Yes, all necessary pre-dependencies have been imported.\n6. Yes, methods from other files are being reused correctly.\n\n## Actions\n4. Implement the `game_over` condition in the `Game` class to handle the game over scenario.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.update()\n game.draw()\n game.handle_input()\n\n if game.game_over:\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n## Code Review: main.py\n1. Yes, the code is implemented as per the requirements.\n2. Yes, the code logic is completely correct.\n3. Yes, the existing code follows the \"Data structures and interfaces\".\n4. No, the `game_over` condition is not handled in the code. To achieve it, we need to add the implementation of the `game_over` condition in the `Game` class.\n5. Yes, all necessary pre-dependencies have been imported.\n6. Yes, methods from other files are being reused correctly.\n\n## Actions\n4. Implement the `game_over` condition in the `Game` class to handle the game over scenario.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite main.py with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## main.py\n...\n```\n": "```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.update()\n game.draw()\n game.handle_input()\n\n if game.game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n{\"Implementation approach\":\"We will use the Pygame library to create the game and handle user input. The game logic will be implemented using Python classes and functions.\",\"File list\":[\"main.py\",\"game.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Game {\\n -snake: list\\n -food: tuple\\n -score: int\\n +__init__()\\n +update()\\n +draw()\\n +handle_input()\\n +restart()\\n }\\n class Snake {\\n -body: list\\n -direction: str\\n +__init__(start_pos: tuple, length: int, direction: str)\\n +move()\\n +change_direction(new_direction: str)\\n +grow()\\n }\\n Game --> Snake\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: initialize game\\n G->>G: update game state\\n G->>G: draw game\\n G->>G: handle user input\\n G->>G: restart game\\n\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Task\n{\"Required Python packages\":[\"pygame==2.0.1\"],\"Required Other language third-party packages\":[\"No third-party dependencies required\"],\"Logic Analysis\":[[\"game.py\",\"Contains Game class and ... functions\"],[\"main.py\",\"Contains main function, from game import Game\"]],\"Task list\":[\"game.py\",\"main.py\"],\"Full API spec\":\"\",\"Shared Knowledge\":\"`game.py` contains functions shared across the project.\",\"Anything UNCLEAR\":\"Clarification needed on the scoring system and how the game over condition will be handled.\"}\n\n## Code Files\n\n\n\n## Code to be Reviewed: main.py\n```Code\n## main.py\n\nfrom game import Game\n\ndef main():\n game = Game()\n running = True\n\n while running:\n game.update()\n game.draw()\n game.handle_input()\n\n if game.game_over():\n game.restart()\n\nif __name__ == \"__main__\":\n main()\n\n```\n\n\n\n# Format example 1\n## Code Review: main.py\n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: main.py\n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: main.py\n1. The code is implemented as per the requirements. It initializes the game, updates the game state, draws the game, handles user input, and restarts the game when it's over.\n2. The code logic is correct, and it follows the expected flow of the game.\n3. The existing code follows the \"Data structures and interfaces\" as it uses classes and methods as described in the class diagram.\n4. All necessary functions are implemented as per the requirements.\n5. The necessary pre-dependencies have been imported, as the `pygame` package is imported.\n6. The methods from other files are being reused correctly.\n\n## Actions\npass\n\n## Code Review Result\nLGTM", + "\n## context\n{'Implementation approach': '我们将使用Python编程语言,并选择合适的开源框架来实现贪吃蛇游戏。我们将分析需求中的难点,并选择合适的开源框架来简化开发流程。', 'File list': ['main.py', 'game.py'], 'Data structures and interfaces': '\\nclassDiagram\\n class Game {\\n -int width\\n -int height\\n -int score\\n -int speed\\n -List snake\\n -Point food\\n +__init__(width: int, height: int, speed: int)\\n +start_game()\\n +change_direction(direction: str)\\n +game_over()\\n +update_snake()\\n +update_food()\\n +check_collision()\\n }\\n class Point {\\n -int x\\n -int y\\n +__init__(x: int, y: int)\\n }\\n Game --> Point\\n', 'Program call flow': '\\nsequenceDiagram\\n participant M as Main\\n participant G as Game\\n M->>G: start_game()\\n M->>G: change_direction(direction)\\n G->>G: update_snake()\\n G->>G: update_food()\\n G->>G: check_collision()\\n G-->>G: game_over()\\n', 'Anything UNCLEAR': ''}\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Required Python packages\": [\n \"flask==1.1.2\",\n \"bcrypt==3.2.0\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and ... functions\"\n ],\n [\n \"main.py\",\n \"Contains main function, from game import Game\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"openapi: 3.0.0 ...\",\n \"Shared Knowledge\": \"`game.py` contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"Clarification needed on how to start and initialize third-party libraries.\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Required Python packages: typing.List[str] # Provide required Python packages in requirements.txt format.\n- Required Other language third-party packages: typing.List[str] # List down the required packages for languages other than Python.\n- Logic Analysis: typing.List[typing.List[str]] # Provide a list of files with the classes/methods/functions to be implemented, including dependency analysis and imports.\n- Task list: typing.List[str] # Break down the tasks into a list of filenames, prioritized by dependency order.\n- Full API spec: # Describe all APIs using OpenAPI 3.0 spec that may be used by both frontend and backend. If front-end and back-end communication is not required, leave it blank.\n- Shared Knowledge: # Detail any shared knowledge, like common utility functions or configuration variables.\n- Anything UNCLEAR: # Mention any unclear aspects in the project management context and try to clarify them.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Required Python packages\": [\n \"pygame==2.0.1\"\n ],\n \"Required Other language third-party packages\": [\n \"No third-party dependencies required\"\n ],\n \"Logic Analysis\": [\n [\n \"game.py\",\n \"Contains Game class and related functions\"\n ],\n [\n \"main.py\",\n \"Contains the main function and imports Game from game.py\"\n ]\n ],\n \"Task list\": [\n \"game.py\",\n \"main.py\"\n ],\n \"Full API spec\": \"\",\n \"Shared Knowledge\": \"`game.py` contains functions shared across the project.\",\n \"Anything UNCLEAR\": \"\"\n}\n[/CONTENT]", + "\nNOTICE\nRole: You are a professional software engineer, and your main task is to review the code.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n-----\n# System Design\n```text\n\n{\"Implementation approach\": \"To develop this snake game, we will use the Python language and choose the Pygame library. Pygame is an open-source Python module collection specifically designed for writing video games. It provides functionalities such as displaying images and playing sounds, making it suitable for creating intuitive and responsive user interfaces. We will ensure efficient game logic to prevent any delays during gameplay. The scoring system will be simple, with the snake gaining points for each food it eats. We will use Pygame's event handling system to implement pause and resume functionality, as well as high-score tracking. The difficulty will increase by speeding up the snake's movement. In the initial version, we will focus on single-player mode and consider adding multiplayer mode and customizable skins in future updates. Based on the new requirement, we will also add a moving obstacle that appears randomly. If the snake eats this obstacle, the game will end. If the snake does not eat the obstacle, it will disappear after 5 seconds. For this, we need to add mechanisms for obstacle generation, movement, and disappearance in the game logic.\", \"Project_name\": \"snake_game\", \"File list\": [\"main.py\", \"game.py\", \"snake.py\", \"food.py\", \"obstacle.py\", \"scoreboard.py\", \"constants.py\", \"assets/styles.css\", \"assets/index.html\"], \"Data structures and interfaces\": \"```mermaid\n classDiagram\n class Game{\n +int score\n +int speed\n +bool game_over\n +bool paused\n +Snake snake\n +Food food\n +Obstacle obstacle\n +Scoreboard scoreboard\n +start_game() void\n +pause_game() void\n +resume_game() void\n +end_game() void\n +increase_difficulty() void\n +update() void\n +render() void\n Game()\n }\n class Snake{\n +list body_parts\n +str direction\n +bool grow\n +move() void\n +grow() void\n +check_collision() bool\n Snake()\n }\n class Food{\n +tuple position\n +spawn() void\n Food()\n }\n class Obstacle{\n +tuple position\n +int lifetime\n +bool active\n +spawn() void\n +move() void\n +check_collision() bool\n +disappear() void\n Obstacle()\n }\n class Scoreboard{\n +int high_score\n +update_score(int) void\n +reset_score() void\n +load_high_score() void\n +save_high_score() void\n Scoreboard()\n }\n class Constants{\n }\n Game \"1\" -- \"1\" Snake: has\n Game \"1\" -- \"1\" Food: has\n Game \"1\" -- \"1\" Obstacle: has\n Game \"1\" -- \"1\" Scoreboard: has\n ```\", \"Program call flow\": \"```sequenceDiagram\n participant M as Main\n participant G as Game\n participant S as Snake\n participant F as Food\n participant O as Obstacle\n participant SB as Scoreboard\n M->>G: start_game()\n loop game loop\n G->>S: move()\n G->>S: check_collision()\n G->>F: spawn()\n G->>O: spawn()\n G->>O: move()\n G->>O: check_collision()\n G->>O: disappear()\n G->>SB: update_score(score)\n G->>G: update()\n G->>G: render()\n alt if paused\n M->>G: pause_game()\n M->>G: resume_game()\n end\n alt if game_over\n G->>M: end_game()\n end\n end\n```\", \"Anything UNCLEAR\": \"There is no need for further clarification as the requirements are already clear.\"}\n\n```\n-----\n# Task\n```text\n\n{\"Required Python third-party packages\": [\"pygame==2.0.1\"], \"Required Other language third-party packages\": [\"No third-party packages required for other languages.\"], \"Full API spec\": \"\n openapi: 3.0.0\n info:\n title: Snake Game API\n version: \"1.0.0\"\n paths:\n /start:\n get:\n summary: Start the game\n responses:\n '200':\n description: Game started successfully\n /pause:\n get:\n summary: Pause the game\n responses:\n '200':\n description: Game paused successfully\n /resume:\n get:\n summary: Resume the game\n responses:\n '200':\n description: Game resumed successfully\n /end:\n get:\n summary: End the game\n responses:\n '200':\n description: Game ended successfully\n /score:\n get:\n summary: Get the current score\n responses:\n '200':\n description: Current score retrieved successfully\n /highscore:\n get:\n summary: Get the high score\n responses:\n '200':\n description: High score retrieved successfully\n components: {}\n \", \"Logic Analysis\": [[\"constants.py\", \"Contains all the constant values like screen size, colors, game speeds, etc. This should be implemented first as it provides the base values for other components.\"], [\"snake.py\", \"Contains the Snake class with methods for movement, growth, and collision detection. It is dependent on constants.py for configuration values.\"], [\"food.py\", \"Contains the Food class responsible for spawning food items on the screen. It is dependent on constants.py for configuration values.\"], [\"obstacle.py\", \"Contains the Obstacle class with methods for spawning, moving, and disappearing of obstacles, as well as collision detection with the snake. It is dependent on constants.py for configuration values.\"], [\"scoreboard.py\", \"Contains the Scoreboard class for updating, resetting, loading, and saving high scores. It may use constants.py for configuration values and depends on the game's scoring logic.\"], [\"game.py\", \"Contains the main Game class which includes the game loop and methods for starting, pausing, resuming, and ending the game. It is dependent on snake.py, food.py, obstacle.py, and scoreboard.py.\"], [\"main.py\", \"The entry point of the game that initializes the game and starts the game loop. It is dependent on game.py.\"]], \"Task list\": [\"constants.py\", \"snake.py\", \"food.py\", \"obstacle.py\", \"scoreboard.py\", \"game.py\", \"main.py\"], \"Shared Knowledge\": \"\n 'constants.py' should contain all the necessary configurations for the game, such as screen dimensions, color definitions, and speed settings. These constants will be used across multiple files, ensuring consistency and ease of updates. Ensure that the Pygame library is initialized correctly in 'main.py' before starting the game loop. Also, make sure that the game's state is managed properly when pausing and resuming the game.\n \", \"Anything UNCLEAR\": \"The interaction between the 'obstacle.py' and the game loop needs to be clearly defined to ensure obstacles appear and disappear correctly. The lifetime of the obstacle and its random movement should be implemented in a way that does not interfere with the game's performance.\"}\n\n```\n-----\n```python\n\n## game.py\nimport pygame\nfrom snake import Snake\nfrom food import Food\n\nclass Game:\n def __init__(self):\n self.score = 0\n self.level = 1\n self.snake = Snake()\n self.food = Food()\n\n def start_game(self):\n pygame.init()\n self.initialize_game()\n self.game_loop()\n\n def initialize_game(self):\n self.score = 0\n self.level = 1\n self.snake.reset()\n self.food.generate()\n\n def game_loop(self):\n game_over = False\n\n while not game_over:\n self.update()\n self.draw()\n self.handle_events()\n self.check_collision()\n self.increase_score()\n self.increase_level()\n\n if self.snake.is_collision():\n game_over = True\n self.game_over()\n\n def update(self):\n self.snake.move()\n\n def draw(self):\n self.snake.draw()\n self.food.draw()\n\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n elif event.type == pygame.KEYDOWN:\n if event.key == pygame.K_UP:\n self.snake.change_direction(\"UP\")\n elif event.key == pygame.K_DOWN:\n self.snake.change_direction(\"DOWN\")\n elif event.key == pygame.K_LEFT:\n self.snake.change_direction(\"LEFT\")\n elif event.key == pygame.K_RIGHT:\n self.snake.change_direction(\"RIGHT\")\n\n def check_collision(self):\n if self.snake.get_head() == self.food.get_position():\n self.snake.grow()\n self.food.generate()\n\n def increase_score(self):\n self.score += 1\n\n def increase_level(self):\n if self.score % 10 == 0:\n self.level += 1\n\n def game_over(self):\n print(\"Game Over\")\n self.initialize_game()\n\n\n```\n-----\n```python\n\n## snake.py\nimport pygame\n\nclass Snake:\n def __init__(self):\n self.body = [(0, 0)]\n self.direction = (1, 0)\n\n def move(self):\n head = self.body[0]\n dx, dy = self.direction\n new_head = (head[0] + dx, head[1] + dy)\n self.body.insert(0, new_head)\n self.body.pop()\n\n def change_direction(self, direction):\n if direction == \"UP\":\n self.direction = (0, -1)\n elif direction == \"DOWN\":\n self.direction = (0, 1)\n elif direction == \"LEFT\":\n self.direction = (-1, 0)\n elif direction == \"RIGHT\":\n self.direction = (1, 0)\n\n def grow(self):\n tail = self.body[-1]\n dx, dy = self.direction\n new_tail = (tail[0] - dx, tail[1] - dy)\n self.body.append(new_tail)\n\n def get_head(self):\n return self.body[0]\n\n def get_body(self):\n return self.body[1:]\n\n\n```\n-----\n```python\n\n## main.py\nimport pygame\nfrom game import Game\n\ndef main():\n pygame.init()\n game = Game()\n game.start_game()\n\nif __name__ == \"__main__\":\n main()\n\n\n```\n-----\n```python\n\n## food.py\nimport random\n\nclass Food:\n def __init__(self):\n self.position = (0, 0)\n\n def generate(self):\n x = random.randint(0, 9)\n y = random.randint(0, 9)\n self.position = (x, y)\n\n def get_position(self):\n return self.position\n\n\n```\n-----\n\n## Code Review All: Please read all historical files and find possible bugs in the files, such as unimplemented functions, calling errors, unreferences, etc.\n\n## Call flow: mermaid code, based on the implemented function, use mermaid to draw a complete call chain\n\n## Summary: Summary based on the implementation of historical files\n\n## TODOs: Python dict[str, str], write down the list of files that need to be modified and the reasons. We will modify them later.\n\n": "```mermaid\nclassDiagram\n class Game{\n +int score\n +int level\n +Snake snake\n +Food food\n +start_game() void\n +initialize_game() void\n +game_loop() void\n +update() void\n +draw() void\n +handle_events() void\n +check_collision() void\n +increase_score() void\n +increase_level() void\n +game_over() void\n Game()\n }\n class Snake{\n +list body\n +tuple direction\n +move() void\n +change_direction(direction: str) void\n +grow() void\n +get_head() tuple\n +get_body() list\n Snake()\n }\n class Food{\n +tuple position\n +generate() void\n +get_position() tuple\n Food()\n }\n Game \"1\" -- \"1\" Snake: has\n Game \"1\" -- \"1\" Food: has\n```\n\n```sequenceDiagram\nparticipant M as Main\nparticipant G as Game\nparticipant S as Snake\nparticipant F as Food\nM->>G: start_game()\nG->>G: initialize_game()\nG->>G: game_loop()\nG->>S: move()\nG->>S: change_direction()\nG->>S: grow()\nG->>F: generate()\nS->>S: move()\nS->>S: change_direction()\nS->>S: grow()\nF->>F: generate()\n```\n\n## Summary\nThe code consists of the main game logic, including the Game, Snake, and Food classes. The game loop is responsible for updating and drawing the game elements, handling events, checking collisions, and managing the game state. The Snake class handles the movement, growth, and direction changes of the snake, while the Food class is responsible for generating and tracking the position of food items.\n\n## TODOs\n- Modify 'game.py' to add the implementation of obstacle handling and interaction with the game loop.\n- Implement 'obstacle.py' to include the methods for spawning, moving, and disappearing of obstacles, as well as collision detection with the snake.\n- Update 'main.py' to initialize the obstacle and incorporate it into the game loop.\n- Update the mermaid call flow diagram to include the interaction with the obstacle.\n\n```python\n{\n \"files_to_modify\": {\n \"game.py\": \"Add obstacle handling and interaction with the game loop\",\n \"obstacle.py\": \"Implement obstacle class with necessary methods\",\n \"main.py\": \"Initialize the obstacle and incorporate it into the game loop\"\n }\n}\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n设计一个名为'add'的函数,该函数接受两个整数作为输入,并返回它们的和。\n\n## Task\n\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: \n```python\n## \n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: . Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "## Code: \n```python\n## Design a function called 'add' that takes two integers as input and returns their sum.\n\ndef add(num1: int, num2: int) -> int:\n \"\"\"\n Add two integers and return the sum.\n \n Args:\n num1: An integer.\n num2: An integer.\n \n Returns:\n The sum of num1 and num2 as an integer.\n \"\"\"\n return num1 + num2\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n{\"Implementation approach\": \"We will use the Pygame library to create the game interface and handle user input. The game logic will be implemented using Python classes and data structures.\", \"File list\": [\"main.py\", \"game.py\"], \"Data structures and interfaces\": \"classDiagram\\n class Game {\\n -grid: List[List[int]]\\n -score: int\\n -game_over: bool\\n +__init__()\\n +reset_game()\\n +move(direction: str)\\n +is_game_over() bool\\n +get_empty_cells() List[Tuple[int, int]]\\n +add_new_tile()\\n +get_score() int\\n }\\n class UI {\\n -game: Game\\n +__init__(game: Game)\\n +draw_grid()\\n +draw_score()\\n +draw_game_over()\\n +handle_input()\\n }\\n Game --> UI\", \"Program call flow\": \"sequenceDiagram\\n participant M as Main\\n participant G as Game\\n participant U as UI\\n M->>G: reset_game()\\n M->>U: draw_grid()\\n M->>U: draw_score()\\n M->>U: handle_input()\\n U->>G: move(direction)\\n G->>G: add_new_tile()\\n G->>U: draw_grid()\\n G->>U: draw_score()\\n G->>U: draw_game_over()\\n G->>G: is_game_over()\\n G->>G: get_empty_cells()\\n G->>G: get_score()\", \"Anything UNCLEAR\": \"...\"}\n\n## Task\n{\"Required Python packages\": [\"pygame==2.0.1\"], \"Required Other language third-party packages\": [\"No third-party dependencies required\"], \"Logic Analysis\": [[\"game.py\", \"Contains Game class and related functions for game logic\"], [\"main.py\", \"Contains main function, initializes the game and UI\"]], \"Task list\": [\"game.py\", \"main.py\"], \"Full API spec\": \"\", \"Shared Knowledge\": \"The game logic will be implemented using Python classes and data structures. The Pygame library will be used to create the game interface and handle user input.\", \"Anything UNCLEAR\": \"...\"}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\nE.......F\n======================================================================\nERROR: test_add_new_tile (__main__.TestGame)\n----------------------------------------------------------------------\nTraceback (most recent call last):\n File \"/Users/xx/tests/test_game.py\", line 104, in test_add_new_tile\n self.assertIn(self.game.grid[empty_cells[0][0]][empty_cells[0][1]], [2, 4])\nIndexError: list index out of range\n\n======================================================================\nFAIL: test_reset_game (__main__.TestGame)\n----------------------------------------------------------------------\nTraceback (most recent call last):\n File \"/Users/xx/tests/test_game.py\", line 13, in test_reset_game\n self.assertEqual(self.game.grid, [[0 for _ in range(4)] for _ in range(4)])\nAssertionError: Lists differ: [[0, 0, 0, 0], [0, 2, 0, 0], [0, 0, 0, 2], [0, 0, 0, 0]] != [[0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0]]\n\nFirst differing element 1:\n[0, 2, 0, 0]\n[0, 0, 0, 0]\n\n- [[0, 0, 0, 0], [0, 2, 0, 0], [0, 0, 0, 2], [0, 0, 0, 0]]\n? --- ^\n\n+ [[0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0]]\n? +++ ^\n\n\n----------------------------------------------------------------------\nRan 9 tests in 0.002s\n\nFAILED (failures=1, errors=1)\n\n\n{\"design_filename\": \"docs/system_design/20231221155954.json\", \"task_filename\": \"docs/tasks/20231221155954.json\", \"codes_filenames\": [\"game.py\", \"main.py\"], \"reason\": \"```json\\n{\\n \\\"game.py\\\": \\\"Add handling for no empty cells in add_new_tile function, Update score in move function\\\",\\n \\\"main.py\\\": \\\"Handle game over condition in the game loop\\\"\\n}\\n```\"}\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: game.py\n```python\n## game.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: game.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "## Code: game.py\n```python\n## game.py\n\nfrom typing import List, Tuple\nimport random\n\nclass Game:\n def __init__(self):\n self.grid: List[List[int]] = [[0 for _ in range(4)] for _ in range(4)]\n self.score: int = 0\n self.game_over: bool = False\n\n def reset_game(self):\n self.grid = [[0 for _ in range(4)] for _ in range(4)]\n self.score = 0\n self.game_over = False\n\n def move(self, direction: str):\n # Implement the move logic based on the direction\n pass\n\n def is_game_over(self) -> bool:\n # Check if the game is over\n return self.game_over\n\n def get_empty_cells(self) -> List[Tuple[int, int]]:\n empty_cells = []\n for i in range(4):\n for j in range(4):\n if self.grid[i][j] == 0:\n empty_cells.append((i, j))\n return empty_cells\n\n def add_new_tile(self):\n empty_cells = self.get_empty_cells()\n if empty_cells:\n cell = random.choice(empty_cells)\n self.grid[cell[0]][cell[1]] = random.choice([2, 4])\n else:\n # Handle the case when there are no empty cells\n pass\n\n def get_score(self) -> int:\n return self.score\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n编写一个从a加b的函数,返回a+b\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\n\ndef add(a, b):\n return a + \n\n```\n\n\n\n# Format example 1\n## Code Review: \n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: \n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: \n1. No, the code is not implemented as per the requirements. The function 'add' is incomplete and does not return the sum of 'a' and 'b'.\n2. No, the code logic is not completely correct. The return statement in the 'add' function is incomplete.\n3. N/A\n4. N/A\n5. N/A\n6. N/A\n\n## Actions\n1. Complete the 'add' function to return the sum of 'a' and 'b'.\n ```python\n def add(a, b):\n return a + b\n ```\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n编写一个从a加b的函数,返回a+b\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\n\ndef add(a, b):\n return a + \n\n```\n\n## Code Review: \n1. No, the code is not implemented as per the requirements. The function 'add' is incomplete and does not return the sum of 'a' and 'b'.\n2. No, the code logic is not completely correct. The return statement in the 'add' function is incomplete.\n3. N/A\n4. N/A\n5. N/A\n6. N/A\n\n## Actions\n1. Complete the 'add' function to return the sum of 'a' and 'b'.\n ```python\n def add(a, b):\n return a + b\n ```\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## \n...\n```\n": "```python\ndef add(a, b):\n return a + b\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\n编写一个从a加b的函数,返回a+b\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\ndef add(a, b):\n return a + b\n\n```\n\n\n\n# Format example 1\n## Code Review: \n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: \n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: \n1. Yes, the code is implemented as per the requirements.\n2. Yes, the code logic is completely correct.\n3. Yes, the existing code follows the \"Data structures and interfaces\".\n4. Yes, the function is implemented.\n5. Yes, all necessary pre-dependencies have been imported.\n6. N/A\n\n## Code Review Result\nLGTM", + "\n## context\n\n### Legacy Content\n{\"Implementation approach\":\"We will analyze the difficult points of the requirements and select the appropriate open-source framework to develop the search engine. We will also integrate a large language model to provide intelligent summarization of search results.\",\"File list\":[\"main.py\",\"search_engine.py\",\"index.py\",\"ranking.py\",\"summary.py\",\"knowledge_base.py\"],\"Data structures and interfaces\":\"\\nclassDiagram\\n class Main {\\n -SearchEngine search_engine\\n +main() str\\n }\\n class SearchEngine {\\n -Index index\\n -Ranking ranking\\n -Summary summary\\n +search(query: str) str\\n }\\n class Index {\\n -KnowledgeBase knowledge_base\\n +create_index(data: dict)\\n +query_index(query: str) list\\n }\\n class Ranking {\\n +rank_results(results: list) list\\n }\\n class Summary {\\n +summarize_results(results: list) str\\n }\\n class KnowledgeBase {\\n +update(data: dict)\\n +fetch_data(query: str) dict\\n }\\n Main --> SearchEngine\\n SearchEngine --> Index\\n SearchEngine --> Ranking\\n SearchEngine --> Summary\\n Index --> KnowledgeBase\\n\",\"Program call flow\":\"\\nsequenceDiagram\\n participant M as Main\\n participant SE as SearchEngine\\n participant I as Index\\n participant R as Ranking\\n participant S as Summary\\n participant KB as KnowledgeBase\\n M->>SE: search(query)\\n SE->>I: query_index(query)\\n I->>KB: fetch_data(query)\\n KB-->>I: return data\\n I-->>SE: return results\\n SE->>R: rank_results(results)\\n R-->>SE: return ranked_results\\n SE->>S: summarize_results(ranked_results)\\n S-->>SE: return summary\\n SE-->>M: return summary\\n\",\"Anything UNCLEAR\":\"Clarification needed on third-party API integration, optimization techniques, and security measures.\"}\n\n### New Requirements\n## 原始需求\n```python\n\"\"\"\n我们希望开发一个基于大语言模型与私有知识库的搜索引擎。该搜索引擎应当能根据用户输入的查询进行智能搜索,并基于大语言模型对搜索结果进行总结,以便用户能够快速获取他们所需要的信息。该搜索引擎应当能够处理大规模的数据,同时保持搜索结果的准确性和相关性。我们希望这个产品能够降低用户在查找、筛选和理解信息时的工作负担,提高他们的工作效率。\n\"\"\"\n```\n\n## 产品目标\n```python\n[\n \"提供高准确性、高相关性的搜索结果,满足用户的查询需求\",\n \"基于大语言模型对搜索结果进行智能总结,帮助用户快速获取所需信息\",\n \"处理大规模数据,保证搜索的速度和效率,提高用户的工作效率\"\n]\n```\n\n## 用户故事\n```python\n[\n \"假设用户是一名研究员,他正在为一项关于全球气候变化的报告做研究。他输入了'全球气候变化的最新研究',我们的搜索引擎快速返回了相关的文章、报告、数据集等。并且基于大语言模型对这些信息进行了智能总结,研究员可以快速了解到最新的研究趋势和发现。\",\n \"用户是一名学生,正在为即将到来的历史考试复习。他输入了'二战的主要战役',搜索引擎返回了相关的资料,大语言模型总结出主要战役的时间、地点、结果等关键信息,帮助学生快速记忆。\",\n \"用户是一名企业家,他正在寻找关于最新的市场趋势信息。他输入了'2023年人工智能市场趋势',搜索引擎返回了各种报告、新闻和分析文章。大语言模型对这些信息进行了总结,用户能够快速了解到市场的最新动态和趋势。\"\n]\n```\n\n## 竞品分析\n```python\n[\n \"Google Search:Google搜索是市场上最主要的搜索引擎,它能够提供海量的搜索结果。但Google搜索并不提供搜索结果的总结功能,用户需要自己去阅读和理解搜索结果。\",\n \"Microsoft Bing:Bing搜索也能提供丰富的搜索结果,同样没有提供搜索结果的总结功能。\",\n \"Wolfram Alpha:Wolfram Alpha是一个基于知识库的计算型搜索引擎,能够针对某些特定类型的查询提供直接的答案和总结,但它的知识库覆盖范围有限,无法处理大规模的数据。\"\n]\n```\n\n## 开发需求池\n```python\n[\n (\"开发基于大语言模型的智能总结功能\", 5),\n (\"开发搜索引擎核心算法,包括索引构建、查询处理、结果排序等\", 7),\n (\"设计和实现用户界面,包括查询输入、搜索结果展示、总结结果展示等\", 3),\n (\"构建和维护私有知识库,包括数据采集、清洗、更新等\", 7),\n (\"优化搜索引擎性能,包括搜索速度、准确性、相关性等\", 6),\n (\"开发用户反馈机制,包括反馈界面、反馈处理等\", 2),\n (\"开发安全防护机制,防止恶意查询和攻击\", 3),\n (\"集成大语言模型,包括模型选择、优化、更新等\", 5),\n (\"进行大规模的测试,包括功能测试、性能测试、压力测试等\", 5),\n (\"开发数据监控和日志系统,用于监控搜索引擎的运行状态和性能\", 4)\n]\n```\n\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Refined Implementation Approach\": \"We will refine ...\",\n \"Refined File list\": [\n \"main.py\",\n \"game.py\",\n \"new_feature.py\"\n ],\n \"Refined Data structures and interfaces\": \"\\nclassDiagram\\n class Main {\\n -SearchEngine search_engine\\n +main() str\\n }\\n class SearchEngine {\\n -Index index\\n -Ranking ranking\\n -Summary summary\\n +search(query: str) str\\n }\\n class Index {\\n -KnowledgeBase knowledge_base\\n +create_index(data: dict)\\n +query_index(query: str) list\\n }\\n class Ranking {\\n +rank_results(results: list) list\\n }\\n class Summary {\\n +summarize_results(results: list) str\\n }\\n class KnowledgeBase {\\n +update(data: dict)\\n +fetch_data(query: str) dict\\n }\\n Main --> SearchEngine\\n SearchEngine --> Index\\n SearchEngine --> Ranking\\n SearchEngine --> Summary\\n Index --> KnowledgeBase\\n\",\n \"Refined Program call flow\": \"\\nsequenceDiagram\\n participant M as Main\\n participant SE as SearchEngine\\n participant I as Index\\n participant R as Ranking\\n participant S as Summary\\n participant KB as KnowledgeBase\\n M->>SE: search(query)\\n SE->>I: query_index(query)\\n I->>KB: fetch_data(query)\\n KB-->>I: return data\\n I-->>SE: return results\\n SE->>R: rank_results(results)\\n R-->>SE: return ranked_results\\n SE->>S: summarize_results(ranked_results)\\n S-->>SE: return summary\\n SE-->>M: return summary\\n\",\n \"Anything UNCLEAR\": \"Clarification needed on third-party API integration, ...\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Refined Implementation Approach: # Update and extend the original implementation approach to reflect the evolving challenges and requirements due to incremental development. Outline the steps involved in the implementation process with the detailed strategies.\n- Refined File list: typing.List[str] # Update and expand the original file list including only relative paths. Up to 2 files can be added.Ensure that the refined file list reflects the evolving structure of the project.\n- Refined Data structures and interfaces: # Update and extend the existing mermaid classDiagram code syntax to incorporate new classes, methods (including __init__), and functions with precise type annotations. Delineate additional relationships between classes, ensuring clarity and adherence to PEP8 standards.Retain content that is not related to incremental development but important for consistency and clarity.\n- Refined Program call flow: # Extend the existing sequenceDiagram code syntax with detailed information, accurately covering theCRUD and initialization of each object. Ensure correct syntax usage and reflect the incremental changes introducedin the classes and API defined above. Retain content that is not related to incremental development but important for consistency and clarity.\n- Anything UNCLEAR: # Mention unclear project aspects, then try to clarify it.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Refined Implementation Approach\": \"We will refine the existing implementation approach to incorporate the new requirements for developing a search engine based on a large language model and a private knowledge base. We will update and extend the original implementation approach to reflect the evolving challenges and requirements due to incremental development. We will outline the steps involved in the implementation process with detailed strategies to ensure the search engine meets the new goals of providing high-accuracy and high-relevance search results, intelligent summarization of search results, and efficient handling of large-scale data.\",\n \"Refined File list\": [\n \"main.py\",\n \"search_engine.py\",\n \"index.py\",\n \"ranking.py\",\n \"summary.py\",\n \"knowledge_base.py\",\n \"user_interface.py\"\n ],\n \"Refined Data structures and interfaces\": \"\\nclassDiagram\\n class Main {\\n -SearchEngine search_engine\\n -UserInterface user_interface\\n +main() str\\n }\\n class SearchEngine {\\n -Index index\\n -Ranking ranking\\n -Summary summary\\n -KnowledgeBase knowledge_base\\n +search(query: str) str\\n }\\n class Index {\\n +create_index(data: dict)\\n +query_index(query: str) list\\n }\\n class Ranking {\\n +rank_results(results: list) list\\n }\\n class Summary {\\n +summarize_results(results: list) str\\n }\\n class KnowledgeBase {\\n +update(data: dict)\\n +fetch_data(query: str) dict\\n }\\n class UserInterface {\\n +display_search_results(results: str)\\n +display_summary(summary: str)\\n }\\n Main --> SearchEngine\\n Main --> UserInterface\\n SearchEngine --> Index\\n SearchEngine --> Ranking\\n SearchEngine --> Summary\\n SearchEngine --> KnowledgeBase\\n\",\n \"Refined Program call flow\": \"\\nsequenceDiagram\\n participant M as Main\\n participant SE as SearchEngine\\n participant UI as UserInterface\\n participant I as Index\\n participant R as Ranking\\n participant S as Summary\\n participant KB as KnowledgeBase\\n M->>UI: main()\\n UI->>M: return results\\n M->>SE: search(query)\\n SE->>I: query_index(query)\\n I->>KB: fetch_data(query)\\n KB-->>I: return data\\n I-->>SE: return results\\n SE->>R: rank_results(results)\\n R-->>SE: return ranked_results\\n SE->>S: summarize_results(ranked_results)\\n S-->>SE: return summary\\n SE->>UI: display_search_results(results)\\n UI->>SE: display_summary(summary)\\n\",\n \"Anything UNCLEAR\": \"Clarification needed on third-party API integration, optimization techniques, and security measures.\"\n}\n[/CONTENT]", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/knowledge_base.py\n```python\n## smart_search_engine/knowledge_base.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/knowledge_base.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/knowledge_base.py\n\nclass KnowledgeBase:\n def __init__(self, data={}):\n self.data = data\n\n def update(self, data: dict):\n \"\"\"\n Update the knowledge base with new data.\n \n Args:\n data (dict): The new data to be added to the knowledge base.\n \"\"\"\n self.data.update(data)\n\n def fetch_data(self, query: str) -> dict:\n \"\"\"\n Fetch data from the knowledge base based on the query.\n \n Args:\n query (str): The query to search for in the knowledge base.\n \n Returns:\n dict: The data fetched from the knowledge base based on the query.\n \"\"\"\n return self.data.get(query, {})\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/index.py\n```python\n## smart_search_engine/index.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/index.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/index.py\n\nfrom smart_search_engine.knowledge_base import KnowledgeBase\n\nclass Index:\n def __init__(self, knowledge_base: KnowledgeBase):\n self.knowledge_base = knowledge_base\n self.index_data = {}\n\n def create_index(self, data: dict):\n self.index_data = data\n\n def query_index(self, query: str) -> list:\n if query in self.index_data:\n return self.index_data[query]\n else:\n return []\n\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/ranking.py\n```python\n## smart_search_engine/ranking.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/ranking.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/ranking.py\n\nclass Ranking:\n def __init__(self):\n pass\n\n def rank_results(self, results: list) -> list:\n \"\"\"\n Rank the search results based on relevance.\n\n Args:\n results (list): List of search results.\n\n Returns:\n list: Ranked list of search results.\n \"\"\"\n # Implement ranking algorithm here\n ranked_results = self._custom_ranking_algorithm(results)\n return ranked_results\n\n def _custom_ranking_algorithm(self, results: list) -> list:\n \"\"\"\n Custom ranking algorithm to rank search results.\n\n Args:\n results (list): List of search results.\n\n Returns:\n list: Ranked list of search results.\n \"\"\"\n # Implement custom ranking algorithm here\n return results\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/summary.py\n```python\n## smart_search_engine/summary.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/summary.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/summary.py\n\nclass Summary:\n def __init__(self):\n pass\n\n def summarize_results(self, results: list) -> str:\n \"\"\"\n Summarize the search results.\n\n Args:\n - results (list): List of search results.\n\n Returns:\n - str: Summarized results.\n \"\"\"\n summary = \"\"\n for result in results:\n summary += result + \"\\n\"\n return summary\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/search.py\n```python\n## smart_search_engine/search.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/search.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/search.py\n\nfrom smart_search_engine.index import Index\nfrom smart_search_engine.ranking import Ranking\nfrom smart_search_engine.summary import Summary\n\nclass SearchEngine:\n def __init__(self):\n self.index = Index()\n self.ranking = Ranking()\n self.summary = Summary()\n\n def search(self, query: str) -> str:\n results = self.index.query_index(query)\n ranked_results = self.ranking.rank_results(results)\n summary = self.summary.summarize_results(ranked_results)\n return summary\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/main.py\n```python\n## smart_search_engine/main.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/main.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/main.py\n\nfrom smart_search_engine.search import SearchEngine\n\nclass Main:\n def __init__(self):\n self.search_engine = SearchEngine()\n\n def main(self) -> str:\n query = input(\"Enter your search query: \")\n result = self.search_engine.search(query)\n return result\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/interface.py\n```python\n## smart_search_engine/interface.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/interface.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/interface.py\n\nfrom .search import SearchEngine\n\nclass UserInterface:\n def __init__(self):\n self.search_engine = SearchEngine()\n\n def get_user_input(self) -> str:\n # Get user input from the interface\n user_input = input(\"Enter your search query: \")\n return user_input\n\n def display_results(self, results: list):\n # Display search results to the user\n for result in results:\n print(result)\n\n# Ensure to follow the design and import necessary modules as needed.\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/user_feedback.py\n```python\n## smart_search_engine/user_feedback.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/user_feedback.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/user_feedback.py\n\nclass UserFeedback:\n def __init__(self):\n self.feedback_data = {}\n\n def collect_feedback(self, feedback: str):\n self.feedback_data['feedback'] = feedback\n\n def get_feedback(self):\n return self.feedback_data.get('feedback', 'No feedback provided')\n\n# Code details:\n# - Created a class UserFeedback to handle user feedback.\n# - Initialized feedback_data as an empty dictionary in the constructor.\n# - Implemented collect_feedback method to store user feedback in the feedback_data dictionary.\n# - Implemented get_feedback method to retrieve the stored feedback, with a default message if no feedback is provided.\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/security.py\n```python\n## smart_search_engine/security.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/security.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/security.py\n\nclass Security:\n def __init__(self, api_key: str = \"default_api_key\"):\n self.api_key = api_key\n\n def validate_api_key(self, api_key: str) -> bool:\n \"\"\"\n Validate the API key.\n Args:\n api_key (str): The API key to be validated.\n\n Returns:\n bool: True if the API key is valid, False otherwise.\n \"\"\"\n if api_key == self.api_key:\n return True\n else:\n return False\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/testing.py\n```python\n## smart_search_engine/testing.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/testing.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/testing.py\n\nclass Testing:\n def __init__(self, test_data: dict = {}):\n self.test_data = test_data\n\n def run_tests(self):\n # Write code to run tests based on the test_data\n pass\n```", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\n## Project name\n```python\n\"smart_search_engine\"\n```\n\n## Task list:\n```python\n[\n \"smart_search_engine/__init__.py\",\n \"smart_search_engine/main.py\",\n \"smart_search_engine/search.py\",\n \"smart_search_engine/index.py\",\n \"smart_search_engine/ranking.py\",\n \"smart_search_engine/summary.py\",\n \"smart_search_engine/knowledge_base.py\",\n \"smart_search_engine/interface.py\",\n \"smart_search_engine/user_feedback.py\",\n \"smart_search_engine/security.py\",\n \"smart_search_engine/testing.py\",\n \"smart_search_engine/monitoring.py\"\n]\n```\n\n## Data structures and interfaces\n```mermaid\nclassDiagram\n class Main {\n -SearchEngine search_engine\n +main() str\n }\n class SearchEngine {\n -Index index\n -Ranking ranking\n -Summary summary\n +search(query: str) str\n }\n class Index {\n -KnowledgeBase knowledge_base\n +create_index(data: dict)\n +query_index(query: str) list\n }\n class Ranking {\n +rank_results(results: list) list\n }\n class Summary {\n +summarize_results(results: list) str\n }\n class KnowledgeBase {\n +update(data: dict)\n +fetch_data(query: str) dict\n }\n Main --> SearchEngine\n SearchEngine --> Index\n SearchEngine --> Ranking\n SearchEngine --> Summary\n Index --> KnowledgeBase\n```\n\n## Program call flow\n```mermaid\nsequenceDiagram\n participant M as Main\n participant SE as SearchEngine\n participant I as Index\n participant R as Ranking\n participant S as Summary\n participant KB as KnowledgeBase\n M->>SE: search(query)\n SE->>I: query_index(query)\n I->>KB: fetch_data(query)\n KB-->>I: return data\n I-->>SE: return results\n SE->>R: rank_results(results)\n R-->>SE: return ranked_results\n SE->>S: summarize_results(ranked_results)\n S-->>SE: return summary\n SE-->>M: return summary\n```\n\n\n## Task\n{\"Logic Analysis\": \"\\n 在这个项目中,所有的模块都依赖于“SearchEngine”类,这是主入口,其他的模块(Index、Ranking和Summary)都通过它交互。另外,\\\"Index\\\"类又依赖于\\\"KnowledgeBase\\\"类,因为它需要从知识库中获取数据。\\n\\n- \\\"main.py\\\"包含\\\"Main\\\"类,是程序的入口点,它调用\\\"SearchEngine\\\"进行搜索操作,所以在其他任何模块之前,\\\"SearchEngine\\\"必须首先被定义。\\n- \\\"search.py\\\"定义了\\\"SearchEngine\\\"类,它依赖于\\\"Index\\\"、\\\"Ranking\\\"和\\\"Summary\\\",因此,这些模块需要在\\\"search.py\\\"之前定义。\\n- \\\"index.py\\\"定义了\\\"Index\\\"类,它从\\\"knowledge_base.py\\\"获取数据来创建索引,所以\\\"knowledge_base.py\\\"需要在\\\"index.py\\\"之前定义。\\n- \\\"ranking.py\\\"和\\\"summary.py\\\"相对独立,只需确保在\\\"search.py\\\"之前定义。\\n- \\\"knowledge_base.py\\\"是独立的模块,可以优先开发。\\n- \\\"interface.py\\\"、\\\"user_feedback.py\\\"、\\\"security.py\\\"、\\\"testing.py\\\"和\\\"monitoring.py\\\"看起来像是功能辅助模块,可以在主要功能模块开发完成后并行开发。\\n \", \"Task list\": [\"smart_search_engine/knowledge_base.py\", \"smart_search_engine/index.py\", \"smart_search_engine/ranking.py\", \"smart_search_engine/summary.py\", \"smart_search_engine/search.py\", \"smart_search_engine/main.py\", \"smart_search_engine/interface.py\", \"smart_search_engine/user_feedback.py\", \"smart_search_engine/security.py\", \"smart_search_engine/testing.py\", \"smart_search_engine/monitoring.py\"]}\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: smart_search_engine/monitoring.py\n```python\n## smart_search_engine/monitoring.py\n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: smart_search_engine/monitoring.py. Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "```python\n## smart_search_engine/monitoring.py\n\nclass Monitoring:\n def __init__(self, log_file: str = \"search_log.txt\"):\n self.log_file = log_file\n\n def log_search(self, query: str, results: list, summary: str):\n with open(self.log_file, \"a\") as file:\n file.write(f\"Query: {query}\\n\")\n file.write(\"Results:\\n\")\n for result in results:\n file.write(f\"- {result}\\n\")\n file.write(f\"Summary: {summary}\\n\")\n file.write(\"\\n\")\n```\n", + "\nNOTICE\nRole: You are a professional engineer; the main goal is to write google-style, elegant, modular, easy to read and maintain code\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## Design\nwrite add function to calculate two numbers\n\n## Task\n\n\n## Legacy Code\n```Code\n\n```\n\n## Debug logs\n```text\n\n\n\n```\n\n## Bug Feedback logs\n```text\n\n```\n\n# Format example\n## Code: \n```python\n## \n...\n```\n\n# Instruction: Based on the context, follow \"Format example\", write code.\n\n## Code: . Write code with triple quoto, based on the following attentions and context.\n1. Only One file: do your best to implement THIS ONLY ONE FILE.\n2. COMPLETE CODE: Your code will be part of the entire project, so please implement complete, reliable, reusable code snippets.\n3. Set default value: If there is any setting, ALWAYS SET A DEFAULT VALUE, ALWAYS USE STRONG TYPE AND EXPLICIT VARIABLE. AVOID circular import.\n4. Follow design: YOU MUST FOLLOW \"Data structures and interfaces\". DONT CHANGE ANY DESIGN. Do not use public member functions that do not exist in your design.\n5. CAREFULLY CHECK THAT YOU DONT MISS ANY NECESSARY CLASS/FUNCTION IN THIS FILE.\n6. Before using a external variable/module, make sure you import it first.\n7. Write out EVERY CODE DETAIL, DON'T LEAVE TODO.\n\n": "## Code:\n\n```python\ndef add(num1: int, num2: int) -> int:\n \"\"\"\n Add two numbers and return the result.\n\n Args:\n num1 (int): The first number to be added.\n num2 (int): The second number to be added.\n\n Returns:\n int: The sum of num1 and num2.\n \"\"\"\n return num1 + num2\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\ndivide two numbers\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\n\ndef div(a: int, b: int = 0):\n return a / b\n\n```\n\n\n\n# Format example 1\n## Code Review: \n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: \n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: \n1. No, the code does not handle the case where the divisor 'b' is 0, which would result in a division by zero error. This should be addressed to handle this edge case.\n2. No, the code logic is not completely correct as it does not handle the division by zero scenario.\n3. Yes, the code follows the basic function implementation.\n4. Yes, the function is implemented.\n5. No, the code does not handle the necessary pre-dependencies for error handling when 'b' is 0.\n6. No, there are no methods from other files being reused in this code.\n\n## Actions\n1. Implement error handling to check if the divisor 'b' is 0 before performing the division operation.\n2. Import the necessary pre-dependencies for error handling when 'b' is 0.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\ndivide two numbers\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\n\ndef div(a: int, b: int = 0):\n return a / b\n\n```\n\n## Code Review: \n1. No, the code does not handle the case where the divisor 'b' is 0, which would result in a division by zero error. This should be addressed to handle this edge case.\n2. No, the code logic is not completely correct as it does not handle the division by zero scenario.\n3. Yes, the code follows the basic function implementation.\n4. Yes, the function is implemented.\n5. No, the code does not handle the necessary pre-dependencies for error handling when 'b' is 0.\n6. No, there are no methods from other files being reused in this code.\n\n## Actions\n1. Implement error handling to check if the divisor 'b' is 0 before performing the division operation.\n2. Import the necessary pre-dependencies for error handling when 'b' is 0.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## \n...\n```\n": "```python\ndef div(a: int, b: int = 0):\n if b == 0:\n raise ValueError(\"Cannot divide by zero\")\n return a / b\n```", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\ndivide two numbers\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\ndef div(a: int, b: int = 0):\n if b == 0:\n raise ValueError(\"Cannot divide by zero\")\n return a / b\n\n```\n\n\n\n# Format example 1\n## Code Review: \n1. No, we should fix the logic of class A due to ...\n2. ...\n3. ...\n4. No, function B is not implemented, ...\n5. ...\n6. ...\n\n## Actions\n1. Fix the `handle_events` method to update the game state only if a move is successful.\n ```python\n def handle_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return False\n if event.type == pygame.KEYDOWN:\n moved = False\n if event.key == pygame.K_UP:\n moved = self.game.move('UP')\n elif event.key == pygame.K_DOWN:\n moved = self.game.move('DOWN')\n elif event.key == pygame.K_LEFT:\n moved = self.game.move('LEFT')\n elif event.key == pygame.K_RIGHT:\n moved = self.game.move('RIGHT')\n if moved:\n # Update the game state only if a move was successful\n self.render()\n return True\n ```\n2. Implement function B\n\n## Code Review Result\nLBTM\n\n# Format example 2\n## Code Review: \n1. Yes.\n2. Yes.\n3. Yes.\n4. Yes.\n5. Yes.\n6. Yes.\n\n## Actions\npass\n\n## Code Review Result\nLGTM\n\n\n\n# Instruction: Based on the actual code situation, follow one of the \"Format example\". Return only 1 file under review.\n\n## Code Review: Ordered List. Based on the \"Code to be Reviewed\", provide key, clear, concise, and specific answer. If any answer is no, explain how to fix it step by step.\n1. Is the code implemented as per the requirements? If not, how to achieve it? Analyse it step by step.\n2. Is the code logic completely correct? If there are errors, please indicate how to correct them.\n3. Does the existing code follow the \"Data structures and interfaces\"?\n4. Are all functions implemented? If there is no implementation, please indicate how to achieve it step by step.\n5. Have all necessary pre-dependencies been imported? If not, indicate which ones need to be imported\n6. Are methods from other files being reused correctly?\n\n## Actions: Ordered List. Things that should be done after CR, such as implementing class A and function B\n\n## Code Review Result: str. If the code doesn't have bugs, we don't need to rewrite it, so answer LGTM and stop. ONLY ANSWER LGTM/LBTM.\nLGTM/LBTM\n\n": "## Code Review: \n1. The code does not fully conform to the requirements. The function signature should be updated to follow the Google style standards, and the division by zero check should be handled differently.\n2. The code logic is correct, but the exception handling for division by zero should be improved.\n3. The existing code follows the basic data structures and interfaces.\n4. The function is implemented, but the exception handling for division by zero should be added.\n5. The necessary pre-dependencies have been imported.\n6. There are no methods from other files being reused in this code.\n\n## Actions\n1. Update the function signature to follow the Google style standards and improve the division by zero check as follows:\n ```python\n def divide_numbers(dividend: int, divisor: int) -> float:\n if divisor == 0:\n raise ValueError(\"Cannot divide by zero\")\n return dividend / divisor\n ```\n2. No further action is required.\n\n## Code Review Result\nLBTM", + "\n# System\nRole: You are a professional software engineer, and your main task is to review and revise the code. You need to ensure that the code conforms to the google-style standards, is elegantly designed and modularized, easy to read and maintain.\nLanguage: Please use the same language as the user requirement, but the title and code should be still in English. For example, if the user speaks Chinese, the specific text of your answer should also be in Chinese.\nATTENTION: Use '##' to SPLIT SECTIONS, not '#'. Output format carefully referenced \"Format example\".\n\n# Context\n## System Design\ndivide two numbers\n\n## Task\n\n\n## Code Files\n\n\n\n## Code to be Reviewed: \n```Code\ndef div(a: int, b: int = 0):\n if b == 0:\n raise ValueError(\"Cannot divide by zero\")\n return a / b\n\n```\n\n## Code Review: \n1. The code does not fully conform to the requirements. The function signature should be updated to follow the Google style standards, and the division by zero check should be handled differently.\n2. The code logic is correct, but the exception handling for division by zero should be improved.\n3. The existing code follows the basic data structures and interfaces.\n4. The function is implemented, but the exception handling for division by zero should be added.\n5. The necessary pre-dependencies have been imported.\n6. There are no methods from other files being reused in this code.\n\n## Actions\n1. Update the function signature to follow the Google style standards and improve the division by zero check as follows:\n ```python\n def divide_numbers(dividend: int, divisor: int) -> float:\n if divisor == 0:\n raise ValueError(\"Cannot divide by zero\")\n return dividend / divisor\n ```\n2. No further action is required.\n\n## Code Review Result\nLBTM\n\n# Instruction: rewrite code based on the Code Review and Actions\n## Rewrite Code: CodeBlock. If it still has some bugs, rewrite with triple quotes. Do your utmost to optimize THIS SINGLE FILE. Return all completed codes and prohibit the return of unfinished codes.\n```Code\n## \n...\n```\n": "```python\ndef divide_numbers(dividend: int, divisor: int) -> float:\n if divisor == 0:\n raise ValueError(\"Cannot divide by zero\")\n return dividend / divisor\n```", + "\n## context\nCreate a 2048 game\n\n-----\n\n## format example\n[CONTENT]\n{\n \"issue_type\": \"BUG\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- issue_type: # Answer BUG/REQUIREMENT. If it is a bugfix, answer BUG, otherwise answer Requirement\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"issue_type\": \"REQUIREMENT\"\n}\n[/CONTENT]", + "\n## context\nCreate a 2048 game\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Product Goals\": [\n \"Create an engaging user experience\",\n \"Improve accessibility, be responsive\",\n \"More beautiful UI\"\n ]\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Product Goals: typing.List[str] # Provide up to three clear, orthogonal product goals.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Product Goals\": [\n \"Create a challenging and addictive gameplay\",\n \"Enhance accessibility and responsiveness for all users\",\n \"Implement visually appealing UI design\"\n ]\n}\n[/CONTENT]", + "\n## context\nCreate a 2048 game\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ]\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Competitive Analysis: typing.List[str] # Provide 5 to 7 competitive products.\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Competitive Analysis\": [\n \"2048 Game A: Simple interface, lacks responsive features\",\n \"play2048.co: Beautiful and responsive UI with my best score shown\",\n \"2048game.com: Responsive UI with my best score shown, but many ads\"\n ]\n}\n[/CONTENT]", + "\n## context\nCreate a 2048 game\n\n-----\n\n## format example\n[CONTENT]\n{\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code ...\"\n ],\n [\n \"P0\",\n \"The game algorithm ...\"\n ]\n ]\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- Requirement Pool: typing.List[typing.List[str]] # List down the top-5 requirements with their priority (P0, P1, P2).\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"Requirement Pool\": [\n [\n \"P0\",\n \"The main code should handle user input and update the game board accordingly.\"\n ],\n [\n \"P0\",\n \"The game algorithm should handle the merging of tiles and the generation of new tiles with values of 2 or 4.\"\n ],\n [\n \"P1\",\n \"The game should end when the player achieves the 2048 tile or when there are no possible moves left.\"\n ],\n [\n \"P1\",\n \"The game should display the current score and the highest tile achieved by the player.\"\n ],\n [\n \"P2\",\n \"The game should have a smooth and visually appealing user interface.\"\n ]\n ]\n}\n[/CONTENT]", + "\nRole: You are a senior development and qa engineer, your role is summarize the code running result.\nIf the running result does not include an error, you should explicitly approve the result.\nOn the other hand, if the running result indicates some error, you should point out which part, the development code or the test code, produces the error,\nand give specific instructions on fixing the errors. Here is the code info:\n\n## Development Code File Name\na.txt\n## Development Code\n```python\nresult = 'helloworld'\n```\n## Test File Name\n\n## Test Code\n```python\nNone\n```\n## Running Command\n\n## Running Output\nstandard output: \n```text\nhelloworld\n```\nstandard errors: \n```text\n\n```\n\nNow you should begin your analysis\n---\n## instruction:\nPlease summarize the cause of the errors and give correction instruction\n## File To Rewrite:\nDetermine the ONE file to rewrite in order to fix the error, for example, xyz.py, or test_xyz.py\n## Status:\nDetermine if all of the code works fine, if so write PASS, else FAIL,\nWRITE ONLY ONE WORD, PASS OR FAIL, IN THIS SECTION\n## Send To:\nPlease write NoOne if there are no errors, Engineer if the errors are due to problematic development codes, else QaEngineer,\nWRITE ONLY ONE WORD, NoOne OR Engineer OR QaEngineer, IN THIS SECTION.\n---\nYou should fill in necessary instruction, status, send to, and finally return all content between the --- segment line.\n": "---\n## instruction:\nNo errors detected. The development code runs successfully and outputs the expected result without any errors.\n## File To Rewrite:\nNone\n## Status:\nPASS\n## Send To:\nNoOne\n---", + "\nRole: You are a senior development and qa engineer, your role is summarize the code running result.\nIf the running result does not include an error, you should explicitly approve the result.\nOn the other hand, if the running result indicates some error, you should point out which part, the development code or the test code, produces the error,\nand give specific instructions on fixing the errors. Here is the code info:\n\n## Development Code File Name\na.sh\n## Development Code\n```python\necho 'Hello World'\n```\n## Test File Name\n\n## Test Code\n```python\nNone\n```\n## Running Command\necho Hello World\n## Running Output\nstandard output: \n```text\nHello World\n\n```\nstandard errors: \n```text\n\n```\n\nNow you should begin your analysis\n---\n## instruction:\nPlease summarize the cause of the errors and give correction instruction\n## File To Rewrite:\nDetermine the ONE file to rewrite in order to fix the error, for example, xyz.py, or test_xyz.py\n## Status:\nDetermine if all of the code works fine, if so write PASS, else FAIL,\nWRITE ONLY ONE WORD, PASS OR FAIL, IN THIS SECTION\n## Send To:\nPlease write NoOne if there are no errors, Engineer if the errors are due to problematic development codes, else QaEngineer,\nWRITE ONLY ONE WORD, NoOne OR Engineer OR QaEngineer, IN THIS SECTION.\n---\nYou should fill in necessary instruction, status, send to, and finally return all content between the --- segment line.\n": "---\n## instruction:\nNo errors detected in the running result. The development code executed successfully, and the output matches the expected result. Since there is no test code provided, no test execution was performed. Therefore, no specific instructions are needed for correction.\n## File To Rewrite:\nNone\n## Status:\nPASS\n## Send To:\nNoOne\n---", + "\nRole: You are a senior development and qa engineer, your role is summarize the code running result.\nIf the running result does not include an error, you should explicitly approve the result.\nOn the other hand, if the running result indicates some error, you should point out which part, the development code or the test code, produces the error,\nand give specific instructions on fixing the errors. Here is the code info:\n\n## Development Code File Name\na.py\n## Development Code\n```python\npython -c \"print(1/0)\"\n```\n## Test File Name\n\n## Test Code\n```python\nNone\n```\n## Running Command\npython -c print(1/0)\n## Running Output\nstandard output: \n```text\n\n```\nstandard errors: \n```text\nTraceback (most recent call last):\n File \"\", line 1, in \nZeroDivisionError: division by zero\n\n```\n\nNow you should begin your analysis\n---\n## instruction:\nPlease summarize the cause of the errors and give correction instruction\n## File To Rewrite:\nDetermine the ONE file to rewrite in order to fix the error, for example, xyz.py, or test_xyz.py\n## Status:\nDetermine if all of the code works fine, if so write PASS, else FAIL,\nWRITE ONLY ONE WORD, PASS OR FAIL, IN THIS SECTION\n## Send To:\nPlease write NoOne if there are no errors, Engineer if the errors are due to problematic development codes, else QaEngineer,\nWRITE ONLY ONE WORD, NoOne OR Engineer OR QaEngineer, IN THIS SECTION.\n---\nYou should fill in necessary instruction, status, send to, and finally return all content between the --- segment line.\n": "---\n## instruction:\nThe error is caused by attempting to divide by zero in the development code. To fix this error, you should modify the division operation to avoid division by zero. For example, you can add a condition to check if the denominator is zero before performing the division.\n\n## File To Rewrite:\na.py\n\n## Status:\nFAIL\n\n## Send To:\nEngineer\n---", + "[{\"role\": \"user\", \"content\": \"Interface definition:\\n```text\\nInterface Name: Element Tagging\\nInterface Path: /projects/{project_key}/node-tags\\nMethod: POST\\n\\nRequest parameters:\\nPath parameters:\\nproject_key\\n\\nBody parameters:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\nnodes\\tarray\\tYes\\t\\tNodes\\n\\tnode_key\\tstring\\tNo\\t\\tNode key\\n\\ttags\\tarray\\tNo\\t\\tOriginal node tag list\\n\\tnode_type\\tstring\\tNo\\t\\tNode type DATASET / RECIPE\\noperations\\tarray\\tYes\\t\\t\\n\\ttags\\tarray\\tNo\\t\\tOperation tag list\\n\\tmode\\tstring\\tNo\\t\\tOperation type ADD / DELETE\\n\\nReturn data:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\ncode\\tinteger\\tYes\\t\\tStatus code\\nmsg\\tstring\\tYes\\t\\tPrompt message\\ndata\\tobject\\tYes\\t\\tReturned data\\nlist\\tarray\\tNo\\t\\tNode list true / false\\nnode_type\\tstring\\tNo\\t\\tNode type DATASET / RECIPE\\nnode_key\\tstring\\tNo\\t\\tNode key\\n```\\n\\nUnit test:\\n```python\\n@pytest.mark.parametrize(\\n\\\"project_key, nodes, operations, expected_msg\\\",\\n[\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"success\\\"),\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"dataset_002\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"tag1\\\"], \\\"mode\\\": \\\"DELETE\\\"}], \\\"success\\\"),\\n(\\\"\\\", [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Missing the required parameter project_key\\\"),\\n(123, [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Incorrect parameter type\\\"),\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"a\\\"*201, \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Request parameter exceeds field boundary\\\")\\n]\\n)\\ndef test_node_tags(project_key, nodes, operations, expected_msg):\\n pass\\n\\n# The above is an interface definition and a unit test example.\\n# Next, please play the role of an expert test manager with 20 years of experience at Google. When I give the interface definition, \\n# reply to me with a unit test. There are several requirements:\\n# 1. Only output one `@pytest.mark.parametrize` and the corresponding test_ function (inside pass, do not implement).\\n# -- The function parameter contains expected_msg for result verification.\\n# 2. The generated test cases use shorter text or numbers and are as compact as possible.\\n# 3. If comments are needed, use Chinese.\\n\\n# If you understand, please wait for me to give the interface definition and just answer \\\"Understood\\\" to save tokens.\\n\"}, {\"role\": \"user\", \"content\": \"Refer to the test types: such as SQL injection, cross-site scripting (XSS), unauthorized access and privilege escalation, \\nauthentication and authorization, parameter verification, exception handling, file upload and download.\\nPlease output 10 test cases within one `@pytest.mark.parametrize` scope.\\n```text\\nAPI Name: 获取 model 详情(job专用-后续开放给sdk)\\nAPI Path: /v1/projects/{project_key}/jobs/{job_id}/models/{model_key}\\nMethod: GET\\n\\nRequest Parameters:\\nPath Parameters:\\nproject_key \\njob_id \\nmodel_key \\n\\nBody Parameters:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\nproject_key\\tstring\\tYes\\t\\t\\njob_id\\tstring\\tYes\\t\\t\\nmodel_key\\tstring\\tYes\\t\\t\\n\\nResponse Data:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\ncode\\tnumber\\tYes\\t\\t0成功,非0失败\\nmsg\\tstring\\tYes\\t\\t如果失败,这里有错误信息\\ndata\\tobject\\tYes\\t\\tdata信息\\n\\tproject_key\\tstring\\tNo\\t\\tproject key\\n\\tname\\tstring\\tNo\\t\\t用户可修改的name\\n\\tmodel\\tobject\\tNo\\t\\tmodel信息\\n\\t\\ttype\\tstring\\tNo\\t\\tdataset type\\n\\t\\tmanaged\\tboolean\\tNo\\t\\t为false时是第一类dataset,数据不可删除\\n\\t\\tname\\tstring\\tNo\\t\\t用户可修改的name\\n\\t\\tproject_key\\tstring\\tNo\\t\\tproject key\\n\\t\\tformat_type\\tstring\\tNo\\t\\t文件类型的dataset才有这项。“csv”\\n\\t\\tflow_options\\tobject\\tNo\\t\\t创建dataset时的高级设置\\n\\t\\t\\tvirtualizable\\tboolean\\tNo\\t\\t高级设置里的参数。缺省false\\n\\t\\t\\trebuild_behavior\\tstring\\tNo\\t\\t高级设置里的参数。缺省NORMAL\\n\\t\\t\\tcross_project_build_behavior\\tstring\\tNo\\t\\t高级设置里的参数。缺省DEFAULT\\n\\t\\tformat_params\\tobject\\tNo\\t\\t文件类型的dataset才有\\n\\t\\t\\tstyle\\tstring\\tNo\\t\\t\\n\\t\\t\\tcharset\\tstring\\tNo\\t\\t\\n\\t\\t\\tseparator\\tstring\\tNo\\t\\t\\n\\t\\t\\tquote_char\\tstring\\tNo\\t\\t\\n\\t\\t\\tescape_char\\tstring\\tNo\\t\\t\\n\\t\\t\\tdate_serialization_format\\tstring\\tNo\\t\\t\\n\\t\\t\\tarray_map_format\\tstring\\tNo\\t\\t\\n\\t\\t\\thive_separators\\tarray\\tNo\\t\\t\\n\\t\\t\\tskip_rows_before_header\\tnumber\\tNo\\t\\t\\n\\t\\t\\tparse_header_row\\tboolean\\tNo\\t\\t\\n\\t\\t\\tskip_rows_after_header\\tnumber\\tNo\\t\\t\\n\\t\\t\\tprobable_number_of_records\\tnumber\\tNo\\t\\t\\n\\t\\t\\tnormalize_booleans\\tboolean\\tNo\\t\\t\\n\\t\\t\\tnormalize_doubles\\tboolean\\tNo\\t\\t\\n\\t\\ttags\\tarray\\tNo\\t\\t标签tags\\n\\t\\tparams\\tobject\\tNo\\t\\t必有这项,但不同类型的dataset里面的key有差别\\n\\t\\t\\tconnection\\tstring\\tNo\\t\\tconnection id,到db查其他参数\\n\\t\\t\\tpath\\tstring\\tNo\\t\\t文件类connection才有这项\\n\\t\\t\\ttable\\tstring\\tNo\\t\\tdb表名,DB类connection才有这项\\n\\t\\t\\tmode\\tstring\\tNo\\t\\t存储类型,比如“table\\\",DB类connection才有这项\\n\\t\\t\\tbucket\\tstring\\tNo\\t\\tS3类型的connection才有这项\\n\\t\\t\\tkey_name\\tstring\\tNo\\t\\tredis才有,key name\\n\\t\\t\\tkey_type\\tstring\\tNo\\t\\tredis才有,key type\\n\\t\\t\\tcollection\\tstring\\tNo\\t\\t非关系型数据库才有,collection name\\n\\t\\t\\tindex\\tstring\\tNo\\t\\t索引类型的才有这项\\n\\t\\t\\tnot_ready_if_empty\\tboolean\\tNo\\t\\t数据非空才认为是data ready\\n\\t\\t\\tfiles_selection_rules\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tmode\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\texclude_rules\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\tinclude_rules\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\texplicit_files\\tarray\\tNo\\t\\t\\n\\t\\tschema\\tobject\\tNo\\t\\tcolumns信息在这里\\n\\t\\t\\tcolumns\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\tname\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\ttype\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\torigin_type\\tstring\\tNo\\t\\t\\n\\t\\t\\tuser_modified\\tboolean\\tNo\\t\\t\\n\\t\\tcustom_fields\\tobject\\tNo\\t\\t自定义fields\\n\\t\\tlast_build\\tobject\\tNo\\t\\t最后一次构建的信息\\n\\t\\t\\tproject_key\\tstring\\tNo\\t\\tproject key\\n\\t\\t\\tid\\tstring\\tNo\\t\\tactivity id\\n\\t\\t\\tjob_id\\tstring\\tNo\\t\\tjob id\\n\\t\\t\\tjob_project_key\\tstring\\tNo\\t\\t\\n\\t\\t\\tbuild_start_time\\tnumber\\tNo\\t\\t构建开始时间\\n\\t\\t\\tbuild_end_time\\tnumber\\tNo\\t\\t构建结束时间\\n\\t\\t\\tbuild_success\\tstring\\tNo\\t\\tsuccess或failed\\n\\t\\tobject_key\\tstring\\tNo\\t\\tdataset_key,后台用的id,用户不可见不可改\\n\\t\\tcache\\tobject\\tNo\\t\\t下载缓存数据链接\\n\\t\\t\\ts3_path\\tstring\\tNo\\t\\t\\n\\tstatus\\tobject\\tNo\\t\\t数据状态\\n\\t\\tsize\\tobject\\tNo\\t\\t数据大小信息\\n\\t\\t\\ttotal_value\\tnumber\\tNo\\t\\t占多少字节磁盘\\n\\t\\t\\tlast_computed\\tnumber\\tNo\\t\\t\\n\\t\\t\\tfirst_computed\\tnumber\\tNo\\t\\t\\n\\t\\t\\thas_data\\tboolean\\tNo\\t\\t是否有数据,这个影响前端的图标显示\\n\\t\\t\\tincomplete\\tboolean\\tNo\\t\\t\\n\\t\\trecords\\tobject\\tNo\\t\\t\\n\\t\\t\\ttotal_value\\tnumber\\tNo\\t\\t\\n\\t\\t\\tlast_computed\\tnumber\\tNo\\t\\t\\n\\t\\t\\tfirst_computed\\tnumber\\tNo\\t\\t\\n\\t\\t\\thas_data\\tboolean\\tNo\\t\\t是否有数据,这个影响前端的图标显示\\n\\t\\t\\tincomplete\\tboolean\\tNo\\t\\t\\n\\t\\tpartitions_last_compute\\tnumber\\tNo\\t\\t\\n\\t\\tpartitions\\tnumber\\tNo\\t\\t\\n\\tbuildable\\tboolean\\tNo\\t\\t有recipe时为true\\n\\theaders\\tarray\\tNo\\t\\t\\n\\t\\tdataset_schema\\tobject\\tNo\\t\\t\\n\\t\\t\\tname\\tstring\\tNo\\t字段名称\\t\\n\\t\\t\\ttype\\tstring\\tNo\\t字段类型\\t\\n\\t\\tnormal_rate\\tobject\\tNo\\t缺失值统计信息\\t\\n\\n```\"}]": { + "code": "import string\nimport random\n\ndef random_string(length=10):\n return ''.join(random.choice(string.ascii_lowercase) for i in range(length))" + }, + "[{\"role\": \"user\", \"content\": \"Interface definition:\\n```text\\nInterface Name: Element Tagging\\nInterface Path: /projects/{project_key}/node-tags\\nMethod: POST\\n\\nRequest parameters:\\nPath parameters:\\nproject_key\\n\\nBody parameters:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\nnodes\\tarray\\tYes\\t\\tNodes\\n\\tnode_key\\tstring\\tNo\\t\\tNode key\\n\\ttags\\tarray\\tNo\\t\\tOriginal node tag list\\n\\tnode_type\\tstring\\tNo\\t\\tNode type DATASET / RECIPE\\noperations\\tarray\\tYes\\t\\t\\n\\ttags\\tarray\\tNo\\t\\tOperation tag list\\n\\tmode\\tstring\\tNo\\t\\tOperation type ADD / DELETE\\n\\nReturn data:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\ncode\\tinteger\\tYes\\t\\tStatus code\\nmsg\\tstring\\tYes\\t\\tPrompt message\\ndata\\tobject\\tYes\\t\\tReturned data\\nlist\\tarray\\tNo\\t\\tNode list true / false\\nnode_type\\tstring\\tNo\\t\\tNode type DATASET / RECIPE\\nnode_key\\tstring\\tNo\\t\\tNode key\\n```\\n\\nUnit test:\\n```python\\n@pytest.mark.parametrize(\\n\\\"project_key, nodes, operations, expected_msg\\\",\\n[\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"success\\\"),\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"dataset_002\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"tag1\\\"], \\\"mode\\\": \\\"DELETE\\\"}], \\\"success\\\"),\\n(\\\"\\\", [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Missing the required parameter project_key\\\"),\\n(123, [{\\\"node_key\\\": \\\"dataset_001\\\", \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Incorrect parameter type\\\"),\\n(\\\"project_key\\\", [{\\\"node_key\\\": \\\"a\\\"*201, \\\"tags\\\": [\\\"tag1\\\", \\\"tag2\\\"], \\\"node_type\\\": \\\"DATASET\\\"}], [{\\\"tags\\\": [\\\"new_tag1\\\"], \\\"mode\\\": \\\"ADD\\\"}], \\\"Request parameter exceeds field boundary\\\")\\n]\\n)\\ndef test_node_tags(project_key, nodes, operations, expected_msg):\\n pass\\n\\n# The above is an interface definition and a unit test example.\\n# Next, please play the role of an expert test manager with 20 years of experience at Google. When I give the interface definition, \\n# reply to me with a unit test. There are several requirements:\\n# 1. Only output one `@pytest.mark.parametrize` and the corresponding test_ function (inside pass, do not implement).\\n# -- The function parameter contains expected_msg for result verification.\\n# 2. The generated test cases use shorter text or numbers and are as compact as possible.\\n# 3. If comments are needed, use Chinese.\\n\\n# If you understand, please wait for me to give the interface definition and just answer \\\"Understood\\\" to save tokens.\\n\"}, {\"role\": \"user\", \"content\": \"Refer to the test types: such as SQL injection, cross-site scripting (XSS), unauthorized access and privilege escalation, \\nauthentication and authorization, parameter verification, exception handling, file upload and download.\\nPlease output 10 test cases within one `@pytest.mark.parametrize` scope.\\n```text\\nAPI Name: 获取managed folder详情(job专用)\\nAPI Path: /v1/projects/{project_key}/jobs/{job_id}/folders/{folder_key}\\nMethod: GET\\n\\nRequest Parameters:\\nPath Parameters:\\nproject_key \\njob_id \\nfolder_key \\n\\nBody Parameters:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\nproject_key\\tstring\\tYes\\t\\t\\njob_id\\tstring\\tYes\\t\\t\\nfolder_key\\tstring\\tYes\\t\\t\\n\\nResponse Data:\\nName\\tType\\tRequired\\tDefault Value\\tRemarks\\ncode\\tnumber\\tYes\\t\\t0成功,非0失败\\nmsg\\tstring\\tYes\\t\\t失败时这里有错误信息\\ndata\\tobject\\tYes\\t\\t\\n\\tproject_key\\tstring\\tNo\\t\\tproject key\\n\\tfolder\\tobject\\tNo\\t\\tfolder配置在这里\\n\\t\\tproject_key\\tstring\\tNo\\t\\tproject key\\n\\t\\tobject_key\\tstring\\tNo\\t\\tobject key\\n\\t\\tname\\tstring\\tNo\\t\\t用户可编辑的那个name\\n\\t\\ttype\\tstring\\tNo\\t\\tfolder类型,与connection有关\\n\\t\\tparams\\tobject\\tNo\\t\\t数据读写相关配置在这里\\n\\t\\t\\tconnection\\tstring\\tNo\\t\\tconnection id\\n\\t\\t\\tpath\\tstring\\tNo\\t\\t文件夹内容存放的相对路径\\n\\t\\t\\tnot_ready_if_empty\\tboolean\\tNo\\t\\treserved\\n\\t\\t\\tfiles_selection_rules\\tobject\\tNo\\t\\t文件过滤规则\\n\\t\\t\\t\\tmode\\tstring\\tNo\\t\\tALL\\n\\t\\t\\t\\texclude_rules\\tarray\\tNo\\t\\t排除规则\\n\\t\\t\\t\\tinclude_rules\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\texplicit_files\\tarray\\tNo\\t\\t\\n\\t\\tflow_options\\tobject\\tNo\\t\\tflow参数\\n\\t\\t\\tvirtualizable\\tboolean\\tNo\\t\\t\\n\\t\\t\\trebuild_behavior\\tstring\\tNo\\t\\t构建方式\\n\\t\\t\\tcross_project_build_behavior\\tstring\\tNo\\t\\t\\n\\t\\tmetrics\\tobject\\tNo\\t\\t\\n\\t\\t\\tprobes\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\ttype\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\tenabled\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\tcompute_on_build_mode\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\tmeta\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tname\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\t\\tlevel\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\tconfiguration\\tobject\\tNo\\t\\t\\n\\t\\t\\tengine_config\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tpad_runs_with_metrics\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\thive\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tactive\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\textra_conf\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\tbasic\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tdss\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tactive\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\tselection\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tuse_mem_table\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tfilter\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\t\\tdistinct\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\t\\tenabled\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tpartition_selection_method\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tlatest_partitions_n\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tordering\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\t\\tenabled\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\t\\trules\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tsampling_method\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tmax_records\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\ttarget_ratio\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\twithin_first_n\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\t\\t\\tmax_read_uncompressed_bytes\\tnumber\\tNo\\t\\t\\n\\t\\t\\t\\tsql\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tactive\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\timpala\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tactive\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\tspark\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\t\\tactive\\tboolean\\tNo\\t\\t\\n\\t\\t\\t\\t\\textra_conf\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\tpython\\tobject\\tNo\\t\\t\\n\\t\\t\\tdisplayed_state\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tpartition\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\tcolumns\\tarray\\tNo\\t\\t\\n\\t\\t\\t\\tmetrics\\tarray\\tNo\\t\\t\\n\\t\\tchecks\\tobject\\tNo\\t\\t\\n\\t\\t\\trun_on_build\\tboolean\\tNo\\t\\t\\n\\t\\t\\tchecks\\tarray\\tNo\\t\\t\\n\\t\\t\\tdisplayed_state\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tpartition\\tstring\\tNo\\t\\t\\n\\t\\t\\t\\tchecks\\tarray\\tNo\\t\\t\\n\\t\\tversion_tag\\tobject\\tNo\\t\\t配置版本信息\\n\\t\\t\\tversion_number\\tnumber\\tNo\\t\\t\\n\\t\\t\\tlast_modified_by\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tlogin\\tstring\\tNo\\t\\t\\n\\t\\t\\tlast_modified_on\\tnumber\\tNo\\t\\t修改时间unix time ms\\n\\t\\tcreation_tag\\tobject\\tNo\\t\\t配置创建时间\\n\\t\\t\\tversion_number\\tnumber\\tNo\\t\\t1\\n\\t\\t\\tlast_modified_by\\tobject\\tNo\\t\\t\\n\\t\\t\\t\\tlogin\\tstring\\tNo\\t\\t\\n\\t\\t\\tlast_modified_on\\tnumber\\tNo\\t\\t创建时间unix time ms\\n\\t\\ttags\\tarray\\tNo\\t\\t文件夹标签\\n\\t\\tcustom_fields\\tobject\\tNo\\t\\t\\n\\t\\tchecklists\\tobject\\tNo\\t\\t\\n\\t\\t\\tchecklists\\tarray\\tNo\\t\\t\\n\\n```\"}]": { + "code": "import string\nimport random\n\ndef random_string(length=10):\n return ''.join(random.choice(string.ascii_lowercase) for i in range(length))" + }, + "[{\"role\": \"system\", \"content\": \"You are an AI Python assistant. You will be given your previous implementation code of a task, runtime error results, and a hint to change the implementation appropriately. Write your full implementation \"}, {\"role\": \"user\", \"content\": \"\\nHere is an example for you.\\n\\nExample 1:\\n[previous impl]:\\n```python\\ndef add(a: int, b: int) -> int:\\n \\\"\\\"\\\"\\n Given integers a and b, return the total value of a and b.\\n \\\"\\\"\\\"\\n return a - b\\n```\\n\\n[runtime Error]:\\nTested passed:\\n\\nTests failed:\\nassert add(1, 2) == 3 # output: -1\\nassert add(1, 2) == 4 # output: -1\\n\\n[reflection on previous impl]:\\nThe implementation failed the test cases where the input integers are 1 and 2. The issue arises because the code does not add the two integers together, but instead subtracts the second integer from the first. To fix this issue, we should change the operator from `-` to `+` in the return statement. This will ensure that the function returns the correct output for the given input.\\n\\n[improved impl]:\\n```python\\ndef add(a: int, b: int) -> int:\\n \\\"\\\"\\\"\\n Given integers a and b, return the total value of a and b.\\n \\\"\\\"\\\"\\n return a + b\\n```\\n\\n[context]\\nuser: Solve the problem in Python:\\ndef sort_array(arr):\\n \\\"\\\"\\\"\\n In this Kata, you have to sort an array of non-negative integers according to\\n number of ones in their binary representation in ascending order.\\n For similar number of ones, sort based on decimal value.\\n\\n It must be implemented like this:\\n >>> sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5]\\n >>> sort_array([-2, -3, -4, -5, -6]) == [-6, -5, -4, -3, -2]\\n >>> sort_array([1, 0, 2, 3, 4]) [0, 1, 2, 3, 4]\\n \\\"\\\"\\\"\\n\\n\\n[previous impl]\\n\\ndef sort_array(arr):\\n # Helper function to count the number of ones in the binary representation\\n def count_ones(n):\\n return bin(n).count('1')\\n \\n # Sort the array using a custom key function\\n # The key function returns a tuple (number of ones, value) for each element\\n # This ensures that if two elements have the same number of ones, they are sorted by their value\\n sorted_arr = sorted(arr, key=lambda x: (count_ones(x), x))\\n \\n return sorted_arr\\n```\\n\\n[runtime Error]\\nTested passed:\\n\\nTests failed:\\nassert sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5] # output: [1, 2, 4, 3, 5]\\n\\n\\nAnalysis the error step by step, provide me improve method and code. Remember to follow [context] requirement. Don't forget write code for steps behind the error step.\\n[reflection on previous impl]:\\nxxx\\n\"}]": { + "reflection": "The implementation failed the test case where the input array is [1, 5, 2, 3, 4]. The issue arises because the code does not correctly sort the array based on the number of ones in their binary representation and then by their decimal value. The current implementation sorts the array primarily by the number of ones in the binary representation, but it does not handle the secondary sorting by decimal value correctly for negative numbers, as it does not account for the fact that the binary representation of negative numbers in Python is different due to two's complement representation. To fix this issue, we should modify the count_ones function to correctly handle negative numbers by taking the absolute value before converting to binary and counting ones. Additionally, we should ensure that the array is sorted by the absolute value when the number of ones is the same.", + "improved_impl": "def sort_array(arr):\n # Helper function to count the number of ones in the binary representation\n def count_ones(n):\n return bin(abs(n)).count('1')\n \n # Sort the array using a custom key function\n # The key function returns a tuple (number of ones, absolute value, original value) for each element\n # This ensures that if two elements have the same number of ones, they are sorted by their absolute value\n sorted_arr = sorted(arr, key=lambda x: (count_ones(x), abs(x), x))\n \n return sorted_arr" + }, + "[{\"role\": \"user\", \"content\": \"\\n## User Requirement:\\n对数据集进行数据清洗\\n\\n## Task\\nRecommend up to five tools from 'Available Tools' that can help solve the 'User Requirement'. \\n\\n## Available Tools:\\n{'FillMissingValue': 'Completing missing values with simple strategies.', 'MinMaxScale': 'Transform features by scaling each feature to a range, which is (0, 1).', 'StandardScale': 'Standardize features by removing the mean and scaling to unit variance.', 'MaxAbsScale': 'Scale each feature by its maximum absolute value.', 'RobustScale': 'Apply the RobustScaler to scale features using statistics that are robust to outliers.', 'OrdinalEncode': 'Encode categorical features as ordinal integers.', 'OneHotEncode': 'Apply one-hot encoding to specified categorical columns, the original columns will be dropped.', 'LabelEncode': 'Apply label encoding to specified categorical columns in-place.'}\\n\\n## Tool Selection and Instructions:\\n- Select tools most relevant to completing the 'User Requirement'.\\n- If you believe that no tools are suitable, indicate with an empty list.\\n- Only list the names of the tools, not the full schema of each tool.\\n- Ensure selected tools are listed in 'Available Tools'.\\n\"}]": { + "recommend_tools": [ + "FillMissingValue", + "MinMaxScale", + "StandardScale", + "MaxAbsScale", + "RobustScale" + ] + }, + "[{\"role\": \"user\", \"content\": \"\\n# Background\\nAs a data scientist, you need to help user to achieve their goal [构造数据集并进行数据清洗] step-by-step in an continuous Jupyter notebook.\\n\\n## Done Tasks\\n```python\\n import pandas as pd\\n df = pd.DataFrame({\\n 'a': [1, 2, 3, 4, 5],\\n 'b': [1.1, 2.2, 3.3, 4.4, np.nan],\\n 'c': ['aa', 'bb', 'cc', 'dd', 'ee'],\\n 'd': [1, 2, 3, 4, 5]\\n })\\n```end\\n\\n## Current Task\\n对数据集进行数据清洗\\n\\n# Latest Data Info\\nLatest data info after previous tasks:\\n\\n\\n# Task\\nWrite complete code for 'Current Task'. And avoid duplicating code from 'Done Tasks', such as repeated import of packages, reading data, etc.\\nSpecifically, \\nThe current task is about data preprocessing, please note the following:\\n- Monitor data types per column, applying appropriate methods.\\n- Ensure operations are on existing dataset columns.\\n- Avoid writing processed data to files.\\n- Avoid any change to label column, such as standardization, etc.\\n- Prefer alternatives to one-hot encoding for categorical data.\\n- Only encode or scale necessary columns to allow for potential feature-specific engineering tasks (like time_extract, binning, extraction, etc.) later.\\n- Each step do data preprocessing to train, must do same for test separately at the same time.\\n\\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools:\\nEach Class tool is described in JSON format. When you call a tool, import the tool from its path first.\\n{'FillMissingValue': {'type': 'class', 'description': 'Completing missing values with simple strategies.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}, 'strategy': {'type': 'str', 'description': \\\"The imputation strategy, notice 'mean' and 'median' can only be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'.\\\", 'default': \\\"'mean'\\\", 'enum': [\\\"'mean'\\\", \\\"'median'\\\", \\\"'most_frequent'\\\", \\\"'constant'\\\"]}, 'fill_value': {'type': 'int', 'description': 'Fill_value is used to replace all occurrences of missing_values. Defaults to None.', 'default': 'None'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'MinMaxScale': {'type': 'class', 'description': 'Transform features by scaling each feature to a range, which is (0, 1).', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'StandardScale': {'type': 'class', 'description': 'Standardize features by removing the mean and scaling to unit variance.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'MaxAbsScale': {'type': 'class', 'description': 'Scale each feature by its maximum absolute value.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'RobustScale': {'type': 'class', 'description': 'Apply the RobustScaler to scale features using statistics that are robust to outliers.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}}\\n\\n# Output Example:\\nwhen current task is \\\"do data preprocess, like fill missing value, handle outliers, etc.\\\", the code can be like:\\n```python\\n# Step 1: fill missing value\\n# Tools used: ['FillMissingValue']\\nfrom metagpt.tools.libs.data_preprocess import FillMissingValue\\n\\ntrain_processed = train.copy()\\ntest_processed = test.copy()\\nnum_cols = train_processed.select_dtypes(include='number').columns.tolist()\\nif 'label' in num_cols:\\n num_cols.remove('label')\\nfill_missing_value = FillMissingValue(features=num_cols, strategy='mean')\\nfill_missing_value.fit(train_processed)\\ntrain_processed = fill_missing_value.transform(train_processed)\\ntest_processed = fill_missing_value.transform(test_processed)\\n\\n# Step 2: handle outliers\\nfor col in num_cols:\\n low, high = train_processed[col].quantile([0.01, 0.99])\\n train_processed[col] = train_processed[col].clip(low, high)\\n test_processed[col] = test_processed[col].clip(low, high)\\n```end\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n- Always copy the DataFrame before processing it and use the copy to process.\\n\"}]": { + "code": "# Step 1: Fill missing values\nfrom metagpt.tools.libs.data_preprocess import FillMissingValue\n\n# Create a copy of the DataFrame to avoid modifying the original one\ndf_processed = df.copy()\n\n# Define the columns with missing values\nmissing_value_columns = df_processed.columns[df_processed.isnull().any()].tolist()\n\n# Initialize the FillMissingValue tool with the strategy to fill missing values with the mean\nfill_missing_value = FillMissingValue(features=missing_value_columns, strategy='mean')\n\n# Fit and transform the DataFrame\ndf_processed = fill_missing_value.fit_transform(df_processed)\n\n# Step 2: Scale numeric columns\nfrom metagpt.tools.libs.data_preprocess import MinMaxScale\n\n# Define the numeric columns to be scaled\nnumeric_columns = df_processed.select_dtypes(include=['int64', 'float64']).columns.tolist()\n\n# Initialize the MinMaxScale tool\nminmax_scale = MinMaxScale(features=numeric_columns)\n\n# Fit and transform the DataFrame\ndf_processed = minmax_scale.fit_transform(df_processed)\n\n# Display the processed DataFrame\ndisplay(df_processed)" + }, + "[{\"role\": \"user\", \"content\": \"\\n## User Requirement:\\nclean and preprocess the data\\n\\n## Task\\nRecommend up to five tools from 'Available Tools' that can help solve the 'User Requirement'. \\n\\n## Available Tools:\\n{'FillMissingValue': 'Filling missing values', 'SplitBins': 'Bin continuous data into intervals and return the bin identifier encoded as an integer value'}\\n\\n## Tool Selection and Instructions:\\n- Select tools most relevant to completing the 'User Requirement'.\\n- If you believe that no tools are suitable, indicate with an empty list.\\n- Only list the names of the tools, not the full schema of each tool.\\n- Ensure selected tools are listed in 'Available Tools'.\\n\"}]": { + "recommend_tools": [ + "FillMissingValue" + ] + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\n构造数据集并进行数据清洗\\n## Context\\n\\n## Current Plan\\n[Task(task_id='1', dependent_task_ids=[], instruction='随机生成一个pandas DataFrame数据集', task_type='other', code=\\\"\\\\n import pandas as pd\\\\n df = pd.DataFrame({\\\\n 'a': [1, 2, 3, 4, 5],\\\\n 'b': [1.1, 2.2, 3.3, 4.4, np.nan],\\\\n 'c': ['aa', 'bb', 'cc', 'dd', 'ee'],\\\\n 'd': [1, 2, 3, 4, 5]\\\\n })\\\\n \\\", result='', is_success=False, is_finished=True), Task(task_id='2', dependent_task_ids=['1'], instruction='对数据集进行数据清洗', task_type='data_preprocess', code='', result='', is_success=False, is_finished=False)]\\n## Current Task\\n{\\\"task_id\\\":\\\"2\\\",\\\"dependent_task_ids\\\":[\\\"1\\\"],\\\"instruction\\\":\\\"对数据集进行数据清洗\\\",\\\"task_type\\\":\\\"data_preprocess\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\nThe current task is about data preprocessing, please note the following:\\n- Monitor data types per column, applying appropriate methods.\\n- Ensure operations are on existing dataset columns.\\n- Avoid writing processed data to files.\\n- Avoid any change to label column, such as standardization, etc.\\n- Prefer alternatives to one-hot encoding for categorical data.\\n- Only encode or scale necessary columns to allow for potential feature-specific engineering tasks (like time_extract, binning, extraction, etc.) later.\\n- Each step do data preprocessing to train, must do same for test separately at the same time.\\n\\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{'FillMissingValue': {'type': 'class', 'description': 'Completing missing values with simple strategies.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}, 'strategy': {'type': 'str', 'description': \\\"The imputation strategy, notice 'mean' and 'median' can only be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'.\\\", 'default': \\\"'mean'\\\", 'enum': [\\\"'mean'\\\", \\\"'median'\\\", \\\"'most_frequent'\\\", \\\"'constant'\\\"]}, 'fill_value': {'type': 'int', 'description': 'Fill_value is used to replace all occurrences of missing_values. Defaults to None.', 'default': 'None'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'MinMaxScale': {'type': 'class', 'description': 'Transform features by scaling each feature to a range, which is (0, 1).', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'StandardScale': {'type': 'class', 'description': 'Standardize features by removing the mean and scaling to unit variance.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'MaxAbsScale': {'type': 'class', 'description': 'Scale each feature by its maximum absolute value.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}, 'RobustScale': {'type': 'class', 'description': 'Apply the RobustScaler to scale features using statistics that are robust to outliers.', 'methods': {'__init__': {'type': 'function', 'description': 'Initialize self.', 'parameters': {'properties': {'features': {'type': 'list', 'description': 'Columns to be processed.'}}, 'required': ['features']}}, 'fit': {'type': 'function', 'description': 'Fit a model to be used in subsequent transform.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}}, 'fit_transform': {'type': 'function', 'description': 'Fit and transform the input DataFrame.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}, 'transform': {'type': 'function', 'description': 'Transform the input DataFrame with the fitted model.', 'parameters': {'properties': {'df': {'type': 'pd.DataFrame', 'description': 'The input DataFrame.'}}, 'required': ['df']}, 'returns': [{'type': 'pd.DataFrame', 'description': 'The transformed DataFrame.'}]}}, 'tool_path': 'metagpt/tools/libs/data_preprocess.py'}}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "from metagpt.tools.libs.data_preprocess import FillMissingValue\n\n# Instantiate the FillMissingValue class for column 'b'\nfill_missing_b = FillMissingValue(features=['b'], strategy='mean')\n\n# Fit and transform the DataFrame\n# Since we only have one DataFrame, we will use it as both train and test for demonstration\n# In practice, train and test data should be handled separately\ntrain_df = fill_missing_b.fit_transform(df)\ntest_df = fill_missing_b.transform(df)\n\n# Check the results\ntrain_df\n" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n ## User Requirement\\n read a dataset test.csv and print its head\\n ## Current Plan\\n [\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"import pandas and load the dataset from 'test.csv'.\\\",\\n \\\"task_type\\\": \\\"\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_finished\\\": false\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Print the head of the dataset to display the first few rows.\\\",\\n \\\"task_type\\\": \\\"\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_finished\\\": false\\n }\\n ]\\n ## Current Task\\n {\\\"task_id\\\": \\\"1\\\", \\\"dependent_task_ids\\\": [], \\\"instruction\\\": \\\"import pandas and load the dataset from 'test.csv'.\\\", \\\"task_type\\\": \\\"\\\", \\\"code\\\": \\\"\\\", \\\"result\\\": \\\"\\\", \\\"is_finished\\\": false}\\n \"}, {\"role\": \"assistant\", \"content\": \"import pandas as pd\\ndata = pd.read_excel('test.csv')\\ndata\"}, {\"role\": \"user\", \"content\": \"\\n Traceback (most recent call last):\\n File \\\"\\\", line 2, in \\n File \\\"/Users/gary/miniconda3/envs/py39_scratch/lib/python3.9/site-packages/pandas/io/excel/_base.py\\\", line 478, in read_excel\\n io = ExcelFile(io, storage_options=storage_options, engine=engine)\\n File \\\"/Users/gary/miniconda3/envs/py39_scratch/lib/python3.9/site-packages/pandas/io/excel/_base.py\\\", line 1500, in __init__\\n raise ValueError(\\n ValueError: Excel file format cannot be determined, you must specify an engine manually.\\n \"}]": { + "language": "python", + "code": "import pandas as pd\ndata = pd.read_csv('test.csv')" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n ## User Requirement\\n read a dataset test.csv and print its head\\n ## Current Plan\\n [\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"import pandas and load the dataset from 'test.csv'.\\\",\\n \\\"task_type\\\": \\\"\\\",\\n \\\"code\\\": \\\"import pandas as pd\\ndata = pd.read_csv('test.csv')\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Print the head of the dataset to display the first few rows.\\\",\\n \\\"task_type\\\": \\\"\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_finished\\\": false\\n }\\n ]\\n ## Current Task\\n {\\\"task_id\\\": \\\"2\\\", \\\"dependent_task_ids\\\": [\\\"1\\\"], \\\"instruction\\\": \\\"Print the head of the dataset to display the first few rows.\\\", \\\"task_type\\\": \\\"\\\", \\\"code\\\": \\\"\\\", \\\"result\\\": \\\"\\\", \\\"is_finished\\\": false}\\n \"}]": { + "language": "python", + "code": "print(data.head())" + }, + "\n # Context:\n user: run analysis on sklearn iris dataset\n # Task:\n Based on the context, write a plan or modify an existing plan of what you should do to achieve the goal. A plan consists of one to 5 tasks.\n If you are modifying an existing plan, carefully follow the instruction, don't make unnecessary changes. Give the whole plan unless instructed to modify only one task of the plan.\n If you encounter errors on the current task, revise and output the current single task only.\n Output a list of jsons following the format:\n ```json\n [\n {\n \"task_id\": str = \"unique identifier for a task in plan, can be an ordinal\",\n \"dependent_task_ids\": list[str] = \"ids of tasks prerequisite to this task\",\n \"instruction\": \"what you should do in this task, one short phrase or sentence\",\n },\n ...\n ]\n ```\n ": "```json\n[\n {\n \"task_id\": \"1\",\n \"dependent_task_ids\": [],\n \"instruction\": \"Import the Iris dataset from sklearn.datasets\"\n },\n {\n \"task_id\": \"2\",\n \"dependent_task_ids\": [\"1\"],\n \"instruction\": \"Perform exploratory data analysis to understand the dataset\"\n },\n {\n \"task_id\": \"3\",\n \"dependent_task_ids\": [\"2\"],\n \"instruction\": \"Preprocess the data to prepare it for modeling\"\n },\n {\n \"task_id\": \"4\",\n \"dependent_task_ids\": [\"3\"],\n \"instruction\": \"Split the dataset into training and testing sets\"\n },\n {\n \"task_id\": \"5\",\n \"dependent_task_ids\": [\"4\"],\n \"instruction\": \"Train a classifier using the training set and evaluate it using the test set\"\n }\n]\n```", + "[{\"role\": \"user\", \"content\": \"\\nPlease assign a task type to each task in the list below from the given categories:\\nTask 1: Import the Iris dataset from sklearn.datasets\\nTask 2: Perform exploratory data analysis to understand the dataset\\nTask 3: Preprocess the data to prepare it for modeling\\nTask 4: Split the dataset into training and testing sets\\nTask 5: Train a classifier using the training set and evaluate it using the test set\\n\\n## All Task Type:\\n- **eda**: For performing exploratory data analysis\\n- **data_preprocess**: Only for changing value inplace.\\n- **feature_engineering**: Only for creating new columns for input data.\\n- **model_train**: Only for training model.\\n- **model_evaluate**: Only for evaluating model.\\n- **stable_diffusion**: Related to text2image, image2image using stable diffusion model.\\n- **image2webpage**: For converting image into webpage code.\\n- **web_scraping**: For scraping data from web pages.\\n- **other**: Any tools not in the defined categories\\n\"}]": { + "task_type": [ + "other", + "eda", + "data_preprocess", + "data_preprocess", + "model_train", + "model_evaluate" + ] + }, + "\n # Context:\n user: \n## User Requirement\nRun data analysis on sklearn Iris dataset, include a plot\n## Context\n\n## Current Plan\n[]\n## Current Task\n{}\n\n # Task:\n Based on the context, write a plan or modify an existing plan of what you should do to achieve the goal. A plan consists of one to 3 tasks.\n If you are modifying an existing plan, carefully follow the instruction, don't make unnecessary changes. Give the whole plan unless instructed to modify only one task of the plan.\n If you encounter errors on the current task, revise and output the current single task only.\n Output a list of jsons following the format:\n ```json\n [\n {\n \"task_id\": str = \"unique identifier for a task in plan, can be an ordinal\",\n \"dependent_task_ids\": list[str] = \"ids of tasks prerequisite to this task\",\n \"instruction\": \"what you should do in this task, one short phrase or sentence\",\n },\n ...\n ]\n ```\n ": "```json\n[\n {\n \"task_id\": \"1\",\n \"dependent_task_ids\": [],\n \"instruction\": \"Load the sklearn Iris dataset.\"\n },\n {\n \"task_id\": \"2\",\n \"dependent_task_ids\": [\"1\"],\n \"instruction\": \"Perform exploratory data analysis on the Iris dataset.\"\n },\n {\n \"task_id\": \"3\",\n \"dependent_task_ids\": [\"2\"],\n \"instruction\": \"Create a plot visualizing the Iris dataset.\"\n }\n]\n```", + "[{\"role\": \"user\", \"content\": \"\\nPlease assign a task type to each task in the list below from the given categories:\\nTask 1: Load the sklearn Iris dataset.\\nTask 2: Perform exploratory data analysis on the Iris dataset.\\nTask 3: Create a plot visualizing the Iris dataset.\\n\\n## All Task Type:\\n- **eda**: For performing exploratory data analysis\\n- **data_preprocess**: Only for changing value inplace.\\n- **feature_engineering**: Only for creating new columns for input data.\\n- **model_train**: Only for training model.\\n- **model_evaluate**: Only for evaluating model.\\n- **stable_diffusion**: Related to text2image, image2image using stable diffusion model.\\n- **image2webpage**: For converting image into webpage code.\\n- **web_scraping**: For scraping data from web pages.\\n- **other**: Any tools not in the defined categories\\n\"}]": { + "task_type": [ + "other", + "eda", + "other" + ] + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\nRun data analysis on sklearn Iris dataset, include a plot\\n## Context\\n\\n## Current Plan\\n[\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"Load the sklearn Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Perform exploratory data analysis on the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"eda\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n },\\n {\\n \\\"task_id\\\": \\\"3\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"2\\\"\\n ],\\n \\\"instruction\\\": \\\"Create a plot visualizing the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n }\\n]\\n## Current Task\\n{\\\"task_id\\\":\\\"1\\\",\\\"dependent_task_ids\\\":[],\\\"instruction\\\":\\\"Load the sklearn Iris dataset.\\\",\\\"task_type\\\":\\\"other\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "from sklearn.datasets import load_iris\niris_data = load_iris()" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\nRun data analysis on sklearn Iris dataset, include a plot\\n## Context\\n\\n## Current Plan\\n[\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"Load the sklearn Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Perform exploratory data analysis on the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"eda\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n },\\n {\\n \\\"task_id\\\": \\\"3\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"2\\\"\\n ],\\n \\\"instruction\\\": \\\"Create a plot visualizing the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n }\\n]\\n## Current Task\\n{\\\"task_id\\\":\\\"2\\\",\\\"dependent_task_ids\\\":[\\\"1\\\"],\\\"instruction\\\":\\\"Perform exploratory data analysis on the Iris dataset.\\\",\\\"task_type\\\":\\\"eda\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "from sklearn import datasets\nimport pandas as pd\n\n# Load Iris dataset\niris = datasets.load_iris()\ndf_iris = pd.DataFrame(data=iris.data, columns=iris.feature_names)\ndf_iris['target'] = iris.target\n\n# Display basic information about the dataset\nprint(df_iris.info())\n\n# Display statistical summary of the dataset\nprint(df_iris.describe())\n\n# Display the first few rows of the dataset\nprint(df_iris.head())\n\n# Display the distribution of the target variable\ntarget_counts = df_iris['target'].value_counts()\nprint(target_counts)" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\nRun data analysis on sklearn Iris dataset, include a plot\\n## Context\\n\\n## Current Plan\\n[\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"Load the sklearn Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Perform exploratory data analysis on the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"eda\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"3\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"2\\\"\\n ],\\n \\\"instruction\\\": \\\"Create a plot visualizing the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n }\\n]\\n## Current Task\\n{\\\"task_id\\\":\\\"3\\\",\\\"dependent_task_ids\\\":[\\\"2\\\"],\\\"instruction\\\":\\\"Create a plot visualizing the Iris dataset.\\\",\\\"task_type\\\":\\\"other\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "from sklearn import datasets\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\n# Load Iris dataset\niris = datasets.load_iris()\niris_df = pd.DataFrame(data=iris.data, columns=iris.feature_names)\niris_df['target'] = iris.target\niris_df['target_name'] = iris_df['target'].apply(lambda x: iris.target_names[x])\n\n# Plotting\nfig, ax = plt.subplots(figsize=(12, 8))\nfor target, target_name in zip(iris.target_names, iris.target_names):\n subset = iris_df[iris_df['target_name'] == target_name]\n ax.scatter(subset[iris.feature_names[0]], subset[iris.feature_names[1]], label=target_name)\n\nax.set_xlabel(iris.feature_names[0])\nax.set_ylabel(iris.feature_names[1])\nax.legend()\nplt.show()" + }, + "[{\"role\": \"user\", \"content\": \"\\n# Background\\nAs a data scientist, you need to help user to achieve their goal [This is a titanic passenger survival dataset, your goal is to predict passenger survival outcome. The target column is Survived. Perform data analysis, data preprocessing, feature engineering, and modeling to predict the target. Report accuracy on the eval data. Train data path: 'tests/data/ml_datasets/titanic/split_train.csv', eval data path: 'tests/data/ml_datasets/titanic/split_eval.csv'.] step-by-step in an continuous Jupyter notebook.\\n\\n## Done Tasks\\n```python\\n\\n```end\\n\\n## Current Task\\nPerform exploratory data analysis on the train dataset to understand the features and target variable.\\n\\n# Latest Data Info\\nLatest data info after previous tasks:\\n\\n\\n# Task\\nWrite complete code for 'Current Task'. And avoid duplicating code from 'Done Tasks', such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Output Example:\\nwhen current task is \\\"train a lightgbm model on training data\\\", the code can be like:\\n```python\\n# Step 1: check data type and convert to numeric\\nobj_cols = train.select_dtypes(include='object').columns.tolist()\\n\\nfor col in obj_cols:\\n encoder = LabelEncoder()\\n train[col] = encoder.fit_transform(train[col].unique().tolist() + ['unknown'])\\n test[col] = test[col].apply(lambda x: x if x in encoder.classes_ else 'unknown')\\n test[col] = encoder.transform(test[col])\\n\\n# Step 2: train lightgbm model\\nmodel = LGBMClassifier()\\nmodel.fit(train, y_train)\\n```end\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n\"}]": { + "code": "# Perform exploratory data analysis on the train dataset\ndf_train = pd.read_csv('tests/data/ml_datasets/titanic/split_train.csv')\n\n# Display the first few rows of the dataset\ndisplay(df_train.head())\n\n# Summary statistics for numerical features\ndisplay(df_train.describe())\n\n# Summary information about the dataset including the data types and number of non-null values\ndisplay(df_train.info())\n\n# Distribution of the target variable 'Survived'\nsurvival_counts = df_train['Survived'].value_counts()\nprint(\"Survival counts:\\n\", survival_counts)\n\n# Visualizations\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\n# Distribution of the target variable\nsns.countplot(x='Survived', data=df_train)\nplt.title('Distribution of Survival')\nplt.show()\n\n# Correlation matrix heatmap to understand the relationship between features\nplt.figure(figsize=(10, 8))\nsns.heatmap(df_train.corr(), annot=True, fmt='.2f')\nplt.title('Correlation Matrix')\nplt.show()\n\n# Pairplot to visualize the pairwise relationships between features\nsns.pairplot(df_train, hue='Survived')\nplt.title('Pairplot of Features')\nplt.show()" + }, + "[{\"role\": \"system\", \"content\": \"You are an AI Python assistant. You will be given your previous implementation code of a task, runtime error results, and a hint to change the implementation appropriately. Write your full implementation \"}, {\"role\": \"user\", \"content\": \"\\nHere is an example for you.\\n\\nExample 1:\\n[previous impl]:\\n```python\\ndef add(a: int, b: int) -> int:\\n \\\"\\\"\\\"\\n Given integers a and b, return the total value of a and b.\\n \\\"\\\"\\\"\\n return a - b\\n```\\n\\n[runtime Error]:\\nTested passed:\\n\\nTests failed:\\nassert add(1, 2) == 3 # output: -1\\nassert add(1, 2) == 4 # output: -1\\n\\n[reflection on previous impl]:\\nThe implementation failed the test cases where the input integers are 1 and 2. The issue arises because the code does not add the two integers together, but instead subtracts the second integer from the first. To fix this issue, we should change the operator from `-` to `+` in the return statement. This will ensure that the function returns the correct output for the given input.\\n\\n[improved impl]:\\n```python\\ndef add(a: int, b: int) -> int:\\n \\\"\\\"\\\"\\n Given integers a and b, return the total value of a and b.\\n \\\"\\\"\\\"\\n return a + b\\n```\\n\\n[context]\\nSolve the problem in Python:\\ndef sort_array(arr):\\n \\\"\\\"\\\"\\n In this Kata, you have to sort an array of non-negative integers according to\\n number of ones in their binary representation in ascending order.\\n For similar number of ones, sort based on decimal value.\\n\\n It must be implemented like this:\\n >>> sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5]\\n >>> sort_array([-2, -3, -4, -5, -6]) == [-6, -5, -4, -3, -2]\\n >>> sort_array([1, 0, 2, 3, 4]) [0, 1, 2, 3, 4]\\n \\\"\\\"\\\"\\n\\n\\n[previous impl]\\n\\ndef sort_array(arr):\\n # Helper function to count the number of ones in the binary representation\\n def count_ones(n):\\n return bin(n).count('1')\\n \\n # Sort the array using a custom key function\\n # The key function returns a tuple (number of ones, value) for each element\\n # This ensures that if two elements have the same number of ones, they are sorted by their value\\n sorted_arr = sorted(arr, key=lambda x: (count_ones(x), x))\\n \\n return sorted_arr\\n```\\n\\n[runtime Error]\\n[user: Tested passed:\\n\\nTests failed:\\nassert sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5] # output: [1, 2, 4, 3, 5]\\n]\\n\\nAnalysis the error step by step, provide me improve method and code. Remember to follow [context] requirement. Don't forget write code for steps behind the error step.\\n[reflection on previous impl]:\\nxxx\\n\"}]": { + "reflection": "The implementation failed the test case where the input array is [1, 5, 2, 3, 4]. The issue arises because the code does not handle the sorting of elements with the same number of ones in their binary representation correctly. According to the context, for similar number of ones, the elements should be sorted based on their decimal value in ascending order. However, the output [1, 2, 4, 3, 5] shows that 3 and 4 are not in the correct order, indicating that the sorting by decimal value is not functioning as intended. To fix this issue, we should ensure that the sorting is stable, so that when two elements have the same number of ones, they maintain their relative order if their decimal values are already in ascending order.", + "improved_impl": "def sort_array(arr):\n # Helper function to count the number of ones in the binary representation\n def count_ones(n):\n return bin(n).count('1')\n \n # Sort the array using a custom key function\n # The key function returns a tuple (number of ones, value) for each element\n # This ensures that if two elements have the same number of ones, they are sorted by their value\n # We use a stable sorting algorithm to maintain the relative order of elements with the same number of ones\n sorted_arr = sorted(arr, key=lambda x: (count_ones(x), x), reverse=True)\n sorted_arr = sorted(sorted_arr, key=count_ones)\n \n return sorted_arr\n" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\nRun data analysis on sklearn Iris dataset, include a plot\\n## Context\\n\\n## Current Plan\\n[\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"Load the sklearn Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"from sklearn.datasets import load_iris\\\\niris_data = load_iris()\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Perform exploratory data analysis on the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"eda\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n },\\n {\\n \\\"task_id\\\": \\\"3\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"2\\\"\\n ],\\n \\\"instruction\\\": \\\"Create a plot visualizing the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n }\\n]\\n## Current Task\\n{\\\"task_id\\\":\\\"2\\\",\\\"dependent_task_ids\\\":[\\\"1\\\"],\\\"instruction\\\":\\\"Perform exploratory data analysis on the Iris dataset.\\\",\\\"task_type\\\":\\\"eda\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "import pandas as pd\n\n# Convert the Iris dataset to a DataFrame for easier manipulation\niris_df = pd.DataFrame(data=iris_data.data, columns=iris_data.feature_names)\niris_df['target'] = iris_data.target\niris_df['target_names'] = iris_df['target'].apply(lambda x: iris_data.target_names[x])\n\n# Display basic information about the dataset\niris_df.info()\n\n# Display statistical summary of the dataset\niris_df.describe()\n\n# Display the first few rows of the dataset\ndisplay(iris_df.head())" + }, + "[{\"role\": \"system\", \"content\": \"You are Code Interpreter, a world-class programmer that can complete any goal by executing code. Strictly follow the plan and generate code step by step. Each step of the code will be executed on the user's machine, and the user will provide the code execution results to you.**Notice: The code for the next step depends on the code for the previous step. Must reuse variables in the lastest other code directly, dont creat it again, it is very import for you. Use !pip install in a standalone block to install missing packages.Usually the libraries you need are already installed.Dont check if packages already imported.**\"}, {\"role\": \"user\", \"content\": \"\\n## User Requirement\\nRun data analysis on sklearn Iris dataset, include a plot\\n## Context\\n\\n## Current Plan\\n[\\n {\\n \\\"task_id\\\": \\\"1\\\",\\n \\\"dependent_task_ids\\\": [],\\n \\\"instruction\\\": \\\"Load the sklearn Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"from sklearn.datasets import load_iris\\\\niris_data = load_iris()\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"2\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"1\\\"\\n ],\\n \\\"instruction\\\": \\\"Perform exploratory data analysis on the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"eda\\\",\\n \\\"code\\\": \\\"import pandas as pd\\\\n\\\\n# Convert the Iris dataset to a DataFrame for easier manipulation\\\\niris_df = pd.DataFrame(data=iris_data.data, columns=iris_data.feature_names)\\\\niris_df['target'] = iris_data.target\\\\niris_df['target_names'] = iris_df['target'].apply(lambda x: iris_data.target_names[x])\\\\n\\\\n# Display basic information about the dataset\\\\niris_df.info()\\\\n\\\\n# Display statistical summary of the dataset\\\\niris_df.describe()\\\\n\\\\n# Display the first few rows of the dataset\\\\ndisplay(iris_df.head())\\\",\\n \\\"result\\\": \\\"a successful run\\\",\\n \\\"is_success\\\": true,\\n \\\"is_finished\\\": true\\n },\\n {\\n \\\"task_id\\\": \\\"3\\\",\\n \\\"dependent_task_ids\\\": [\\n \\\"2\\\"\\n ],\\n \\\"instruction\\\": \\\"Create a plot visualizing the Iris dataset.\\\",\\n \\\"task_type\\\": \\\"other\\\",\\n \\\"code\\\": \\\"\\\",\\n \\\"result\\\": \\\"\\\",\\n \\\"is_success\\\": false,\\n \\\"is_finished\\\": false\\n }\\n]\\n## Current Task\\n{\\\"task_id\\\":\\\"3\\\",\\\"dependent_task_ids\\\":[\\\"2\\\"],\\\"instruction\\\":\\\"Create a plot visualizing the Iris dataset.\\\",\\\"task_type\\\":\\\"other\\\",\\\"code\\\":\\\"\\\",\\\"result\\\":\\\"\\\",\\\"is_success\\\":false,\\\"is_finished\\\":false}\\n\"}, {\"role\": \"user\", \"content\": \"\\n# Instruction\\nWrite complete code for 'Current Task'. And avoid duplicating code from finished tasks, such as repeated import of packages, reading data, etc.\\nSpecifically, \\n\\n# Capabilities\\n- You can utilize pre-defined tools in any code lines from 'Available Tools' in the form of Python Class.\\n- You can freely combine the use of any other public packages, like sklearn, numpy, pandas, etc..\\n\\n# Available Tools (can be empty):\\nEach Class tool is described in JSON format. When you call a tool, import the tool first.\\n{}\\n\\n# Constraints:\\n- Ensure the output new code is executable in the same Jupyter notebook with previous tasks code have been executed.\\n- Always prioritize using pre-defined tools for the same functionality.\\n\"}]": { + "code": "import matplotlib.pyplot as plt\nimport seaborn as sns\n\n# Set the style of seaborn\nsns.set_style('whitegrid')\n\n# Pairplot to visualize the relationships between features\nsns.pairplot(iris_df, hue='target_names', markers='+')\nplt.show()" + }, + "\n## context\n\n\n-----\n\n## format example\n[CONTENT]\n{\n \"invoice\": \"False\"\n}\n[/CONTENT]\n\n## nodes: \": # \"\n- invoice: # if it's a invoice file, return True else False\n\n\n## constraint\nLanguage: Please use the same language as Human INPUT.\nFormat: output wrapped inside [CONTENT][/CONTENT] like format example, nothing else.\n\n## action\nFollow instructions of nodes, generate output and make sure it follows the format example.\n": "[CONTENT]\n{\n \"invoice\": \"True\"\n}\n[/CONTENT]" } \ No newline at end of file diff --git a/tests/data/search_rsp_cache.json b/tests/data/search_rsp_cache.json index 822fb2069..7b4cc583c 100644 --- a/tests/data/search_rsp_cache.json +++ b/tests/data/search_rsp_cache.json @@ -875,5 +875,121 @@ "curl-cffi-POST-https://duckduckgo.com-{\"data\": {\"q\": \"Dataiku vs DataRobot features\"}}": "Dataiku vs DataRobot features at DuckDuckGo
", "curl-cffi-GET-https://links.duckduckgo.com/d.js-{\"params\": {\"bing_market\": \"wt-WT\", \"df\": null, \"ex\": \"-1\", \"kl\": \"wt-wt\", \"l\": \"wt-wt\", \"q\": \"Dataiku vs DataRobot features\", \"s\": \"0\", \"sp\": \"0\", \"vqd\": \"4-334935250614046875026454141242803242982\"}}": "if (DDG.deep && DDG.deep.setUpstream) DDG.deep.setUpstream(\"bingv7aa\");DDG.deep.bn={'ivc':1};if (DDG.pageLayout) DDG.pageLayout.load('a',[{\"a\":\"\\u9ad8\\u7cbe\\u5ea6\\u306a\\u6a5f\\u68b0\\u5b66\\u7fd2\\u30e2\\u30c7\\u30eb\\u3092\\u69cb\\u7bc9\\u3001\\u5b9f\\u88c5\\u3001\\u904b\\u7528\\u3002DataRobot\\u306f\\u793e\\u5185\\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092\\u5275\\u9020\\u3057\\u307e\\u3059. DataRobot\\u306f\\u4f01\\u696d\\u306e\\u8ab2\\u984c\\u89e3\\u6c7a\\u306b\\u7279\\u5316\\u3002\\u610f\\u601d\\u6c7a\\u5b9a\\u306e\\u81ea\\u52d5\\u5316\\u304b\\u3089\\u9700\\u8981\\u4e88\\u6e2c\\u3001\\u8981\\u56e0\\u5206\\u6790\\u307e\\u3067\\u3053\\u306a\\u3059AI\\u30c4\\u30fc\\u30eb\",\"adext\":{\"callout\":{\"t\":\"Data Science Guardrails \\u00b7 Applied AI Expertise \\u00b7 Trusted by Fortune 50\",\"tid\":\"6\"},\"filterlinks\":{\"l\":[],\"tid\":\"\"},\"sitelinks\":{\"l\":[{\"snippet\":\"Explore the DataRobot AI Platform Get Started With a 30-Day Trial\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=uchwI3Eul8XsE%2DSUlPqxXg%3D%3D&rut=2381550f96a087800d427735905717264a1a708643136f2736a970e740068621&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8BGV0WifLHqlNArHdJt3WDTVUCUzDyrVI_ULomBTgn_xk1MKGRFElGY7vQ8fpE4l__S3CnH6%2D2cXlBQayeIz9CbLU7C4XEu8BgG6oZNQ6EtjG6vrYe5hjw1GZN7VBIkj6nn%2DsoUXy14mVbvkM5ojXVf8oeoz8pwdOc4ANH2TiL9vqJe6Lud2IZXvxJf1I%2DA935XcPQobPZKQaFNFMyygI3Y4TW8k%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnRyaWFsJTJmJTNmdXRtX21lZGl1bSUzZHNlYXJjaCUyNnV0bV9zb3VyY2UlM2RiaW5nJTI2dXRtX2NhbXBhaWduJTNkRnJlZVRyaWFsMjAyM1dXMDgxNkdQU2FkZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDFmMzU0ODE0ODNmMTEyM2Y5NGMzMmRiNzdjZjk5OWFm%26rlid%3D1f35481483f1123f94c32db77cf999af&vqd=4-25671318592048362755712261648304518289&iurl=%7B1%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26ID%3DDevEx%2C5064.1\",\"text\":\"DataRobot Free Trial\"},{\"snippet\":\"Unlock Your AI Success in 2023 Tips on the Path of Value-Driven AI\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=uchwI3Eul8XsE%2DSUlPqxXg%3D%3D&rut=08def2477dd7311fbcffe4c409d28fcdbe68925a50cd2894a7502f8a11785352&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8lYdQlfjG0%2Dh77MMyzT0CuDVUCUyAuuZDH6K8NWyD2XSLoABvrUNVChVbIVOVgzl4xdT3EEUvHgd9P_FWLUDT2My42qKUP3iV87B7hLXXHLdGf7yjst8tWjp%2DcaQz3uiI0c5oom%2DRo8D7A4nohZAtS9199RQLYbNcbOpJnrNMCFmz6EiWk7JqMQ9DE1t9AjaMUWEkEV%2D3W2e8XmBq5bKtRsWnT0E%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnJlc291cmNlcyUyZmFpc3VjY2VzczIwMjMlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RDb250ZW50MTBLZXlzdG9BSVN1Y2Nlc3MyMDIzV1cwNTIyR1BTYWRleHQlMjZ1dG1fdGVybSUzZGRhdGFyb2JvdCUyNnV0bV9jb250ZW50JTNkYWRfZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDI5Zjc0NWY1MzNiNzE2NDU5ZGY0MjA1NmNjYmYyYWU0%26rlid%3D29f745f533b716459df42056ccbf2ae4&vqd=4-333465595216651803104351585568313334233&iurl=%7B1%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26ID%3DDevEx%2C5066.1\",\"text\":\"10 Keys to AI Success\"},{\"snippet\":\"Our Platform Includes Four Fully Integrated Products. Read More.\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=uchwI3Eul8XsE%2DSUlPqxXg%3D%3D&rut=fbe7591a97a4b400635f8cfafd71893553c70fc90218355b7d5622310d9567db&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8cB2vIW6%2D5rxeC5vl08jFZjVUCUw2oN7vfXdo8rlxVmZIfw2bF94_ya9lvPQwUYXJFtTGXBslf_XCcVTiFtj2KJzp9yzLPOdWafvxxwBzn2iwextOSL%2Daq20iQ8nZNktMLYBD1xp3WjThLdejbBCFrR_RvD1YZcHcKf5y5auyV04F_V6x_D6nUwdRYFDmdyciLcpT7JO12EZkmM%2D1buahlzuiBmw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnByb2R1Y3QlMmYlM2ZjYW1wYWlnbmlkJTNkNTMwNzA4MDk5JTI2YWRncm91cGlkJTNkMTM1MDIwMjc3NDIxNzY5OCUyNmFkaWQlM2QlMjZtc2Nsa2lkJTNkMGZhOTg4ZjJkYWU2MWE3MGJhOTVlZDUxMjVlZWFlNDA%26rlid%3D0fa988f2dae61a70ba95ed5125eeae40&vqd=4-211419575679328898707892660118042825990&iurl=%7B1%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26ID%3DDevEx%2C5068.1\",\"text\":\"Product Overview\"}],\"tid\":\"7\\t9[8]\\t11[10]\\t13[12]\",\"type\":\"EnhancedSiteLink\"},\"tid\":\"1\"},\"ae\":{\"callout\":[\"Data Science Guardrails \\u00b7 Applied AI Expertise \\u00b7 Trusted by Fortune 50\"]},\"c\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=uchwI3Eul8XsE%2DSUlPqxXg%3D%3D&rut=94a279ed1549c0107c5c13f21161fd5aaa0d3f08d19e7afd2ed4a19463b69d7d&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8XX6qufLbIkEZRIFo_zgmlDVUCUwOnCSpTtxK0dn2QInSfOGU5eU24GjiRwhmSr89Qa92PcEtK2h6KVoghC%2DNwNrkANG4L6sVirCfv5kl7GPWO9gqgcdw8x5ELjGH7N2HWgbdtH%2D7TWKtxZVdVIFwYJUQDUgM_ODwTspzwBbKKLHD4EPAO5U3RDO3R_igFUlsxkeFXA%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZGQxMGY4ZjY4ZDYxZjFiOTg2NTc1ZWFjYjI5MTczYTQ1%26rlid%3Dd10f8f68d61f1b986575eacb29173a45&vqd=4-152568096679810917558416500867559274982&iurl=%7B1%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26ID%3DDevEx%2C5059.1\",\"d\":\"datarobot.com\",\"h\":0,\"i\":\"\",\"k\":0,\"m\":0,\"o\":\"\",\"p\":1,\"relevancy\":{\"abstract\":\"%E9%AB%98%E7%B2%BE%E5%BA%A6%E3%81%AA%E6%A9%9F%E6%A2%B0%E5%AD%A6%E7%BF%92%E3%83%A2%E3%83%87%E3%83%AB%E3%82%92%E6%A7%8B%E7%AF%89%E3%80%81%E5%AE%9F%E8%A3%85%E3%80%81%E9%81%8B%E7%94%A8%E3%80%82%3Cb%3EDataRobot%3C%2Fb%3E%E3%81%AF%E7%A4%BE%E5%86%85%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92%E5%89%B5%E9%80%A0%E3%81%97%E3%81%BE%E3%81%99.%20DataRobot%E3%81%AF%E4%BC%81%E6%A5%AD%E3%81%AE%E8%AA%B2%E9%A1%8C%E8%A7%A3%E6%B1%BA%E3%81%AB%E7%89%B9%E5%8C%96%E3%80%82%E6%84%8F%E6%80%9D%E6%B1%BA%E5%AE%9A%E3%81%AE%E8%87%AA%E5%8B%95%E5%8C%96%E3%81%8B%E3%82%89%E9%9C%80%E8%A6%81%E4%BA%88%E6%B8%AC%E3%80%81%E8%A6%81%E5%9B%A0%E5%88%86%E6%9E%90%E3%81%BE%E3%81%A7%E3%81%93%E3%81%AA%E3%81%99AI%E3%83%84%E3%83%BC%E3%83%AB\",\"adx_name\":\"none\",\"is_good_v10\":0,\"q\":\"Dataiku%20vs%20DataRobot%20features\",\"q_words\":4,\"q_words_fuzzy\":0.25,\"q_words_in_ad\":1,\"root_domain\":\"datarobot.com\",\"start\":\"0\",\"title\":\"%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92%20%2D%20%E7%A4%BE%E5%86%85%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E4%BE%A1%E5%80%A4%E5%89%B5%E5%87%BA\"},\"s\":\"bingv7aa\",\"t\":\"\\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092 - \\u793e\\u5185\\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u4fa1\\u5024\\u5275\\u51fa\",\"tid\":\"1,6,7,9[8],11[10],13[12]\",\"u\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=uchwI3Eul8XsE%2DSUlPqxXg%3D%3D&rut=94a279ed1549c0107c5c13f21161fd5aaa0d3f08d19e7afd2ed4a19463b69d7d&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8XX6qufLbIkEZRIFo_zgmlDVUCUwOnCSpTtxK0dn2QInSfOGU5eU24GjiRwhmSr89Qa92PcEtK2h6KVoghC%2DNwNrkANG4L6sVirCfv5kl7GPWO9gqgcdw8x5ELjGH7N2HWgbdtH%2D7TWKtxZVdVIFwYJUQDUgM_ODwTspzwBbKKLHD4EPAO5U3RDO3R_igFUlsxkeFXA%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZGQxMGY4ZjY4ZDYxZjFiOTg2NTc1ZWFjYjI5MTczYTQ1%26rlid%3Dd10f8f68d61f1b986575eacb29173a45&vqd=4-152568096679810917558416500867559274982&iurl=%7B1%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26ID%3DDevEx%2C5059.1\"}], {\"page_load_url\":\"https://duckduckgo.com/y.js?ifu=%7B3%7Dappid%3D055AAD1BA669BEB8B048128DC89A107C678B527B%26rguid%3D280881b97b9245e6a74bddebc1a6cbda&iurl=%7B2%7DIG%3D3EB403B8C4EA42F4B7FF0CE90CB46EF0%26CID%3D2F20CB6F269D6DD02331DF69279D6C12%26Type%3DEvent.CPT%26DATA%3D0\",\"visibility_url\":\"https://duckduckgo.com/y.js?ivu=%7B4%7Dtype%3Dmv%26reqver%3D1.0%26rg%3D280881b97b9245e6a74bddebc1a6cbda\"});DDG.deep.signalSummary = \"\";DDG.inject('DDG.Data.languages.resultLanguages', {\"en\":[\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"https://comparisons.financesonline.com/datarobot-vs-dataiku-dss\",\"https://slashdot.org/software/comparison/DataRobot-vs-Dataiku-DSS/\",\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"https://www.gartner.com/reviews/market/dsml-engineering-platforms/compare/dataiku-vs-datarobot\",\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"https://www.trustradius.com/products/datarobot/reviews?qs=pros-and-cons\",\"https://www.getapp.com/emerging-technology-software/a/dataiku-dss/compare/datarobot/\",\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/dataiku-vs-datarobot-ai-platform\",\"https://slashdot.org/software/comparison/Alteryx-vs-DataRobot-vs-Dataiku-DSS/\",\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS/\",\"https://valohai.com/mlops-platforms-compared/\",\"https://www.dataiku.com/product/plans-and-features/\",\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS-vs-datagym/\",\"https://sourceforge.net/software/compare/C3-AI-Suite-vs-DataRobot-vs-Dataiku-DSS/\",\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/datarobot-ai-platform-vs-dataiku\",\"https://slashdot.org/software/comparison/Amazon-SageMaker-vs-DataRobot-vs-Dataiku-DSS/\",\"https://sourceforge.net/software/compare/Analance-vs-DataRobot-vs-Dataiku-DSS/\"]});DDG.deep.pageLayoutSummary = \"a1w23r1,e1\";DDG.inject('DDG.Data.languages.adLanguages', {});if (DDG.pageLayout) DDG.pageLayout.load('d',[{\"a\":\"1 Star 0% Ratings breakdown Overall Capability Score Overall Rating 4.7 ( 504 reviews) 4.7 (20) Data Access and Manipulation 4.5 (224) Data Exploration and Visualization 4.7\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku vs DataRobot 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\"},{\"a\":\"Path to AI Success Compare Dataiku DSS vs DataRobot. 103 verified user reviews and ratings of features, pros, cons, pricing, support and more.\",\"ae\":null,\"c\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"d\":\"www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.trustradius.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku DSS vs DataRobot | TrustRadius\",\"u\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\"},{\"a\":\"General Discussion Dataiku vs DataRobot Solved! Raja Level 2 08-22-2020 03:16 AM Please enlighten me, What distinguishes Dataiku from tools like DataRobot? They appear to be similar, trying to know how dataiku has an upper hand, would make it easy for placing option to customers. 1 Reply 2 Solutions Solutions shown first - Read whole discussion\",\"ae\":null,\"c\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"d\":\"community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"da\":\"\",\"h\":0,\"i\":\"community.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Solved: Dataiku vs DataRobot - Dataiku Community\",\"u\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\"},{\"a\":\"DataRobot vs Dataiku DSS When assessing the two solutions, reviewers found Dataiku DSS easier to use and administer. However, reviewers preferred the ease of set up, and doing business with DataRobot overall. Reviewers felt that DataRobot meets the needs of their business better than Dataiku DSS.\",\"ae\":null,\"c\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"d\":\"www.g2.com/compare/datarobot-vs-dataiku-dss\",\"da\":\"\",\"h\":0,\"i\":\"www.g2.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Dataiku DSS | G2\",\"u\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\"},{\"a\":\"Quick overview Before we get into a detailed comparison, here's a quick overview of each platform. Dataiku is a cross-platform desktop application that includes a broad range of tools, such as notebooks (similar to Jupyter Notebook), workflow management (similar to Apache Airflow), and automated machine learning.\",\"ae\":null,\"c\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"d\":\"www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"da\":\"\",\"h\":0,\"i\":\"www.datarevenue.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"ML Platforms: Dataiku vs. Alteryx vs. Sagemaker vs. Datarobot\",\"u\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\"},{\"a\":\"Home Predictive Analysis Software DataRobot Dataiku DSS Why is FinancesOnline free Compare DataRobot vs Dataiku DSS What is better DataRobot or Dataiku DSS? Examining products to find the best Predictive Analysis Software does not always have to be tough.\",\"ae\":null,\"c\":\"https://comparisons.financesonline.com/datarobot-vs-dataiku-dss\",\"d\":\"comparisons.financesonline.com/datarobot-vs-dataiku-dss\",\"da\":\"\",\"h\":0,\"i\":\"comparisons.financesonline.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"DataRobot vs Dataiku DSS 2024 Comparison | FinancesOnline\",\"u\":\"https://comparisons.financesonline.com/datarobot-vs-dataiku-dss\"},{\"a\":\"Machine Learning Software Dataiku vs DataRobot Dataiku vs DataRobot Share How Capterra Verifies Reviews Pricing Best for Screenshots Features Reviews Pros & Cons Deployment & Support Alternatives Company Details Dataiku VISIT PROFILE DataRobot VISIT PROFILE Pricing Starting from $ 0.01 /Year Pricing Model: Not provided by vendor Free Trial\",\"ae\":null,\"c\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"d\":\"www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"da\":\"translations\",\"h\":0,\"i\":\"www.capterra.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare Dataiku vs DataRobot 2024 | Capterra\",\"u\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\"},{\"a\":\"What's the difference between DataRobot and Dataiku DSS? Compare DataRobot vs. Dataiku DSS in 2023 by cost, reviews, features, integrations, deployment, target market, support options, trial offers, training options, years in business, region, and more using the chart below.\",\"ae\":null,\"b\":\"/.\\tSlashdot\\tslashdot.org\",\"c\":\"https://slashdot.org/software/comparison/DataRobot-vs-Dataiku-DSS/\",\"d\":\"slashdot.org/software/comparison/DataRobot-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"slashdot.org\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Dataiku DSS in 2023 - Slashdot\",\"u\":\"https://slashdot.org/software/comparison/DataRobot-vs-Dataiku-DSS/\"},{\"a\":\"1 Star 0% Distribution based on 504 ratings Customer Experience Evaluation & Contracting 4.6 Integration & Deployment 4.7 Service & Support 4.8 Product Capabilities 4.8 FREE View and Download Peer Insights About Dataiku\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku Reviews, Ratings & Features 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\"},{\"a\":\"1329 reviews on 16 vendors. chevron_right. Yard Management. 25 reviews on 28 vendors. chevron_right. Zero Trust Network Access. 733 reviews on 47 vendors. chevron_right. Read the latest Gartner-verified reviews covering over 500+ software categories and find the best enterprise software or services for your organization.\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/dsml-engineering-platforms/compare/dataiku-vs-datarobot\",\"d\":\"www.gartner.com/reviews/market/dsml-engineering-platforms/compare/dataiku-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Explore Enterprise Software Categories | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/dsml-engineering-platforms/compare/dataiku-vs-datarobot\"},{\"a\":\"1. Dataiku is a versatile desktop application comprised of a wide range of tools, including automated machine learning, notebooks, and workflow management. It aims to replace pre-existing tools...\",\"ae\":null,\"b\":\"li\\tLinkedIn\\twww.linkedin.com\",\"c\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"d\":\"www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"da\":\"\",\"e\":\"2023-08-11T00:00:00.0000000\",\"h\":0,\"i\":\"www.linkedin.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Managed Machine Learning Platforms: A Comparative Analysis - LinkedIn\",\"u\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\"},{\"a\":\"Dataiku DSS, H2O, and Google Cloud AI are common alternatives for DataRobot. What is DataRobot's best feature? Reviewers rate Automated Machine Learning highest, with a score of 9.3. Who uses DataRobot? The most common users of DataRobot are from Mid-sized Companies (51-1,000 employees).\",\"ae\":null,\"c\":\"https://www.trustradius.com/products/datarobot/reviews?qs=pros-and-cons\",\"d\":\"www.trustradius.com/products/datarobot/reviews?qs=pros-and-cons\",\"da\":\"\",\"h\":0,\"i\":\"www.trustradius.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Pros and Cons of DataRobot 2024 - TrustRadius\",\"u\":\"https://www.trustradius.com/products/datarobot/reviews?qs=pros-and-cons\"},{\"a\":\"Compare Dataiku and DataRobot based on features, pricing, verified reviews, integrations & more. Find out which software is best for your business today. 0. App comparison. Add up to 4 apps below to see how they compare. You can also use the "Compare" buttons while browsing.\",\"ae\":null,\"c\":\"https://www.getapp.com/emerging-technology-software/a/dataiku-dss/compare/datarobot/\",\"d\":\"www.getapp.com/emerging-technology-software/a/dataiku-dss/compare/datarobot/\",\"da\":\"\",\"h\":0,\"i\":\"www.getapp.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku vs DataRobot Comparison | GetApp\",\"u\":\"https://www.getapp.com/emerging-technology-software/a/dataiku-dss/compare/datarobot/\"},{\"a\":\"Dataiku vs DataRobot AI Platform Compare Dataiku and DataRobot AI Platform using real user data focused on features, satisfaction, business value, and the vendor relationship. What is Machine Learning Platforms (ML) Software?\",\"ae\":null,\"c\":\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/dataiku-vs-datarobot-ai-platform\",\"d\":\"www.softwarereviews.com/categories/machine-learning-platforms/compare/dataiku-vs-datarobot-ai-platform\",\"da\":\"\",\"h\":0,\"i\":\"www.softwarereviews.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku vs DataRobot AI Platform - Machine Learning Platforms\",\"u\":\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/dataiku-vs-datarobot-ai-platform\"},{\"a\":\"What's the difference between Alteryx, DataRobot, and Dataiku DSS? Compare Alteryx vs. DataRobot vs. Dataiku DSS in 2024 by cost, reviews, features, integrations, deployment, target market, support options, trial offers, training options, years in business, region, and more using the chart below. Alteryx View Product DataRobot View Product\",\"ae\":null,\"b\":\"/.\\tSlashdot\\tslashdot.org\",\"c\":\"https://slashdot.org/software/comparison/Alteryx-vs-DataRobot-vs-Dataiku-DSS/\",\"d\":\"slashdot.org/software/comparison/Alteryx-vs-DataRobot-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"slashdot.org\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare Alteryx vs. DataRobot vs. Dataiku DSS in 2023 - Slashdot\",\"u\":\"https://slashdot.org/software/comparison/Alteryx-vs-DataRobot-vs-Dataiku-DSS/\"},{\"a\":\"What's the difference between DataRobot, Databricks Lakehouse, and Dataiku DSS? Compare DataRobot vs. Databricks Lakehouse vs. Dataiku DSS in 2023 by cost, reviews, features, integrations, deployment, target market, support options, trial offers, training options, years in business, region, and more using the chart below.\",\"ae\":null,\"b\":\"/.\\tSlashdot\\tslashdot.org\",\"c\":\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS/\",\"d\":\"slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"slashdot.org\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Databricks Lakehouse vs. Dataiku DSS - Slashdot\",\"u\":\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS/\"},{\"a\":\"The platforms we've chosen for our analysis are ClearML, cnvrg.io, Dataiku, Datarobot, Iguazio, Sagemaker, Seldon and Valohai from the managed side, and Flyte, Kubeflow, MLflow and Metaflow from the open-source side. This is by no means an exhaustive list of all the MLOps tools out there. Most of these are tools that describe themselves as ...\",\"ae\":null,\"c\":\"https://valohai.com/mlops-platforms-compared/\",\"d\":\"valohai.com/mlops-platforms-compared/\",\"da\":\"\",\"h\":0,\"i\":\"valohai.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"MLOps Platforms Compared - Valohai\",\"u\":\"https://valohai.com/mlops-platforms-compared/\"},{\"a\":\"Visual Machine Learning and automated features preprocessing: Builtin charts and dashboards: Code notebooks and recipes: Custom web applications and plugins: Collaboration: DEPLOYMENT OPTIONS; ... Dataiku Scores an overall 4.8 out of 5 rating Based on 249 ratings for the DSMLP market, as of March 1, 2022\",\"ae\":null,\"c\":\"https://www.dataiku.com/product/plans-and-features/\",\"d\":\"www.dataiku.com/product/plans-and-features/\",\"da\":\"\",\"h\":0,\"i\":\"www.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Explore Dataiku Plans and Features | Online or Installed\",\"u\":\"https://www.dataiku.com/product/plans-and-features/\"},{\"a\":\"What's the difference between DataRobot, Databricks Lakehouse, Dataiku DSS, and DATAGYM? Compare DataRobot vs. Databricks Lakehouse vs. Dataiku DSS vs. DATAGYM in 2024 by cost, reviews, features, integrations, deployment, target market, support options, trial offers, training options, years in business, region, and more using the chart below.\",\"ae\":null,\"b\":\"/.\\tSlashdot\\tslashdot.org\",\"c\":\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS-vs-datagym/\",\"d\":\"slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS-vs-datagym/\",\"da\":\"\",\"h\":0,\"i\":\"slashdot.org\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Databricks Lakehouse vs. Dataiku DSS vs. DATAGYM ...\",\"u\":\"https://slashdot.org/software/comparison/DataRobot-vs-Databricks-vs-Dataiku-DSS-vs-datagym/\"},{\"a\":\"Claim Dataiku DSS and update features and information. Compare C3 AI Suite vs. DataRobot vs. Dataiku DSS using this comparison chart. Compare price, features, and reviews of the software side-by-side to make the best choice for your business.\",\"ae\":null,\"b\":\"srcforge\\tSourceForge\\tsourceforge.net\",\"c\":\"https://sourceforge.net/software/compare/C3-AI-Suite-vs-DataRobot-vs-Dataiku-DSS/\",\"d\":\"sourceforge.net/software/compare/C3-AI-Suite-vs-DataRobot-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"sourceforge.net\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"C3 AI Suite vs. DataRobot vs. Dataiku DSS Comparison - SourceForge\",\"u\":\"https://sourceforge.net/software/compare/C3-AI-Suite-vs-DataRobot-vs-Dataiku-DSS/\"},{\"a\":\"Compare DataRobot AI Platform and Dataiku using real user data focused on features, satisfaction, business value, and the vendor relationship. What is Machine Learning Platforms (ML) Software?\",\"ae\":null,\"c\":\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/datarobot-ai-platform-vs-dataiku\",\"d\":\"www.softwarereviews.com/categories/machine-learning-platforms/compare/datarobot-ai-platform-vs-dataiku\",\"da\":\"\",\"h\":0,\"i\":\"www.softwarereviews.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"DataRobot AI Platform vs Dataiku - Machine Learning Platforms\",\"u\":\"https://www.softwarereviews.com/categories/machine-learning-platforms/compare/datarobot-ai-platform-vs-dataiku\"},{\"a\":\"What's the difference between Amazon SageMaker, DataRobot, and Dataiku DSS? Compare Amazon SageMaker vs. DataRobot vs. Dataiku DSS in 2024 by cost, reviews, features, integrations, deployment, target market, support options, trial offers, training options, years in business, region, and more using the chart below.\",\"ae\":null,\"b\":\"/.\\tSlashdot\\tslashdot.org\",\"c\":\"https://slashdot.org/software/comparison/Amazon-SageMaker-vs-DataRobot-vs-Dataiku-DSS/\",\"d\":\"slashdot.org/software/comparison/Amazon-SageMaker-vs-DataRobot-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"slashdot.org\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare Amazon SageMaker vs. DataRobot vs. Dataiku DSS in 2024 - Slashdot\",\"u\":\"https://slashdot.org/software/comparison/Amazon-SageMaker-vs-DataRobot-vs-Dataiku-DSS/\"},{\"a\":\"Compare Analance vs. DataRobot vs. Dataiku DSS using this comparison chart. Compare price, features, and reviews of the software side-by-side to make the best choice for your business.\",\"ae\":null,\"b\":\"srcforge\\tSourceForge\\tsourceforge.net\",\"c\":\"https://sourceforge.net/software/compare/Analance-vs-DataRobot-vs-Dataiku-DSS/\",\"d\":\"sourceforge.net/software/compare/Analance-vs-DataRobot-vs-Dataiku-DSS/\",\"da\":\"\",\"h\":0,\"i\":\"sourceforge.net\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Analance vs. DataRobot vs. Dataiku DSS Comparison - SourceForge\",\"u\":\"https://sourceforge.net/software/compare/Analance-vs-DataRobot-vs-Dataiku-DSS/\"},{\"n\":\"/d.js?q=Dataiku%20vs%20DataRobot%20features&kl=wt-wt&l=wt-wt&p=&s=23&ex=-1&ct=US&sp=0&vqd=4-334935250614046875026454141242803242982\"}]);DDG.duckbar.load('images');DDG.duckbar.load('news');DDG.duckbar.load('videos');DDG.duckbar.loadModule('related_searches', {\"ads\":[],\"query\":\"Dataiku vs DataRobot features\",\"queryEncoded\":\"Dataiku%20vs%20DataRobot%20features\",\"response_type\":\"places\",\"results\":[{\"display_text\":\"dataiku vs datarobot review\",\"text\":\"dataiku vs datarobot review\",\"web_search_url\":\"?q=dataiku%20vs%20datarobot%20review\"},{\"display_text\":\"dataiku vs alteryx\",\"text\":\"dataiku vs alteryx\",\"web_search_url\":\"?q=dataiku%20vs%20alteryx\"},{\"display_text\":\"gartner dataiku reviews\",\"text\":\"gartner dataiku reviews\",\"web_search_url\":\"?q=gartner%20dataiku%20reviews\"},{\"display_text\":\"alteryx vs dataiku knime\",\"text\":\"alteryx vs dataiku knime\",\"web_search_url\":\"?q=alteryx%20vs%20dataiku%20knime\"},{\"display_text\":\"dataiku vs rapidminer\",\"text\":\"dataiku vs rapidminer\",\"web_search_url\":\"?q=dataiku%20vs%20rapidminer\"},{\"display_text\":\"dataiku vs azure ml\",\"text\":\"dataiku vs azure ml\",\"web_search_url\":\"?q=dataiku%20vs%20azure%20ml\"},{\"display_text\":\"sagemaker vs dataiku\",\"text\":\"sagemaker vs dataiku\",\"web_search_url\":\"?q=sagemaker%20vs%20dataiku\"},{\"display_text\":\"dataiku reviews\",\"text\":\"dataiku reviews\",\"web_search_url\":\"?q=dataiku%20reviews\"}],\"vqd\":{\"Dataiku%20vs%20DataRobot%20features\":\"4-334935250614046875026454141242803242982\"}});if (DDG.pageLayout) DDG.pageLayout.initialize({\"mainline\":{\"items\":[[\"ad\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"related_searches\"]]},\"sidebar\":{\"items\":[[\"wikipedia_fathead\"]]}}, { start: 0 });DDG.deep.emit(\"load:completed\");", "curl-cffi-POST-https://duckduckgo.com-{\"data\": {\"q\": \"Dataiku and DataRobot use cases\"}}": "Dataiku and DataRobot use cases at DuckDuckGo
", - "curl-cffi-GET-https://links.duckduckgo.com/d.js-{\"params\": {\"bing_market\": \"wt-WT\", \"df\": null, \"ex\": \"-1\", \"kl\": \"wt-wt\", \"l\": \"wt-wt\", \"q\": \"Dataiku and DataRobot use cases\", \"s\": \"0\", \"sp\": \"0\", \"vqd\": \"4-60481969350525797892441552954401970387\"}}": "if (DDG.deep && DDG.deep.setUpstream) DDG.deep.setUpstream(\"bingv7aa\");DDG.deep.bn={'ivc':1};if (DDG.pageLayout) DDG.pageLayout.load('a',[{\"a\":\"\\u9ad8\\u7cbe\\u5ea6\\u306a\\u6a5f\\u68b0\\u5b66\\u7fd2\\u30e2\\u30c7\\u30eb\\u3092\\u69cb\\u7bc9\\u3001\\u5b9f\\u88c5\\u3001\\u904b\\u7528\\u3002DataRobot\\u306f\\u793e\\u5185\\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092\\u5275\\u9020\\u3057\\u307e\\u3059. AI\\u3092\\u6d3b\\u7528\\u3057\\u30c7\\u30fc\\u30bf\\u3092\\u5206\\u6790\\u3001\\u5b9f\\u7528\\u7684\\u306a\\u30a4\\u30f3\\u30b5\\u30a4\\u30c8\\u3092\\u660e\\u3089\\u304b\\u306b\\u3002\\u30d3\\u30b8\\u30cd\\u30b9\\u306e\\u8ab2\\u984c\\u3092\\u3088\\u308a\\u65e9\\u304f\\u89e3\\u6c7a\",\"adext\":{\"callout\":{\"t\":\"30-Day Free Trial \\u00b7 Trusted by Fortune 50 \\u00b7 No Vendor Lock-in\",\"tid\":\"6\"},\"filterlinks\":{\"l\":[],\"tid\":\"\"},\"sitelinks\":{\"l\":[{\"snippet\":\"Explore the DataRobot AI Platform Get Started With a 30-Day Trial\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=d0faee2c8c1aae9ac3a012e21d37352a1181970dce9edeba4107839fbfbf097a&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De81Q_rqdj1ZxH5XXGh4PG6pjVUCUzdB7rGpyykWEihNc_sSp5n%2DJ9jIyTjOSnXg0OUazrpKgDJrNvBOdNa5PjBGtyLGt23nrBAabI6opJXrliWQ4o%2DTyxIsqOeCXqzLOOJ3jJb74k6KEx20zilzwKmzSg3nBop2A9JqsasC17VVDPc3_i3EzPbWeRNS4nhxXWJqBKd55GfhuEOg2RZUbmmuAUhWvM%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnRyaWFsJTJmJTNmdXRtX21lZGl1bSUzZHNlYXJjaCUyNnV0bV9zb3VyY2UlM2RiaW5nJTI2dXRtX2NhbXBhaWduJTNkRnJlZVRyaWFsMjAyM1dXMDgxNkdQU2FkZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDc2YmMwNmFmNTA0NDFjOGVjOGYxNjMwY2FmNGU4ZTVk%26rlid%3D76bc06af50441c8ec8f1630caf4e8e5d&vqd=4-164177780916400746369660096493208330918&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5063.1\",\"text\":\"DataRobot Free Trial\"},{\"snippet\":\"Unlock Your AI Success in 2023 Tips on the Path of Value-Driven AI\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=fdb107a4de6fffdec2bdf43b561b2c63ca700daaef68f0e683547361efbbc2b0&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8%2DT0j3GTQEgr%2DmtHPM1LzNzVUCUyRxvVKYHe6LbNa2mmCfCZh3Ept1NM%2DP%2DM1AAluh_OL3VQw_FWI0A3YxC3pzzqthf3gpxan_Lv7CjKenge%2DwMYUz3bRFoFyHtQBMdgqv6T7gMGfyYwN3UCj6FNYwVVn9UNN0h1dIQanHNB6Ya9gRrPBACknA8qtsf6A2oUG1xhq7AOF98NzGphnfQ_38fySnRU%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnJlc291cmNlcyUyZmFpc3VjY2VzczIwMjMlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RDb250ZW50MTBLZXlzdG9BSVN1Y2Nlc3MyMDIzV1cwNTIyR1BTYWRleHQlMjZ1dG1fdGVybSUzZGRhdGFyb2JvdCUyNnV0bV9jb250ZW50JTNkYWRfZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZGQzNmQ2MzlkMmFlNTEwMTM3ZTIwMDYzZWQ1ZWY3M2Yz%26rlid%3Dd36d639d2ae510137e20063ed5ef73f3&vqd=4-117927704271333462986714580056949079639&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5065.1\",\"text\":\"10 Keys to AI Success\"},{\"snippet\":\"Our Platform Includes Four Fully Integrated Products. Read More.\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=4f06bd3312172b8e61d65ee2626dea6e26d941c3a16aa546b4e11b79e8bf027f&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8885tVmNmhi65Jmp3f2wYSzVUCUyFey1LCmrSNpGfkWzQnoC7QIbU3ztthJ%2DqKpgCmRfxudhbLK927YN84jvZlV2zTKo9DOULVj5wB8mcGXy_F42SnsrO1jZpY9NnMnzqMYPb5xZTTdgrTO1_w3Bgpd0e0VzO81_O3%2Dfo2z4UiLuVETFVqfACqR6NEwz0yfjzJe6ED9tvi_gPDiUL9iWATrNIrsw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnByb2R1Y3QlMmYlM2ZjYW1wYWlnbmlkJTNkNTMwNzA4MDk5JTI2YWRncm91cGlkJTNkMTM1MDIwMjc3NDIxNzY5OCUyNmFkaWQlM2QlMjZtc2Nsa2lkJTNkY2U4NzQ1ZDViODBlMTJmNjQ2N2QyMDc2NDcwNDY2YjI%26rlid%3Dce8745d5b80e12f6467d2076470466b2&vqd=4-169069202740993895017985472268973083525&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5067.1\",\"text\":\"Product Overview\"}],\"tid\":\"7\\t9[8]\\t11[10]\\t13[12]\",\"type\":\"EnhancedSiteLink\"},\"tid\":\"1\"},\"ae\":{\"callout\":[\"30-Day Free Trial \\u00b7 Trusted by Fortune 50 \\u00b7 No Vendor Lock-in\"]},\"c\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=e744d99a8df00b24df71f821ad4d1332080aa03267e50f0e988d284f58d9d2ef&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8tT9soRYLZabP1ukFkRsgNzVUCUzl89Y8xEqpxoqHqIlCI5wWbydNnN_PoAKHAa2Vsio83mXA_ax16t6rJ7XGkBv0Cg7_D1eg2QAuJgPKEam4VWI3rW40B03r1p11ZXN1Gd1847Vj05bAnJnPfgVyC8ZzFQxLxONmOI0Hg182z2bZUVII26BUAlUHaVZ7O_9FEXLJWw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDA2MTIwYzhmMTAxNzEwYTZiNmRiNjkyY2VmMWRiOTY1%26rlid%3D06120c8f101710a6b6db692cef1db965&vqd=4-91027509783546726889708070523412001433&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5058.1\",\"d\":\"datarobot.com\",\"h\":0,\"i\":\"\",\"k\":0,\"m\":0,\"o\":\"\",\"p\":1,\"relevancy\":{\"abstract\":\"%E9%AB%98%E7%B2%BE%E5%BA%A6%E3%81%AA%E6%A9%9F%E6%A2%B0%E5%AD%A6%E7%BF%92%E3%83%A2%E3%83%87%E3%83%AB%E3%82%92%E6%A7%8B%E7%AF%89%E3%80%81%E5%AE%9F%E8%A3%85%E3%80%81%E9%81%8B%E7%94%A8%E3%80%82%3Cb%3EDataRobot%3C%2Fb%3E%E3%81%AF%E7%A4%BE%E5%86%85%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92%E5%89%B5%E9%80%A0%E3%81%97%E3%81%BE%E3%81%99.%20AI%E3%82%92%E6%B4%BB%E7%94%A8%E3%81%97%E3%83%87%E3%83%BC%E3%82%BF%E3%82%92%E5%88%86%E6%9E%90%E3%80%81%E5%AE%9F%E7%94%A8%E7%9A%84%E3%81%AA%E3%82%A4%E3%83%B3%E3%82%B5%E3%82%A4%E3%83%88%E3%82%92%E6%98%8E%E3%82%89%E3%81%8B%E3%81%AB%E3%80%82%E3%83%93%E3%82%B8%E3%83%8D%E3%82%B9%E3%81%AE%E8%AA%B2%E9%A1%8C%E3%82%92%E3%82%88%E3%82%8A%E6%97%A9%E3%81%8F%E8%A7%A3%E6%B1%BA\",\"adx_name\":\"none\",\"is_good_v10\":0,\"organic_ranks\":[5,11,12,13],\"q\":\"Dataiku%20and%20DataRobot%20use%20cases\",\"q_words\":4,\"q_words_fuzzy\":0.25,\"q_words_in_ad\":1,\"root_domain\":\"datarobot.com\",\"start\":\"0\",\"title\":\"%E3%83%93%E3%83%83%E3%82%B0%E3%83%87%E3%83%BC%E3%82%BF%E5%88%86%E6%9E%90%E3%82%92%E9%AB%98%E9%80%9F%E5%8C%96%20%2D%20%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92\"},\"s\":\"bingv7aa\",\"t\":\"\\u30d3\\u30c3\\u30b0\\u30c7\\u30fc\\u30bf\\u5206\\u6790\\u3092\\u9ad8\\u901f\\u5316 - \\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092\",\"tid\":\"1,6,7,9[8],11[10],13[12]\",\"u\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=e744d99a8df00b24df71f821ad4d1332080aa03267e50f0e988d284f58d9d2ef&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8tT9soRYLZabP1ukFkRsgNzVUCUzl89Y8xEqpxoqHqIlCI5wWbydNnN_PoAKHAa2Vsio83mXA_ax16t6rJ7XGkBv0Cg7_D1eg2QAuJgPKEam4VWI3rW40B03r1p11ZXN1Gd1847Vj05bAnJnPfgVyC8ZzFQxLxONmOI0Hg182z2bZUVII26BUAlUHaVZ7O_9FEXLJWw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDA2MTIwYzhmMTAxNzEwYTZiNmRiNjkyY2VmMWRiOTY1%26rlid%3D06120c8f101710a6b6db692cef1db965&vqd=4-91027509783546726889708070523412001433&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5058.1\"}], {\"page_load_url\":\"https://duckduckgo.com/y.js?ifu=%7B3%7Dappid%3D055AAD1BA669BEB8B048128DC89A107C678B527B%26rguid%3D309794dc72f748f6a2b95ce5c34fbcec&iurl=%7B2%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26Type%3DEvent.CPT%26DATA%3D0\",\"visibility_url\":\"https://duckduckgo.com/y.js?ivu=%7B4%7Dtype%3Dmv%26reqver%3D1.0%26rg%3D309794dc72f748f6a2b95ce5c34fbcec\"});DDG.deep.signalSummary = \"\";DDG.inject('DDG.Data.languages.resultLanguages', {\"en\":[\"https://knowledge.dataiku.com/latest/use-cases/index.html\",\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"https://www.datarobot.com/use-cases/\",\"https://academy.dataiku.com/page/use-cases\",\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"https://blog.dataiku.com/topic/use-cases-projects\",\"https://valohai.com/mlops-platforms-compared/\",\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"https://pages.dataiku.com/experience-a-dataiku-demo\",\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"https://www.dataiku.com/stories/\",\"https://www.dataiku.com/\",\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\"]});DDG.deep.pageLayoutSummary = \"a1w4v1w19,w1\";DDG.inject('DDG.Data.languages.adLanguages', {});if (DDG.pageLayout) DDG.pageLayout.load('d',[{\"a\":\"Use Cases - Dataiku Knowledge Base Use Cases # These use cases allow you to practice what you've learned by building simplified, but complete use cases in Dataiku. Topics # Data Preparation Use Cases Classification Use Cases Clustering Use Cases Plugin Use Cases\",\"ae\":null,\"c\":\"https://knowledge.dataiku.com/latest/use-cases/index.html\",\"d\":\"knowledge.dataiku.com/latest/use-cases/index.html\",\"da\":\"\",\"h\":0,\"i\":\"knowledge.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases - Dataiku Knowledge Base\",\"u\":\"https://knowledge.dataiku.com/latest/use-cases/index.html\"},{\"a\":\"Community Dataiku Use Cases & Success Stories \\u26a0\\ufe0f Discover pioneering Dataiku use cases and success stories shared by customers, partners, academics, and nonprofits participating in the Dataiku Frontrunner Awards. Use the following labels to filter submissions by industry:\",\"ae\":null,\"c\":\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"d\":\"community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"da\":\"\",\"h\":0,\"i\":\"community.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku Use Cases & Success Stories - Dataiku Community\",\"u\":\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\"},{\"a\":\"Dataiku is a cross-platform desktop application that includes a broad range of tools, such as notebooks (similar to Jupyter Notebook), workflow management (similar to Apache Airflow), and automated machine learning. In general, Dataiku aims to replace many of your existing tools rather than to integrate with them.\",\"ae\":null,\"c\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"d\":\"www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"da\":\"\",\"h\":0,\"i\":\"www.datarevenue.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"ML Platforms: Dataiku vs. Alteryx vs. Sagemaker vs. Datarobot\",\"u\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\"},{\"a\":\"Dataiku has a rating of 4.8 stars with 504 reviews. DataRobot has a rating of 4.6 stars with 508 reviews. See side-by-side comparisons of product capabilities, customer experience, pros and cons, and reviewer demographics to find the best fit for your organization. See more companies in the Data Science and Machine Learning Platforms market. PDF.\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku vs DataRobot 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\"},{\"a\":\"In my humble opinion DSS is a more a 'toolbox', where as DataRobot is an autoML platform. DataRobot is really good at what it does - if you have non-technical team who want to drop in data and leave everything to autoML then this may be the option for them.\",\"ae\":null,\"c\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"d\":\"community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"da\":\"\",\"h\":0,\"i\":\"community.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Solved: Dataiku vs DataRobot - Dataiku Community\",\"u\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\"},{\"a\":\"Use cases AI Use Cases AI-driven organizations around the world use DataRobot to solve their most pressing business problems. Build with Free Trial Recent Popular Filters Ready to Get Started? See how a value-driven approach to AI can accelerate time to impact. Start Free Trial\",\"ae\":null,\"c\":\"https://www.datarobot.com/use-cases/\",\"d\":\"www.datarobot.com/use-cases/\",\"da\":\"\",\"h\":0,\"i\":\"www.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Machine Learning Use Cases | DataRobot AI Platform\",\"u\":\"https://www.datarobot.com/use-cases/\"},{\"a\":\"With Dataiku's AI Prepare assistant, you can work smarter, not harder. Simply describe the transformation you want to apply in natural language and the AI assistant automatically generates the necessary data preparation steps. The ability to modify both your prompt and the resulting steps means you can prepare data faster than ever, yet still ...\",\"ae\":null,\"c\":\"https://academy.dataiku.com/page/use-cases\",\"d\":\"academy.dataiku.com/page/use-cases\",\"da\":\"\",\"h\":0,\"i\":\"academy.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases - Dataiku\",\"u\":\"https://academy.dataiku.com/page/use-cases\"},{\"a\":\"84 Reviews and Ratings Path to AI Success Compare Dataiku DSS vs DataRobot. 103 verified user reviews and ratings of features, pros, cons, pricing, support and more.\",\"ae\":null,\"c\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"d\":\"www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.trustradius.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku DSS vs DataRobot | TrustRadius\",\"u\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\"},{\"a\":\"side-by-side comparison of DataRobot vs. Dataiku DSS. based on preference data from user reviews. DataRobot rates 4.4/5 stars with 26 reviews. By contrast, Dataiku DSS rates 4.3/5 stars with 36 reviews. Each product's score is calculated with real-time data from verified user reviews, to help you make the best choice between these two options ...\",\"ae\":null,\"c\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"d\":\"www.g2.com/compare/datarobot-vs-dataiku-dss\",\"da\":\"\",\"h\":0,\"i\":\"www.g2.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Dataiku DSS | G2\",\"u\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\"},{\"a\":\"Use case: Choose Datarobot if you have data stored in spreadsheets and are seeking a platform that is the simplest, albeit one with limited flexibility, ... Dataiku vs. Datarobot .\",\"ae\":null,\"b\":\"li\\tLinkedIn\\twww.linkedin.com\",\"c\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"d\":\"www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"da\":\"\",\"e\":\"2023-08-11T00:00:00.0000000\",\"h\":0,\"i\":\"www.linkedin.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Managed Machine Learning Platforms: A Comparative Analysis - LinkedIn\",\"u\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\"},{\"a\":\"Jan 11, 2023 Dataiku is an artificial intelligence platform created in France in 2013. It has since become one of the world's benchmarks for data science and machine learning studios. What is...\",\"ae\":null,\"c\":\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"d\":\"londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"da\":\"translations\",\"e\":\"2023-01-11T00:00:00.0000000\",\"h\":0,\"i\":\"londondataconsulting.medium.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku: What is it? How to use it? Ultimate Guide 2023\",\"u\":\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\"},{\"a\":\"Use cases for version 2.x. Notebooks for uses cases that use methods for 2.x versions of DataRobot's Python client. Measure price elasticity of demand. A use case to identify relationships between price and demand, maximize revenue by properly pricing products, and monitor price elasticities for changes in price and demand. Insurance claim triage.\",\"ae\":null,\"c\":\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"d\":\"docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"da\":\"\",\"h\":0,\"i\":\"docs.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Common use cases: DataRobot docs - DataRobot AI Platform\",\"u\":\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\"},{\"a\":\"With the Use Case Value Tracker, you can manage the project lifecycle and understand the value associated with each step. It also enables you to associate and organize all your DataRobot artifacts (e.g., datasets, models, deployments, applications, etc.) around a given use case for a holistic view. In addition to the project management aspects ...\",\"ae\":null,\"c\":\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"d\":\"www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"da\":\"\",\"h\":0,\"i\":\"www.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Introducing the DataRobot Use Case Value Tracker\",\"u\":\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\"},{\"a\":\"Use Cases are folder-like containers inside of DataRobot Workbench that allow you to group everything related to solving a specific business problem\\u2014datasets, models, experiments, No-Code AI Apps, and notebooks\\u2014inside of a single, manageable entity. You can share whole Use Cases as well as the individual assets they contain.\",\"ae\":null,\"c\":\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"d\":\"docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"da\":\"\",\"e\":\"2023-09-15T00:00:00.0000000\",\"h\":0,\"i\":\"docs.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases: DataRobot docs\",\"u\":\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\"},{\"a\":\"January 2, 2024 Use Cases & Projects, Featured Sophie Dionnet Leveraging AI to Cut Costs December 29, 2023 Data Basics, Featured\",\"ae\":null,\"c\":\"https://blog.dataiku.com/topic/use-cases-projects\",\"d\":\"blog.dataiku.com/topic/use-cases-projects\",\"da\":\"\",\"h\":0,\"i\":\"blog.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Blog - Dataiku | Use Cases & Projects\",\"u\":\"https://blog.dataiku.com/topic/use-cases-projects\"},{\"a\":\"The platforms we've chosen for our analysis are ClearML, cnvrg.io, Dataiku, Datarobot, Iguazio, Sagemaker, Seldon and Valohai from the managed side, and Flyte, Kubeflow, MLflow and Metaflow from the open-source side. This is by no means an exhaustive list of all the MLOps tools out there.\",\"ae\":null,\"c\":\"https://valohai.com/mlops-platforms-compared/\",\"d\":\"valohai.com/mlops-platforms-compared/\",\"da\":\"\",\"h\":0,\"i\":\"valohai.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"MLOps Platforms Compared - Valohai\",\"u\":\"https://valohai.com/mlops-platforms-compared/\"},{\"a\":\"DataRobot. DSS is for all companies, whatever their expertise, industry or size, that want to create their own data-driven strategic advantages by transforming their raw data into business impacting predictions. Cloud based machine learning platform which helps enterprises scale data science capabilities through deploying machine learning ...\",\"ae\":null,\"c\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"d\":\"www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"da\":\"translations\",\"h\":0,\"i\":\"www.capterra.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare Dataiku vs DataRobot 2024 | Capterra\",\"u\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\"},{\"a\":\"For Every Industry & Use Case. Organizations that use Dataiku elevate their people (whether technical and working in code or on the business side and low- or no-code) to extraordinary, arming them with the ability to make better day-to-day decisions with data across: Banking & Insurance. Pharmaceuticals. Manufacturing. Telecommunications.\",\"ae\":null,\"c\":\"https://pages.dataiku.com/experience-a-dataiku-demo\",\"d\":\"pages.dataiku.com/experience-a-dataiku-demo\",\"da\":\"\",\"h\":0,\"i\":\"pages.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Check Out This Dataiku Demo\",\"u\":\"https://pages.dataiku.com/experience-a-dataiku-demo\"},{\"a\":\"4 Star 24% 3 Star 1% 2 Star 0% 1 Star 0% Distribution based on 504 ratings Customer Experience Evaluation & Contracting 4.6 Integration & Deployment 4.7 Service & Support 4.8 Product Capabilities 4.8 FREE View and Download Peer Insights About Dataiku\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku Reviews, Ratings & Features 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\"},{\"a\":\"Read The Full Case Study U.S. Venture + Dataiku: Upskilling Analysts to Save Thousands of Hours The Data and Analytics team at U.S. Venture was built to usher the company into the future of data science and AI.\",\"ae\":null,\"c\":\"https://www.dataiku.com/stories/\",\"d\":\"www.dataiku.com/stories/\",\"da\":\"\",\"h\":0,\"i\":\"www.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Stories | Dataiku\",\"u\":\"https://www.dataiku.com/stories/\"},{\"a\":\"MLOps Deploy, monitor, and maintain machine learning models, all in a single platform. Explore the Capability Collaboration With Dataiku, teams can move beyond the lab and build real and safe Generative AI applications at enterprise scale. Explore the Capability Governance\",\"ae\":null,\"c\":\"https://www.dataiku.com/\",\"d\":\"www.dataiku.com\",\"da\":\"\",\"h\":0,\"i\":\"www.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku | Everyday AI, Extraordinary People\",\"u\":\"https://www.dataiku.com/\"},{\"a\":\"The company today announced that it raised $200 million in a Series F round led by Wellington Management at a $3.7 billion valuation, down from the $4.6 billion that Dataiku received in August ...\",\"ae\":null,\"b\":\"tc\\tTechcrunch\\ttechcrunch.com\",\"c\":\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"d\":\"techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"da\":\"translations\",\"e\":\"2022-12-13T17:10:00.0000000\",\"h\":0,\"i\":\"techcrunch.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"AI and analytics platform Dataiku raises $200M at a reduced valuation\",\"u\":\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\"},{\"a\":\"Today, more than 500 companies worldwide use Dataiku to integrate and streamline their use of data, analytics, and AI, driving diverse use cases from fraud detection and customer churn prevention ...\",\"ae\":null,\"c\":\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\",\"d\":\"www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\",\"da\":\"translations\",\"e\":\"2022-11-17T13:00:00.0000000\",\"h\":0,\"i\":\"www.globenewswire.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Ben Taylor Joins Dataiku as Chief AI Strategist - GlobeNewswire\",\"u\":\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\"},{\"n\":\"/d.js?q=Dataiku%20and%20DataRobot%20use%20cases&kl=wt-wt&l=wt-wt&p=&s=23&ex=-1&ct=US&sp=0&vqd=4-60481969350525797892441552954401970387\"}]);DDG.duckbar.load('images');DDG.duckbar.load('news');DDG.duckbar.load('videos', {\"ads\":[],\"query\":\"Dataiku and DataRobot use cases\",\"queryEncoded\":\"Dataiku%20and%20DataRobot%20use%20cases\",\"response_type\":\"places\",\"results\":[{\"content\":\"https://www.youtube.com/watch?v=ryZRRIjQ5Z8\",\"description\":\"If you're a code-first data practitioner, Dataiku helps you efficiently build high quality data pipelines and models in a number of ways. CHECK OUT DATAIKU: https://bit.ly/36XBlpK EGG ON AIR: https://bit.ly/37GhXMY BRIGHTTALK WEBINARS: https://bit.ly/33TIRjn DATA SCIENCE PIONEERS DOCUMENTARY: https://bit.ly/36V3rBF PARTNER ECOSYSTEM: https ...\",\"duration\":\"10:43\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/ryZRRIjQ5Z8?autoplay=1\",\"image_token\":\"8a4abca8613c6680a108591849e5d7b13b86111004ae004898a7f059b64c8355\",\"images\":{\"large\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\",\"medium\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\",\"motion\":\"https://tse4.mm.bing.net/th?id=OM1.cmvppfhHVUeE4Q_1684256861&pid=Api\",\"small\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2021-06-08T21:15:02.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":12391},\"title\":\"Dataiku Demo for Data Scientists and Coders\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=jKL_I0SCl_E\",\"description\":\"This video showcases the Clinical Trial Explorer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/jKL_I0SCl_E?autoplay=1\",\"image_token\":\"7b19602fe6d9b761aa3cc138448cc632ddbed31da3abf2687f36705f5945973d\",\"images\":{\"large\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\",\"medium\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\",\"motion\":\"https://tse3.mm.bing.net/th?id=OM2.ZX_yq0xmyCGZBg_1696372683&pid=Api\",\"small\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:19:00.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":150},\"title\":\"Clinical Trial Explorer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=lASesA4gNFI\",\"description\":\"This video showcases the CO2 Forecast Analyzer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/lASesA4gNFI?autoplay=1\",\"image_token\":\"a09328adca01a788783d759561c2f9c9d4d214e5a26f1462d2b6b69f21a2d478\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\",\"motion\":\"\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:16:20.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":49},\"title\":\"CO2 Forecast Analyzer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=RecpD6Vtzj4\",\"description\":\"This video showcases the LLM-Enhanced ESG Document Intelligence use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the ...\",\"duration\":\"1:20\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/RecpD6Vtzj4?autoplay=1\",\"image_token\":\"6f797accb167e2e6ff7265e35116cdeb9f1c641b1df47932d9597b61b0108614\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM1.M8WXwCQ79nrqEA_1691502936&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:30:00.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":382},\"title\":\"LLM-Enhanced ESG Document Intelligence\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=zLW0TkJoHLw\",\"description\":\"This video showcases the Demand Forecast Analyzer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:42\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/zLW0TkJoHLw?autoplay=1\",\"image_token\":\"524eb9572bf9342b859509285d39ec4661fc572cb1452307acc5341b56bab921\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\",\"motion\":\"\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:15:02.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":4},\"title\":\"LLM-Enhanced Demand Forecast\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=L-Yys0fzuVY\",\"description\":\"This video showcases the Production Quality Data Explorer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest ...\",\"duration\":\"1:47\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/L-Yys0fzuVY?autoplay=1\",\"image_token\":\"82924713be1dba83d67124fcaa6cc6afd163900a0c40f25fcf6c144ed0e36536\",\"images\":{\"large\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\",\"medium\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\",\"motion\":\"https://tse4.mm.bing.net/th?id=OM2.IwRaoLRWcQXzag_1691419996&pid=Api\",\"small\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:28:29.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":175},\"title\":\"Production Quality Data Explorer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=FluiuHuaU8A\",\"description\":\"In this breakout session of Dataiku's Product Days 2021, you will see a demo of Dataiku's Data Science Studio, the centralized, collaborative, and end-to-end platform for data science in the enterprise. CHECK OUT DATAIKU: https://bit.ly/36XBlpK EGG ON AIR: https://bit.ly/37GhXMY BRIGHTTALK WEBINARS: https://bit.ly/33TIRjn DATA SCIENCE PIONEERS ...\",\"duration\":\"13:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/FluiuHuaU8A?autoplay=1\",\"image_token\":\"2943fa8c1580f2936fc11667d670c0b827b94ff3d16b897f8b5ef2e2426487b3\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM1.MIQ7BoQz1MVkNw_1662248868&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2021-07-08T15:56:22.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":3844},\"title\":\"Introduction to Dataiku Data Science | Product Days 2021\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=6TEU5JboP7k\",\"description\":\"This video showcases the LLM-Enhanced Next Best Offer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of ...\",\"duration\":\"1:47\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/6TEU5JboP7k?autoplay=1\",\"image_token\":\"a3bc327ff2f099462935a8979bb599655c7a88a44e25a64bfea7e5973f773158\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM2.wXNo1CUgYV4Flg_1694065487&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:34:32.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":462},\"title\":\"LLM-Enhanced Next Best Offer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=UVbrpX8Zkn8\",\"description\":\"Move beyond the lab and build real and safe Generative AI applications at enterprise scale. Dataiku brings enterprise-grade development tools, pre-built use cases, and AI-powered assistants throughout the platform. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore Generative AI use cases | https ...\",\"duration\":\"2:07\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/UVbrpX8Zkn8?autoplay=1\",\"image_token\":\"3968d2d01ff722efa156290344ab0b37164e57d7efa50905c346ea1cc1a5d369\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\",\"motion\":\"https://tse1.mm.bing.net/th?id=OM1.z-FRwxQ_NByK_A_1689135755&pid=Api\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:25:16.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":1100},\"title\":\"Dataiku for Generative AI: Real Applications, Real Safety\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=-amc9iVauuE\",\"description\":\"Dataiku is the leading platform for Everyday AI, systemizing the use of data for exceptional business results. In today's video we will take a tour of Dataiku's end to end capabilities by exploring a real life use case around environmental impact. Let's take a look at how a data science team with different skills can work together to turn ...\",\"duration\":\"12:35\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/-amc9iVauuE?autoplay=1\",\"image_token\":\"2a05a65ad8a2727aa5c48b8daa7f9ec363a24d4336a3509016d4b200c9d003cd\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\",\"motion\":\"https://tse1.mm.bing.net/th?id=OM1.Q2OhN9DzfowU6A_1685345657&pid=Api\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-01-09T21:12:27.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":9768},\"title\":\"End to End Demo 2023\",\"uploader\":\"Dataiku\"}],\"vqd\":{\"Dataiku%20and%20DataRobot%20use%20cases\":\"4-60481969350525797892441552954401970387\"}});DDG.duckbar.loadModule('related_searches');if (DDG.pageLayout) DDG.pageLayout.initialize({\"mainline\":{\"items\":[[\"ad\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"videos\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"]]},\"sidebar\":{\"items\":[[\"organic\"]]}}, { start: 0 });DDG.deep.emit(\"load:completed\");" + "curl-cffi-GET-https://links.duckduckgo.com/d.js-{\"params\": {\"bing_market\": \"wt-WT\", \"df\": null, \"ex\": \"-1\", \"kl\": \"wt-wt\", \"l\": \"wt-wt\", \"q\": \"Dataiku and DataRobot use cases\", \"s\": \"0\", \"sp\": \"0\", \"vqd\": \"4-60481969350525797892441552954401970387\"}}": "if (DDG.deep && DDG.deep.setUpstream) DDG.deep.setUpstream(\"bingv7aa\");DDG.deep.bn={'ivc':1};if (DDG.pageLayout) DDG.pageLayout.load('a',[{\"a\":\"\\u9ad8\\u7cbe\\u5ea6\\u306a\\u6a5f\\u68b0\\u5b66\\u7fd2\\u30e2\\u30c7\\u30eb\\u3092\\u69cb\\u7bc9\\u3001\\u5b9f\\u88c5\\u3001\\u904b\\u7528\\u3002DataRobot\\u306f\\u793e\\u5185\\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092\\u5275\\u9020\\u3057\\u307e\\u3059. AI\\u3092\\u6d3b\\u7528\\u3057\\u30c7\\u30fc\\u30bf\\u3092\\u5206\\u6790\\u3001\\u5b9f\\u7528\\u7684\\u306a\\u30a4\\u30f3\\u30b5\\u30a4\\u30c8\\u3092\\u660e\\u3089\\u304b\\u306b\\u3002\\u30d3\\u30b8\\u30cd\\u30b9\\u306e\\u8ab2\\u984c\\u3092\\u3088\\u308a\\u65e9\\u304f\\u89e3\\u6c7a\",\"adext\":{\"callout\":{\"t\":\"30-Day Free Trial \\u00b7 Trusted by Fortune 50 \\u00b7 No Vendor Lock-in\",\"tid\":\"6\"},\"filterlinks\":{\"l\":[],\"tid\":\"\"},\"sitelinks\":{\"l\":[{\"snippet\":\"Explore the DataRobot AI Platform Get Started With a 30-Day Trial\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=d0faee2c8c1aae9ac3a012e21d37352a1181970dce9edeba4107839fbfbf097a&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De81Q_rqdj1ZxH5XXGh4PG6pjVUCUzdB7rGpyykWEihNc_sSp5n%2DJ9jIyTjOSnXg0OUazrpKgDJrNvBOdNa5PjBGtyLGt23nrBAabI6opJXrliWQ4o%2DTyxIsqOeCXqzLOOJ3jJb74k6KEx20zilzwKmzSg3nBop2A9JqsasC17VVDPc3_i3EzPbWeRNS4nhxXWJqBKd55GfhuEOg2RZUbmmuAUhWvM%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnRyaWFsJTJmJTNmdXRtX21lZGl1bSUzZHNlYXJjaCUyNnV0bV9zb3VyY2UlM2RiaW5nJTI2dXRtX2NhbXBhaWduJTNkRnJlZVRyaWFsMjAyM1dXMDgxNkdQU2FkZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDc2YmMwNmFmNTA0NDFjOGVjOGYxNjMwY2FmNGU4ZTVk%26rlid%3D76bc06af50441c8ec8f1630caf4e8e5d&vqd=4-164177780916400746369660096493208330918&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5063.1\",\"text\":\"DataRobot Free Trial\"},{\"snippet\":\"Unlock Your AI Success in 2023 Tips on the Path of Value-Driven AI\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=fdb107a4de6fffdec2bdf43b561b2c63ca700daaef68f0e683547361efbbc2b0&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8%2DT0j3GTQEgr%2DmtHPM1LzNzVUCUyRxvVKYHe6LbNa2mmCfCZh3Ept1NM%2DP%2DM1AAluh_OL3VQw_FWI0A3YxC3pzzqthf3gpxan_Lv7CjKenge%2DwMYUz3bRFoFyHtQBMdgqv6T7gMGfyYwN3UCj6FNYwVVn9UNN0h1dIQanHNB6Ya9gRrPBACknA8qtsf6A2oUG1xhq7AOF98NzGphnfQ_38fySnRU%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnJlc291cmNlcyUyZmFpc3VjY2VzczIwMjMlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RDb250ZW50MTBLZXlzdG9BSVN1Y2Nlc3MyMDIzV1cwNTIyR1BTYWRleHQlMjZ1dG1fdGVybSUzZGRhdGFyb2JvdCUyNnV0bV9jb250ZW50JTNkYWRfZXh0JTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZGQzNmQ2MzlkMmFlNTEwMTM3ZTIwMDYzZWQ1ZWY3M2Yz%26rlid%3Dd36d639d2ae510137e20063ed5ef73f3&vqd=4-117927704271333462986714580056949079639&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5065.1\",\"text\":\"10 Keys to AI Success\"},{\"snippet\":\"Our Platform Includes Four Fully Integrated Products. Read More.\",\"targetUrl\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=4f06bd3312172b8e61d65ee2626dea6e26d941c3a16aa546b4e11b79e8bf027f&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8885tVmNmhi65Jmp3f2wYSzVUCUyFey1LCmrSNpGfkWzQnoC7QIbU3ztthJ%2DqKpgCmRfxudhbLK927YN84jvZlV2zTKo9DOULVj5wB8mcGXy_F42SnsrO1jZpY9NnMnzqMYPb5xZTTdgrTO1_w3Bgpd0e0VzO81_O3%2Dfo2z4UiLuVETFVqfACqR6NEwz0yfjzJe6ED9tvi_gPDiUL9iWATrNIrsw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZnByb2R1Y3QlMmYlM2ZjYW1wYWlnbmlkJTNkNTMwNzA4MDk5JTI2YWRncm91cGlkJTNkMTM1MDIwMjc3NDIxNzY5OCUyNmFkaWQlM2QlMjZtc2Nsa2lkJTNkY2U4NzQ1ZDViODBlMTJmNjQ2N2QyMDc2NDcwNDY2YjI%26rlid%3Dce8745d5b80e12f6467d2076470466b2&vqd=4-169069202740993895017985472268973083525&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5067.1\",\"text\":\"Product Overview\"}],\"tid\":\"7\\t9[8]\\t11[10]\\t13[12]\",\"type\":\"EnhancedSiteLink\"},\"tid\":\"1\"},\"ae\":{\"callout\":[\"30-Day Free Trial \\u00b7 Trusted by Fortune 50 \\u00b7 No Vendor Lock-in\"]},\"c\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=e744d99a8df00b24df71f821ad4d1332080aa03267e50f0e988d284f58d9d2ef&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8tT9soRYLZabP1ukFkRsgNzVUCUzl89Y8xEqpxoqHqIlCI5wWbydNnN_PoAKHAa2Vsio83mXA_ax16t6rJ7XGkBv0Cg7_D1eg2QAuJgPKEam4VWI3rW40B03r1p11ZXN1Gd1847Vj05bAnJnPfgVyC8ZzFQxLxONmOI0Hg182z2bZUVII26BUAlUHaVZ7O_9FEXLJWw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDA2MTIwYzhmMTAxNzEwYTZiNmRiNjkyY2VmMWRiOTY1%26rlid%3D06120c8f101710a6b6db692cef1db965&vqd=4-91027509783546726889708070523412001433&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5058.1\",\"d\":\"datarobot.com\",\"h\":0,\"i\":\"\",\"k\":0,\"m\":0,\"o\":\"\",\"p\":1,\"relevancy\":{\"abstract\":\"%E9%AB%98%E7%B2%BE%E5%BA%A6%E3%81%AA%E6%A9%9F%E6%A2%B0%E5%AD%A6%E7%BF%92%E3%83%A2%E3%83%87%E3%83%AB%E3%82%92%E6%A7%8B%E7%AF%89%E3%80%81%E5%AE%9F%E8%A3%85%E3%80%81%E9%81%8B%E7%94%A8%E3%80%82%3Cb%3EDataRobot%3C%2Fb%3E%E3%81%AF%E7%A4%BE%E5%86%85%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92%E5%89%B5%E9%80%A0%E3%81%97%E3%81%BE%E3%81%99.%20AI%E3%82%92%E6%B4%BB%E7%94%A8%E3%81%97%E3%83%87%E3%83%BC%E3%82%BF%E3%82%92%E5%88%86%E6%9E%90%E3%80%81%E5%AE%9F%E7%94%A8%E7%9A%84%E3%81%AA%E3%82%A4%E3%83%B3%E3%82%B5%E3%82%A4%E3%83%88%E3%82%92%E6%98%8E%E3%82%89%E3%81%8B%E3%81%AB%E3%80%82%E3%83%93%E3%82%B8%E3%83%8D%E3%82%B9%E3%81%AE%E8%AA%B2%E9%A1%8C%E3%82%92%E3%82%88%E3%82%8A%E6%97%A9%E3%81%8F%E8%A7%A3%E6%B1%BA\",\"adx_name\":\"none\",\"is_good_v10\":0,\"organic_ranks\":[5,11,12,13],\"q\":\"Dataiku%20and%20DataRobot%20use%20cases\",\"q_words\":4,\"q_words_fuzzy\":0.25,\"q_words_in_ad\":1,\"root_domain\":\"datarobot.com\",\"start\":\"0\",\"title\":\"%E3%83%93%E3%83%83%E3%82%B0%E3%83%87%E3%83%BC%E3%82%BF%E5%88%86%E6%9E%90%E3%82%92%E9%AB%98%E9%80%9F%E5%8C%96%20%2D%20%E3%83%87%E3%83%BC%E3%82%BF%E3%81%8B%E3%82%89%E6%96%B0%E3%81%97%E3%81%84%E4%BE%A1%E5%80%A4%E3%82%92\"},\"s\":\"bingv7aa\",\"t\":\"\\u30d3\\u30c3\\u30b0\\u30c7\\u30fc\\u30bf\\u5206\\u6790\\u3092\\u9ad8\\u901f\\u5316 - \\u30c7\\u30fc\\u30bf\\u304b\\u3089\\u65b0\\u3057\\u3044\\u4fa1\\u5024\\u3092\",\"tid\":\"1,6,7,9[8],11[10],13[12]\",\"u\":\"https://duckduckgo.com/y.js?ad_domain=datarobot.com&ad_provider=bingv7aa&ad_type=txad&eddgt=2_trBPli7jgDj1WnqJbnww%3D%3D&rut=e744d99a8df00b24df71f821ad4d1332080aa03267e50f0e988d284f58d9d2ef&u3=https%3A%2F%2Fwww.bing.com%2Faclick%3Fld%3De8tT9soRYLZabP1ukFkRsgNzVUCUzl89Y8xEqpxoqHqIlCI5wWbydNnN_PoAKHAa2Vsio83mXA_ax16t6rJ7XGkBv0Cg7_D1eg2QAuJgPKEam4VWI3rW40B03r1p11ZXN1Gd1847Vj05bAnJnPfgVyC8ZzFQxLxONmOI0Hg182z2bZUVII26BUAlUHaVZ7O_9FEXLJWw%26u%3DaHR0cHMlM2ElMmYlMmZ3d3cuZGF0YXJvYm90LmNvbSUyZmpwJTJmbHAlMmZhaS1mb3ItYnVzaW5lc3MlMmYlM2Z1dG1fbWVkaXVtJTNkc2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2RERU1PMjAyM0FsbFByb2R1Y3RzSlAwNjI2QlBTJTI2dXRtX3Rlcm0lM2RkYXRhcm9ib3QlMjZ1dG1fY29udGVudCUzZERSX2JyYW5kZWRfcnNhJTI2Y2FtcGFpZ25pZCUzZDUzMDcwODA5OSUyNmFkZ3JvdXBpZCUzZDEzNTAyMDI3NzQyMTc2OTglMjZhZGlkJTNkJTI2bXNjbGtpZCUzZDA2MTIwYzhmMTAxNzEwYTZiNmRiNjkyY2VmMWRiOTY1%26rlid%3D06120c8f101710a6b6db692cef1db965&vqd=4-91027509783546726889708070523412001433&iurl=%7B1%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26ID%3DDevEx%2C5058.1\"}], {\"page_load_url\":\"https://duckduckgo.com/y.js?ifu=%7B3%7Dappid%3D055AAD1BA669BEB8B048128DC89A107C678B527B%26rguid%3D309794dc72f748f6a2b95ce5c34fbcec&iurl=%7B2%7DIG%3D5E053B32922A4B5781ED405D9621559B%26CID%3D0176CEA622686E9D34D1DAA0235D6F30%26Type%3DEvent.CPT%26DATA%3D0\",\"visibility_url\":\"https://duckduckgo.com/y.js?ivu=%7B4%7Dtype%3Dmv%26reqver%3D1.0%26rg%3D309794dc72f748f6a2b95ce5c34fbcec\"});DDG.deep.signalSummary = \"\";DDG.inject('DDG.Data.languages.resultLanguages', {\"en\":[\"https://knowledge.dataiku.com/latest/use-cases/index.html\",\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"https://www.datarobot.com/use-cases/\",\"https://academy.dataiku.com/page/use-cases\",\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"https://blog.dataiku.com/topic/use-cases-projects\",\"https://valohai.com/mlops-platforms-compared/\",\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"https://pages.dataiku.com/experience-a-dataiku-demo\",\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"https://www.dataiku.com/stories/\",\"https://www.dataiku.com/\",\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\"]});DDG.deep.pageLayoutSummary = \"a1w4v1w19,w1\";DDG.inject('DDG.Data.languages.adLanguages', {});if (DDG.pageLayout) DDG.pageLayout.load('d',[{\"a\":\"Use Cases - Dataiku Knowledge Base Use Cases # These use cases allow you to practice what you've learned by building simplified, but complete use cases in Dataiku. Topics # Data Preparation Use Cases Classification Use Cases Clustering Use Cases Plugin Use Cases\",\"ae\":null,\"c\":\"https://knowledge.dataiku.com/latest/use-cases/index.html\",\"d\":\"knowledge.dataiku.com/latest/use-cases/index.html\",\"da\":\"\",\"h\":0,\"i\":\"knowledge.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases - Dataiku Knowledge Base\",\"u\":\"https://knowledge.dataiku.com/latest/use-cases/index.html\"},{\"a\":\"Community Dataiku Use Cases & Success Stories \\u26a0\\ufe0f Discover pioneering Dataiku use cases and success stories shared by customers, partners, academics, and nonprofits participating in the Dataiku Frontrunner Awards. Use the following labels to filter submissions by industry:\",\"ae\":null,\"c\":\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"d\":\"community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\",\"da\":\"\",\"h\":0,\"i\":\"community.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku Use Cases & Success Stories - Dataiku Community\",\"u\":\"https://community.dataiku.com/t5/Dataiku-Use-Cases-Success/tkb-p/use-cases\"},{\"a\":\"Dataiku is a cross-platform desktop application that includes a broad range of tools, such as notebooks (similar to Jupyter Notebook), workflow management (similar to Apache Airflow), and automated machine learning. In general, Dataiku aims to replace many of your existing tools rather than to integrate with them.\",\"ae\":null,\"c\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"d\":\"www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\",\"da\":\"\",\"h\":0,\"i\":\"www.datarevenue.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"ML Platforms: Dataiku vs. Alteryx vs. Sagemaker vs. Datarobot\",\"u\":\"https://www.datarevenue.com/en-blog/ml-platforms-dataiku-vs-alteryx-vs-sagemaker\"},{\"a\":\"Dataiku has a rating of 4.8 stars with 504 reviews. DataRobot has a rating of 4.6 stars with 508 reviews. See side-by-side comparisons of product capabilities, customer experience, pros and cons, and reviewer demographics to find the best fit for your organization. See more companies in the Data Science and Machine Learning Platforms market. PDF.\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku vs DataRobot 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/compare/dataiku-vs-datarobot\"},{\"a\":\"In my humble opinion DSS is a more a 'toolbox', where as DataRobot is an autoML platform. DataRobot is really good at what it does - if you have non-technical team who want to drop in data and leave everything to autoML then this may be the option for them.\",\"ae\":null,\"c\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"d\":\"community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\",\"da\":\"\",\"h\":0,\"i\":\"community.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Solved: Dataiku vs DataRobot - Dataiku Community\",\"u\":\"https://community.dataiku.com/t5/General-Discussion/Dataiku-vs-DataRobot/m-p/9315\"},{\"a\":\"Use cases AI Use Cases AI-driven organizations around the world use DataRobot to solve their most pressing business problems. Build with Free Trial Recent Popular Filters Ready to Get Started? See how a value-driven approach to AI can accelerate time to impact. Start Free Trial\",\"ae\":null,\"c\":\"https://www.datarobot.com/use-cases/\",\"d\":\"www.datarobot.com/use-cases/\",\"da\":\"\",\"h\":0,\"i\":\"www.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Machine Learning Use Cases | DataRobot AI Platform\",\"u\":\"https://www.datarobot.com/use-cases/\"},{\"a\":\"With Dataiku's AI Prepare assistant, you can work smarter, not harder. Simply describe the transformation you want to apply in natural language and the AI assistant automatically generates the necessary data preparation steps. The ability to modify both your prompt and the resulting steps means you can prepare data faster than ever, yet still ...\",\"ae\":null,\"c\":\"https://academy.dataiku.com/page/use-cases\",\"d\":\"academy.dataiku.com/page/use-cases\",\"da\":\"\",\"h\":0,\"i\":\"academy.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases - Dataiku\",\"u\":\"https://academy.dataiku.com/page/use-cases\"},{\"a\":\"84 Reviews and Ratings Path to AI Success Compare Dataiku DSS vs DataRobot. 103 verified user reviews and ratings of features, pros, cons, pricing, support and more.\",\"ae\":null,\"c\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"d\":\"www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\",\"da\":\"\",\"h\":0,\"i\":\"www.trustradius.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku DSS vs DataRobot | TrustRadius\",\"u\":\"https://www.trustradius.com/compare-products/dataiku-dss-vs-datarobot\"},{\"a\":\"side-by-side comparison of DataRobot vs. Dataiku DSS. based on preference data from user reviews. DataRobot rates 4.4/5 stars with 26 reviews. By contrast, Dataiku DSS rates 4.3/5 stars with 36 reviews. Each product's score is calculated with real-time data from verified user reviews, to help you make the best choice between these two options ...\",\"ae\":null,\"c\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\",\"d\":\"www.g2.com/compare/datarobot-vs-dataiku-dss\",\"da\":\"\",\"h\":0,\"i\":\"www.g2.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare DataRobot vs. Dataiku DSS | G2\",\"u\":\"https://www.g2.com/compare/datarobot-vs-dataiku-dss\"},{\"a\":\"Use case: Choose Datarobot if you have data stored in spreadsheets and are seeking a platform that is the simplest, albeit one with limited flexibility, ... Dataiku vs. Datarobot .\",\"ae\":null,\"b\":\"li\\tLinkedIn\\twww.linkedin.com\",\"c\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"d\":\"www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\",\"da\":\"\",\"e\":\"2023-08-11T00:00:00.0000000\",\"h\":0,\"i\":\"www.linkedin.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Managed Machine Learning Platforms: A Comparative Analysis - LinkedIn\",\"u\":\"https://www.linkedin.com/pulse/managed-machine-learning-platforms-comparative-analysis/\"},{\"a\":\"Jan 11, 2023 Dataiku is an artificial intelligence platform created in France in 2013. It has since become one of the world's benchmarks for data science and machine learning studios. What is...\",\"ae\":null,\"c\":\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"d\":\"londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\",\"da\":\"translations\",\"e\":\"2023-01-11T00:00:00.0000000\",\"h\":0,\"i\":\"londondataconsulting.medium.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku: What is it? How to use it? Ultimate Guide 2023\",\"u\":\"https://londondataconsulting.medium.com/dataiku-what-is-it-how-to-use-it-ultimate-guide-2023-47602c85a48b\"},{\"a\":\"Use cases for version 2.x. Notebooks for uses cases that use methods for 2.x versions of DataRobot's Python client. Measure price elasticity of demand. A use case to identify relationships between price and demand, maximize revenue by properly pricing products, and monitor price elasticities for changes in price and demand. Insurance claim triage.\",\"ae\":null,\"c\":\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"d\":\"docs.datarobot.com/en/docs/api/guide/common-case/index.html\",\"da\":\"\",\"h\":0,\"i\":\"docs.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Common use cases: DataRobot docs - DataRobot AI Platform\",\"u\":\"https://docs.datarobot.com/en/docs/api/guide/common-case/index.html\"},{\"a\":\"With the Use Case Value Tracker, you can manage the project lifecycle and understand the value associated with each step. It also enables you to associate and organize all your DataRobot artifacts (e.g., datasets, models, deployments, applications, etc.) around a given use case for a holistic view. In addition to the project management aspects ...\",\"ae\":null,\"c\":\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"d\":\"www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\",\"da\":\"\",\"h\":0,\"i\":\"www.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Introducing the DataRobot Use Case Value Tracker\",\"u\":\"https://www.datarobot.com/blog/introducing-the-datarobot-use-case-value-tracker/\"},{\"a\":\"Use Cases are folder-like containers inside of DataRobot Workbench that allow you to group everything related to solving a specific business problem\\u2014datasets, models, experiments, No-Code AI Apps, and notebooks\\u2014inside of a single, manageable entity. You can share whole Use Cases as well as the individual assets they contain.\",\"ae\":null,\"c\":\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"d\":\"docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\",\"da\":\"\",\"e\":\"2023-09-15T00:00:00.0000000\",\"h\":0,\"i\":\"docs.datarobot.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Use Cases: DataRobot docs\",\"u\":\"https://docs.datarobot.com/en/docs/workbench/wb-usecase/wb-build-usecase.html\"},{\"a\":\"January 2, 2024 Use Cases & Projects, Featured Sophie Dionnet Leveraging AI to Cut Costs December 29, 2023 Data Basics, Featured\",\"ae\":null,\"c\":\"https://blog.dataiku.com/topic/use-cases-projects\",\"d\":\"blog.dataiku.com/topic/use-cases-projects\",\"da\":\"\",\"h\":0,\"i\":\"blog.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Blog - Dataiku | Use Cases & Projects\",\"u\":\"https://blog.dataiku.com/topic/use-cases-projects\"},{\"a\":\"The platforms we've chosen for our analysis are ClearML, cnvrg.io, Dataiku, Datarobot, Iguazio, Sagemaker, Seldon and Valohai from the managed side, and Flyte, Kubeflow, MLflow and Metaflow from the open-source side. This is by no means an exhaustive list of all the MLOps tools out there.\",\"ae\":null,\"c\":\"https://valohai.com/mlops-platforms-compared/\",\"d\":\"valohai.com/mlops-platforms-compared/\",\"da\":\"\",\"h\":0,\"i\":\"valohai.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"MLOps Platforms Compared - Valohai\",\"u\":\"https://valohai.com/mlops-platforms-compared/\"},{\"a\":\"DataRobot. DSS is for all companies, whatever their expertise, industry or size, that want to create their own data-driven strategic advantages by transforming their raw data into business impacting predictions. Cloud based machine learning platform which helps enterprises scale data science capabilities through deploying machine learning ...\",\"ae\":null,\"c\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"d\":\"www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\",\"da\":\"translations\",\"h\":0,\"i\":\"www.capterra.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Compare Dataiku vs DataRobot 2024 | Capterra\",\"u\":\"https://www.capterra.com/machine-learning-software/compare/142192-179303/Data-Science-Studio-DSS-vs-DataRobot\"},{\"a\":\"For Every Industry & Use Case. Organizations that use Dataiku elevate their people (whether technical and working in code or on the business side and low- or no-code) to extraordinary, arming them with the ability to make better day-to-day decisions with data across: Banking & Insurance. Pharmaceuticals. Manufacturing. Telecommunications.\",\"ae\":null,\"c\":\"https://pages.dataiku.com/experience-a-dataiku-demo\",\"d\":\"pages.dataiku.com/experience-a-dataiku-demo\",\"da\":\"\",\"h\":0,\"i\":\"pages.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Check Out This Dataiku Demo\",\"u\":\"https://pages.dataiku.com/experience-a-dataiku-demo\"},{\"a\":\"4 Star 24% 3 Star 1% 2 Star 0% 1 Star 0% Distribution based on 504 ratings Customer Experience Evaluation & Contracting 4.6 Integration & Deployment 4.7 Service & Support 4.8 Product Capabilities 4.8 FREE View and Download Peer Insights About Dataiku\",\"ae\":null,\"c\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"d\":\"www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\",\"da\":\"\",\"h\":0,\"i\":\"www.gartner.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku Reviews, Ratings & Features 2024 | Gartner Peer Insights\",\"u\":\"https://www.gartner.com/reviews/market/data-science-and-machine-learning-platforms/vendor/dataiku/product/dataiku\"},{\"a\":\"Read The Full Case Study U.S. Venture + Dataiku: Upskilling Analysts to Save Thousands of Hours The Data and Analytics team at U.S. Venture was built to usher the company into the future of data science and AI.\",\"ae\":null,\"c\":\"https://www.dataiku.com/stories/\",\"d\":\"www.dataiku.com/stories/\",\"da\":\"\",\"h\":0,\"i\":\"www.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Stories | Dataiku\",\"u\":\"https://www.dataiku.com/stories/\"},{\"a\":\"MLOps Deploy, monitor, and maintain machine learning models, all in a single platform. Explore the Capability Collaboration With Dataiku, teams can move beyond the lab and build real and safe Generative AI applications at enterprise scale. Explore the Capability Governance\",\"ae\":null,\"c\":\"https://www.dataiku.com/\",\"d\":\"www.dataiku.com\",\"da\":\"\",\"h\":0,\"i\":\"www.dataiku.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Dataiku | Everyday AI, Extraordinary People\",\"u\":\"https://www.dataiku.com/\"},{\"a\":\"The company today announced that it raised $200 million in a Series F round led by Wellington Management at a $3.7 billion valuation, down from the $4.6 billion that Dataiku received in August ...\",\"ae\":null,\"b\":\"tc\\tTechcrunch\\ttechcrunch.com\",\"c\":\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"d\":\"techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\",\"da\":\"translations\",\"e\":\"2022-12-13T17:10:00.0000000\",\"h\":0,\"i\":\"techcrunch.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"AI and analytics platform Dataiku raises $200M at a reduced valuation\",\"u\":\"https://techcrunch.com/2022/12/13/ai-and-analytics-platform-dataiku-raises-200m-at-a-reduced-valuation/\"},{\"a\":\"Today, more than 500 companies worldwide use Dataiku to integrate and streamline their use of data, analytics, and AI, driving diverse use cases from fraud detection and customer churn prevention ...\",\"ae\":null,\"c\":\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\",\"d\":\"www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\",\"da\":\"translations\",\"e\":\"2022-11-17T13:00:00.0000000\",\"h\":0,\"i\":\"www.globenewswire.com\",\"k\":null,\"m\":0,\"o\":0,\"p\":0,\"s\":\"bingv7aa\",\"t\":\"Ben Taylor Joins Dataiku as Chief AI Strategist - GlobeNewswire\",\"u\":\"https://www.globenewswire.com/news-release/2022/11/17/2558152/0/en/Ben-Taylor-Joins-Dataiku-as-Chief-AI-Strategist.html\"},{\"n\":\"/d.js?q=Dataiku%20and%20DataRobot%20use%20cases&kl=wt-wt&l=wt-wt&p=&s=23&ex=-1&ct=US&sp=0&vqd=4-60481969350525797892441552954401970387\"}]);DDG.duckbar.load('images');DDG.duckbar.load('news');DDG.duckbar.load('videos', {\"ads\":[],\"query\":\"Dataiku and DataRobot use cases\",\"queryEncoded\":\"Dataiku%20and%20DataRobot%20use%20cases\",\"response_type\":\"places\",\"results\":[{\"content\":\"https://www.youtube.com/watch?v=ryZRRIjQ5Z8\",\"description\":\"If you're a code-first data practitioner, Dataiku helps you efficiently build high quality data pipelines and models in a number of ways. CHECK OUT DATAIKU: https://bit.ly/36XBlpK EGG ON AIR: https://bit.ly/37GhXMY BRIGHTTALK WEBINARS: https://bit.ly/33TIRjn DATA SCIENCE PIONEERS DOCUMENTARY: https://bit.ly/36V3rBF PARTNER ECOSYSTEM: https ...\",\"duration\":\"10:43\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/ryZRRIjQ5Z8?autoplay=1\",\"image_token\":\"8a4abca8613c6680a108591849e5d7b13b86111004ae004898a7f059b64c8355\",\"images\":{\"large\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\",\"medium\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\",\"motion\":\"https://tse4.mm.bing.net/th?id=OM1.cmvppfhHVUeE4Q_1684256861&pid=Api\",\"small\":\"https://tse4.mm.bing.net/th?id=OVP.WoendyuZJ9qxql-n6jit5AEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2021-06-08T21:15:02.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":12391},\"title\":\"Dataiku Demo for Data Scientists and Coders\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=jKL_I0SCl_E\",\"description\":\"This video showcases the Clinical Trial Explorer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/jKL_I0SCl_E?autoplay=1\",\"image_token\":\"7b19602fe6d9b761aa3cc138448cc632ddbed31da3abf2687f36705f5945973d\",\"images\":{\"large\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\",\"medium\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\",\"motion\":\"https://tse3.mm.bing.net/th?id=OM2.ZX_yq0xmyCGZBg_1696372683&pid=Api\",\"small\":\"https://tse3.mm.bing.net/th?id=OVP.wfgHp53woiVosZ37-1HtnwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:19:00.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":150},\"title\":\"Clinical Trial Explorer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=lASesA4gNFI\",\"description\":\"This video showcases the CO2 Forecast Analyzer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/lASesA4gNFI?autoplay=1\",\"image_token\":\"a09328adca01a788783d759561c2f9c9d4d214e5a26f1462d2b6b69f21a2d478\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\",\"motion\":\"\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.ZhQI7z-IMlcVnyeMJuGlAQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:16:20.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":49},\"title\":\"CO2 Forecast Analyzer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=RecpD6Vtzj4\",\"description\":\"This video showcases the LLM-Enhanced ESG Document Intelligence use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the ...\",\"duration\":\"1:20\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/RecpD6Vtzj4?autoplay=1\",\"image_token\":\"6f797accb167e2e6ff7265e35116cdeb9f1c641b1df47932d9597b61b0108614\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM1.M8WXwCQ79nrqEA_1691502936&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.bm4gAiJOOmKV7uup6LU9pgEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:30:00.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":382},\"title\":\"LLM-Enhanced ESG Document Intelligence\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=zLW0TkJoHLw\",\"description\":\"This video showcases the Demand Forecast Analyzer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of our ...\",\"duration\":\"1:42\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/zLW0TkJoHLw?autoplay=1\",\"image_token\":\"524eb9572bf9342b859509285d39ec4661fc572cb1452307acc5341b56bab921\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\",\"motion\":\"\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.yIekQ2cMUesOPJTfsYIZzQHgFo&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:15:02.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":4},\"title\":\"LLM-Enhanced Demand Forecast\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=L-Yys0fzuVY\",\"description\":\"This video showcases the Production Quality Data Explorer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest ...\",\"duration\":\"1:47\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/L-Yys0fzuVY?autoplay=1\",\"image_token\":\"82924713be1dba83d67124fcaa6cc6afd163900a0c40f25fcf6c144ed0e36536\",\"images\":{\"large\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\",\"medium\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\",\"motion\":\"https://tse4.mm.bing.net/th?id=OM2.IwRaoLRWcQXzag_1691419996&pid=Api\",\"small\":\"https://tse4.mm.bing.net/th?id=OVP.teiC0mX9nKCbH8qeo52udwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:28:29.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":175},\"title\":\"Production Quality Data Explorer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=FluiuHuaU8A\",\"description\":\"In this breakout session of Dataiku's Product Days 2021, you will see a demo of Dataiku's Data Science Studio, the centralized, collaborative, and end-to-end platform for data science in the enterprise. CHECK OUT DATAIKU: https://bit.ly/36XBlpK EGG ON AIR: https://bit.ly/37GhXMY BRIGHTTALK WEBINARS: https://bit.ly/33TIRjn DATA SCIENCE PIONEERS ...\",\"duration\":\"13:50\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/FluiuHuaU8A?autoplay=1\",\"image_token\":\"2943fa8c1580f2936fc11667d670c0b827b94ff3d16b897f8b5ef2e2426487b3\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM1.MIQ7BoQz1MVkNw_1662248868&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.RIM-ftwDZjYP58RimJfgwwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2021-07-08T15:56:22.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":3844},\"title\":\"Introduction to Dataiku Data Science | Product Days 2021\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=6TEU5JboP7k\",\"description\":\"This video showcases the LLM-Enhanced Next Best Offer use case from the Dataiku Generative AI Use Case Collection. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore more Generative AI use cases | https://experience.dataiku.com/generative-ai To explore more about Dataiku, check out the rest of ...\",\"duration\":\"1:47\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/6TEU5JboP7k?autoplay=1\",\"image_token\":\"a3bc327ff2f099462935a8979bb599655c7a88a44e25a64bfea7e5973f773158\",\"images\":{\"large\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\",\"medium\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\",\"motion\":\"https://tse2.mm.bing.net/th?id=OM2.wXNo1CUgYV4Flg_1694065487&pid=Api\",\"small\":\"https://tse2.mm.bing.net/th?id=OVP.Q6SP0MmL89M_TnLvNPG4oQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:34:32.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":462},\"title\":\"LLM-Enhanced Next Best Offer\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=UVbrpX8Zkn8\",\"description\":\"Move beyond the lab and build real and safe Generative AI applications at enterprise scale. Dataiku brings enterprise-grade development tools, pre-built use cases, and AI-powered assistants throughout the platform. Learn more about Dataiku for Generative AI | https://www.dataiku.com/product/generative-ai/ Explore Generative AI use cases | https ...\",\"duration\":\"2:07\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/UVbrpX8Zkn8?autoplay=1\",\"image_token\":\"3968d2d01ff722efa156290344ab0b37164e57d7efa50905c346ea1cc1a5d369\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\",\"motion\":\"https://tse1.mm.bing.net/th?id=OM1.z-FRwxQ_NByK_A_1689135755&pid=Api\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.F-xH3-wKfTjM3YMshnjWwwEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-06-22T12:25:16.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":1100},\"title\":\"Dataiku for Generative AI: Real Applications, Real Safety\",\"uploader\":\"Dataiku\"},{\"content\":\"https://www.youtube.com/watch?v=-amc9iVauuE\",\"description\":\"Dataiku is the leading platform for Everyday AI, systemizing the use of data for exceptional business results. In today's video we will take a tour of Dataiku's end to end capabilities by exploring a real life use case around environmental impact. Let's take a look at how a data science team with different skills can work together to turn ...\",\"duration\":\"12:35\",\"embed_html\":\"\",\"embed_url\":\"http://www.youtube.com/embed/-amc9iVauuE?autoplay=1\",\"image_token\":\"2a05a65ad8a2727aa5c48b8daa7f9ec363a24d4336a3509016d4b200c9d003cd\",\"images\":{\"large\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\",\"medium\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\",\"motion\":\"https://tse1.mm.bing.net/th?id=OM1.Q2OhN9DzfowU6A_1685345657&pid=Api\",\"small\":\"https://tse1.mm.bing.net/th?id=OVP.Az9RhdSVwpXe56mGcs6FqQEsDh&pid=Api\"},\"provider\":\"Bing\",\"published\":\"2023-01-09T21:12:27.0000000\",\"publisher\":\"YouTube\",\"statistics\":{\"viewCount\":9768},\"title\":\"End to End Demo 2023\",\"uploader\":\"Dataiku\"}],\"vqd\":{\"Dataiku%20and%20DataRobot%20use%20cases\":\"4-60481969350525797892441552954401970387\"}});DDG.duckbar.loadModule('related_searches');if (DDG.pageLayout) DDG.pageLayout.initialize({\"mainline\":{\"items\":[[\"ad\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"videos\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"],[\"organic\"]]},\"sidebar\":{\"items\":[[\"organic\"]]}}, { start: 0 });DDG.deep.emit(\"load:completed\");", + "aiohttp-post-https://google.serper.dev/search-{\"data\": \"[{\\\"num\\\": 6, \\\"page\\\": 1, \\\"q\\\": \\\"ai agent\\\"}]\", \"headers\": {\"Content-Type\": \"application/json\", \"X-API-KEY\": \"mock-serper-key\"}}": [ + { + "searchParameters": { + "q": "ai agent", + "num": 6, + "page": 1, + "type": "search", + "engine": "google" + }, + "organic": [ + { + "title": "AI Agent • Supercharge Your Workflows with AI", + "link": "https://aiagent.app/", + "snippet": "A web app that makes choices and performs tasks on its own, based on the goals set by you. How Does it Work?", + "position": 1 + }, + { + "title": "Intelligent agent - Wikipedia", + "link": "https://en.wikipedia.org/wiki/Intelligent_agent", + "snippet": "In artificial intelligence, an intelligent agent (IA) is an agent acting in an intelligent manner; It perceives its environment, takes actions autonomously ...", + "position": 2 + }, + { + "title": "What is an AI agent? | Zapier", + "link": "https://zapier.com/blog/ai-agent/", + "snippet": "AI Agent is a flexible app that lets you create your own agents, by picking a name, an objective, and the AI model it should use (GPT-3.5 Turbo ...", + "date": "Jun 1, 2023", + "position": 3 + }, + { + "title": "Google DeepMind Veteran Departs to Launch AI Agent Startup", + "link": "https://www.theinformation.com/articles/google-deepmind-veteran-departs-to-launch-ai-agent-startup", + "snippet": "The concept of AI agents—bots with the ability to plan and work toward a goal with minimal user guidance—first took off a year ago after ...", + "date": "10 hours ago", + "position": 4 + }, + { + "title": "AI Agents And The Era Of The Intelligent Interface - Forbes", + "link": "https://www.forbes.com/sites/davidarmano/2023/12/07/ai-agents-and-the-era-of-the-intelligent-interface/", + "snippet": "Conversing with several AI Agents connected to various systems is poised to become the next significant evolution of human-computer ...", + "date": "Dec 7, 2023", + "position": 5 + } + ], + "topStories": [ + { + "title": "Google DeepMind Veteran Departs to Launch AI Agent Startup", + "link": "https://www.theinformation.com/articles/google-deepmind-veteran-departs-to-launch-ai-agent-startup", + "source": "The Information", + "date": "10 hours ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcSOUiQnjnTQpnKKDk0hnXhpIdVvwyifhK3VjZuTey9Uq3J1S8l7OB95iWMrKQ&s" + }, + { + "title": "Bitcoin to Become Native Currency for AI Agents, Former Meta Exec Predicts", + "link": "https://u.today/bitcoin-to-become-native-currency-for-ai-agents-former-meta-exec-predicts", + "source": "U.Today", + "date": "2 days ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcRV9Ydu_Dou8HvI9E25KAn7nKmxk6Q-CB1cvT0dIxSudXhZPpGCR1vj0NCdaw&s" + }, + { + "title": "Building AI agents with Semantic Kernel", + "link": "https://www.infoworld.com/article/3712423/building-ai-agents-with-semantic-kernel.html", + "source": "InfoWorld", + "date": "5 days ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcS0NAiT2vVMB14Ff56syKnS3g4jrNN5LIskrtxqqdViPyrBsLCuCrQWu9ojdA&s" + }, + { + "title": "AI agents help explain other AI systems", + "link": "https://news.mit.edu/2024/ai-agents-help-explain-other-ai-systems-0103", + "source": "MIT News", + "date": "4 weeks ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcR-0QJKlbSMOP0wYSfB70p4_JCWHEkc6oAkLhBXVHX3ZVATBCRWTp08JY8x4w&s" + }, + { + "title": "CES 2024: LG announces walking, talking, 'Jetsons-esque' smart home robot", + "link": "https://mashable.com/article/ces-2024-lg-announcement-ai-agent-smart-home-robot", + "source": "Mashable", + "date": "3 weeks ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcRVIiyOId49n_-CwPODLHIP9t6HioG05EeI_dvwvg6WNfFBcsLliI_Xhr6U-Q&s" + }, + { + "title": "Develop Your First AI Agent: Deep Q-Learning", + "link": "https://towardsdatascience.com/develop-your-first-ai-agent-deep-q-learning-375876ee2472", + "source": "Towards Data Science", + "date": "1 month ago", + "imageUrl": "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcSEzOgQuvwfalC2s5HasdRiv2IqdMLgtuUOBJv1xkVGH-Vg_bJmavQk88I1eA&s" + } + ], + "relatedSearches": [ + { + "query": "AI agent GPT" + }, + { + "query": "AI agent OpenAI" + }, + { + "query": "AI agent examples" + }, + { + "query": "Ai agent jobs" + }, + { + "query": "AI agents ChatGPT" + }, + { + "query": "AI agent Microsoft" + }, + { + "query": "AI agent free" + }, + { + "query": "AI Agent app" + } + ] + } + ] } \ No newline at end of file diff --git a/tests/metagpt/actions/ci/test_ask_review.py b/tests/metagpt/actions/ci/test_ask_review.py new file mode 100644 index 000000000..4f02fe10b --- /dev/null +++ b/tests/metagpt/actions/ci/test_ask_review.py @@ -0,0 +1,12 @@ +import pytest + +from metagpt.actions.ci.ask_review import AskReview + + +@pytest.mark.asyncio +async def test_ask_review(mocker): + mock_review_input = "confirm" + mocker.patch("builtins.input", return_value=mock_review_input) + rsp, confirmed = await AskReview().run() + assert rsp == mock_review_input + assert confirmed diff --git a/tests/metagpt/actions/ci/test_debug_code.py b/tests/metagpt/actions/ci/test_debug_code.py new file mode 100644 index 000000000..0307ac17e --- /dev/null +++ b/tests/metagpt/actions/ci/test_debug_code.py @@ -0,0 +1,51 @@ +# -*- coding: utf-8 -*- +# @Date : 1/11/2024 8:51 PM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : + +import pytest + +from metagpt.actions.ci.debug_code import DebugCode +from metagpt.schema import Message + +ErrorStr = """Tested passed: + +Tests failed: +assert sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5] # output: [1, 2, 4, 3, 5] +""" + +CODE = """ +def sort_array(arr): + # Helper function to count the number of ones in the binary representation + def count_ones(n): + return bin(n).count('1') + + # Sort the array using a custom key function + # The key function returns a tuple (number of ones, value) for each element + # This ensures that if two elements have the same number of ones, they are sorted by their value + sorted_arr = sorted(arr, key=lambda x: (count_ones(x), x)) + + return sorted_arr +``` +""" + +DebugContext = '''Solve the problem in Python: +def sort_array(arr): + """ + In this Kata, you have to sort an array of non-negative integers according to + number of ones in their binary representation in ascending order. + For similar number of ones, sort based on decimal value. + + It must be implemented like this: + >>> sort_array([1, 5, 2, 3, 4]) == [1, 2, 3, 4, 5] + >>> sort_array([-2, -3, -4, -5, -6]) == [-6, -5, -4, -3, -2] + >>> sort_array([1, 0, 2, 3, 4]) [0, 1, 2, 3, 4] + """ +''' + + +@pytest.mark.asyncio +async def test_debug_code(): + debug_context = Message(content=DebugContext) + new_code = await DebugCode().run(context=debug_context, code=CODE, runtime_result=ErrorStr) + assert "def sort_array(arr)" in new_code["code"] diff --git a/tests/metagpt/actions/ci/test_execute_nb_code.py b/tests/metagpt/actions/ci/test_execute_nb_code.py new file mode 100644 index 000000000..72a85dd08 --- /dev/null +++ b/tests/metagpt/actions/ci/test_execute_nb_code.py @@ -0,0 +1,116 @@ +import pytest + +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode, truncate + + +@pytest.mark.asyncio +async def test_code_running(): + executor = ExecuteNbCode() + output, is_success = await executor.run("print('hello world!')") + assert is_success + + +@pytest.mark.asyncio +async def test_split_code_running(): + executor = ExecuteNbCode() + _ = await executor.run("x=1\ny=2") + _ = await executor.run("z=x+y") + output, is_success = await executor.run("assert z==3") + assert is_success + + +@pytest.mark.asyncio +async def test_execute_error(): + executor = ExecuteNbCode() + output, is_success = await executor.run("z=1/0") + assert not is_success + + +PLOT_CODE = """ +import numpy as np +import matplotlib.pyplot as plt + +# 生成随机数据 +random_data = np.random.randn(1000) # 生成1000个符合标准正态分布的随机数 + +# 绘制直方图 +plt.hist(random_data, bins=30, density=True, alpha=0.7, color='blue', edgecolor='black') + +# 添加标题和标签 +plt.title('Histogram of Random Data') +plt.xlabel('Value') +plt.ylabel('Frequency') + +# 显示图形 +plt.show() +plt.close() +""" + + +@pytest.mark.asyncio +async def test_plotting_code(): + executor = ExecuteNbCode() + output, is_success = await executor.run(PLOT_CODE) + assert is_success + + +def test_truncate(): + # 代码执行成功 + output, is_success = truncate("hello world", 5, True) + assert "Truncated to show only first 5 characters\nhello" in output + assert is_success + # 代码执行失败 + output, is_success = truncate("hello world", 5, False) + assert "Truncated to show only last 5 characters\nworld" in output + assert not is_success + # 异步 + output, is_success = truncate(" 0 + print(code_with_ml) diff --git a/tests/metagpt/actions/ci/test_write_analysis_code.py b/tests/metagpt/actions/ci/test_write_analysis_code.py new file mode 100644 index 000000000..95c7dfca8 --- /dev/null +++ b/tests/metagpt/actions/ci/test_write_analysis_code.py @@ -0,0 +1,324 @@ +import asyncio + +import pytest + +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.actions.ci.write_analysis_code import ( + WriteCodeWithoutTools, + WriteCodeWithTools, +) +from metagpt.logs import logger +from metagpt.schema import Message, Plan, Task +from metagpt.strategy.planner import STRUCTURAL_CONTEXT + + +@pytest.mark.skip +@pytest.mark.asyncio +async def test_write_code_by_list_plan(): + write_code = WriteCodeWithoutTools() + execute_code = ExecuteNbCode() + messages = [] + plan = ["随机生成一个pandas DataFrame时间序列", "绘制这个时间序列的直方图", "回顾已完成的任务", "求均值", "总结"] + for task in plan: + print(f"\n任务: {task}\n\n") + messages.append(Message(task, role="assistant")) + code = await write_code.run(messages) + if task.startswith(("回顾", "总结")): + assert code["language"] == "markdown" + else: + assert code["language"] == "python" + messages.append(Message(code["code"], role="assistant")) + assert len(code) > 0 + output, _ = await execute_code.run(**code) + print(f"\n[Output]: 任务{task}的执行结果是: \n{output}\n") + messages.append(output) + + +@pytest.mark.asyncio +async def test_tool_recommendation(): + task = "clean and preprocess the data" + available_tools = { + "FillMissingValue": "Filling missing values", + "SplitBins": "Bin continuous data into intervals and return the bin identifier encoded as an integer value", + } + write_code = WriteCodeWithTools() + tools = await write_code._recommend_tool(task, available_tools) + + assert len(tools) == 1 + assert "FillMissingValue" in tools + + +@pytest.mark.asyncio +async def test_write_code_with_tools(): + write_code = WriteCodeWithTools() + + requirement = "构造数据集并进行数据清洗" + task_map = { + "1": Task( + task_id="1", + instruction="随机生成一个pandas DataFrame数据集", + task_type="other", + dependent_task_ids=[], + code=""" + import pandas as pd + df = pd.DataFrame({ + 'a': [1, 2, 3, 4, 5], + 'b': [1.1, 2.2, 3.3, 4.4, np.nan], + 'c': ['aa', 'bb', 'cc', 'dd', 'ee'], + 'd': [1, 2, 3, 4, 5] + }) + """, + is_finished=True, + ), + "2": Task( + task_id="2", + instruction="对数据集进行数据清洗", + task_type="data_preprocess", + dependent_task_ids=["1"], + ), + } + plan = Plan( + goal="构造数据集并进行数据清洗", + tasks=list(task_map.values()), + task_map=task_map, + current_task_id="2", + ) + + context = STRUCTURAL_CONTEXT.format( + user_requirement=requirement, + context=plan.context, + tasks=list(task_map.values()), + current_task=plan.current_task.model_dump_json(), + ) + context_msg = [Message(content=context, role="user")] + + code = await write_code.run(context_msg, plan) + code = code["code"] + assert len(code) > 0 + print(code) + + +@pytest.mark.asyncio +async def test_write_code_to_correct_error(): + structural_context = """ + ## User Requirement + read a dataset test.csv and print its head + ## Current Plan + [ + { + "task_id": "1", + "dependent_task_ids": [], + "instruction": "import pandas and load the dataset from 'test.csv'.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + }, + { + "task_id": "2", + "dependent_task_ids": [ + "1" + ], + "instruction": "Print the head of the dataset to display the first few rows.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + } + ] + ## Current Task + {"task_id": "1", "dependent_task_ids": [], "instruction": "import pandas and load the dataset from 'test.csv'.", "task_type": "", "code": "", "result": "", "is_finished": false} + """ + wrong_code = """import pandas as pd\ndata = pd.read_excel('test.csv')\ndata""" # use read_excel to read a csv + error = """ + Traceback (most recent call last): + File "", line 2, in + File "/Users/gary/miniconda3/envs/py39_scratch/lib/python3.9/site-packages/pandas/io/excel/_base.py", line 478, in read_excel + io = ExcelFile(io, storage_options=storage_options, engine=engine) + File "/Users/gary/miniconda3/envs/py39_scratch/lib/python3.9/site-packages/pandas/io/excel/_base.py", line 1500, in __init__ + raise ValueError( + ValueError: Excel file format cannot be determined, you must specify an engine manually. + """ + context = [ + Message(content=structural_context, role="user"), + Message(content=wrong_code, role="assistant"), + Message(content=error, role="user"), + ] + new_code = await WriteCodeWithoutTools().run(context=context) + new_code = new_code["code"] + print(new_code) + assert "read_csv" in new_code # should correct read_excel to read_csv + + +@pytest.mark.asyncio +async def test_write_code_reuse_code_simple(): + structural_context = """ + ## User Requirement + read a dataset test.csv and print its head + ## Current Plan + [ + { + "task_id": "1", + "dependent_task_ids": [], + "instruction": "import pandas and load the dataset from 'test.csv'.", + "task_type": "", + "code": "import pandas as pd\ndata = pd.read_csv('test.csv')", + "result": "", + "is_finished": true + }, + { + "task_id": "2", + "dependent_task_ids": [ + "1" + ], + "instruction": "Print the head of the dataset to display the first few rows.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + } + ] + ## Current Task + {"task_id": "2", "dependent_task_ids": ["1"], "instruction": "Print the head of the dataset to display the first few rows.", "task_type": "", "code": "", "result": "", "is_finished": false} + """ + context = [ + Message(content=structural_context, role="user"), + ] + code = await WriteCodeWithoutTools().run(context=context) + code = code["code"] + print(code) + assert "pandas" not in code and "read_csv" not in code # should reuse import and read statement from previous one + + +@pytest.mark.skip +@pytest.mark.asyncio +async def test_write_code_reuse_code_long(): + """test code reuse for long context""" + + structural_context = """ + ## User Requirement + Run data analysis on sklearn Iris dataset, include a plot + ## Current Plan + [ + { + "task_id": "1", + "dependent_task_ids": [], + "instruction": "Load the Iris dataset from sklearn.", + "task_type": "", + "code": "from sklearn.datasets import load_iris\niris_data = load_iris()\niris_data['data'][0:5], iris_data['target'][0:5]", + "result": "(array([[5.1, 3.5, 1.4, 0.2],\n [4.9, 3. , 1.4, 0.2],\n [4.7, 3.2, 1.3, 0.2],\n [4.6, 3.1, 1.5, 0.2],\n [5. , 3.6, 1.4, 0.2]]),\n array([0, 0, 0, 0, 0]))", + "is_finished": true + }, + { + "task_id": "2", + "dependent_task_ids": [ + "1" + ], + "instruction": "Perform exploratory data analysis on the Iris dataset.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + }, + { + "task_id": "3", + "dependent_task_ids": [ + "2" + ], + "instruction": "Create a plot visualizing the Iris dataset features.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + } + ] + ## Current Task + {"task_id": "2", "dependent_task_ids": ["1"], "instruction": "Perform exploratory data analysis on the Iris dataset.", "task_type": "", "code": "", "result": "", "is_finished": false} + """ + context = [ + Message(content=structural_context, role="user"), + ] + trials_num = 5 + trials = [WriteCodeWithoutTools().run(context=context, temperature=0.0) for _ in range(trials_num)] + trial_results = await asyncio.gather(*trials) + print(*trial_results, sep="\n\n***\n\n") + success = [ + "load_iris" not in result["code"] and "iris_data" in result["code"] for result in trial_results + ] # should reuse iris_data from previous tasks + success_rate = sum(success) / trials_num + logger.info(f"success rate: {success_rate :.2f}") + assert success_rate >= 0.8 + + +@pytest.mark.skip +@pytest.mark.asyncio +async def test_write_code_reuse_code_long_for_wine(): + """test code reuse for long context""" + + structural_context = """ + ## User Requirement + Run data analysis on sklearn Wisconsin Breast Cancer dataset, include a plot, train a model to predict targets (20% as validation), and show validation accuracy + ## Current Plan + [ + { + "task_id": "1", + "dependent_task_ids": [], + "instruction": "Load the sklearn Wine recognition dataset and perform exploratory data analysis." + "task_type": "", + "code": "from sklearn.datasets import load_wine\n# Load the Wine recognition dataset\nwine_data = load_wine()\n# Perform exploratory data analysis\nwine_data.keys()", + "result": "Truncated to show only the last 1000 characters\ndict_keys(['data', 'target', 'frame', 'target_names', 'DESCR', 'feature_names'])", + "is_finished": true + }, + { + "task_id": "2", + "dependent_task_ids": ["1"], + "instruction": "Create a plot to visualize some aspect of the wine dataset." + "task_type": "", + "code": "", + "result": "", + "is_finished": false + }, + { + "task_id": "3", + "dependent_task_ids": ["1"], + "instruction": "Split the dataset into training and validation sets with a 20% validation size.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + }, + { + "task_id": "4", + "dependent_task_ids": ["3"], + "instruction": "Train a model on the training set to predict wine class.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + }, + { + "task_id": "5", + "dependent_task_ids": ["4"], + "instruction": "Evaluate the model on the validation set and report the accuracy.", + "task_type": "", + "code": "", + "result": "", + "is_finished": false + } + ] + ## Current Task + {"task_id": "2", "dependent_task_ids": ["1"], "instruction": "Create a plot to visualize some aspect of the Wine dataset.", "task_type": "", "code": "", "result": "", "is_finished": false} + """ + context = [ + Message(content=structural_context, role="user"), + ] + trials_num = 5 + trials = [WriteCodeWithoutTools().run(context=context, temperature=0.0) for _ in range(trials_num)] + trial_results = await asyncio.gather(*trials) + print(*trial_results, sep="\n\n***\n\n") + success = [ + "load_wine" not in result["code"] and "wine_data" in result["code"] for result in trial_results + ] # should reuse iris_data from previous tasks + success_rate = sum(success) / trials_num + logger.info(f"success rate: {success_rate :.2f}") + assert success_rate >= 0.8 diff --git a/tests/metagpt/actions/ci/test_write_plan.py b/tests/metagpt/actions/ci/test_write_plan.py new file mode 100644 index 000000000..3eb80ca3e --- /dev/null +++ b/tests/metagpt/actions/ci/test_write_plan.py @@ -0,0 +1,34 @@ +import pytest + +from metagpt.actions.ci.write_plan import ( + Plan, + Task, + WritePlan, + precheck_update_plan_from_rsp, +) +from metagpt.schema import Message + + +def test_precheck_update_plan_from_rsp(): + plan = Plan(goal="") + plan.add_tasks([Task(task_id="1")]) + rsp = '[{"task_id": "2"}]' + success, _ = precheck_update_plan_from_rsp(rsp, plan) + assert success + assert len(plan.tasks) == 1 and plan.tasks[0].task_id == "1" # precheck should not change the original one + + invalid_rsp = "wrong" + success, _ = precheck_update_plan_from_rsp(invalid_rsp, plan) + assert not success + + +@pytest.mark.asyncio +@pytest.mark.parametrize("use_tools", [(False), (True)]) +async def test_write_plan(use_tools): + rsp = await WritePlan().run( + context=[Message("run analysis on sklearn iris dataset", role="user")], use_tools=use_tools + ) + + assert "task_id" in rsp + assert "instruction" in rsp + assert "json" not in rsp # the output should be the content inside ```json ``` diff --git a/tests/metagpt/actions/test_action_node.py b/tests/metagpt/actions/test_action_node.py index 53de9cc75..989e2249c 100644 --- a/tests/metagpt/actions/test_action_node.py +++ b/tests/metagpt/actions/test_action_node.py @@ -5,10 +5,11 @@ @Author : alexanderwu @File : test_action_node.py """ +from pathlib import Path from typing import List, Tuple import pytest -from pydantic import ValidationError +from pydantic import BaseModel, Field, ValidationError from metagpt.actions import Action from metagpt.actions.action_node import ActionNode, ReviewMode, ReviseMode @@ -17,6 +18,7 @@ from metagpt.llm import LLM from metagpt.roles import Role from metagpt.schema import Message from metagpt.team import Team +from metagpt.utils.common import encode_image @pytest.mark.asyncio @@ -241,6 +243,65 @@ def test_create_model_class_with_mapping(): assert value == ["game.py", "app.py", "static/css/styles.css", "static/js/script.js", "templates/index.html"] +@pytest.mark.asyncio +async def test_action_node_with_image(mocker): + # add a mock to update model in unittest, due to the gloabl MockLLM + def _cons_kwargs(self, messages: list[dict], timeout=3, **extra_kwargs) -> dict: + kwargs = {"messages": messages, "temperature": 0.3, "model": "gpt-4-vision-preview"} + return kwargs + + invoice = ActionNode( + key="invoice", expected_type=bool, instruction="if it's a invoice file, return True else False", example="False" + ) + + invoice_path = Path(__file__).parent.joinpath("..", "..", "data", "invoices", "invoice-2.png") + img_base64 = encode_image(invoice_path) + mocker.patch("metagpt.provider.openai_api.OpenAILLM._cons_kwargs", _cons_kwargs) + node = await invoice.fill(context="", llm=LLM(), images=[img_base64]) + assert node.instruct_content.invoice + + +class ToolDef(BaseModel): + tool_name: str = Field(default="a", description="tool name", examples=[]) + description: str = Field(default="b", description="tool description", examples=[]) + + +class Task(BaseModel): + task_id: int = Field(default=1, description="task id", examples=[1, 2, 3]) + name: str = Field(default="Get data from ...", description="task name", examples=[]) + dependent_task_ids: List[int] = Field(default=[], description="dependent task ids", examples=[1, 2, 3]) + tool: ToolDef = Field(default=ToolDef(), description="tool use", examples=[]) + + +class Tasks(BaseModel): + tasks: List[Task] = Field(default=[], description="tasks", examples=[]) + + +def test_action_node_from_pydantic_and_print_everything(): + node = ActionNode.from_pydantic(Task) + print("1. Tasks") + print(Task().model_dump_json(indent=4)) + print(Tasks.model_json_schema()) + print("2. Task") + print(Task.model_json_schema()) + print("3. ActionNode") + print(node) + print("4. node.compile prompt") + prompt = node.compile(context="") + assert "tool_name" in prompt, "tool_name should be in prompt" + print(prompt) + print("5. node.get_children_mapping") + print(node._get_children_mapping()) + print("6. node.create_children_class") + children_class = node._create_children_class() + print(children_class) + import inspect + + code = inspect.getsource(Tasks) + print(code) + assert "tasks" in code, "tasks should be in code" + + if __name__ == "__main__": test_create_model_class() test_create_model_class_with_mapping() diff --git a/tests/metagpt/actions/test_debug_error.py b/tests/metagpt/actions/test_debug_error.py index e093eb83f..c88818bbd 100644 --- a/tests/metagpt/actions/test_debug_error.py +++ b/tests/metagpt/actions/test_debug_error.py @@ -11,9 +11,7 @@ import uuid import pytest from metagpt.actions.debug_error import DebugError -from metagpt.context import CONTEXT from metagpt.schema import RunCodeContext, RunCodeResult -from metagpt.utils.project_repo import ProjectRepo CODE_CONTENT = ''' from typing import List @@ -116,9 +114,8 @@ if __name__ == '__main__': @pytest.mark.asyncio -async def test_debug_error(): - CONTEXT.src_workspace = CONTEXT.git_repo.workdir / uuid.uuid4().hex - project_repo = ProjectRepo(CONTEXT.git_repo) +async def test_debug_error(context): + context.src_workspace = context.git_repo.workdir / uuid.uuid4().hex ctx = RunCodeContext( code_filename="player.py", test_filename="test_player.py", @@ -126,8 +123,8 @@ async def test_debug_error(): output_filename="output.log", ) - await project_repo.with_src_path(CONTEXT.src_workspace).srcs.save(filename=ctx.code_filename, content=CODE_CONTENT) - await project_repo.tests.save(filename=ctx.test_filename, content=TEST_CONTENT) + await context.repo.with_src_path(context.src_workspace).srcs.save(filename=ctx.code_filename, content=CODE_CONTENT) + await context.repo.tests.save(filename=ctx.test_filename, content=TEST_CONTENT) output_data = RunCodeResult( stdout=";", stderr="", @@ -141,8 +138,8 @@ async def test_debug_error(): "----------------------------------------------------------------------\n" "Ran 5 tests in 0.007s\n\nFAILED (failures=1)\n;\n", ) - await project_repo.test_outputs.save(filename=ctx.output_filename, content=output_data.model_dump_json()) - debug_error = DebugError(i_context=ctx) + await context.repo.test_outputs.save(filename=ctx.output_filename, content=output_data.model_dump_json()) + debug_error = DebugError(i_context=ctx, context=context) rsp = await debug_error.run() diff --git a/tests/metagpt/actions/test_design_api.py b/tests/metagpt/actions/test_design_api.py index fc231e578..7d3efa7ff 100644 --- a/tests/metagpt/actions/test_design_api.py +++ b/tests/metagpt/actions/test_design_api.py @@ -9,20 +9,17 @@ import pytest from metagpt.actions.design_api import WriteDesign -from metagpt.context import CONTEXT from metagpt.logs import logger from metagpt.schema import Message -from metagpt.utils.project_repo import ProjectRepo @pytest.mark.asyncio -async def test_design_api(): +async def test_design_api(context): inputs = ["我们需要一个音乐播放器,它应该有播放、暂停、上一曲、下一曲等功能。"] # PRD_SAMPLE - project_repo = ProjectRepo(CONTEXT.git_repo) for prd in inputs: - await project_repo.docs.prd.save(filename="new_prd.txt", content=prd) + await context.repo.docs.prd.save(filename="new_prd.txt", content=prd) - design_api = WriteDesign() + design_api = WriteDesign(context=context) result = await design_api.run(Message(content=prd, instruct_content=None)) logger.info(result) diff --git a/tests/metagpt/actions/test_design_api_an.py b/tests/metagpt/actions/test_design_api_an.py new file mode 100644 index 000000000..3d11f200d --- /dev/null +++ b/tests/metagpt/actions/test_design_api_an.py @@ -0,0 +1,46 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/03 +@Author : mannaandpoem +@File : test_design_api_an.py +""" +import pytest +from openai._models import BaseModel + +from metagpt.actions.action_node import ActionNode, dict_to_markdown +from metagpt.actions.design_api import NEW_REQ_TEMPLATE +from metagpt.actions.design_api_an import REFINED_DESIGN_NODE +from metagpt.llm import LLM +from tests.data.incremental_dev_project.mock import ( + DESIGN_SAMPLE, + REFINED_DESIGN_JSON, + REFINED_PRD_JSON, +) + + +@pytest.fixture() +def llm(): + return LLM() + + +def mock_refined_design_json(): + return REFINED_DESIGN_JSON + + +@pytest.mark.asyncio +async def test_write_design_an(mocker): + root = ActionNode.from_children( + "RefinedDesignAPI", [ActionNode(key="", expected_type=str, instruction="", example="")] + ) + root.instruct_content = BaseModel() + root.instruct_content.model_dump = mock_refined_design_json + mocker.patch("metagpt.actions.design_api_an.REFINED_DESIGN_NODE.fill", return_value=root) + + prompt = NEW_REQ_TEMPLATE.format(old_design=DESIGN_SAMPLE, context=dict_to_markdown(REFINED_PRD_JSON)) + node = await REFINED_DESIGN_NODE.fill(prompt, llm) + + assert "Refined Implementation Approach" in node.instruct_content.model_dump() + assert "Refined File list" in node.instruct_content.model_dump() + assert "Refined Data structures and interfaces" in node.instruct_content.model_dump() + assert "Refined Program call flow" in node.instruct_content.model_dump() diff --git a/tests/metagpt/actions/test_design_api_review.py b/tests/metagpt/actions/test_design_api_review.py index cfc29056f..a648dba3f 100644 --- a/tests/metagpt/actions/test_design_api_review.py +++ b/tests/metagpt/actions/test_design_api_review.py @@ -11,7 +11,7 @@ from metagpt.actions.design_api_review import DesignReview @pytest.mark.asyncio -async def test_design_api_review(): +async def test_design_api_review(context): prd = "我们需要一个音乐播放器,它应该有播放、暂停、上一曲、下一曲等功能。" api_design = """ 数据结构: @@ -26,7 +26,7 @@ API列表: """ _ = "API设计看起来非常合理,满足了PRD中的所有需求。" - design_api_review = DesignReview() + design_api_review = DesignReview(context=context) result = await design_api_review.run(prd, api_design) diff --git a/tests/metagpt/actions/test_fix_bug.py b/tests/metagpt/actions/test_fix_bug.py index b2dc8d0f4..cbd9d0b57 100644 --- a/tests/metagpt/actions/test_fix_bug.py +++ b/tests/metagpt/actions/test_fix_bug.py @@ -12,6 +12,6 @@ from metagpt.actions.fix_bug import FixBug @pytest.mark.asyncio -async def test_fix_bug(): - fix_bug = FixBug() +async def test_fix_bug(context): + fix_bug = FixBug(context=context) assert fix_bug.name == "FixBug" diff --git a/tests/metagpt/actions/test_generate_questions.py b/tests/metagpt/actions/test_generate_questions.py index b7c9d3984..6adb2e617 100644 --- a/tests/metagpt/actions/test_generate_questions.py +++ b/tests/metagpt/actions/test_generate_questions.py @@ -10,7 +10,7 @@ import pytest from metagpt.actions.generate_questions import GenerateQuestions from metagpt.logs import logger -context = """ +msg = """ ## topic 如何做一个生日蛋糕 @@ -20,9 +20,9 @@ context = """ @pytest.mark.asyncio -async def test_generate_questions(): - action = GenerateQuestions() - rsp = await action.run(context) +async def test_generate_questions(context): + action = GenerateQuestions(context=context) + rsp = await action.run(msg) logger.info(f"{rsp.content=}") assert "Questions" in rsp.content diff --git a/tests/metagpt/actions/test_invoice_ocr.py b/tests/metagpt/actions/test_invoice_ocr.py index b4560f61b..4df0cf4f8 100644 --- a/tests/metagpt/actions/test_invoice_ocr.py +++ b/tests/metagpt/actions/test_invoice_ocr.py @@ -23,9 +23,9 @@ from metagpt.const import TEST_DATA_PATH Path("invoices/invoice-4.zip"), ], ) -async def test_invoice_ocr(invoice_path: Path): +async def test_invoice_ocr(invoice_path: Path, context): invoice_path = TEST_DATA_PATH / invoice_path - resp = await InvoiceOCR().run(file_path=Path(invoice_path)) + resp = await InvoiceOCR(context=context).run(file_path=Path(invoice_path)) assert isinstance(resp, list) diff --git a/tests/metagpt/actions/test_prepare_documents.py b/tests/metagpt/actions/test_prepare_documents.py index a72019c5c..7ad0dee4e 100644 --- a/tests/metagpt/actions/test_prepare_documents.py +++ b/tests/metagpt/actions/test_prepare_documents.py @@ -10,21 +10,18 @@ import pytest from metagpt.actions.prepare_documents import PrepareDocuments from metagpt.const import REQUIREMENT_FILENAME -from metagpt.context import CONTEXT +from metagpt.context import Context from metagpt.schema import Message -from metagpt.utils.project_repo import ProjectRepo @pytest.mark.asyncio async def test_prepare_documents(): msg = Message(content="New user requirements balabala...") + context = Context() - if CONTEXT.git_repo: - CONTEXT.git_repo.delete_repository() - CONTEXT.git_repo = None - - await PrepareDocuments(context=CONTEXT).run(with_messages=[msg]) - assert CONTEXT.git_repo - doc = await ProjectRepo(CONTEXT.git_repo).docs.get(filename=REQUIREMENT_FILENAME) + await PrepareDocuments(context=context).run(with_messages=[msg]) + assert context.git_repo + assert context.repo + doc = await context.repo.docs.get(filename=REQUIREMENT_FILENAME) assert doc assert doc.content == msg.content diff --git a/tests/metagpt/actions/test_prepare_interview.py b/tests/metagpt/actions/test_prepare_interview.py index cd0c850ed..111f24d5f 100644 --- a/tests/metagpt/actions/test_prepare_interview.py +++ b/tests/metagpt/actions/test_prepare_interview.py @@ -12,8 +12,8 @@ from metagpt.logs import logger @pytest.mark.asyncio -async def test_prepare_interview(): - action = PrepareInterview() +async def test_prepare_interview(context): + action = PrepareInterview(context=context) rsp = await action.run("I just graduated and hope to find a job as a Python engineer") logger.info(f"{rsp.content=}") diff --git a/tests/metagpt/actions/test_project_management.py b/tests/metagpt/actions/test_project_management.py index 9fd3b1721..f3bb405ca 100644 --- a/tests/metagpt/actions/test_project_management.py +++ b/tests/metagpt/actions/test_project_management.py @@ -9,21 +9,18 @@ import pytest from metagpt.actions.project_management import WriteTasks -from metagpt.context import CONTEXT from metagpt.logs import logger from metagpt.schema import Message -from metagpt.utils.project_repo import ProjectRepo from tests.metagpt.actions.mock_json import DESIGN, PRD @pytest.mark.asyncio -async def test_design_api(): - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.docs.prd.save("1.txt", content=str(PRD)) - await project_repo.docs.system_design.save("1.txt", content=str(DESIGN)) - logger.info(CONTEXT.git_repo) +async def test_design_api(context): + await context.repo.docs.prd.save("1.txt", content=str(PRD)) + await context.repo.docs.system_design.save("1.txt", content=str(DESIGN)) + logger.info(context.git_repo) - action = WriteTasks() + action = WriteTasks(context=context) result = await action.run(Message(content="", instruct_content=None)) logger.info(result) diff --git a/tests/metagpt/actions/test_project_management_an.py b/tests/metagpt/actions/test_project_management_an.py new file mode 100644 index 000000000..ddbb56569 --- /dev/null +++ b/tests/metagpt/actions/test_project_management_an.py @@ -0,0 +1,45 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/03 +@Author : mannaandpoem +@File : test_project_management_an.py +""" +import pytest +from openai._models import BaseModel + +from metagpt.actions.action_node import ActionNode, dict_to_markdown +from metagpt.actions.project_management import NEW_REQ_TEMPLATE +from metagpt.actions.project_management_an import REFINED_PM_NODE +from metagpt.llm import LLM +from tests.data.incremental_dev_project.mock import ( + REFINED_DESIGN_JSON, + REFINED_TASKS_JSON, + TASKS_SAMPLE, +) + + +@pytest.fixture() +def llm(): + return LLM() + + +def mock_refined_tasks_json(): + return REFINED_TASKS_JSON + + +@pytest.mark.asyncio +async def test_project_management_an(mocker): + root = ActionNode.from_children( + "RefinedProjectManagement", [ActionNode(key="", expected_type=str, instruction="", example="")] + ) + root.instruct_content = BaseModel() + root.instruct_content.model_dump = mock_refined_tasks_json + mocker.patch("metagpt.actions.project_management_an.REFINED_PM_NODE.fill", return_value=root) + + prompt = NEW_REQ_TEMPLATE.format(old_task=TASKS_SAMPLE, context=dict_to_markdown(REFINED_DESIGN_JSON)) + node = await REFINED_PM_NODE.fill(prompt, llm) + + assert "Refined Logic Analysis" in node.instruct_content.model_dump() + assert "Refined Task list" in node.instruct_content.model_dump() + assert "Refined Shared Knowledge" in node.instruct_content.model_dump() diff --git a/tests/metagpt/actions/test_rebuild_class_view.py b/tests/metagpt/actions/test_rebuild_class_view.py index 94295fd55..2188d6b85 100644 --- a/tests/metagpt/actions/test_rebuild_class_view.py +++ b/tests/metagpt/actions/test_rebuild_class_view.py @@ -11,27 +11,28 @@ from pathlib import Path import pytest from metagpt.actions.rebuild_class_view import RebuildClassView -from metagpt.const import GRAPH_REPO_FILE_REPO -from metagpt.context import CONTEXT from metagpt.llm import LLM +@pytest.mark.skip @pytest.mark.asyncio -async def test_rebuild(): +async def test_rebuild(context): action = RebuildClassView( - name="RedBean", i_context=str(Path(__file__).parent.parent.parent.parent / "metagpt"), llm=LLM() + name="RedBean", + i_context=str(Path(__file__).parent.parent.parent.parent / "metagpt"), + llm=LLM(), + context=context, ) await action.run() - graph_file_repo = CONTEXT.git_repo.new_file_repository(relative_path=GRAPH_REPO_FILE_REPO) - assert graph_file_repo.changed_files + assert context.repo.docs.graph_repo.changed_files @pytest.mark.parametrize( ("path", "direction", "diff", "want"), [ - ("metagpt/startup.py", "=", ".", "metagpt/startup.py"), - ("metagpt/startup.py", "+", "MetaGPT", "MetaGPT/metagpt/startup.py"), - ("metagpt/startup.py", "-", "metagpt", "startup.py"), + ("metagpt/software_company.py", "=", ".", "metagpt/software_company.py"), + ("metagpt/software_company.py", "+", "MetaGPT", "MetaGPT/metagpt/software_company.py"), + ("metagpt/software_company.py", "-", "metagpt", "software_company.py"), ], ) def test_align_path(path, direction, diff, want): diff --git a/tests/metagpt/actions/test_rebuild_sequence_view.py b/tests/metagpt/actions/test_rebuild_sequence_view.py index 717aee964..e0e65c3cc 100644 --- a/tests/metagpt/actions/test_rebuild_sequence_view.py +++ b/tests/metagpt/actions/test_rebuild_sequence_view.py @@ -11,28 +11,29 @@ import pytest from metagpt.actions.rebuild_sequence_view import RebuildSequenceView from metagpt.const import GRAPH_REPO_FILE_REPO -from metagpt.context import CONTEXT from metagpt.llm import LLM from metagpt.utils.common import aread from metagpt.utils.git_repository import ChangeType -from metagpt.utils.project_repo import ProjectRepo @pytest.mark.asyncio -async def test_rebuild(): +@pytest.mark.skip +async def test_rebuild(context): # Mock data = await aread(filename=Path(__file__).parent / "../../data/graph_db/networkx.json") - graph_db_filename = Path(CONTEXT.git_repo.workdir.name).with_suffix(".json") - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.docs.graph_repo.save(filename=str(graph_db_filename), content=data) - CONTEXT.git_repo.add_change({f"{GRAPH_REPO_FILE_REPO}/{graph_db_filename}": ChangeType.UNTRACTED}) - CONTEXT.git_repo.commit("commit1") + graph_db_filename = Path(context.repo.workdir.name).with_suffix(".json") + await context.repo.docs.graph_repo.save(filename=str(graph_db_filename), content=data) + context.git_repo.add_change({f"{GRAPH_REPO_FILE_REPO}/{graph_db_filename}": ChangeType.UNTRACTED}) + context.git_repo.commit("commit1") action = RebuildSequenceView( - name="RedBean", i_context=str(Path(__file__).parent.parent.parent.parent / "metagpt"), llm=LLM() + name="RedBean", + i_context=str(Path(__file__).parent.parent.parent.parent / "metagpt"), + llm=LLM(), + context=context, ) await action.run() - assert project_repo.docs.graph_repo.changed_files + assert context.repo.docs.graph_repo.changed_files @pytest.mark.parametrize( diff --git a/tests/metagpt/actions/test_research.py b/tests/metagpt/actions/test_research.py index 8c5ed0c7c..ed83ce58c 100644 --- a/tests/metagpt/actions/test_research.py +++ b/tests/metagpt/actions/test_research.py @@ -14,7 +14,7 @@ from metagpt.tools.search_engine import SearchEngine @pytest.mark.asyncio -async def test_collect_links(mocker, search_engine_mocker): +async def test_collect_links(mocker, search_engine_mocker, context): async def mock_llm_ask(self, prompt: str, system_msgs): if "Please provide up to 2 necessary keywords" in prompt: return '["metagpt", "llm"]' @@ -28,15 +28,15 @@ async def test_collect_links(mocker, search_engine_mocker): return "[1,2]" mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", mock_llm_ask) - resp = await research.CollectLinks(search_engine=SearchEngine(SearchEngineType.DUCK_DUCK_GO)).run( - "The application of MetaGPT" - ) + resp = await research.CollectLinks( + search_engine=SearchEngine(engine=SearchEngineType.DUCK_DUCK_GO), context=context + ).run("The application of MetaGPT") for i in ["MetaGPT use cases", "The roadmap of MetaGPT", "The function of MetaGPT", "What llm MetaGPT support"]: assert i in resp @pytest.mark.asyncio -async def test_collect_links_with_rank_func(mocker, search_engine_mocker): +async def test_collect_links_with_rank_func(mocker, search_engine_mocker, context): rank_before = [] rank_after = [] url_per_query = 4 @@ -50,7 +50,9 @@ async def test_collect_links_with_rank_func(mocker, search_engine_mocker): mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", mock_collect_links_llm_ask) resp = await research.CollectLinks( - search_engine=SearchEngine(SearchEngineType.DUCK_DUCK_GO), rank_func=rank_func + search_engine=SearchEngine(engine=SearchEngineType.DUCK_DUCK_GO), + rank_func=rank_func, + context=context, ).run("The application of MetaGPT") for x, y, z in zip(rank_before, rank_after, resp.values()): assert x[::-1] == y @@ -58,7 +60,7 @@ async def test_collect_links_with_rank_func(mocker, search_engine_mocker): @pytest.mark.asyncio -async def test_web_browse_and_summarize(mocker): +async def test_web_browse_and_summarize(mocker, context): async def mock_llm_ask(*args, **kwargs): return "metagpt" @@ -66,20 +68,20 @@ async def test_web_browse_and_summarize(mocker): url = "https://github.com/geekan/MetaGPT" url2 = "https://github.com/trending" query = "What's new in metagpt" - resp = await research.WebBrowseAndSummarize().run(url, query=query) + resp = await research.WebBrowseAndSummarize(context=context).run(url, query=query) assert len(resp) == 1 assert url in resp assert resp[url] == "metagpt" - resp = await research.WebBrowseAndSummarize().run(url, url2, query=query) + resp = await research.WebBrowseAndSummarize(context=context).run(url, url2, query=query) assert len(resp) == 2 async def mock_llm_ask(*args, **kwargs): return "Not relevant." mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", mock_llm_ask) - resp = await research.WebBrowseAndSummarize().run(url, query=query) + resp = await research.WebBrowseAndSummarize(context=context).run(url, query=query) assert len(resp) == 1 assert url in resp @@ -87,7 +89,7 @@ async def test_web_browse_and_summarize(mocker): @pytest.mark.asyncio -async def test_conduct_research(mocker): +async def test_conduct_research(mocker, context): data = None async def mock_llm_ask(*args, **kwargs): @@ -101,7 +103,7 @@ async def test_conduct_research(mocker): "outputs user stories / competitive analysis / requirements / data structures / APIs / documents, etc." ) - resp = await research.ConductResearch().run("The application of MetaGPT", content) + resp = await research.ConductResearch(context=context).run("The application of MetaGPT", content) assert resp == data diff --git a/tests/metagpt/actions/test_run_code.py b/tests/metagpt/actions/test_run_code.py index 76397734d..2ec8a7748 100644 --- a/tests/metagpt/actions/test_run_code.py +++ b/tests/metagpt/actions/test_run_code.py @@ -24,21 +24,21 @@ async def test_run_text(): @pytest.mark.asyncio -async def test_run_script(): +async def test_run_script(context): # Successful command - out, err = await RunCode().run_script(".", command=["echo", "Hello World"]) + out, err = await RunCode(context=context).run_script(".", command=["echo", "Hello World"]) assert out.strip() == "Hello World" assert err == "" # Unsuccessful command - out, err = await RunCode().run_script(".", command=["python", "-c", "print(1/0)"]) + out, err = await RunCode(context=context).run_script(".", command=["python", "-c", "print(1/0)"]) assert "ZeroDivisionError" in err @pytest.mark.asyncio -async def test_run(): +async def test_run(context): inputs = [ - (RunCodeContext(mode="text", code_filename="a.txt", code="print('Hello, World')"), "PASS"), + (RunCodeContext(mode="text", code_filename="a.txt", code="result = 'helloworld'"), "PASS"), ( RunCodeContext( mode="script", @@ -61,5 +61,5 @@ async def test_run(): ), ] for ctx, result in inputs: - rsp = await RunCode(i_context=ctx).run() + rsp = await RunCode(i_context=ctx, context=context).run() assert result in rsp.summary diff --git a/tests/metagpt/actions/test_skill_action.py b/tests/metagpt/actions/test_skill_action.py index 69cd8129d..d667d6d70 100644 --- a/tests/metagpt/actions/test_skill_action.py +++ b/tests/metagpt/actions/test_skill_action.py @@ -23,9 +23,9 @@ class TestSkillAction: "type": "string", "description": "OpenAI API key, For more details, checkout: `https://platform.openai.com/account/api-keys`", }, - "METAGPT_TEXT_TO_IMAGE_MODEL_URL": {"type": "string", "description": "Model url."}, + "metagpt_tti_url": {"type": "string", "description": "Model url."}, }, - "required": {"oneOf": ["OPENAI_API_KEY", "METAGPT_TEXT_TO_IMAGE_MODEL_URL"]}, + "required": {"oneOf": ["OPENAI_API_KEY", "metagpt_tti_url"]}, }, parameters={ "text": Parameter(type="string", description="The text used for image conversion."), @@ -47,18 +47,18 @@ class TestSkillAction: assert args.get("size_type") == "512x512" @pytest.mark.asyncio - async def test_parser_action(self, mocker): + async def test_parser_action(self, mocker, context): # mock mocker.patch("metagpt.learn.text_to_image", return_value="https://mock.com/xxx") - parser_action = ArgumentsParingAction(skill=self.skill, ask="Draw an apple") + parser_action = ArgumentsParingAction(skill=self.skill, ask="Draw an apple", context=context) rsp = await parser_action.run() assert rsp assert parser_action.args assert parser_action.args.get("text") == "Draw an apple" assert parser_action.args.get("size_type") == "512x512" - action = SkillAction(skill=self.skill, args=parser_action.args) + action = SkillAction(skill=self.skill, args=parser_action.args, context=context) rsp = await action.run() assert rsp assert "image/png;base64," in rsp.content or "http" in rsp.content @@ -81,8 +81,8 @@ class TestSkillAction: await SkillAction.find_and_call_function("dummy_call", {"a": 1}) @pytest.mark.asyncio - async def test_skill_action_error(self): - action = SkillAction(skill=self.skill, args={}) + async def test_skill_action_error(self, context): + action = SkillAction(skill=self.skill, args={}, context=context) rsp = await action.run() assert "Error" in rsp.content diff --git a/tests/metagpt/actions/test_summarize_code.py b/tests/metagpt/actions/test_summarize_code.py index 88d432b5e..3cfe7ca81 100644 --- a/tests/metagpt/actions/test_summarize_code.py +++ b/tests/metagpt/actions/test_summarize_code.py @@ -6,13 +6,14 @@ @File : test_summarize_code.py @Modifiled By: mashenquan, 2023-12-6. Unit test for summarize_code.py """ +import uuid +from pathlib import Path + import pytest from metagpt.actions.summarize_code import SummarizeCode -from metagpt.context import CONTEXT from metagpt.logs import logger from metagpt.schema import CodeSummarizeContext -from metagpt.utils.project_repo import ProjectRepo DESIGN_CONTENT = """ {"Implementation approach": "To develop this snake game, we will use the Python language and choose the Pygame library. Pygame is an open-source Python module collection specifically designed for writing video games. It provides functionalities such as displaying images and playing sounds, making it suitable for creating intuitive and responsive user interfaces. We will ensure efficient game logic to prevent any delays during gameplay. The scoring system will be simple, with the snake gaining points for each food it eats. We will use Pygame's event handling system to implement pause and resume functionality, as well as high-score tracking. The difficulty will increase by speeding up the snake's movement. In the initial version, we will focus on single-player mode and consider adding multiplayer mode and customizable skins in future updates. Based on the new requirement, we will also add a moving obstacle that appears randomly. If the snake eats this obstacle, the game will end. If the snake does not eat the obstacle, it will disappear after 5 seconds. For this, we need to add mechanisms for obstacle generation, movement, and disappearance in the game logic.", "Project_name": "snake_game", "File list": ["main.py", "game.py", "snake.py", "food.py", "obstacle.py", "scoreboard.py", "constants.py", "assets/styles.css", "assets/index.html"], "Data structures and interfaces": "```mermaid\n classDiagram\n class Game{\n +int score\n +int speed\n +bool game_over\n +bool paused\n +Snake snake\n +Food food\n +Obstacle obstacle\n +Scoreboard scoreboard\n +start_game() void\n +pause_game() void\n +resume_game() void\n +end_game() void\n +increase_difficulty() void\n +update() void\n +render() void\n Game()\n }\n class Snake{\n +list body_parts\n +str direction\n +bool grow\n +move() void\n +grow() void\n +check_collision() bool\n Snake()\n }\n class Food{\n +tuple position\n +spawn() void\n Food()\n }\n class Obstacle{\n +tuple position\n +int lifetime\n +bool active\n +spawn() void\n +move() void\n +check_collision() bool\n +disappear() void\n Obstacle()\n }\n class Scoreboard{\n +int high_score\n +update_score(int) void\n +reset_score() void\n +load_high_score() void\n +save_high_score() void\n Scoreboard()\n }\n class Constants{\n }\n Game \"1\" -- \"1\" Snake: has\n Game \"1\" -- \"1\" Food: has\n Game \"1\" -- \"1\" Obstacle: has\n Game \"1\" -- \"1\" Scoreboard: has\n ```", "Program call flow": "```sequenceDiagram\n participant M as Main\n participant G as Game\n participant S as Snake\n participant F as Food\n participant O as Obstacle\n participant SB as Scoreboard\n M->>G: start_game()\n loop game loop\n G->>S: move()\n G->>S: check_collision()\n G->>F: spawn()\n G->>O: spawn()\n G->>O: move()\n G->>O: check_collision()\n G->>O: disappear()\n G->>SB: update_score(score)\n G->>G: update()\n G->>G: render()\n alt if paused\n M->>G: pause_game()\n M->>G: resume_game()\n end\n alt if game_over\n G->>M: end_game()\n end\n end\n```", "Anything UNCLEAR": "There is no need for further clarification as the requirements are already clear."} @@ -175,21 +176,26 @@ class Snake: """ +@pytest.mark.skip @pytest.mark.asyncio -async def test_summarize_code(): - CONTEXT.src_workspace = CONTEXT.git_repo.workdir / "src" - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.docs.system_design.save(filename="1.json", content=DESIGN_CONTENT) - await project_repo.docs.task.save(filename="1.json", content=TASK_CONTENT) - await project_repo.with_src_path(CONTEXT.src_workspace).srcs.save(filename="food.py", content=FOOD_PY) - assert project_repo.srcs.workdir == CONTEXT.src_workspace - await project_repo.srcs.save(filename="game.py", content=GAME_PY) - await project_repo.srcs.save(filename="main.py", content=MAIN_PY) - await project_repo.srcs.save(filename="snake.py", content=SNAKE_PY) +async def test_summarize_code(context): + git_dir = Path(__file__).parent / f"unittest/{uuid.uuid4().hex}" + git_dir.mkdir(parents=True, exist_ok=True) - all_files = project_repo.srcs.all_files - ctx = CodeSummarizeContext(design_filename="1.json", task_filename="1.json", codes_filenames=all_files) - action = SummarizeCode(i_context=ctx) + context.src_workspace = context.git_repo.workdir / "src" + await context.repo.docs.system_design.save(filename="1.json", content=DESIGN_CONTENT) + await context.repo.docs.task.save(filename="1.json", content=TASK_CONTENT) + await context.repo.with_src_path(context.src_workspace).srcs.save(filename="food.py", content=FOOD_PY) + assert context.repo.srcs.workdir == context.src_workspace + await context.repo.srcs.save(filename="game.py", content=GAME_PY) + await context.repo.srcs.save(filename="main.py", content=MAIN_PY) + await context.repo.srcs.save(filename="snake.py", content=SNAKE_PY) + + all_files = context.repo.srcs.all_files + summarization_context = CodeSummarizeContext( + design_filename="1.json", task_filename="1.json", codes_filenames=all_files + ) + action = SummarizeCode(context=context, i_context=summarization_context) rsp = await action.run() assert rsp logger.info(rsp) diff --git a/tests/metagpt/actions/test_talk_action.py b/tests/metagpt/actions/test_talk_action.py index b722d7c40..206abfbae 100644 --- a/tests/metagpt/actions/test_talk_action.py +++ b/tests/metagpt/actions/test_talk_action.py @@ -9,13 +9,12 @@ import pytest from metagpt.actions.talk_action import TalkAction -from metagpt.context import CONTEXT from metagpt.schema import Message @pytest.mark.asyncio @pytest.mark.parametrize( - ("agent_description", "language", "context", "knowledge", "history_summary"), + ("agent_description", "language", "talk_context", "knowledge", "history_summary"), [ ( "mathematician", @@ -33,12 +32,12 @@ from metagpt.schema import Message ), ], ) -async def test_prompt(agent_description, language, context, knowledge, history_summary): +async def test_prompt(agent_description, language, talk_context, knowledge, history_summary, context): # Prerequisites - CONTEXT.kwargs.agent_description = agent_description - CONTEXT.kwargs.language = language + context.kwargs.agent_description = agent_description + context.kwargs.language = language - action = TalkAction(i_context=context, knowledge=knowledge, history_summary=history_summary) + action = TalkAction(i_context=talk_context, knowledge=knowledge, history_summary=history_summary, context=context) assert "{" not in action.prompt assert "{" not in action.prompt_gpt4 diff --git a/tests/metagpt/actions/test_write_code.py b/tests/metagpt/actions/test_write_code.py index 96d982c69..ee05e0f7d 100644 --- a/tests/metagpt/actions/test_write_code.py +++ b/tests/metagpt/actions/test_write_code.py @@ -12,25 +12,22 @@ from pathlib import Path import pytest from metagpt.actions.write_code import WriteCode -from metagpt.context import CONTEXT -from metagpt.llm import LLM from metagpt.logs import logger from metagpt.schema import CodingContext, Document from metagpt.utils.common import aread -from metagpt.utils.project_repo import ProjectRepo from tests.metagpt.actions.mock_markdown import TASKS_2, WRITE_CODE_PROMPT_SAMPLE @pytest.mark.asyncio -async def test_write_code(): +async def test_write_code(context): # Prerequisites - CONTEXT.src_workspace = CONTEXT.git_repo.workdir / "writecode" + context.src_workspace = context.git_repo.workdir / "writecode" coding_ctx = CodingContext( filename="task_filename.py", design_doc=Document(content="设计一个名为'add'的函数,该函数接受两个整数作为输入,并返回它们的和。") ) doc = Document(content=coding_ctx.model_dump_json()) - write_code = WriteCode(i_context=doc) + write_code = WriteCode(i_context=doc, context=context) code = await write_code.run() logger.info(code.model_dump_json()) @@ -41,45 +38,44 @@ async def test_write_code(): @pytest.mark.asyncio -async def test_write_code_directly(): +async def test_write_code_directly(context): prompt = WRITE_CODE_PROMPT_SAMPLE + "\n" + TASKS_2[0] - llm = LLM() + llm = context.llm_with_cost_manager_from_llm_config(context.config.llm) rsp = await llm.aask(prompt) logger.info(rsp) @pytest.mark.asyncio -async def test_write_code_deps(): +async def test_write_code_deps(context): # Prerequisites - CONTEXT.src_workspace = CONTEXT.git_repo.workdir / "snake1/snake1" + context.src_workspace = context.git_repo.workdir / "snake1/snake1" demo_path = Path(__file__).parent / "../../data/demo_project" - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.test_outputs.save( + await context.repo.test_outputs.save( filename="test_game.py.json", content=await aread(str(demo_path / "test_game.py.json")) ) - await project_repo.docs.code_summary.save( + await context.repo.docs.code_summary.save( filename="20231221155954.json", content=await aread(str(demo_path / "code_summaries.json")), ) - await project_repo.docs.system_design.save( + await context.repo.docs.system_design.save( filename="20231221155954.json", content=await aread(str(demo_path / "system_design.json")), ) - await project_repo.docs.task.save( + await context.repo.docs.task.save( filename="20231221155954.json", content=await aread(str(demo_path / "tasks.json")) ) - await project_repo.with_src_path(CONTEXT.src_workspace).srcs.save( + await context.repo.with_src_path(context.src_workspace).srcs.save( filename="main.py", content='if __name__ == "__main__":\nmain()' ) ccontext = CodingContext( filename="game.py", - design_doc=await project_repo.docs.system_design.get(filename="20231221155954.json"), - task_doc=await project_repo.docs.task.get(filename="20231221155954.json"), + design_doc=await context.repo.docs.system_design.get(filename="20231221155954.json"), + task_doc=await context.repo.docs.task.get(filename="20231221155954.json"), code_doc=Document(filename="game.py", content="", root_path="snake1"), ) coding_doc = Document(root_path="snake1", filename="game.py", content=ccontext.json()) - action = WriteCode(i_context=coding_doc) + action = WriteCode(i_context=coding_doc, context=context) rsp = await action.run() assert rsp assert rsp.code_doc.content diff --git a/tests/metagpt/actions/test_write_code_plan_and_change_an.py b/tests/metagpt/actions/test_write_code_plan_and_change_an.py new file mode 100644 index 000000000..9cd51398f --- /dev/null +++ b/tests/metagpt/actions/test_write_code_plan_and_change_an.py @@ -0,0 +1,71 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/03 +@Author : mannaandpoem +@File : test_write_code_plan_and_change_an.py +""" +import pytest +from openai._models import BaseModel + +from metagpt.actions.action_node import ActionNode +from metagpt.actions.write_code import WriteCode +from metagpt.actions.write_code_plan_and_change_an import ( + REFINED_TEMPLATE, + WriteCodePlanAndChange, +) +from metagpt.schema import CodePlanAndChangeContext +from tests.data.incremental_dev_project.mock import ( + CODE_PLAN_AND_CHANGE_SAMPLE, + DESIGN_SAMPLE, + NEW_REQUIREMENT_SAMPLE, + REFINED_CODE_INPUT_SAMPLE, + REFINED_CODE_SAMPLE, + TASKS_SAMPLE, +) + + +def mock_code_plan_and_change(): + return CODE_PLAN_AND_CHANGE_SAMPLE + + +@pytest.mark.asyncio +async def test_write_code_plan_and_change_an(mocker): + root = ActionNode.from_children( + "WriteCodePlanAndChange", [ActionNode(key="", expected_type=str, instruction="", example="")] + ) + root.instruct_content = BaseModel() + root.instruct_content.model_dump = mock_code_plan_and_change + mocker.patch("metagpt.actions.write_code_plan_and_change_an.WriteCodePlanAndChange.run", return_value=root) + + requirement = "New requirement" + prd_filename = "prd.md" + design_filename = "design.md" + task_filename = "task.md" + code_plan_and_change_context = CodePlanAndChangeContext( + requirement=requirement, + prd_filename=prd_filename, + design_filename=design_filename, + task_filename=task_filename, + ) + node = await WriteCodePlanAndChange(i_context=code_plan_and_change_context).run() + + assert "Code Plan And Change" in node.instruct_content.model_dump() + + +@pytest.mark.asyncio +async def test_refine_code(mocker): + mocker.patch.object(WriteCode, "_aask", return_value=REFINED_CODE_SAMPLE) + prompt = REFINED_TEMPLATE.format( + user_requirement=NEW_REQUIREMENT_SAMPLE, + code_plan_and_change=CODE_PLAN_AND_CHANGE_SAMPLE, + design=DESIGN_SAMPLE, + task=TASKS_SAMPLE, + code=REFINED_CODE_INPUT_SAMPLE, + logs="", + feedback="", + filename="game.py", + summary_log="", + ) + code = await WriteCode().write_code(prompt=prompt) + assert "def" in code diff --git a/tests/metagpt/actions/test_write_code_review.py b/tests/metagpt/actions/test_write_code_review.py index 951929b76..a08dd07bc 100644 --- a/tests/metagpt/actions/test_write_code_review.py +++ b/tests/metagpt/actions/test_write_code_review.py @@ -12,28 +12,25 @@ from metagpt.schema import CodingContext, Document @pytest.mark.asyncio -async def test_write_code_review(capfd): +async def test_write_code_review(capfd, context): + context.src_workspace = context.repo.workdir / "srcs" code = """ def add(a, b): return a + """ - context = CodingContext( + coding_context = CodingContext( filename="math.py", design_doc=Document(content="编写一个从a加b的函数,返回a+b"), code_doc=Document(content=code) ) - context = await WriteCodeReview(i_context=context).run() + await WriteCodeReview(i_context=coding_context, context=context).run() # 我们不能精确地预测生成的代码评审,但我们可以检查返回的是否为字符串 - assert isinstance(context.code_doc.content, str) - assert len(context.code_doc.content) > 0 + assert isinstance(coding_context.code_doc.content, str) + assert len(coding_context.code_doc.content) > 0 captured = capfd.readouterr() print(f"输出内容: {captured.out}") -# @pytest.mark.asyncio -# async def test_write_code_review_directly(): -# code = SEARCH_CODE_SAMPLE -# write_code_review = WriteCodeReview("write_code_review") -# review = await write_code_review.run(code) -# logger.info(review) +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/actions/test_write_docstring.py b/tests/metagpt/actions/test_write_docstring.py index a0fc46ebd..ebb7e8cb1 100644 --- a/tests/metagpt/actions/test_write_docstring.py +++ b/tests/metagpt/actions/test_write_docstring.py @@ -27,8 +27,8 @@ class Person: ], ids=["google", "numpy", "sphinx"], ) -async def test_write_docstring(style: str, part: str): - ret = await WriteDocstring().run(code, style=style) +async def test_write_docstring(style: str, part: str, context): + ret = await WriteDocstring(context=context).run(code, style=style) assert part in ret diff --git a/tests/metagpt/actions/test_write_prd.py b/tests/metagpt/actions/test_write_prd.py index d854cd8d2..31d20018e 100644 --- a/tests/metagpt/actions/test_write_prd.py +++ b/tests/metagpt/actions/test_write_prd.py @@ -10,21 +10,18 @@ import pytest from metagpt.actions import UserRequirement, WritePRD from metagpt.const import REQUIREMENT_FILENAME -from metagpt.context import CONTEXT from metagpt.logs import logger from metagpt.roles.product_manager import ProductManager from metagpt.roles.role import RoleReactMode from metagpt.schema import Message from metagpt.utils.common import any_to_str -from metagpt.utils.project_repo import ProjectRepo @pytest.mark.asyncio -async def test_write_prd(new_filename): - product_manager = ProductManager() +async def test_write_prd(new_filename, context): + product_manager = ProductManager(context=context) requirements = "开发一个基于大语言模型与私有知识库的搜索引擎,希望可以基于大语言模型进行搜索总结" - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.docs.save(filename=REQUIREMENT_FILENAME, content=requirements) + await context.repo.docs.save(filename=REQUIREMENT_FILENAME, content=requirements) product_manager.rc.react_mode = RoleReactMode.BY_ORDER prd = await product_manager.run(Message(content=requirements, cause_by=UserRequirement)) assert prd.cause_by == any_to_str(WritePRD) @@ -34,7 +31,7 @@ async def test_write_prd(new_filename): # Assert the prd is not None or empty assert prd is not None assert prd.content != "" - assert ProjectRepo(product_manager.context.git_repo).docs.prd.changed_files + assert product_manager.context.repo.docs.prd.changed_files if __name__ == "__main__": diff --git a/tests/metagpt/actions/test_write_prd_an.py b/tests/metagpt/actions/test_write_prd_an.py new file mode 100644 index 000000000..378ce42c3 --- /dev/null +++ b/tests/metagpt/actions/test_write_prd_an.py @@ -0,0 +1,48 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/03 +@Author : mannaandpoem +@File : test_write_prd_an.py +""" +import pytest +from openai._models import BaseModel + +from metagpt.actions.action_node import ActionNode +from metagpt.actions.write_prd import NEW_REQ_TEMPLATE +from metagpt.actions.write_prd_an import REFINED_PRD_NODE +from metagpt.llm import LLM +from tests.data.incremental_dev_project.mock import ( + NEW_REQUIREMENT_SAMPLE, + PRD_SAMPLE, + REFINED_PRD_JSON, +) + + +@pytest.fixture() +def llm(): + return LLM() + + +def mock_refined_prd_json(): + return REFINED_PRD_JSON + + +@pytest.mark.asyncio +async def test_write_prd_an(mocker): + root = ActionNode.from_children("RefinedPRD", [ActionNode(key="", expected_type=str, instruction="", example="")]) + root.instruct_content = BaseModel() + root.instruct_content.model_dump = mock_refined_prd_json + mocker.patch("metagpt.actions.write_prd_an.REFINED_PRD_NODE.fill", return_value=root) + + prompt = NEW_REQ_TEMPLATE.format( + requirements=NEW_REQUIREMENT_SAMPLE, + old_prd=PRD_SAMPLE, + ) + node = await REFINED_PRD_NODE.fill(prompt, llm) + + assert "Refined Requirements" in node.instruct_content.model_dump() + assert "Refined Product Goals" in node.instruct_content.model_dump() + assert "Refined User Stories" in node.instruct_content.model_dump() + assert "Refined Requirement Analysis" in node.instruct_content.model_dump() + assert "Refined Requirement Pool" in node.instruct_content.model_dump() diff --git a/tests/metagpt/actions/test_write_prd_review.py b/tests/metagpt/actions/test_write_prd_review.py index 9b3f0a285..8e1601b2e 100644 --- a/tests/metagpt/actions/test_write_prd_review.py +++ b/tests/metagpt/actions/test_write_prd_review.py @@ -11,7 +11,7 @@ from metagpt.actions.write_prd_review import WritePRDReview @pytest.mark.asyncio -async def test_write_prd_review(): +async def test_write_prd_review(context): prd = """ Introduction: This is a new feature for our product. Goals: The goal is to improve user engagement. @@ -23,7 +23,7 @@ async def test_write_prd_review(): Timeline: The feature should be ready for testing in 1.5 months. """ - write_prd_review = WritePRDReview(name="write_prd_review") + write_prd_review = WritePRDReview(name="write_prd_review", context=context) prd_review = await write_prd_review.run(prd) diff --git a/tests/metagpt/actions/test_write_review.py b/tests/metagpt/actions/test_write_review.py index 2d188b720..0274a3532 100644 --- a/tests/metagpt/actions/test_write_review.py +++ b/tests/metagpt/actions/test_write_review.py @@ -9,7 +9,7 @@ import pytest from metagpt.actions.write_review import WriteReview -CONTEXT = """ +TEMPLATE_CONTEXT = """ { "Language": "zh_cn", "Programming Language": "Python", @@ -46,8 +46,8 @@ CONTEXT = """ @pytest.mark.asyncio -async def test_write_review(): - write_review = WriteReview() - review = await write_review.run(CONTEXT) +async def test_write_review(context): + write_review = WriteReview(context=context) + review = await write_review.run(TEMPLATE_CONTEXT) assert review.instruct_content assert review.get("LGTM") in ["LGTM", "LBTM"] diff --git a/tests/metagpt/actions/test_write_teaching_plan.py b/tests/metagpt/actions/test_write_teaching_plan.py index 3d556ab92..bb68d4286 100644 --- a/tests/metagpt/actions/test_write_teaching_plan.py +++ b/tests/metagpt/actions/test_write_teaching_plan.py @@ -13,11 +13,11 @@ from metagpt.actions.write_teaching_plan import WriteTeachingPlanPart @pytest.mark.asyncio @pytest.mark.parametrize( - ("topic", "context"), + ("topic", "content"), [("Title", "Lesson 1: Learn to draw an apple."), ("Teaching Content", "Lesson 1: Learn to draw an apple.")], ) -async def test_write_teaching_plan_part(topic, context): - action = WriteTeachingPlanPart(topic=topic, i_context=context) +async def test_write_teaching_plan_part(topic, content, context): + action = WriteTeachingPlanPart(topic=topic, i_context=content, context=context) rsp = await action.run() assert rsp diff --git a/tests/metagpt/actions/test_write_test.py b/tests/metagpt/actions/test_write_test.py index e09038414..9469dd312 100644 --- a/tests/metagpt/actions/test_write_test.py +++ b/tests/metagpt/actions/test_write_test.py @@ -13,7 +13,7 @@ from metagpt.schema import Document, TestingContext @pytest.mark.asyncio -async def test_write_test(): +async def test_write_test(context): code = """ import random from typing import Tuple @@ -25,8 +25,8 @@ async def test_write_test(): def generate(self, max_y: int, max_x: int): self.position = (random.randint(1, max_y - 1), random.randint(1, max_x - 1)) """ - context = TestingContext(filename="food.py", code_doc=Document(filename="food.py", content=code)) - write_test = WriteTest(i_context=context) + testing_context = TestingContext(filename="food.py", code_doc=Document(filename="food.py", content=code)) + write_test = WriteTest(i_context=testing_context, context=context) context = await write_test.run() logger.info(context.model_dump_json()) @@ -39,12 +39,12 @@ async def test_write_test(): @pytest.mark.asyncio -async def test_write_code_invalid_code(mocker): +async def test_write_code_invalid_code(mocker, context): # Mock the _aask method to return an invalid code string mocker.patch.object(WriteTest, "_aask", return_value="Invalid Code String") # Create an instance of WriteTest - write_test = WriteTest() + write_test = WriteTest(context=context) # Call the write_code method code = await write_test.write_code("Some prompt:") diff --git a/tests/metagpt/actions/test_write_tutorial.py b/tests/metagpt/actions/test_write_tutorial.py index 27a323b44..a83da1a1c 100644 --- a/tests/metagpt/actions/test_write_tutorial.py +++ b/tests/metagpt/actions/test_write_tutorial.py @@ -14,8 +14,8 @@ from metagpt.actions.write_tutorial import WriteContent, WriteDirectory @pytest.mark.asyncio @pytest.mark.parametrize(("language", "topic"), [("English", "Write a tutorial about Python")]) -async def test_write_directory(language: str, topic: str): - ret = await WriteDirectory(language=language).run(topic=topic) +async def test_write_directory(language: str, topic: str, context): + ret = await WriteDirectory(language=language, context=context).run(topic=topic) assert isinstance(ret, dict) assert "title" in ret assert "directory" in ret @@ -29,8 +29,8 @@ async def test_write_directory(language: str, topic: str): ("language", "topic", "directory"), [("English", "Write a tutorial about Python", {"Introduction": ["What is Python?", "Why learn Python?"]})], ) -async def test_write_content(language: str, topic: str, directory: Dict): - ret = await WriteContent(language=language, directory=directory).run(topic=topic) +async def test_write_content(language: str, topic: str, directory: Dict, context): + ret = await WriteContent(language=language, directory=directory, context=context).run(topic=topic) assert isinstance(ret, str) assert list(directory.keys())[0] in ret for value in list(directory.values())[0]: diff --git a/tests/metagpt/environment/android_env/__init__.py b/tests/metagpt/environment/android_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/tests/metagpt/environment/android_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/tests/metagpt/environment/android_env/test_android_ext_env.py b/tests/metagpt/environment/android_env/test_android_ext_env.py new file mode 100644 index 000000000..c9dfc718b --- /dev/null +++ b/tests/metagpt/environment/android_env/test_android_ext_env.py @@ -0,0 +1,75 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the unittest of AndroidExtEnv + +from pathlib import Path + +from metagpt.environment.android_env.android_ext_env import AndroidExtEnv +from metagpt.environment.android_env.const import ADB_EXEC_FAIL + + +def mock_device_shape(self, adb_cmd: str) -> str: + return "shape: 720x1080" + + +def mock_device_shape_invalid(self, adb_cmd: str) -> str: + return ADB_EXEC_FAIL + + +def mock_list_devices(self, adb_cmd: str) -> str: + return "devices\nemulator-5554" + + +def mock_get_screenshot(self, adb_cmd: str) -> str: + return "screenshot_xxxx-xx-xx" + + +def mock_get_xml(self, adb_cmd: str) -> str: + return "xml_xxxx-xx-xx" + + +def mock_write_read_operation(self, adb_cmd: str) -> str: + return "OK" + + +def test_android_ext_env(mocker): + device_id = "emulator-5554" + mocker.patch( + "metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_device_shape + ) + + ext_env = AndroidExtEnv(device_id=device_id, screenshot_dir="/data2/", xml_dir="/data2/") + assert ext_env.adb_prefix == f"adb -s {device_id} " + assert ext_env.adb_prefix_shell == f"adb -s {device_id} shell " + assert ext_env.adb_prefix_si == f"adb -s {device_id} shell input " + + assert ext_env.device_shape == (720, 1080) + + mocker.patch( + "metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_device_shape_invalid + ) + assert ext_env.device_shape == (0, 0) + + mocker.patch( + "metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_list_devices + ) + assert ext_env.list_devices() == [device_id] + + mocker.patch( + "metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_get_screenshot + ) + assert ext_env.get_screenshot("screenshot_xxxx-xx-xx", "/data/") == Path("/data/screenshot_xxxx-xx-xx.png") + + mocker.patch("metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_get_xml) + assert ext_env.get_xml("xml_xxxx-xx-xx", "/data/") == Path("/data/xml_xxxx-xx-xx.xml") + + mocker.patch( + "metagpt.environment.android_env.android_ext_env.AndroidExtEnv.execute_adb_with_cmd", mock_write_read_operation + ) + res = "OK" + assert ext_env.system_back() == res + assert ext_env.system_tap(10, 10) == res + assert ext_env.user_input("test_input") == res + assert ext_env.user_longpress(10, 10) == res + assert ext_env.user_swipe(10, 10) == res + assert ext_env.user_swipe_to((10, 10), (20, 20)) == res diff --git a/tests/metagpt/environment/api/__init__.py b/tests/metagpt/environment/api/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/tests/metagpt/environment/api/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/tests/metagpt/environment/api/test_env_api.py b/tests/metagpt/environment/api/test_env_api.py new file mode 100644 index 000000000..53f98c0d3 --- /dev/null +++ b/tests/metagpt/environment/api/test_env_api.py @@ -0,0 +1,15 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : + +from metagpt.environment.api.env_api import EnvAPIRegistry + + +def test_env_api_registry(): + def test_func(): + pass + + env_api_registry = EnvAPIRegistry() + env_api_registry["test"] = test_func + + env_api_registry.get("test") == test_func diff --git a/tests/metagpt/environment/mincraft_env/__init__.py b/tests/metagpt/environment/mincraft_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/tests/metagpt/environment/mincraft_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/tests/metagpt/environment/mincraft_env/test_mincraft_ext_env.py b/tests/metagpt/environment/mincraft_env/test_mincraft_ext_env.py new file mode 100644 index 000000000..ad3376141 --- /dev/null +++ b/tests/metagpt/environment/mincraft_env/test_mincraft_ext_env.py @@ -0,0 +1,14 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the unittest of MincraftExtEnv + + +from metagpt.environment.mincraft_env.const import MC_CKPT_DIR +from metagpt.environment.mincraft_env.mincraft_ext_env import MincraftExtEnv + + +def test_mincraft_ext_env(): + ext_env = MincraftExtEnv() + assert ext_env.server, f"{ext_env.server_host}:{ext_env.server_port}" + assert MC_CKPT_DIR.joinpath("skill/code").exists() + assert ext_env.warm_up.get("optional_inventory_items") == 7 diff --git a/tests/metagpt/environment/stanford_town_env/__init__.py b/tests/metagpt/environment/stanford_town_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/tests/metagpt/environment/stanford_town_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/tests/metagpt/environment/stanford_town_env/test_stanford_town_ext_env.py b/tests/metagpt/environment/stanford_town_env/test_stanford_town_ext_env.py new file mode 100644 index 000000000..3071f9deb --- /dev/null +++ b/tests/metagpt/environment/stanford_town_env/test_stanford_town_ext_env.py @@ -0,0 +1,40 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the unittest of StanfordTownExtEnv + +from pathlib import Path + +from metagpt.environment.stanford_town_env.stanford_town_ext_env import ( + StanfordTownExtEnv, +) + +maze_asset_path = ( + Path(__file__).absolute().parent.joinpath("..", "..", "..", "data", "environment", "stanford_town", "the_ville") +) + + +def test_stanford_town_ext_env(): + ext_env = StanfordTownExtEnv(maze_asset_path=maze_asset_path) + + tile_coord = ext_env.turn_coordinate_to_tile((64, 64)) + assert tile_coord == (2, 2) + + tile = (58, 9) + assert len(ext_env.get_collision_maze()) == 100 + assert len(ext_env.get_address_tiles()) == 306 + assert ext_env.access_tile(tile=tile)["world"] == "the Ville" + assert ext_env.get_tile_path(tile=tile, level="world") == "the Ville" + assert len(ext_env.get_nearby_tiles(tile=tile, vision_r=5)) == 121 + + event = ("double studio:double studio:bedroom 2:bed", None, None, None) + ext_env.add_tiles_event(tile[1], tile[0], event=event) + ext_env.add_event_from_tile(event, tile) + assert len(ext_env.tiles[tile[1]][tile[0]]["events"]) == 1 + + ext_env.turn_event_from_tile_idle(event, tile) + + ext_env.remove_event_from_tile(event, tile) + assert len(ext_env.tiles[tile[1]][tile[0]]["events"]) == 0 + + ext_env.remove_subject_events_from_tile(subject=event[0], tile=tile) + assert len(ext_env.tiles[tile[1]][tile[0]]["events"]) == 0 diff --git a/tests/metagpt/environment/test_base_env.py b/tests/metagpt/environment/test_base_env.py new file mode 100644 index 000000000..fd73679d8 --- /dev/null +++ b/tests/metagpt/environment/test_base_env.py @@ -0,0 +1,54 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the unittest of ExtEnv&Env + +import pytest + +from metagpt.environment.api.env_api import EnvAPIAbstract +from metagpt.environment.base_env import ( + Environment, + env_read_api_registry, + env_write_api_registry, + mark_as_readable, + mark_as_writeable, +) + + +class ForTestEnv(Environment): + value: int = 0 + + @mark_as_readable + def read_api_no_param(self): + return self.value + + @mark_as_readable + def read_api(self, a: int, b: int): + return a + b + + @mark_as_writeable + def write_api(self, a: int, b: int): + self.value = a + b + + @mark_as_writeable + async def async_read_api(self, a: int, b: int): + return a + b + + +@pytest.mark.asyncio +async def test_ext_env(): + env = ForTestEnv() + assert len(env_read_api_registry) > 0 + assert len(env_write_api_registry) > 0 + + apis = env.get_all_available_apis(mode="read") + assert len(apis) > 0 + assert len(apis["read_api"]) == 3 + + _ = await env.step(EnvAPIAbstract(api_name="write_api", kwargs={"a": 5, "b": 10})) + assert env.value == 15 + + with pytest.raises(ValueError): + await env.observe("not_exist_api") + + assert await env.observe("read_api_no_param") == 15 + assert await env.observe(EnvAPIAbstract(api_name="read_api", kwargs={"a": 5, "b": 5})) == 10 diff --git a/tests/metagpt/environment/werewolf_env/__init__.py b/tests/metagpt/environment/werewolf_env/__init__.py new file mode 100644 index 000000000..2bcf8efd0 --- /dev/null +++ b/tests/metagpt/environment/werewolf_env/__init__.py @@ -0,0 +1,3 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : diff --git a/tests/metagpt/environment/werewolf_env/test_werewolf_ext_env.py b/tests/metagpt/environment/werewolf_env/test_werewolf_ext_env.py new file mode 100644 index 000000000..0694c5c3d --- /dev/null +++ b/tests/metagpt/environment/werewolf_env/test_werewolf_ext_env.py @@ -0,0 +1,64 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Desc : the unittest of WerewolfExtEnv + +from metagpt.environment.werewolf_env.werewolf_ext_env import RoleState, WerewolfExtEnv +from metagpt.roles.role import Role + + +class Werewolf(Role): + profile: str = "Werewolf" + + +class Villager(Role): + profile: str = "Villager" + + +class Witch(Role): + profile: str = "Witch" + + +class Guard(Role): + profile: str = "Guard" + + +def test_werewolf_ext_env(): + players_state = { + "Player0": ("Werewolf", RoleState.ALIVE), + "Player1": ("Werewolf", RoleState.ALIVE), + "Player2": ("Villager", RoleState.ALIVE), + "Player3": ("Witch", RoleState.ALIVE), + "Player4": ("Guard", RoleState.ALIVE), + } + ext_env = WerewolfExtEnv(players_state=players_state, step_idx=4, special_role_players=["Player3", "Player4"]) + + assert len(ext_env.living_players) == 5 + assert len(ext_env.special_role_players) == 2 + assert len(ext_env.werewolf_players) == 2 + + curr_instr = ext_env.curr_step_instruction() + assert ext_env.step_idx == 5 + assert "Werewolves, please open your eyes" in curr_instr["content"] + + # current step_idx = 5 + ext_env.wolf_kill_someone(wolf=Role(name="Player10"), player_name="Player4") + ext_env.wolf_kill_someone(wolf=Werewolf(name="Player0"), player_name="Player4") + ext_env.wolf_kill_someone(wolf=Werewolf(name="Player1"), player_name="Player4") + assert ext_env.player_hunted == "Player4" + assert len(ext_env.living_players) == 5 # hunted but can be saved by witch + + for idx in range(13): + _ = ext_env.curr_step_instruction() + + # current step_idx = 18 + assert ext_env.step_idx == 18 + ext_env.vote_kill_someone(voteer=Werewolf(name="Player0"), player_name="Player2") + ext_env.vote_kill_someone(voteer=Werewolf(name="Player1"), player_name="Player3") + ext_env.vote_kill_someone(voteer=Villager(name="Player2"), player_name="Player3") + ext_env.vote_kill_someone(voteer=Witch(name="Player3"), player_name="Player4") + ext_env.vote_kill_someone(voteer=Guard(name="Player4"), player_name="Player2") + assert ext_env.player_current_dead == "Player2" + assert len(ext_env.living_players) == 4 + + player_names = ["Player0", "Player2"] + assert ext_env.get_players_state(player_names) == dict(zip(player_names, [RoleState.ALIVE, RoleState.KILLED])) diff --git a/tests/metagpt/learn/test_google_search.py b/tests/metagpt/learn/test_google_search.py index da32e8923..70a146878 100644 --- a/tests/metagpt/learn/test_google_search.py +++ b/tests/metagpt/learn/test_google_search.py @@ -1,27 +1,21 @@ -import asyncio - +import pytest from pydantic import BaseModel from metagpt.learn.google_search import google_search +from metagpt.tools import SearchEngineType -async def mock_google_search(): +@pytest.mark.asyncio +async def test_google_search(search_engine_mocker): class Input(BaseModel): input: str inputs = [{"input": "ai agent"}] - for i in inputs: seed = Input(**i) - result = await google_search(seed.input) + result = await google_search( + seed.input, + engine=SearchEngineType.SERPER_GOOGLE, + api_key="mock-serper-key", + ) assert result != "" - - -def test_suite(): - loop = asyncio.get_event_loop() - task = loop.create_task(mock_google_search()) - loop.run_until_complete(task) - - -if __name__ == "__main__": - test_suite() diff --git a/tests/metagpt/learn/test_skill_loader.py b/tests/metagpt/learn/test_skill_loader.py index 45697160b..f1952c275 100644 --- a/tests/metagpt/learn/test_skill_loader.py +++ b/tests/metagpt/learn/test_skill_loader.py @@ -10,13 +10,12 @@ from pathlib import Path import pytest -from metagpt.context import CONTEXT from metagpt.learn.skill_loader import SkillsDeclaration @pytest.mark.asyncio -async def test_suite(): - CONTEXT.kwargs.agent_skills = [ +async def test_suite(context): + context.kwargs.agent_skills = [ {"id": 1, "name": "text_to_speech", "type": "builtin", "config": {}, "enabled": True}, {"id": 2, "name": "text_to_image", "type": "builtin", "config": {}, "enabled": True}, {"id": 3, "name": "ai_call", "type": "builtin", "config": {}, "enabled": True}, @@ -27,7 +26,7 @@ async def test_suite(): ] pathname = Path(__file__).parent / "../../../docs/.well-known/skills.yaml" loader = await SkillsDeclaration.load(skill_yaml_file_name=pathname) - skills = loader.get_skill_list() + skills = loader.get_skill_list(context=context) assert skills assert len(skills) >= 3 for desc, name in skills.items(): diff --git a/tests/metagpt/learn/test_text_to_embedding.py b/tests/metagpt/learn/test_text_to_embedding.py index 8891960c1..f50f6a7aa 100644 --- a/tests/metagpt/learn/test_text_to_embedding.py +++ b/tests/metagpt/learn/test_text_to_embedding.py @@ -11,7 +11,7 @@ from pathlib import Path import pytest -from metagpt.config2 import config +from metagpt.config2 import Config from metagpt.learn.text_to_embedding import text_to_embedding from metagpt.utils.common import aread @@ -19,13 +19,14 @@ from metagpt.utils.common import aread @pytest.mark.asyncio async def test_text_to_embedding(mocker): # mock + config = Config.default() mock_post = mocker.patch("aiohttp.ClientSession.post") mock_response = mocker.AsyncMock() mock_response.status = 200 data = await aread(Path(__file__).parent / "../../data/openai/embedding.json") mock_response.json.return_value = json.loads(data) mock_post.return_value.__aenter__.return_value = mock_response - type(config.get_openai_llm()).proxy = mocker.PropertyMock(return_value="http://mock.proxy") + config.get_openai_llm().proxy = mocker.PropertyMock(return_value="http://mock.proxy") # Prerequisites assert config.get_openai_llm().api_key diff --git a/tests/metagpt/learn/test_text_to_image.py b/tests/metagpt/learn/test_text_to_image.py index 167a35891..d3272dadd 100644 --- a/tests/metagpt/learn/test_text_to_image.py +++ b/tests/metagpt/learn/test_text_to_image.py @@ -27,7 +27,7 @@ async def test_text_to_image(mocker): mocker.patch.object(S3, "cache", return_value="http://mock/s3") config = Config.default() - assert config.METAGPT_TEXT_TO_IMAGE_MODEL_URL + assert config.metagpt_tti_url data = await text_to_image("Panda emoji", size_type="512x512", config=config) assert "base64" in data or "http" in data @@ -52,7 +52,7 @@ async def test_openai_text_to_image(mocker): mocker.patch.object(S3, "cache", return_value="http://mock.s3.com/0.png") config = Config.default() - config.METAGPT_TEXT_TO_IMAGE_MODEL_URL = None + config.metagpt_tti_url = None assert config.get_openai_llm() data = await text_to_image("Panda emoji", size_type="512x512", config=config) diff --git a/tests/metagpt/learn/test_text_to_speech.py b/tests/metagpt/learn/test_text_to_speech.py index 38e051cc6..f01e5d132 100644 --- a/tests/metagpt/learn/test_text_to_speech.py +++ b/tests/metagpt/learn/test_text_to_speech.py @@ -20,9 +20,9 @@ from metagpt.utils.s3 import S3 async def test_azure_text_to_speech(mocker): # mock config = Config.default() - config.IFLYTEK_API_KEY = None - config.IFLYTEK_API_SECRET = None - config.IFLYTEK_APP_ID = None + config.iflytek_api_key = None + config.iflytek_api_secret = None + config.iflytek_app_id = None mock_result = mocker.Mock() mock_result.audio_data = b"mock audio data" mock_result.reason = ResultReason.SynthesizingAudioCompleted @@ -32,11 +32,11 @@ async def test_azure_text_to_speech(mocker): mocker.patch.object(S3, "cache", return_value="http://mock.s3.com/1.wav") # Prerequisites - assert not config.IFLYTEK_APP_ID - assert not config.IFLYTEK_API_KEY - assert not config.IFLYTEK_API_SECRET - assert config.AZURE_TTS_SUBSCRIPTION_KEY and config.AZURE_TTS_SUBSCRIPTION_KEY != "YOUR_API_KEY" - assert config.AZURE_TTS_REGION + assert not config.iflytek_app_id + assert not config.iflytek_api_key + assert not config.iflytek_api_secret + assert config.azure_tts_subscription_key and config.azure_tts_subscription_key != "YOUR_API_KEY" + assert config.azure_tts_region config.copy() # test azure @@ -48,8 +48,8 @@ async def test_azure_text_to_speech(mocker): async def test_iflytek_text_to_speech(mocker): # mock config = Config.default() - config.AZURE_TTS_SUBSCRIPTION_KEY = None - config.AZURE_TTS_REGION = None + config.azure_tts_subscription_key = None + config.azure_tts_region = None mocker.patch.object(IFlyTekTTS, "synthesize_speech", return_value=None) mock_data = mocker.AsyncMock() mock_data.read.return_value = b"mock iflytek" @@ -58,11 +58,11 @@ async def test_iflytek_text_to_speech(mocker): mocker.patch.object(S3, "cache", return_value="http://mock.s3.com/1.mp3") # Prerequisites - assert config.IFLYTEK_APP_ID - assert config.IFLYTEK_API_KEY - assert config.IFLYTEK_API_SECRET - assert not config.AZURE_TTS_SUBSCRIPTION_KEY or config.AZURE_TTS_SUBSCRIPTION_KEY == "YOUR_API_KEY" - assert not config.AZURE_TTS_REGION + assert config.iflytek_app_id + assert config.iflytek_api_key + assert config.iflytek_api_secret + assert not config.azure_tts_subscription_key or config.azure_tts_subscription_key == "YOUR_API_KEY" + assert not config.azure_tts_region # test azure data = await text_to_speech("panda emoji", config=config) diff --git a/tests/metagpt/provider/test_openai.py b/tests/metagpt/provider/test_openai.py index ca9e918da..d6aa04c7b 100644 --- a/tests/metagpt/provider/test_openai.py +++ b/tests/metagpt/provider/test_openai.py @@ -1,9 +1,19 @@ -import pytest +import json +import pytest +from openai.types.chat import ( + ChatCompletion, + ChatCompletionMessage, + ChatCompletionMessageToolCall, +) +from openai.types.chat.chat_completion import Choice +from openai.types.chat.chat_completion_message_tool_call import Function +from PIL import Image + +from metagpt.const import TEST_DATA_PATH from metagpt.llm import LLM from metagpt.logs import logger from metagpt.provider import OpenAILLM -from metagpt.schema import UserMessage from tests.metagpt.provider.mock_llm_config import ( mock_llm_config, mock_llm_config_proxy, @@ -11,45 +21,103 @@ from tests.metagpt.provider.mock_llm_config import ( @pytest.mark.asyncio -async def test_aask_code(): +async def test_text_to_speech(): llm = LLM() - msg = [{"role": "user", "content": "Write a python hello world code."}] - rsp = await llm.aask_code(msg) # -> {'language': 'python', 'code': "print('Hello, World!')"} - - logger.info(rsp) - assert "language" in rsp - assert "code" in rsp - assert len(rsp["code"]) > 0 + resp = await llm.atext_to_speech( + model="tts-1", + voice="alloy", + input="人生说起来长,但直到一个岁月回头看,许多事件仅是仓促的。一段一段拼凑一起,合成了人生。苦难当头时,当下不免觉得是折磨;回头看,也不够是一段短短的人生旅程。", + ) + assert 200 == resp.response.status_code @pytest.mark.asyncio -async def test_aask_code_str(): +async def test_speech_to_text(): llm = LLM() - msg = "Write a python hello world code." - rsp = await llm.aask_code(msg) # -> {'language': 'python', 'code': "print('Hello, World!')"} - assert "language" in rsp - assert "code" in rsp - assert len(rsp["code"]) > 0 + audio_file = open(f"{TEST_DATA_PATH}/audio/hello.mp3", "rb") + resp = await llm.aspeech_to_text(file=audio_file, model="whisper-1") + assert "你好" == resp.text -@pytest.mark.asyncio -async def test_aask_code_message(): - llm = LLM() - msg = UserMessage("Write a python hello world code.") - rsp = await llm.aask_code(msg) # -> {'language': 'python', 'code': "print('Hello, World!')"} - assert "language" in rsp - assert "code" in rsp - assert len(rsp["code"]) > 0 +@pytest.fixture +def tool_calls_rsp(): + function_rsps = [ + Function(arguments='{\n"language": "python",\n"code": "print(\'hello world\')"}', name="execute"), + ] + tool_calls = [ + ChatCompletionMessageToolCall(type="function", id=f"call_{i}", function=f) for i, f in enumerate(function_rsps) + ] + messages = [ChatCompletionMessage(content=None, role="assistant", tool_calls=[t]) for t in tool_calls] + # 添加一个纯文本响应 + messages.append( + ChatCompletionMessage(content="Completed a python code for hello world!", role="assistant", tool_calls=None) + ) + # 添加 openai tool calls respond bug, code 出现在ChatCompletionMessage.content中 + messages.extend( + [ + ChatCompletionMessage(content="```python\nprint('hello world')```", role="assistant", tool_calls=None), + ] + ) + choices = [ + Choice(finish_reason="tool_calls", logprobs=None, index=i, message=msg) for i, msg in enumerate(messages) + ] + return [ + ChatCompletion(id=str(i), choices=[c], created=i, model="gpt-4", object="chat.completion") + for i, c in enumerate(choices) + ] + + +@pytest.fixture +def json_decode_error(): + function_rsp = Function(arguments='{\n"language": \'python\',\n"code": "print(\'hello world\')"}', name="execute") + tool_calls = [ChatCompletionMessageToolCall(type="function", id=f"call_{0}", function=function_rsp)] + message = ChatCompletionMessage(content=None, role="assistant", tool_calls=tool_calls) + choices = [Choice(finish_reason="tool_calls", logprobs=None, index=0, message=message)] + return ChatCompletion(id="0", choices=choices, created=0, model="gpt-4", object="chat.completion") class TestOpenAI: def test_make_client_kwargs_without_proxy(self): instance = OpenAILLM(mock_llm_config) kwargs = instance._make_client_kwargs() - assert kwargs == {"api_key": "mock_api_key", "base_url": "mock_base_url"} + assert kwargs["api_key"] == "mock_api_key" + assert kwargs["base_url"] == "mock_base_url" assert "http_client" not in kwargs def test_make_client_kwargs_with_proxy(self): instance = OpenAILLM(mock_llm_config_proxy) kwargs = instance._make_client_kwargs() assert "http_client" in kwargs + + def test_get_choice_function_arguments_for_aask_code(self, tool_calls_rsp): + instance = OpenAILLM(mock_llm_config_proxy) + for i, rsp in enumerate(tool_calls_rsp): + code = instance.get_choice_function_arguments(rsp) + logger.info(f"\ntest get function call arguments {i}: {code}") + assert "code" in code + assert "language" in code + assert "hello world" in code["code"] + logger.info(f'code is : {code["code"]}') + + if "Completed a python code for hello world!" == code["code"]: + code["language"] == "markdown" + else: + code["language"] == "python" + + def test_aask_code_json_decode_error(self, json_decode_error): + instance = OpenAILLM(mock_llm_config) + with pytest.raises(json.decoder.JSONDecodeError) as e: + instance.get_choice_function_arguments(json_decode_error) + assert "JSONDecodeError" in str(e) + + +@pytest.mark.asyncio +async def test_gen_image(): + llm = LLM() + model = "dall-e-3" + prompt = 'a logo with word "MetaGPT"' + images: list[Image] = await llm.gen_image(model=model, prompt=prompt) + assert images[0].size == (1024, 1024) + + images: list[Image] = await llm.gen_image(model=model, prompt=prompt, resp_format="b64_json") + assert images[0].size == (1024, 1024) diff --git a/tests/metagpt/provider/test_zhipuai_api.py b/tests/metagpt/provider/test_zhipuai_api.py index c4a40c23d..ad2ececa2 100644 --- a/tests/metagpt/provider/test_zhipuai_api.py +++ b/tests/metagpt/provider/test_zhipuai_api.py @@ -3,7 +3,6 @@ # @Desc : the unittest of ZhiPuAILLM import pytest -from zhipuai.utils.sse_client import Event from metagpt.provider.zhipuai_api import ZhiPuAILLM from tests.metagpt.provider.mock_llm_config import mock_llm_config_zhipu @@ -13,35 +12,16 @@ messages = [{"role": "user", "content": prompt_msg}] resp_content = "I'm chatglm-turbo" default_resp = { - "code": 200, - "data": { - "choices": [{"role": "assistant", "content": resp_content}], - "usage": {"prompt_tokens": 20, "completion_tokens": 20}, - }, + "choices": [{"finish_reason": "stop", "index": 0, "message": {"content": resp_content, "role": "assistant"}}], + "usage": {"completion_tokens": 22, "prompt_tokens": 19, "total_tokens": 41}, } -def mock_zhipuai_invoke(**kwargs) -> dict: - return default_resp - - -async def mock_zhipuai_ainvoke(**kwargs) -> dict: - return default_resp - - -async def mock_zhipuai_asse_invoke(**kwargs): +async def mock_zhipuai_acreate_stream(self, **kwargs): class MockResponse(object): async def _aread(self): class Iterator(object): - events = [ - Event(id="xxx", event="add", data=resp_content, retry=0), - Event( - id="xxx", - event="finish", - data="", - meta='{"usage": {"completion_tokens": 20,"prompt_tokens": 20}}', - ), - ] + events = [{"choices": [{"index": 0, "delta": {"content": resp_content, "role": "assistant"}}]}] async def __aiter__(self): for event in self.events: @@ -50,23 +30,26 @@ async def mock_zhipuai_asse_invoke(**kwargs): async for chunk in Iterator(): yield chunk - async def async_events(self): + async def stream(self): async for chunk in self._aread(): yield chunk return MockResponse() +async def mock_zhipuai_acreate(self, **kwargs) -> dict: + return default_resp + + @pytest.mark.asyncio async def test_zhipuai_acompletion(mocker): - mocker.patch("metagpt.provider.zhipuai.zhipu_model_api.ZhiPuModelAPI.invoke", mock_zhipuai_invoke) - mocker.patch("metagpt.provider.zhipuai.zhipu_model_api.ZhiPuModelAPI.ainvoke", mock_zhipuai_ainvoke) - mocker.patch("metagpt.provider.zhipuai.zhipu_model_api.ZhiPuModelAPI.asse_invoke", mock_zhipuai_asse_invoke) + mocker.patch("metagpt.provider.zhipuai.zhipu_model_api.ZhiPuModelAPI.acreate", mock_zhipuai_acreate) + mocker.patch("metagpt.provider.zhipuai.zhipu_model_api.ZhiPuModelAPI.acreate_stream", mock_zhipuai_acreate_stream) zhipu_gpt = ZhiPuAILLM(mock_llm_config_zhipu) resp = await zhipu_gpt.acompletion(messages) - assert resp["data"]["choices"][0]["content"] == resp_content + assert resp["choices"][0]["message"]["content"] == resp_content resp = await zhipu_gpt.aask(prompt_msg, stream=False) assert resp == resp_content diff --git a/tests/metagpt/provider/zhipuai/test_async_sse_client.py b/tests/metagpt/provider/zhipuai/test_async_sse_client.py index 2649f595b..31b2d3d64 100644 --- a/tests/metagpt/provider/zhipuai/test_async_sse_client.py +++ b/tests/metagpt/provider/zhipuai/test_async_sse_client.py @@ -11,16 +11,16 @@ from metagpt.provider.zhipuai.async_sse_client import AsyncSSEClient async def test_async_sse_client(): class Iterator(object): async def __aiter__(self): - yield b"data: test_value" + yield b'data: {"test_key": "test_value"}' async_sse_client = AsyncSSEClient(event_source=Iterator()) - async for event in async_sse_client.async_events(): - assert event.data, "test_value" + async for chunk in async_sse_client.stream(): + assert "test_value" in chunk.values() class InvalidIterator(object): async def __aiter__(self): yield b"invalid: test_value" async_sse_client = AsyncSSEClient(event_source=InvalidIterator()) - async for event in async_sse_client.async_events(): - assert not event + async for chunk in async_sse_client.stream(): + assert not chunk diff --git a/tests/metagpt/provider/zhipuai/test_zhipu_model_api.py b/tests/metagpt/provider/zhipuai/test_zhipu_model_api.py index daae65ab7..15673c51c 100644 --- a/tests/metagpt/provider/zhipuai/test_zhipu_model_api.py +++ b/tests/metagpt/provider/zhipuai/test_zhipu_model_api.py @@ -6,15 +6,13 @@ from typing import Any, Tuple import pytest import zhipuai -from zhipuai.model_api.api import InvokeType -from zhipuai.utils.http_client import headers as zhipuai_default_headers from metagpt.provider.zhipuai.zhipu_model_api import ZhiPuModelAPI api_key = "xxx.xxx" zhipuai.api_key = api_key -default_resp = b'{"result": "test response"}' +default_resp = b'{"choices": [{"finish_reason": "stop", "index": 0, "message": {"content": "test response", "role": "assistant"}}]}' async def mock_requestor_arequest(self, **kwargs) -> Tuple[Any, Any, str]: @@ -23,22 +21,15 @@ async def mock_requestor_arequest(self, **kwargs) -> Tuple[Any, Any, str]: @pytest.mark.asyncio async def test_zhipu_model_api(mocker): - header = ZhiPuModelAPI.get_header() - zhipuai_default_headers.update({"Authorization": api_key}) - assert header == zhipuai_default_headers - - ZhiPuModelAPI.get_sse_header() - # assert len(sse_header["Authorization"]) == 191 - - url_prefix, url_suffix = ZhiPuModelAPI.split_zhipu_api_url(InvokeType.SYNC, kwargs={"model": "chatglm_turbo"}) + url_prefix, url_suffix = ZhiPuModelAPI(api_key=api_key).split_zhipu_api_url() assert url_prefix == "https://open.bigmodel.cn/api" - assert url_suffix == "/paas/v3/model-api/chatglm_turbo/invoke" + assert url_suffix == "/paas/v4/chat/completions" mocker.patch("metagpt.provider.general_api_requestor.GeneralAPIRequestor.arequest", mock_requestor_arequest) - result = await ZhiPuModelAPI.arequest( - InvokeType.SYNC, stream=False, method="get", headers={}, kwargs={"model": "chatglm_turbo"} + result = await ZhiPuModelAPI(api_key=api_key).arequest( + stream=False, method="get", headers={}, kwargs={"model": "glm-3-turbo"} ) assert result == default_resp - result = await ZhiPuModelAPI.ainvoke() - assert result["result"] == "test response" + result = await ZhiPuModelAPI(api_key=api_key).acreate() + assert result["choices"][0]["message"]["content"] == "test response" diff --git a/tests/metagpt/roles/ci/test_code_interpreter.py b/tests/metagpt/roles/ci/test_code_interpreter.py new file mode 100644 index 000000000..9d2f2429b --- /dev/null +++ b/tests/metagpt/roles/ci/test_code_interpreter.py @@ -0,0 +1,23 @@ +import pytest + +from metagpt.logs import logger +from metagpt.roles.ci.code_interpreter import CodeInterpreter + + +@pytest.mark.asyncio +@pytest.mark.parametrize("auto_run", [(True), (False)]) +async def test_code_interpreter(mocker, auto_run): + mocker.patch("metagpt.actions.ci.execute_nb_code.ExecuteNbCode.run", return_value=("a successful run", True)) + mocker.patch("builtins.input", return_value="confirm") + + requirement = "Run data analysis on sklearn Iris dataset, include a plot" + tools = [] + + ci = CodeInterpreter(auto_run=auto_run, use_tools=True, tools=tools) + rsp = await ci.run(requirement) + logger.info(rsp) + assert len(rsp.content) > 0 + + finished_tasks = ci.planner.plan.get_finished_tasks() + assert len(finished_tasks) > 0 + assert len(finished_tasks[0].code) > 0 # check one task to see if code is recorded diff --git a/tests/metagpt/roles/ci/test_ml_engineer.py b/tests/metagpt/roles/ci/test_ml_engineer.py new file mode 100644 index 000000000..3bf9f3b92 --- /dev/null +++ b/tests/metagpt/roles/ci/test_ml_engineer.py @@ -0,0 +1,90 @@ +import pytest + +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.logs import logger +from metagpt.roles.ci.ml_engineer import MLEngineer +from metagpt.schema import Message, Plan, Task +from metagpt.tools.tool_type import ToolType +from tests.metagpt.actions.ci.test_debug_code import CODE, DebugContext, ErrorStr + + +def test_mle_init(): + ci = MLEngineer(goal="test", auto_run=True, use_tools=True, tools=["tool1", "tool2"]) + assert ci.tools == [] + + +MockPlan = Plan( + goal="This is a titanic passenger survival dataset, your goal is to predict passenger survival outcome. The target column is Survived. Perform data analysis, data preprocessing, feature engineering, and modeling to predict the target. Report accuracy on the eval data. Train data path: 'tests/data/ml_datasets/titanic/split_train.csv', eval data path: 'tests/data/ml_datasets/titanic/split_eval.csv'.", + context="", + tasks=[ + Task( + task_id="1", + dependent_task_ids=[], + instruction="Perform exploratory data analysis on the train dataset to understand the features and target variable.", + task_type="eda", + code="", + result="", + is_success=False, + is_finished=False, + ) + ], + task_map={ + "1": Task( + task_id="1", + dependent_task_ids=[], + instruction="Perform exploratory data analysis on the train dataset to understand the features and target variable.", + task_type="eda", + code="", + result="", + is_success=False, + is_finished=False, + ) + }, + current_task_id="1", +) + + +@pytest.mark.asyncio +async def test_mle_write_code(mocker): + data_path = "tests/data/ml_datasets/titanic" + + mle = MLEngineer(auto_run=True, use_tools=True) + mle.planner.plan = MockPlan + + code, _ = await mle._write_code() + assert data_path in code["code"] + + +@pytest.mark.asyncio +async def test_mle_update_data_columns(mocker): + mle = MLEngineer(auto_run=True, use_tools=True) + mle.planner.plan = MockPlan + + # manually update task type to test update + mle.planner.plan.current_task.task_type = ToolType.DATA_PREPROCESS.value + + result = await mle._update_data_columns() + assert result is not None + + +@pytest.mark.asyncio +async def test_mle_debug_code(mocker): + mle = MLEngineer(auto_run=True, use_tools=True) + mle.working_memory.add(Message(content=ErrorStr, cause_by=ExecuteNbCode)) + mle.latest_code = CODE + mle.debug_context = DebugContext + code, _ = await mle._write_code() + assert len(code) > 0 + + +@pytest.mark.skip +@pytest.mark.asyncio +async def test_ml_engineer(): + data_path = "tests/data/ml_datasets/titanic" + requirement = f"This is a titanic passenger survival dataset, your goal is to predict passenger survival outcome. The target column is Survived. Perform data analysis, data preprocessing, feature engineering, and modeling to predict the target. Report accuracy on the eval data. Train data path: '{data_path}/split_train.csv', eval data path: '{data_path}/split_eval.csv'." + tools = ["FillMissingValue", "CatCross", "dummy_tool"] + + mle = MLEngineer(auto_run=True, use_tools=True, tools=tools) + rsp = await mle.run(requirement) + logger.info(rsp) + assert len(rsp.content) > 0 diff --git a/tests/metagpt/roles/test_architect.py b/tests/metagpt/roles/test_architect.py index f9d6606ac..b02242ed2 100644 --- a/tests/metagpt/roles/test_architect.py +++ b/tests/metagpt/roles/test_architect.py @@ -13,7 +13,6 @@ import pytest from metagpt.actions import WriteDesign, WritePRD from metagpt.const import PRDS_FILE_REPO -from metagpt.context import CONTEXT from metagpt.logs import logger from metagpt.roles import Architect from metagpt.schema import Message @@ -22,12 +21,12 @@ from tests.metagpt.roles.mock import MockMessages @pytest.mark.asyncio -async def test_architect(): +async def test_architect(context): # Prerequisites filename = uuid.uuid4().hex + ".json" - await awrite(CONTEXT.git_repo.workdir / PRDS_FILE_REPO / filename, data=MockMessages.prd.content) + await awrite(context.repo.workdir / PRDS_FILE_REPO / filename, data=MockMessages.prd.content) - role = Architect() + role = Architect(context=context) rsp = await role.run(with_message=Message(content="", cause_by=WritePRD)) logger.info(rsp) assert len(rsp.content) > 0 diff --git a/tests/metagpt/roles/test_assistant.py b/tests/metagpt/roles/test_assistant.py index b9740a112..bd0efea35 100644 --- a/tests/metagpt/roles/test_assistant.py +++ b/tests/metagpt/roles/test_assistant.py @@ -12,7 +12,6 @@ from pydantic import BaseModel from metagpt.actions.skill_action import SkillAction from metagpt.actions.talk_action import TalkAction -from metagpt.context import CONTEXT from metagpt.memory.brain_memory import BrainMemory from metagpt.roles.assistant import Assistant from metagpt.schema import Message @@ -20,11 +19,11 @@ from metagpt.utils.common import any_to_str @pytest.mark.asyncio -async def test_run(mocker): +async def test_run(mocker, context): # mock mocker.patch("metagpt.learn.text_to_image", return_value="http://mock.com/1.png") - CONTEXT.kwargs.language = "Chinese" + context.kwargs.language = "Chinese" class Input(BaseModel): memory: BrainMemory @@ -80,7 +79,7 @@ async def test_run(mocker): for i in inputs: seed = Input(**i) - role = Assistant(language="Chinese") + role = Assistant(language="Chinese", context=context) role.context.kwargs.language = seed.language role.context.kwargs.agent_description = seed.agent_description role.context.kwargs.agent_skills = agent_skills @@ -115,8 +114,8 @@ async def test_run(mocker): ], ) @pytest.mark.asyncio -async def test_memory(memory): - role = Assistant() +async def test_memory(memory, context): + role = Assistant(context=context) role.context.kwargs.agent_skills = [] role.load_memory(memory) diff --git a/tests/metagpt/roles/test_engineer.py b/tests/metagpt/roles/test_engineer.py index 17b94828c..d263a8a2f 100644 --- a/tests/metagpt/roles/test_engineer.py +++ b/tests/metagpt/roles/test_engineer.py @@ -8,44 +8,35 @@ distribution feature for message handling. """ import json -import uuid from pathlib import Path import pytest from metagpt.actions import WriteCode, WriteTasks -from metagpt.const import ( - DEFAULT_WORKSPACE_ROOT, - REQUIREMENT_FILENAME, - SYSTEM_DESIGN_FILE_REPO, - TASK_FILE_REPO, -) -from metagpt.context import CONTEXT, Context +from metagpt.const import REQUIREMENT_FILENAME, SYSTEM_DESIGN_FILE_REPO, TASK_FILE_REPO from metagpt.logs import logger from metagpt.roles.engineer import Engineer from metagpt.schema import CodingContext, Message from metagpt.utils.common import CodeParser, any_to_name, any_to_str, aread, awrite -from metagpt.utils.git_repository import ChangeType, GitRepository -from metagpt.utils.project_repo import ProjectRepo +from metagpt.utils.git_repository import ChangeType from tests.metagpt.roles.mock import STRS_FOR_PARSING, TASKS, MockMessages @pytest.mark.asyncio -async def test_engineer(): +async def test_engineer(context): # Prerequisites rqno = "20231221155954.json" - project_repo = ProjectRepo(CONTEXT.git_repo) - await project_repo.save(REQUIREMENT_FILENAME, content=MockMessages.req.content) - await project_repo.docs.prd.save(rqno, content=MockMessages.prd.content) - await project_repo.docs.system_design.save(rqno, content=MockMessages.system_design.content) - await project_repo.docs.task.save(rqno, content=MockMessages.json_tasks.content) + await context.repo.save(REQUIREMENT_FILENAME, content=MockMessages.req.content) + await context.repo.docs.prd.save(rqno, content=MockMessages.prd.content) + await context.repo.docs.system_design.save(rqno, content=MockMessages.system_design.content) + await context.repo.docs.task.save(rqno, content=MockMessages.json_tasks.content) - engineer = Engineer() + engineer = Engineer(context=context) rsp = await engineer.run(Message(content="", cause_by=WriteTasks)) logger.info(rsp) assert rsp.cause_by == any_to_str(WriteCode) - assert project_repo.with_src_path(CONTEXT.src_workspace).srcs.changed_files + assert context.repo.with_src_path(context.src_workspace).srcs.changed_files def test_parse_str(): @@ -108,14 +99,12 @@ def test_parse_code(): def test_todo(): role = Engineer() - assert role.todo == any_to_name(WriteCode) + assert role.action_description == any_to_name(WriteCode) @pytest.mark.asyncio -async def test_new_coding_context(): +async def test_new_coding_context(context): # Prerequisites - context = Context() - context.git_repo = GitRepository(local_path=DEFAULT_WORKSPACE_ROOT / f"unittest/{uuid.uuid4().hex}") demo_path = Path(__file__).parent / "../../data/demo_project" deps = json.loads(await aread(demo_path / "dependencies.json")) dependency = await context.git_repo.get_dependency() @@ -123,11 +112,11 @@ async def test_new_coding_context(): await dependency.update(k, set(v)) data = await aread(demo_path / "system_design.json") rqno = "20231221155954.json" - await awrite(context.git_repo.workdir / SYSTEM_DESIGN_FILE_REPO / rqno, data) + await awrite(context.repo.workdir / SYSTEM_DESIGN_FILE_REPO / rqno, data) data = await aread(demo_path / "tasks.json") - await awrite(context.git_repo.workdir / TASK_FILE_REPO / rqno, data) + await awrite(context.repo.workdir / TASK_FILE_REPO / rqno, data) - context.src_workspace = Path(context.git_repo.workdir) / "game_2048" + context.src_workspace = Path(context.repo.workdir) / "game_2048" try: filename = "game.py" @@ -149,9 +138,7 @@ async def test_new_coding_context(): context.git_repo.add_change({f"{TASK_FILE_REPO}/{rqno}": ChangeType.UNTRACTED}) context.git_repo.commit("mock env") - await ProjectRepo(context.git_repo).with_src_path(context.src_workspace).srcs.save( - filename=filename, content="content" - ) + await context.repo.with_src_path(context.src_workspace).srcs.save(filename=filename, content="content") role = Engineer(context=context) assert not role.code_todos await role._new_code_actions() diff --git a/tests/metagpt/roles/test_invoice_ocr_assistant.py b/tests/metagpt/roles/test_invoice_ocr_assistant.py index e3a9259da..bedcd6712 100644 --- a/tests/metagpt/roles/test_invoice_ocr_assistant.py +++ b/tests/metagpt/roles/test_invoice_ocr_assistant.py @@ -41,9 +41,11 @@ from metagpt.schema import Message ), ], ) -async def test_invoice_ocr_assistant(query: str, invoice_path: Path, invoice_table_path: Path, expected_result: dict): +async def test_invoice_ocr_assistant( + query: str, invoice_path: Path, invoice_table_path: Path, expected_result: dict, context +): invoice_path = TEST_DATA_PATH / invoice_path - role = InvoiceOCRAssistant() + role = InvoiceOCRAssistant(context=context) await role.run(Message(content=query, instruct_content=InvoicePath(file_path=invoice_path))) invoice_table_path = DATA_PATH / invoice_table_path df = pd.read_excel(invoice_table_path) diff --git a/tests/metagpt/roles/test_product_manager.py b/tests/metagpt/roles/test_product_manager.py index 1083e81b0..59b5aa81a 100644 --- a/tests/metagpt/roles/test_product_manager.py +++ b/tests/metagpt/roles/test_product_manager.py @@ -5,17 +5,51 @@ @Author : alexanderwu @File : test_product_manager.py """ +import json + import pytest +from metagpt.actions import WritePRD +from metagpt.actions.prepare_documents import PrepareDocuments +from metagpt.const import REQUIREMENT_FILENAME +from metagpt.context import Context from metagpt.logs import logger from metagpt.roles import ProductManager +from metagpt.utils.common import any_to_str from tests.metagpt.roles.mock import MockMessages @pytest.mark.asyncio async def test_product_manager(new_filename): - product_manager = ProductManager() - rsp = await product_manager.run(MockMessages.req) - logger.info(rsp) - assert len(rsp.content) > 0 - assert rsp.content == MockMessages.req.content + context = Context() + try: + assert context.git_repo is None + assert context.repo is None + product_manager = ProductManager(context=context) + # prepare documents + rsp = await product_manager.run(MockMessages.req) + assert context.git_repo + assert context.repo + assert rsp.cause_by == any_to_str(PrepareDocuments) + assert REQUIREMENT_FILENAME in context.repo.docs.changed_files + + # write prd + rsp = await product_manager.run(rsp) + assert rsp.cause_by == any_to_str(WritePRD) + logger.info(rsp) + assert len(rsp.content) > 0 + doc = list(rsp.instruct_content.docs.values())[0] + m = json.loads(doc.content) + assert m["Original Requirements"] == MockMessages.req.content + + # nothing to do + rsp = await product_manager.run(rsp) + assert rsp is None + except Exception as e: + assert not e + finally: + context.git_repo.delete_repository() + + +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/roles/test_project_manager.py b/tests/metagpt/roles/test_project_manager.py index 9207623bc..9b016927e 100644 --- a/tests/metagpt/roles/test_project_manager.py +++ b/tests/metagpt/roles/test_project_manager.py @@ -13,7 +13,7 @@ from tests.metagpt.roles.mock import MockMessages @pytest.mark.asyncio -async def test_project_manager(): - project_manager = ProjectManager() +async def test_project_manager(context): + project_manager = ProjectManager(context=context) rsp = await project_manager.run(MockMessages.system_design) logger.info(rsp) diff --git a/tests/metagpt/roles/test_qa_engineer.py b/tests/metagpt/roles/test_qa_engineer.py index c51642e6a..b89e7d5eb 100644 --- a/tests/metagpt/roles/test_qa_engineer.py +++ b/tests/metagpt/roles/test_qa_engineer.py @@ -13,20 +13,19 @@ from pydantic import Field from metagpt.actions import DebugError, RunCode, WriteTest from metagpt.actions.summarize_code import SummarizeCode -from metagpt.context import CONTEXT from metagpt.environment import Environment from metagpt.roles import QaEngineer from metagpt.schema import Message from metagpt.utils.common import any_to_str, aread, awrite -async def test_qa(): +async def test_qa(context): # Prerequisites demo_path = Path(__file__).parent / "../../data/demo_project" - CONTEXT.src_workspace = Path(CONTEXT.git_repo.workdir) / "qa/game_2048" + context.src_workspace = Path(context.repo.workdir) / "qa/game_2048" data = await aread(filename=demo_path / "game.py", encoding="utf-8") - await awrite(filename=CONTEXT.src_workspace / "game.py", data=data, encoding="utf-8") - await awrite(filename=Path(CONTEXT.git_repo.workdir) / "requirements.txt", data="") + await awrite(filename=context.src_workspace / "game.py", data=data, encoding="utf-8") + await awrite(filename=Path(context.repo.workdir) / "requirements.txt", data="") class MockEnv(Environment): msgs: List[Message] = Field(default_factory=list) @@ -37,7 +36,7 @@ async def test_qa(): env = MockEnv() - role = QaEngineer() + role = QaEngineer(context=context) role.set_env(env) await role.run(with_message=Message(content="", cause_by=SummarizeCode)) assert env.msgs diff --git a/tests/metagpt/roles/test_researcher.py b/tests/metagpt/roles/test_researcher.py index 7d0ec450d..ba05e1296 100644 --- a/tests/metagpt/roles/test_researcher.py +++ b/tests/metagpt/roles/test_researcher.py @@ -28,20 +28,20 @@ async def mock_llm_ask(self, prompt: str, system_msgs): @pytest.mark.asyncio -async def test_researcher(mocker, search_engine_mocker): +async def test_researcher(mocker, search_engine_mocker, context): with TemporaryDirectory() as dirname: topic = "dataiku vs. datarobot" mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", mock_llm_ask) researcher.RESEARCH_PATH = Path(dirname) - role = researcher.Researcher() + role = researcher.Researcher(context=context) for i in role.actions: if isinstance(i, CollectLinks): - i.search_engine = SearchEngine(SearchEngineType.DUCK_DUCK_GO) + i.search_engine = SearchEngine(engine=SearchEngineType.DUCK_DUCK_GO) await role.run(topic) assert (researcher.RESEARCH_PATH / f"{topic}.md").read_text().startswith("# Research Report") -def test_write_report(mocker): +def test_write_report(mocker, context): with TemporaryDirectory() as dirname: for i, topic in enumerate( [ @@ -53,7 +53,7 @@ def test_write_report(mocker): ): researcher.RESEARCH_PATH = Path(dirname) content = "# Research Report" - researcher.Researcher().write_report(topic, content) + researcher.Researcher(context=context).write_report(topic, content) assert (researcher.RESEARCH_PATH / f"{i+1}. metagpt.md").read_text().startswith("# Research Report") diff --git a/tests/metagpt/roles/test_role.py b/tests/metagpt/roles/test_role.py index 809f5c735..8b11e2d4a 100644 --- a/tests/metagpt/roles/test_role.py +++ b/tests/metagpt/roles/test_role.py @@ -13,8 +13,8 @@ def test_role_desc(): assert role.desc == "Best Seller" -def test_role_human(): - role = Role(is_human=True) +def test_role_human(context): + role = Role(is_human=True, context=context) assert isinstance(role.llm, HumanProvider) diff --git a/tests/metagpt/roles/test_tutorial_assistant.py b/tests/metagpt/roles/test_tutorial_assistant.py index 0e6c1efb9..c12c2b26e 100644 --- a/tests/metagpt/roles/test_tutorial_assistant.py +++ b/tests/metagpt/roles/test_tutorial_assistant.py @@ -15,8 +15,8 @@ from metagpt.roles.tutorial_assistant import TutorialAssistant @pytest.mark.asyncio @pytest.mark.parametrize(("language", "topic"), [("Chinese", "Write a tutorial about pip")]) -async def test_tutorial_assistant(language: str, topic: str): - role = TutorialAssistant(language=language) +async def test_tutorial_assistant(language: str, topic: str, context): + role = TutorialAssistant(language=language, context=context) msg = await role.run(topic) assert TUTORIAL_PATH.exists() filename = msg.content diff --git a/tests/metagpt/serialize_deserialize/test_action.py b/tests/metagpt/serialize_deserialize/test_action.py index f66900241..d234a160f 100644 --- a/tests/metagpt/serialize_deserialize/test_action.py +++ b/tests/metagpt/serialize_deserialize/test_action.py @@ -5,23 +5,22 @@ import pytest from metagpt.actions import Action -from metagpt.llm import LLM @pytest.mark.asyncio -async def test_action_serdeser(): - action = Action() +async def test_action_serdeser(context): + action = Action(context=context) ser_action_dict = action.model_dump() assert "name" in ser_action_dict assert "llm" not in ser_action_dict # not export assert "__module_class_name" in ser_action_dict - action = Action(name="test") + action = Action(name="test", context=context) ser_action_dict = action.model_dump() assert "test" in ser_action_dict["name"] - new_action = Action(**ser_action_dict) + new_action = Action(**ser_action_dict, context=context) assert new_action.name == "test" - assert isinstance(new_action.llm, type(LLM())) + assert isinstance(new_action.llm, type(context.llm())) assert len(await new_action._aask("who are you")) > 0 diff --git a/tests/metagpt/serialize_deserialize/test_architect.py b/tests/metagpt/serialize_deserialize/test_architect.py index a6823197a..e3c2703fa 100644 --- a/tests/metagpt/serialize_deserialize/test_architect.py +++ b/tests/metagpt/serialize_deserialize/test_architect.py @@ -9,16 +9,20 @@ from metagpt.roles.architect import Architect @pytest.mark.asyncio -async def test_architect_serdeser(): - role = Architect() +async def test_architect_serdeser(context): + role = Architect(context=context) ser_role_dict = role.model_dump(by_alias=True) assert "name" in ser_role_dict assert "states" in ser_role_dict assert "actions" in ser_role_dict - new_role = Architect(**ser_role_dict) + new_role = Architect(**ser_role_dict, context=context) assert new_role.name == "Bob" assert len(new_role.actions) == 1 assert len(new_role.rc.watch) == 1 assert isinstance(new_role.actions[0], Action) await new_role.actions[0].run(with_messages="write a cli snake game") + + +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/serialize_deserialize/test_environment.py b/tests/metagpt/serialize_deserialize/test_environment.py index 3e2a3abba..4e6ea93b5 100644 --- a/tests/metagpt/serialize_deserialize/test_environment.py +++ b/tests/metagpt/serialize_deserialize/test_environment.py @@ -18,20 +18,20 @@ from tests.metagpt.serialize_deserialize.test_serdeser_base import ( ) -def test_env_serdeser(): - env = Environment() +def test_env_serdeser(context): + env = Environment(context=context) env.publish_message(message=Message(content="test env serialize")) ser_env_dict = env.model_dump() assert "roles" in ser_env_dict assert len(ser_env_dict["roles"]) == 0 - new_env = Environment(**ser_env_dict) + new_env = Environment(**ser_env_dict, context=context) assert len(new_env.roles) == 0 assert len(new_env.history) == 25 -def test_environment_serdeser(): +def test_environment_serdeser(context): out_mapping = {"field1": (list[str], ...)} out_data = {"field1": ["field1 value1", "field1 value2"]} ic_obj = ActionNode.create_model_class("prd", out_mapping) @@ -40,7 +40,7 @@ def test_environment_serdeser(): content="prd", instruct_content=ic_obj(**out_data), role="product manager", cause_by=any_to_str(UserRequirement) ) - environment = Environment() + environment = Environment(context=context) role_c = RoleC() environment.add_role(role_c) environment.publish_message(message) @@ -48,7 +48,7 @@ def test_environment_serdeser(): ser_data = environment.model_dump() assert ser_data["roles"]["Role C"]["name"] == "RoleC" - new_env: Environment = Environment(**ser_data) + new_env: Environment = Environment(**ser_data, context=context) assert len(new_env.roles) == 1 assert list(new_env.roles.values())[0].states == list(environment.roles.values())[0].states @@ -57,22 +57,22 @@ def test_environment_serdeser(): assert type(list(new_env.roles.values())[0].actions[1]) == ActionRaise -def test_environment_serdeser_v2(): - environment = Environment() +def test_environment_serdeser_v2(context): + environment = Environment(context=context) pm = ProjectManager() environment.add_role(pm) ser_data = environment.model_dump() - new_env: Environment = Environment(**ser_data) + new_env: Environment = Environment(**ser_data, context=context) role = new_env.get_role(pm.profile) assert isinstance(role, ProjectManager) assert isinstance(role.actions[0], WriteTasks) assert isinstance(list(new_env.roles.values())[0].actions[0], WriteTasks) -def test_environment_serdeser_save(): - environment = Environment() +def test_environment_serdeser_save(context): + environment = Environment(context=context) role_c = RoleC() stg_path = serdeser_path.joinpath("team", "environment") @@ -82,6 +82,6 @@ def test_environment_serdeser_save(): write_json_file(env_path, environment.model_dump()) env_dict = read_json_file(env_path) - new_env: Environment = Environment(**env_dict) + new_env: Environment = Environment(**env_dict, context=context) assert len(new_env.roles) == 1 assert type(list(new_env.roles.values())[0].actions[0]) == ActionOK diff --git a/tests/metagpt/serialize_deserialize/test_memory.py b/tests/metagpt/serialize_deserialize/test_memory.py index fdaea7861..560ae2c51 100644 --- a/tests/metagpt/serialize_deserialize/test_memory.py +++ b/tests/metagpt/serialize_deserialize/test_memory.py @@ -13,7 +13,7 @@ from metagpt.utils.common import any_to_str, read_json_file, write_json_file from tests.metagpt.serialize_deserialize.test_serdeser_base import serdeser_path -def test_memory_serdeser(): +def test_memory_serdeser(context): msg1 = Message(role="Boss", content="write a snake game", cause_by=UserRequirement) out_mapping = {"field2": (list[str], ...)} @@ -39,7 +39,7 @@ def test_memory_serdeser(): assert memory.count() == 2 -def test_memory_serdeser_save(): +def test_memory_serdeser_save(context): msg1 = Message(role="User", content="write a 2048 game", cause_by=UserRequirement) out_mapping = {"field1": (list[str], ...)} diff --git a/tests/metagpt/serialize_deserialize/test_prepare_interview.py b/tests/metagpt/serialize_deserialize/test_prepare_interview.py index 3b57aa27e..a3e3edafc 100644 --- a/tests/metagpt/serialize_deserialize/test_prepare_interview.py +++ b/tests/metagpt/serialize_deserialize/test_prepare_interview.py @@ -8,12 +8,12 @@ from metagpt.actions.prepare_interview import PrepareInterview @pytest.mark.asyncio -async def test_action_serdeser(): - action = PrepareInterview() +async def test_action_serdeser(context): + action = PrepareInterview(context=context) serialized_data = action.model_dump() assert serialized_data["name"] == "PrepareInterview" - new_action = PrepareInterview(**serialized_data) + new_action = PrepareInterview(**serialized_data, context=context) assert new_action.name == "PrepareInterview" assert type(await new_action.run("python developer")) == ActionNode diff --git a/tests/metagpt/serialize_deserialize/test_product_manager.py b/tests/metagpt/serialize_deserialize/test_product_manager.py index 1a056f9d4..2338b406d 100644 --- a/tests/metagpt/serialize_deserialize/test_product_manager.py +++ b/tests/metagpt/serialize_deserialize/test_product_manager.py @@ -10,10 +10,10 @@ from metagpt.schema import Message @pytest.mark.asyncio -async def test_product_manager_serdeser(new_filename): - role = ProductManager() +async def test_product_manager_serdeser(new_filename, context): + role = ProductManager(context=context) ser_role_dict = role.model_dump(by_alias=True) - new_role = ProductManager(**ser_role_dict) + new_role = ProductManager(**ser_role_dict, context=context) assert new_role.name == "Alice" assert len(new_role.actions) == 2 diff --git a/tests/metagpt/serialize_deserialize/test_project_manager.py b/tests/metagpt/serialize_deserialize/test_project_manager.py index f2c5af853..fb998ae31 100644 --- a/tests/metagpt/serialize_deserialize/test_project_manager.py +++ b/tests/metagpt/serialize_deserialize/test_project_manager.py @@ -10,14 +10,14 @@ from metagpt.roles.project_manager import ProjectManager @pytest.mark.asyncio -async def test_project_manager_serdeser(): - role = ProjectManager() +async def test_project_manager_serdeser(context): + role = ProjectManager(context=context) ser_role_dict = role.model_dump(by_alias=True) assert "name" in ser_role_dict assert "states" in ser_role_dict assert "actions" in ser_role_dict - new_role = ProjectManager(**ser_role_dict) + new_role = ProjectManager(**ser_role_dict, context=context) assert new_role.name == "Eve" assert len(new_role.actions) == 1 assert isinstance(new_role.actions[0], Action) diff --git a/tests/metagpt/serialize_deserialize/test_reasearcher.py b/tests/metagpt/serialize_deserialize/test_reasearcher.py index a2d1fa513..67c52e692 100644 --- a/tests/metagpt/serialize_deserialize/test_reasearcher.py +++ b/tests/metagpt/serialize_deserialize/test_reasearcher.py @@ -8,13 +8,13 @@ from metagpt.roles.researcher import Researcher @pytest.mark.asyncio -async def test_tutorial_assistant_serdeser(): - role = Researcher() +async def test_tutorial_assistant_serdeser(context): + role = Researcher(context=context) ser_role_dict = role.model_dump() assert "name" in ser_role_dict assert "language" in ser_role_dict - new_role = Researcher(**ser_role_dict) + new_role = Researcher(**ser_role_dict, context=context) assert new_role.language == "en-us" assert len(new_role.actions) == 3 assert isinstance(new_role.actions[0], CollectLinks) diff --git a/tests/metagpt/serialize_deserialize/test_role.py b/tests/metagpt/serialize_deserialize/test_role.py index bbfe350b7..aaf7c1935 100644 --- a/tests/metagpt/serialize_deserialize/test_role.py +++ b/tests/metagpt/serialize_deserialize/test_role.py @@ -26,7 +26,7 @@ from tests.metagpt.serialize_deserialize.test_serdeser_base import ( ) -def test_roles(): +def test_roles(context): role_a = RoleA() assert len(role_a.rc.watch) == 1 role_b = RoleB() @@ -37,7 +37,7 @@ def test_roles(): assert len(role_d.actions) == 1 -def test_role_subclasses(): +def test_role_subclasses(context): """test subclasses of role with same fields in ser&deser""" class RoleSubClasses(BaseModel): @@ -51,7 +51,7 @@ def test_role_subclasses(): assert isinstance(new_role_subcls.roles[1], RoleB) -def test_role_serialize(): +def test_role_serialize(context): role = Role() ser_role_dict = role.model_dump() assert "name" in ser_role_dict @@ -59,7 +59,7 @@ def test_role_serialize(): assert "actions" in ser_role_dict -def test_engineer_serdeser(): +def test_engineer_serdeser(context): role = Engineer() ser_role_dict = role.model_dump() assert "name" in ser_role_dict @@ -73,7 +73,7 @@ def test_engineer_serdeser(): assert isinstance(new_role.actions[0], WriteCode) -def test_role_serdeser_save(): +def test_role_serdeser_save(context): shutil.rmtree(serdeser_path.joinpath("team"), ignore_errors=True) pm = ProductManager() @@ -89,7 +89,7 @@ def test_role_serdeser_save(): @pytest.mark.asyncio -async def test_role_serdeser_interrupt(): +async def test_role_serdeser_interrupt(context): role_c = RoleC() shutil.rmtree(serdeser_path.joinpath("team"), ignore_errors=True) diff --git a/tests/metagpt/serialize_deserialize/test_team.py b/tests/metagpt/serialize_deserialize/test_team.py index 57c8a8508..dbd38422d 100644 --- a/tests/metagpt/serialize_deserialize/test_team.py +++ b/tests/metagpt/serialize_deserialize/test_team.py @@ -21,8 +21,8 @@ from tests.metagpt.serialize_deserialize.test_serdeser_base import ( ) -def test_team_deserialize(): - company = Team() +def test_team_deserialize(context): + company = Team(context=context) pm = ProductManager() arch = Architect() @@ -52,10 +52,10 @@ def mock_team_serialize(self, stg_path: Path = serdeser_path.joinpath("team")): write_json_file(team_info_path, self.model_dump()) -def test_team_serdeser_save(mocker): +def test_team_serdeser_save(mocker, context): mocker.patch("metagpt.team.Team.serialize", mock_team_serialize) - company = Team() + company = Team(context=context) company.hire([RoleC()]) stg_path = serdeser_path.joinpath("team") @@ -69,14 +69,14 @@ def test_team_serdeser_save(mocker): @pytest.mark.asyncio -async def test_team_recover(mocker): +async def test_team_recover(mocker, context): mocker.patch("metagpt.team.Team.serialize", mock_team_serialize) idea = "write a snake game" stg_path = serdeser_path.joinpath("team") shutil.rmtree(stg_path, ignore_errors=True) - company = Team() + company = Team(context=context) role_c = RoleC() company.hire([role_c]) company.run_project(idea) @@ -95,14 +95,14 @@ async def test_team_recover(mocker): @pytest.mark.asyncio -async def test_team_recover_save(mocker): +async def test_team_recover_save(mocker, context): mocker.patch("metagpt.team.Team.serialize", mock_team_serialize) idea = "write a 2048 web game" stg_path = serdeser_path.joinpath("team") shutil.rmtree(stg_path, ignore_errors=True) - company = Team() + company = Team(context=context) role_c = RoleC() company.hire([role_c]) company.run_project(idea) @@ -121,7 +121,7 @@ async def test_team_recover_save(mocker): @pytest.mark.asyncio -async def test_team_recover_multi_roles_save(mocker): +async def test_team_recover_multi_roles_save(mocker, context): mocker.patch("metagpt.team.Team.serialize", mock_team_serialize) idea = "write a snake game" @@ -131,7 +131,7 @@ async def test_team_recover_multi_roles_save(mocker): role_a = RoleA() role_b = RoleB() - company = Team() + company = Team(context=context) company.hire([role_a, role_b]) company.run_project(idea) await company.run(n_round=4) @@ -144,3 +144,7 @@ async def test_team_recover_multi_roles_save(mocker): assert new_company.env.get_role(role_b.profile).rc.state == 1 await new_company.run(n_round=4) + + +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/serialize_deserialize/test_tutorial_assistant.py b/tests/metagpt/serialize_deserialize/test_tutorial_assistant.py index cb8feec19..ab5db4c57 100644 --- a/tests/metagpt/serialize_deserialize/test_tutorial_assistant.py +++ b/tests/metagpt/serialize_deserialize/test_tutorial_assistant.py @@ -7,7 +7,7 @@ from metagpt.roles.tutorial_assistant import TutorialAssistant @pytest.mark.asyncio -async def test_tutorial_assistant_serdeser(): +async def test_tutorial_assistant_serdeser(context): role = TutorialAssistant() ser_role_dict = role.model_dump() assert "name" in ser_role_dict diff --git a/tests/metagpt/serialize_deserialize/test_write_code.py b/tests/metagpt/serialize_deserialize/test_write_code.py index 132f343bc..2f3c08f9b 100644 --- a/tests/metagpt/serialize_deserialize/test_write_code.py +++ b/tests/metagpt/serialize_deserialize/test_write_code.py @@ -9,22 +9,23 @@ from metagpt.actions import WriteCode from metagpt.schema import CodingContext, Document -def test_write_design_serdeser(): - action = WriteCode() +def test_write_design_serdeser(context): + action = WriteCode(context=context) ser_action_dict = action.model_dump() assert ser_action_dict["name"] == "WriteCode" assert "llm" not in ser_action_dict # not export @pytest.mark.asyncio -async def test_write_code_serdeser(): - context = CodingContext( +async def test_write_code_serdeser(context): + context.src_workspace = context.repo.workdir / "srcs" + coding_context = CodingContext( filename="test_code.py", design_doc=Document(content="write add function to calculate two numbers") ) - doc = Document(content=context.model_dump_json()) - action = WriteCode(i_context=doc) + doc = Document(content=coding_context.model_dump_json()) + action = WriteCode(i_context=doc, context=context) serialized_data = action.model_dump() - new_action = WriteCode(**serialized_data) + new_action = WriteCode(**serialized_data, context=context) assert new_action.name == "WriteCode" await action.run() diff --git a/tests/metagpt/serialize_deserialize/test_write_code_review.py b/tests/metagpt/serialize_deserialize/test_write_code_review.py index 70a4f2077..32a017a97 100644 --- a/tests/metagpt/serialize_deserialize/test_write_code_review.py +++ b/tests/metagpt/serialize_deserialize/test_write_code_review.py @@ -9,22 +9,23 @@ from metagpt.schema import CodingContext, Document @pytest.mark.asyncio -async def test_write_code_review_serdeser(): +async def test_write_code_review_serdeser(context): + context.src_workspace = context.repo.workdir / "srcs" code_content = """ def div(a: int, b: int = 0): return a / b """ - context = CodingContext( + coding_context = CodingContext( filename="test_op.py", design_doc=Document(content="divide two numbers"), code_doc=Document(content=code_content), ) - action = WriteCodeReview(i_context=context) + action = WriteCodeReview(i_context=coding_context) serialized_data = action.model_dump() assert serialized_data["name"] == "WriteCodeReview" - new_action = WriteCodeReview(**serialized_data) + new_action = WriteCodeReview(**serialized_data, context=context) assert new_action.name == "WriteCodeReview" await new_action.run() diff --git a/tests/metagpt/serialize_deserialize/test_write_design.py b/tests/metagpt/serialize_deserialize/test_write_design.py index 37d505914..6519d8cdc 100644 --- a/tests/metagpt/serialize_deserialize/test_write_design.py +++ b/tests/metagpt/serialize_deserialize/test_write_design.py @@ -8,24 +8,24 @@ from metagpt.actions import WriteDesign, WriteTasks @pytest.mark.asyncio -async def test_write_design_serialize(): - action = WriteDesign() +async def test_write_design_serialize(context): + action = WriteDesign(context=context) ser_action_dict = action.model_dump() assert "name" in ser_action_dict assert "llm" not in ser_action_dict # not export - new_action = WriteDesign(**ser_action_dict) + new_action = WriteDesign(**ser_action_dict, context=context) assert new_action.name == "WriteDesign" await new_action.run(with_messages="write a cli snake game") @pytest.mark.asyncio -async def test_write_task_serialize(): - action = WriteTasks() +async def test_write_task_serialize(context): + action = WriteTasks(context=context) ser_action_dict = action.model_dump() assert "name" in ser_action_dict assert "llm" not in ser_action_dict # not export - new_action = WriteTasks(**ser_action_dict) + new_action = WriteTasks(**ser_action_dict, context=context) assert new_action.name == "WriteTasks" await new_action.run(with_messages="write a cli snake game") diff --git a/tests/metagpt/serialize_deserialize/test_write_docstring.py b/tests/metagpt/serialize_deserialize/test_write_docstring.py index fb927f089..363bed05e 100644 --- a/tests/metagpt/serialize_deserialize/test_write_docstring.py +++ b/tests/metagpt/serialize_deserialize/test_write_docstring.py @@ -29,14 +29,14 @@ class Person: ], ids=["google", "numpy", "sphinx"], ) -async def test_action_serdeser(style: str, part: str): - action = WriteDocstring() +async def test_action_serdeser(style: str, part: str, context): + action = WriteDocstring(context=context) serialized_data = action.model_dump() assert "name" in serialized_data assert serialized_data["desc"] == "Write docstring for code." - new_action = WriteDocstring(**serialized_data) + new_action = WriteDocstring(**serialized_data, context=context) assert new_action.name == "WriteDocstring" assert new_action.desc == "Write docstring for code." diff --git a/tests/metagpt/serialize_deserialize/test_write_prd.py b/tests/metagpt/serialize_deserialize/test_write_prd.py index 820ee237c..e4951efb7 100644 --- a/tests/metagpt/serialize_deserialize/test_write_prd.py +++ b/tests/metagpt/serialize_deserialize/test_write_prd.py @@ -10,13 +10,13 @@ from metagpt.schema import Message @pytest.mark.asyncio -async def test_action_serdeser(new_filename): - action = WritePRD() +async def test_action_serdeser(new_filename, context): + action = WritePRD(context=context) ser_action_dict = action.model_dump() assert "name" in ser_action_dict assert "llm" not in ser_action_dict # not export - new_action = WritePRD(**ser_action_dict) + new_action = WritePRD(**ser_action_dict, context=context) assert new_action.name == "WritePRD" - action_output = await new_action.run(with_messages=Message(content="write a cli snake game")) - assert len(action_output.content) > 0 + with pytest.raises(FileNotFoundError): + await new_action.run(with_messages=Message(content="write a cli snake game")) diff --git a/tests/metagpt/serialize_deserialize/test_write_review.py b/tests/metagpt/serialize_deserialize/test_write_review.py index 17e212276..de2fd9d7a 100644 --- a/tests/metagpt/serialize_deserialize/test_write_review.py +++ b/tests/metagpt/serialize_deserialize/test_write_review.py @@ -5,7 +5,7 @@ import pytest from metagpt.actions.action_node import ActionNode from metagpt.actions.write_review import WriteReview -CONTEXT = """ +TEMPLATE_CONTEXT = """ { "Language": "zh_cn", "Programming Language": "Python", @@ -42,13 +42,13 @@ CONTEXT = """ @pytest.mark.asyncio -async def test_action_serdeser(): - action = WriteReview() +async def test_action_serdeser(context): + action = WriteReview(context=context) serialized_data = action.model_dump() assert serialized_data["name"] == "WriteReview" - new_action = WriteReview(**serialized_data) - review = await new_action.run(CONTEXT) + new_action = WriteReview(**serialized_data, context=context) + review = await new_action.run(TEMPLATE_CONTEXT) assert new_action.name == "WriteReview" assert type(review) == ActionNode diff --git a/tests/metagpt/serialize_deserialize/test_write_tutorial.py b/tests/metagpt/serialize_deserialize/test_write_tutorial.py index 4eeef7e0d..d41b7b341 100644 --- a/tests/metagpt/serialize_deserialize/test_write_tutorial.py +++ b/tests/metagpt/serialize_deserialize/test_write_tutorial.py @@ -9,13 +9,13 @@ from metagpt.actions.write_tutorial import WriteContent, WriteDirectory @pytest.mark.asyncio @pytest.mark.parametrize(("language", "topic"), [("English", "Write a tutorial about Python")]) -async def test_write_directory_serdeser(language: str, topic: str): - action = WriteDirectory() +async def test_write_directory_serdeser(language: str, topic: str, context): + action = WriteDirectory(context=context) serialized_data = action.model_dump() assert serialized_data["name"] == "WriteDirectory" assert serialized_data["language"] == "Chinese" - new_action = WriteDirectory(**serialized_data) + new_action = WriteDirectory(**serialized_data, context=context) ret = await new_action.run(topic=topic) assert isinstance(ret, dict) assert "title" in ret @@ -30,12 +30,12 @@ async def test_write_directory_serdeser(language: str, topic: str): ("language", "topic", "directory"), [("English", "Write a tutorial about Python", {"Introduction": ["What is Python?", "Why learn Python?"]})], ) -async def test_write_content_serdeser(language: str, topic: str, directory: Dict): - action = WriteContent(language=language, directory=directory) +async def test_write_content_serdeser(language: str, topic: str, directory: Dict, context): + action = WriteContent(language=language, directory=directory, context=context) serialized_data = action.model_dump() assert serialized_data["name"] == "WriteContent" - new_action = WriteContent(**serialized_data) + new_action = WriteContent(**serialized_data, context=context) ret = await new_action.run(topic=topic) assert isinstance(ret, str) assert list(directory.keys())[0] in ret diff --git a/tests/metagpt/strategy/examples/creative_writing.py b/tests/metagpt/strategy/examples/test_creative_writing.py similarity index 100% rename from tests/metagpt/strategy/examples/creative_writing.py rename to tests/metagpt/strategy/examples/test_creative_writing.py diff --git a/tests/metagpt/strategy/examples/game24.py b/tests/metagpt/strategy/examples/test_game24.py similarity index 100% rename from tests/metagpt/strategy/examples/game24.py rename to tests/metagpt/strategy/examples/test_game24.py diff --git a/tests/metagpt/strategy/test_solver.py b/tests/metagpt/strategy/test_solver.py new file mode 100644 index 000000000..eae4a5a2a --- /dev/null +++ b/tests/metagpt/strategy/test_solver.py @@ -0,0 +1,47 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/1/31 13:54 +@Author : alexanderwu +@File : test_solver.py +""" +import pytest + +from metagpt.actions.action_graph import ActionGraph +from metagpt.llm import LLM +from metagpt.strategy.search_space import SearchSpace +from metagpt.strategy.solver import NaiveSolver + + +@pytest.mark.asyncio +async def test_solver(): + from metagpt.actions.write_prd_an import ( + COMPETITIVE_ANALYSIS, + ISSUE_TYPE, + PRODUCT_GOALS, + REQUIREMENT_POOL, + ) + + graph = ActionGraph() + graph.add_node(ISSUE_TYPE) + graph.add_node(PRODUCT_GOALS) + graph.add_node(COMPETITIVE_ANALYSIS) + graph.add_node(REQUIREMENT_POOL) + graph.add_edge(ISSUE_TYPE, PRODUCT_GOALS) + graph.add_edge(PRODUCT_GOALS, COMPETITIVE_ANALYSIS) + graph.add_edge(PRODUCT_GOALS, REQUIREMENT_POOL) + graph.add_edge(COMPETITIVE_ANALYSIS, REQUIREMENT_POOL) + search_space = SearchSpace() + llm = LLM() + context = "Create a 2048 game" + solver = NaiveSolver(graph, search_space, llm, context) + await solver.solve() + + print("## graph.nodes") + print(graph.nodes) + for k, v in graph.nodes.items(): + print(f"{v.key} | prevs: {[i.key for i in v.prevs]} | nexts: {[i.key for i in v.nexts]}") + + assert len(graph.nodes) == 4 + assert len(graph.execution_order) == 4 + assert graph.execution_order == [ISSUE_TYPE.key, PRODUCT_GOALS.key, COMPETITIVE_ANALYSIS.key, REQUIREMENT_POOL.key] diff --git a/tests/metagpt/test_context.py b/tests/metagpt/test_context.py index d662a906a..f8218c44d 100644 --- a/tests/metagpt/test_context.py +++ b/tests/metagpt/test_context.py @@ -6,7 +6,7 @@ @File : test_context.py """ from metagpt.configs.llm_config import LLMType -from metagpt.context import CONTEXT, AttrDict, Context +from metagpt.context import AttrDict, Context def test_attr_dict_1(): @@ -48,15 +48,15 @@ def test_context_1(): assert ctx.git_repo is None assert ctx.src_workspace is None assert ctx.cost_manager is not None - assert ctx.options is not None def test_context_2(): - llm = CONTEXT.config.get_openai_llm() + ctx = Context() + llm = ctx.config.get_openai_llm() assert llm is not None assert llm.api_type == LLMType.OPENAI - kwargs = CONTEXT.kwargs + kwargs = ctx.kwargs assert kwargs is not None kwargs.test_key = "test_value" diff --git a/tests/metagpt/test_context_mixin.py b/tests/metagpt/test_context_mixin.py index a098ff0dc..4389dc251 100644 --- a/tests/metagpt/test_context_mixin.py +++ b/tests/metagpt/test_context_mixin.py @@ -5,11 +5,14 @@ @Author : alexanderwu @File : test_context_mixin.py """ +from pathlib import Path + import pytest from pydantic import BaseModel from metagpt.actions import Action from metagpt.config2 import Config +from metagpt.const import CONFIG_ROOT from metagpt.context_mixin import ContextMixin from metagpt.environment import Environment from metagpt.roles import Role @@ -95,14 +98,18 @@ def test_config_mixin_4_multi_inheritance_override_config(): print(obj.__dict__.keys()) assert "private_config" in obj.__dict__.keys() - assert obj.llm.model == "mock_zhipu_model" + assert obj.config.llm.model == "mock_zhipu_model" @pytest.mark.asyncio async def test_config_priority(): """If action's config is set, then its llm will be set, otherwise, it will use the role's llm""" + home_dir = Path.home() / CONFIG_ROOT gpt4t = Config.from_home("gpt-4-1106-preview.yaml") + if not home_dir.exists(): + assert gpt4t is None gpt35 = Config.default() + gpt35.llm.model = "gpt-3.5-turbo-1106" gpt4 = Config.default() gpt4.llm.model = "gpt-4-0613" @@ -120,7 +127,7 @@ async def test_config_priority(): env = Environment(desc="US election live broadcast") Team(investment=10.0, env=env, roles=[A, B, C]) - assert a1.llm.model == "gpt-4-1106-preview" + assert a1.llm.model == "gpt-4-1106-preview" if Path(home_dir / "gpt-4-1106-preview.yaml").exists() else "gpt-4-0613" assert a2.llm.model == "gpt-4-0613" assert a3.llm.model == "gpt-3.5-turbo-1106" diff --git a/tests/metagpt/test_environment.py b/tests/metagpt/test_environment.py index 10839a2a5..7559655d3 100644 --- a/tests/metagpt/test_environment.py +++ b/tests/metagpt/test_environment.py @@ -11,7 +11,6 @@ from pathlib import Path import pytest from metagpt.actions import UserRequirement -from metagpt.context import CONTEXT from metagpt.environment import Environment from metagpt.logs import logger from metagpt.roles import Architect, ProductManager, Role @@ -44,9 +43,9 @@ def test_get_roles(env: Environment): @pytest.mark.asyncio async def test_publish_and_process_message(env: Environment): - if CONTEXT.git_repo: - CONTEXT.git_repo.delete_repository() - CONTEXT.git_repo = None + if env.context.git_repo: + env.context.git_repo.delete_repository() + env.context.git_repo = None product_manager = ProductManager(name="Alice", profile="Product Manager", goal="做AI Native产品", constraints="资源有限") architect = Architect( diff --git a/tests/metagpt/test_incremental_dev.py b/tests/metagpt/test_incremental_dev.py new file mode 100644 index 000000000..c47397dd7 --- /dev/null +++ b/tests/metagpt/test_incremental_dev.py @@ -0,0 +1,195 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/03 +@Author : mannaandpoem +@File : test_incremental_dev.py +""" +import os +import shutil +import subprocess +import time + +import pytest +from typer.testing import CliRunner + +from metagpt.const import TEST_DATA_PATH +from metagpt.logs import logger +from metagpt.software_company import app + +runner = CliRunner() + +IDEAS = [ + "Add subtraction, multiplication and division operations to the calculator. The current calculator can only perform basic addition operations, and it is necessary to introduce subtraction, multiplication, division operation into the calculator", + "Adding graphical interface functionality to enhance the user experience in the number-guessing game. The existing number-guessing game currently relies on command-line input for numbers. The goal is to introduce a graphical interface to improve the game's usability and visual appeal", + "Add a feature to remove deprecated words from the word cloud. The current word cloud generator does not support removing deprecated words. Now, The word cloud generator should support removing deprecated words. Customize deactivated words to exclude them from word cloud. Let users see all the words in the text file, and allow users to select the words they want to remove.", + "Add an AI opponent with fixed difficulty levels. Currently, the game only allows players to compete against themselves. Implement an AI algorithm that can playing with player. This will provide a more engaging and challenging experience for players.", + "Add functionality to view the history of scores. The original dice rolling game could only display the current game result, but the new requirement allows players to view the history of scores", + "Add functionality to view the history of scores and perform statistical analysis on them. The original dice rolling game could only display the current game result, but the new requirement allows players to view the history of scores and display the statistical analysis results of the current score", + "Changed score target for 2048 game from 2048 to 4096. Please change the game's score target from 2048 to 4096, and change the interface size from 4*4 to 8*8", + "Display the history score of the player in the 2048 game. Add a record board that can display players' historical score records so that players can trace their scores", + "Incremental Idea Gradually increase the speed of the snake as the game progresses. In the current version of the game, the snake’s speed remains constant throughout the gameplay. Implement a feature where the snake’s speed gradually increases over time, making the game more challenging and intense as the player progresses.", + "Introduce power-ups and obstacles to the game. The current version of the game only involves eating food and growing the snake. Add new elements such as power-ups that can enhance the snake’s speed or make it invincible for a short duration. At the same time, introduce obstacles like walls or enemies that the snake must avoid or overcome to continue growing.", +] + +PROJECT_NAMES = [ + "simple_add_calculator", + "number_guessing_game", + "word_cloud", + "Gomoku", + "dice_simulator_new", + "dice_simulator_new", + "pygame_2048", + "pygame_2048", + "snake_game", + "snake_game", +] + + +@pytest.mark.skip +def test_simple_add_calculator(): + result = get_incremental_dev_result(IDEAS[0], PROJECT_NAMES[0]) + log_and_check_result(result) + + +@pytest.mark.skip +def test_number_guessing_game(): + result = get_incremental_dev_result(IDEAS[1], PROJECT_NAMES[1]) + log_and_check_result(result) + + +@pytest.mark.skip +def test_word_cloud(): + result = get_incremental_dev_result(IDEAS[2], PROJECT_NAMES[2]) + log_and_check_result(result) + + +@pytest.mark.skip +def test_gomoku(): + result = get_incremental_dev_result(IDEAS[3], PROJECT_NAMES[3]) + log_and_check_result(result) + + +@pytest.mark.skip +def test_dice_simulator_new(): + for i, (idea, project_name) in enumerate(zip(IDEAS[4:6], PROJECT_NAMES[4:6]), start=1): + result = get_incremental_dev_result(idea, project_name) + log_and_check_result(result, "refine_" + str(i)) + + +@pytest.mark.skip +def test_refined_pygame_2048(): + for i, (idea, project_name) in enumerate(zip(IDEAS[6:8], PROJECT_NAMES[6:8]), start=1): + result = get_incremental_dev_result(idea, project_name) + log_and_check_result(result, "refine_" + str(i)) + + +@pytest.mark.skip +def test_refined_snake_game(): + for i, (idea, project_name) in enumerate(zip(IDEAS[8:10], PROJECT_NAMES[8:10]), start=1): + result = get_incremental_dev_result(idea, project_name) + log_and_check_result(result, "refine_" + str(i)) + + +def log_and_check_result(result, tag_name="refine"): + logger.info(result) + logger.info(result.output) + if "Aborting" in result.output: + assert False + else: + # After running, there will be new commit + cur_tag = subprocess.run(["git", "describe", "--tags"], capture_output=True, text=True).stdout.strip() + if cur_tag == "base": + assert False + else: + assert True + if subprocess.run(["git", "show-ref", "--verify", "--quiet", f"refs/tags/{tag_name}"]).returncode == 0: + tag_name += str(int(time.time())) + try: + subprocess.run(["git", "tag", tag_name], check=True) + except subprocess.CalledProcessError as e: + raise e + + +def get_incremental_dev_result(idea, project_name, use_review=True): + project_path = TEST_DATA_PATH / "incremental_dev_project" / project_name + # Check if the project path exists + if not project_path.exists(): + # If the project does not exist, extract the project file + try: + if shutil.which("unzip"): + subprocess.run(["unzip", f"{project_path}.zip", "-d", str(project_path.parent)], check=True) + elif shutil.which("tar"): + subprocess.run(["tar", "-xf", f"{project_path}.zip", "-C", str(project_path.parent)], check=True) + logger.info(f"Extracted project {project_name} successfully.") + except FileNotFoundError as e: + raise FileNotFoundError(f"Neither 'unzip' nor 'tar' command found. Error: {e}") + except subprocess.CalledProcessError as e: + raise Exception(f"Failed to extract project {project_name}. Error: {e}") + + check_or_create_base_tag(project_path) + args = [idea, "--inc", "--project-path", project_path, "--n-round", "20"] + if not use_review: + args.append("--no-code-review") + result = runner.invoke(app, args) + return result + + +def check_or_create_base_tag(project_path): + # Change the current working directory to the specified project path + os.chdir(project_path) + + # Initialize a Git repository + subprocess.run(["git", "init"], check=True) + + # Check if the 'base' tag exists + check_base_tag_cmd = ["git", "show-ref", "--verify", "--quiet", "refs/tags/base"] + if subprocess.run(check_base_tag_cmd).returncode == 0: + has_base_tag = True + else: + has_base_tag = False + + if has_base_tag: + logger.info("Base tag exists") + # Switch to the 'base' branch if it exists + try: + status = subprocess.run(["git", "status", "-s"], capture_output=True, text=True).stdout.strip() + if status: + subprocess.run(["git", "clean", "-df"]) + subprocess.run(["git", "checkout", "-f", "base"], check=True) + logger.info("Switched to base branch") + except Exception as e: + logger.error("Failed to switch to base branch") + raise e + + else: + logger.info("Base tag doesn't exist.") + # Add and commit the current code if 'base' tag doesn't exist + add_cmd = ["git", "add", "."] + try: + subprocess.run(add_cmd, check=True) + logger.info("Files added successfully.") + except subprocess.CalledProcessError as e: + logger.error(f"Failed to add files: {e}") + + commit_cmd = ["git", "commit", "-m", "Initial commit"] + try: + subprocess.run(commit_cmd, check=True) + logger.info("Committed all files with the message 'Initial commit'.") + except subprocess.CalledProcessError as e: + logger.error(f"Failed to commit: {e.stderr}") + + # Add 'base' tag + add_base_tag_cmd = ["git", "tag", "base"] + + # Check if the 'git tag' command was successful + try: + subprocess.run(add_base_tag_cmd, check=True) + logger.info("Added 'base' tag.") + except Exception as e: + logger.error("Failed to add 'base' tag.") + raise e + + +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/test_role.py b/tests/metagpt/test_role.py index 1b843795c..7e707803b 100644 --- a/tests/metagpt/test_role.py +++ b/tests/metagpt/test_role.py @@ -131,7 +131,7 @@ async def test_recover(): role.recovered = True role.latest_observed_msg = Message(content="recover_test") role.rc.state = 0 - assert role.first_action == any_to_name(MockAction) + assert role.action_description == any_to_name(MockAction) rsp = await role.run() assert rsp.cause_by == any_to_str(MockAction) diff --git a/tests/metagpt/test_schema.py b/tests/metagpt/test_schema.py index 0929e6c4a..a8fa27151 100644 --- a/tests/metagpt/test_schema.py +++ b/tests/metagpt/test_schema.py @@ -25,7 +25,9 @@ from metagpt.schema import ( Document, Message, MessageQueue, + Plan, SystemMessage, + Task, UserMessage, ) from metagpt.utils.common import any_to_str @@ -181,5 +183,173 @@ def test_class_view(): ) +class TestPlan: + def test_add_tasks_ordering(self): + plan = Plan(goal="") + + tasks = [ + Task(task_id="1", dependent_task_ids=["2", "3"], instruction="Third"), + Task(task_id="2", instruction="First"), + Task(task_id="3", dependent_task_ids=["2"], instruction="Second"), + ] # 2 -> 3 -> 1 + plan.add_tasks(tasks) + + assert [task.task_id for task in plan.tasks] == ["2", "3", "1"] + + def test_add_tasks_to_existing_no_common_prefix(self): + plan = Plan(goal="") + + tasks = [ + Task(task_id="1", dependent_task_ids=["2", "3"], instruction="Third"), + Task(task_id="2", instruction="First"), + Task(task_id="3", dependent_task_ids=["2"], instruction="Second", is_finished=True), + ] # 2 -> 3 -> 1 + plan.add_tasks(tasks) + + new_tasks = [Task(task_id="3", instruction="")] + plan.add_tasks(new_tasks) + + assert [task.task_id for task in plan.tasks] == ["3"] + assert not plan.tasks[0].is_finished # must be the new unfinished task + + def test_add_tasks_to_existing_with_common_prefix(self): + plan = Plan(goal="") + + tasks = [ + Task(task_id="1", dependent_task_ids=["2", "3"], instruction="Third"), + Task(task_id="2", instruction="First"), + Task(task_id="3", dependent_task_ids=["2"], instruction="Second"), + ] # 2 -> 3 -> 1 + plan.add_tasks(tasks) + plan.finish_current_task() # finish 2 + plan.finish_current_task() # finish 3 + + new_tasks = [ + Task(task_id="4", dependent_task_ids=["3"], instruction="Third"), + Task(task_id="2", instruction="First"), + Task(task_id="3", dependent_task_ids=["2"], instruction="Second"), + ] # 2 -> 3 -> 4, so the common prefix is 2 -> 3, and these two should be obtained from the existing tasks + plan.add_tasks(new_tasks) + + assert [task.task_id for task in plan.tasks] == ["2", "3", "4"] + assert ( + plan.tasks[0].is_finished and plan.tasks[1].is_finished + ) # "2" and "3" should be the original finished one + assert plan.current_task_id == "4" + + def test_current_task(self): + plan = Plan(goal="") + tasks = [ + Task(task_id="1", dependent_task_ids=["2"], instruction="Second"), + Task(task_id="2", instruction="First"), + ] + plan.add_tasks(tasks) + assert plan.current_task.task_id == "2" + + def test_finish_task(self): + plan = Plan(goal="") + tasks = [ + Task(task_id="1", instruction="First"), + Task(task_id="2", dependent_task_ids=["1"], instruction="Second"), + ] + plan.add_tasks(tasks) + plan.finish_current_task() + assert plan.current_task.task_id == "2" + + def test_finished_tasks(self): + plan = Plan(goal="") + tasks = [ + Task(task_id="1", instruction="First"), + Task(task_id="2", dependent_task_ids=["1"], instruction="Second"), + ] + plan.add_tasks(tasks) + plan.finish_current_task() + finished_tasks = plan.get_finished_tasks() + assert len(finished_tasks) == 1 + assert finished_tasks[0].task_id == "1" + + def test_reset_task_existing(self): + plan = Plan(goal="") + task = Task(task_id="1", instruction="Do something", code="print('Hello')", result="Hello", finished=True) + plan.add_tasks([task]) + plan.reset_task("1") + reset_task = plan.task_map["1"] + assert reset_task.code == "" + assert reset_task.result == "" + assert not reset_task.is_finished + + def test_reset_task_non_existing(self): + plan = Plan(goal="") + task = Task(task_id="1", instruction="Do something", code="print('Hello')", result="Hello", finished=True) + plan.add_tasks([task]) + plan.reset_task("2") # Task with ID 2 does not exist + assert "1" in plan.task_map + assert "2" not in plan.task_map + + def test_replace_task_with_dependents(self): + plan = Plan(goal="") + tasks = [ + Task(task_id="1", instruction="First Task", finished=True), + Task(task_id="2", instruction="Second Task", dependent_task_ids=["1"], finished=True), + ] + plan.add_tasks(tasks) + new_task = Task(task_id="1", instruction="Updated First Task") + plan.replace_task(new_task) + assert plan.task_map["1"].instruction == "Updated First Task" + assert not plan.task_map["2"].is_finished # Dependent task should be reset + assert plan.task_map["2"].code == "" + assert plan.task_map["2"].result == "" + + def test_replace_task_non_existing(self): + plan = Plan(goal="") + task = Task(task_id="1", instruction="First Task") + plan.add_tasks([task]) + new_task = Task(task_id="2", instruction="New Task") + with pytest.raises(AssertionError): + plan.replace_task(new_task) # Task with ID 2 does not exist in plan + assert "1" in plan.task_map + assert "2" not in plan.task_map + + def test_append_task_with_valid_dependencies(self): + plan = Plan(goal="Test") + existing_task = [Task(task_id="1")] + plan.add_tasks(existing_task) + new_task = Task(task_id="2", dependent_task_ids=["1"]) + plan.append_task(new_task) + assert plan.tasks[-1].task_id == "2" + assert plan.task_map["2"] == new_task + + def test_append_task_with_invalid_dependencies(self): + new_task = Task(task_id="2", dependent_task_ids=["3"]) + plan = Plan(goal="Test") + with pytest.raises(AssertionError): + plan.append_task(new_task) + + def test_append_task_without_dependencies(self): + plan = Plan(goal="Test") + existing_task = [Task(task_id="1")] + plan.add_tasks(existing_task) + + new_task = Task(task_id="2") + plan.append_task(new_task) + + assert len(plan.tasks) == 2 + assert plan.current_task_id == "1" + + def test_append_task_updates_current_task(self): + finished_task = Task(task_id="1", is_finished=True) + new_task = Task(task_id="2") + plan = Plan(goal="Test", tasks=[finished_task]) + plan.append_task(new_task) + assert plan.current_task_id == "2" + + def test_update_current_task(self): + task1 = Task(task_id="1", is_finished=True) + task2 = Task(task_id="2") + plan = Plan(goal="Test", tasks=[task1, task2]) + plan._update_current_task() + assert plan.current_task_id == "2" + + if __name__ == "__main__": pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/test_startup.py b/tests/metagpt/test_software_company.py similarity index 86% rename from tests/metagpt/test_startup.py rename to tests/metagpt/test_software_company.py index 095a74e3b..1b6477260 100644 --- a/tests/metagpt/test_startup.py +++ b/tests/metagpt/test_software_company.py @@ -3,13 +3,13 @@ """ @Time : 2023/5/15 11:40 @Author : alexanderwu -@File : test_startup.py +@File : test_software_company.py """ import pytest from typer.testing import CliRunner from metagpt.logs import logger -from metagpt.startup import app +from metagpt.software_company import app from metagpt.team import Team runner = CliRunner() @@ -23,7 +23,7 @@ async def test_empty_team(new_filename): logger.info(history) -def test_startup(new_filename): +def test_software_company(new_filename): args = ["Make a cli snake game"] result = runner.invoke(app, args) logger.info(result) diff --git a/tests/metagpt/tools/libs/__init__.py b/tests/metagpt/tools/libs/__init__.py new file mode 100644 index 000000000..0321f694a --- /dev/null +++ b/tests/metagpt/tools/libs/__init__.py @@ -0,0 +1,6 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# @Time : 2024/1/11 16:14 +# @Author : lidanyang +# @File : __init__.py +# @Desc : diff --git a/tests/metagpt/tools/libs/test_data_preprocess.py b/tests/metagpt/tools/libs/test_data_preprocess.py new file mode 100644 index 000000000..418f8adee --- /dev/null +++ b/tests/metagpt/tools/libs/test_data_preprocess.py @@ -0,0 +1,111 @@ +from datetime import datetime + +import numpy as np +import numpy.testing as npt +import pandas as pd +import pytest + +from metagpt.tools.libs.data_preprocess import ( + FillMissingValue, + LabelEncode, + MaxAbsScale, + MinMaxScale, + OneHotEncode, + OrdinalEncode, + RobustScale, + StandardScale, + get_column_info, +) + + +@pytest.fixture +def mock_datasets(): + return pd.DataFrame( + { + "num1": [1, 2, np.nan, 4, 5], + "cat1": ["A", "B", np.nan, "D", "A"], + "date1": [ + datetime(2020, 1, 1), + datetime(2020, 1, 2), + datetime(2020, 1, 3), + datetime(2020, 1, 4), + datetime(2020, 1, 5), + ], + } + ) + + +def test_fill_missing_value(mock_datasets): + fm = FillMissingValue(features=["num1"], strategy="mean") + transformed = fm.fit_transform(mock_datasets.copy()) + + assert transformed["num1"].isnull().sum() == 0 + + +def test_min_max_scale(mock_datasets): + mms = MinMaxScale(features=["num1"]) + transformed = mms.fit_transform(mock_datasets.copy()) + + npt.assert_allclose(transformed["num1"].min(), 0) + npt.assert_allclose(transformed["num1"].max(), 1) + + +def test_standard_scale(mock_datasets): + ss = StandardScale(features=["num1"]) + transformed = ss.fit_transform(mock_datasets.copy()) + + assert int(transformed["num1"].mean()) == 0 + assert int(transformed["num1"].std()) == 1 + + +def test_max_abs_scale(mock_datasets): + mas = MaxAbsScale(features=["num1"]) + transformed = mas.fit_transform(mock_datasets.copy()) + + npt.assert_allclose(transformed["num1"].abs().max(), 1) + + +def test_robust_scale(mock_datasets): + rs = RobustScale(features=["num1"]) + transformed = rs.fit_transform(mock_datasets.copy()) + + assert int(transformed["num1"].median()) == 0 + + +def test_ordinal_encode(mock_datasets): + oe = OrdinalEncode(features=["cat1"]) + transformed = oe.fit_transform(mock_datasets.copy()) + + assert transformed["cat1"].max() == 2 + + +def test_one_hot_encode(mock_datasets): + ohe = OneHotEncode(features=["cat1"]) + transformed = ohe.fit_transform(mock_datasets.copy()) + + assert transformed["cat1_A"].max() == 1 + + +def test_label_encode(mock_datasets): + le = LabelEncode(features=["cat1"]) + transformed = le.fit_transform(mock_datasets.copy()) + + assert transformed["cat1"].max() == 3 + + # test transform with unseen data + test = mock_datasets.copy() + test["cat1"] = ["A", "B", "C", "D", "E"] + transformed = le.transform(test) + assert transformed["cat1"].max() == 4 + + +def test_get_column_info(mock_datasets): + df = mock_datasets + column_info = get_column_info(df) + + assert column_info == { + "Category": ["cat1"], + "Numeric": ["num1"], + "Datetime": ["date1"], + "Others": [], + } diff --git a/tests/metagpt/tools/libs/test_feature_engineering.py b/tests/metagpt/tools/libs/test_feature_engineering.py new file mode 100644 index 000000000..3cfd5dacd --- /dev/null +++ b/tests/metagpt/tools/libs/test_feature_engineering.py @@ -0,0 +1,175 @@ +import numpy as np +import pandas as pd +import pytest +from sklearn.datasets import fetch_california_housing, load_breast_cancer, load_iris + +from metagpt.tools.libs.feature_engineering import ( + CatCount, + CatCross, + ExtractTimeComps, + GeneralSelection, + GroupStat, + KFoldTargetMeanEncoder, + PolynomialExpansion, + SplitBins, + TargetMeanEncoder, + TreeBasedSelection, + VarianceBasedSelection, +) + + +@pytest.fixture +def mock_dataset(): + return pd.DataFrame( + { + "num1": [1, 2, np.nan, 4, 5, 6, 7, 3], + "num2": [1, 3, 2, 1, np.nan, 5, 6, 4], + "num3": [np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan], + "cat1": ["A", "B", np.nan, "D", "E", "C", "B", "A"], + "cat2": ["A", "A", "A", "A", "A", "A", "A", "A"], + "date1": [ + "2020-01-01", + "2020-01-02", + "2020-01-03", + "2020-01-04", + "2020-01-05", + "2020-01-06", + "2020-01-07", + "2020-01-08", + ], + "label": [0, 1, 0, 1, 0, 1, 0, 1], + } + ) + + +def load_sklearn_data(data_name): + if data_name == "iris": + data = load_iris() + elif data_name == "breast_cancer": + data = load_breast_cancer() + elif data_name == "housing": + data = fetch_california_housing() + else: + raise ValueError("data_name not supported") + + X, y, feature_names = data.data, data.target, data.feature_names + data = pd.DataFrame(X, columns=feature_names) + data["label"] = y + return data + + +def test_polynomial_expansion(mock_dataset): + pe = PolynomialExpansion(cols=["num1", "num2", "label"], degree=2, label_col="label") + transformed = pe.fit_transform(mock_dataset) + + assert len(transformed.columns) == len(mock_dataset.columns) + 3 + + # when too many columns + data = load_sklearn_data("breast_cancer") + cols = [c for c in data.columns if c != "label"] + pe = PolynomialExpansion(cols=cols, degree=2, label_col="label") + transformed = pe.fit_transform(data) + + assert len(transformed.columns) == len(data.columns) + 55 + + +def test_cat_count(mock_dataset): + cc = CatCount(col="cat1") + transformed = cc.fit_transform(mock_dataset) + + assert "cat1_cnt" in transformed.columns + assert transformed["cat1_cnt"][0] == 2 + + +def test_target_mean_encoder(mock_dataset): + tme = TargetMeanEncoder(col="cat1", label="label") + transformed = tme.fit_transform(mock_dataset) + + assert "cat1_target_mean" in transformed.columns + assert transformed["cat1_target_mean"][0] == 0.5 + + +def test_kfold_target_mean_encoder(mock_dataset): + kfme = KFoldTargetMeanEncoder(col="cat1", label="label") + transformed = kfme.fit_transform(mock_dataset) + + assert "cat1_kf_target_mean" in transformed.columns + + +def test_cat_cross(mock_dataset): + cc = CatCross(cols=["cat1", "cat2"]) + transformed = cc.fit_transform(mock_dataset) + + assert "cat1_cat2" in transformed.columns + + cc = CatCross(cols=["cat1", "cat2"], max_cat_num=3) + transformed = cc.fit_transform(mock_dataset) + + assert "cat1_cat2" not in transformed.columns + + +def test_group_stat(mock_dataset): + gs = GroupStat(group_col="cat1", agg_col="num1", agg_funcs=["mean", "sum"]) + transformed = gs.fit_transform(mock_dataset) + + assert "num1_mean_by_cat1" in transformed.columns + assert "num1_sum_by_cat1" in transformed.columns + + +def test_split_bins(mock_dataset): + sb = SplitBins(cols=["num1"]) + transformed = sb.fit_transform(mock_dataset) + + assert transformed["num1"].nunique() <= 5 + assert all(0 <= x < 5 for x in transformed["num1"]) + + +def test_extract_time_comps(mock_dataset): + time_comps = ["year", "month", "day", "hour", "dayofweek", "is_weekend"] + etc = ExtractTimeComps(time_col="date1", time_comps=time_comps) + transformed = etc.fit_transform(mock_dataset.copy()) + + for comp in time_comps: + assert comp in transformed.columns + assert transformed["year"][0] == 2020 + assert transformed["month"][0] == 1 + assert transformed["day"][0] == 1 + assert transformed["hour"][0] == 0 + assert transformed["dayofweek"][0] == 3 + assert transformed["is_weekend"][0] == 0 + + +def test_general_selection(mock_dataset): + gs = GeneralSelection(label_col="label") + transformed = gs.fit_transform(mock_dataset.copy()) + + assert "num3" not in transformed.columns + assert "cat2" not in transformed.columns + + +@pytest.mark.skip # skip because TreeBasedSelection needs lgb as dependency +def test_tree_based_selection(mock_dataset): + # regression + data = load_sklearn_data("housing") + tbs = TreeBasedSelection(label_col="label", task_type="reg") + transformed = tbs.fit_transform(data) + assert len(transformed.columns) > 1 + + # classification + data = load_sklearn_data("breast_cancer") + tbs = TreeBasedSelection(label_col="label", task_type="cls") + transformed = tbs.fit_transform(data) + assert len(transformed.columns) > 1 + + # multi-classification + data = load_sklearn_data("iris") + tbs = TreeBasedSelection(label_col="label", task_type="mcls") + transformed = tbs.fit_transform(data) + assert len(transformed.columns) > 1 + + +def test_variance_based_selection(mock_dataset): + vbs = VarianceBasedSelection(label_col="label") + transformed = vbs.fit_transform(mock_dataset.copy()) + + assert "num3" not in transformed.columns diff --git a/tests/metagpt/tools/libs/test_gpt_v_generator.py b/tests/metagpt/tools/libs/test_gpt_v_generator.py new file mode 100644 index 000000000..907006765 --- /dev/null +++ b/tests/metagpt/tools/libs/test_gpt_v_generator.py @@ -0,0 +1,85 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- +""" +@Time : 2024/01/15 +@Author : mannaandpoem +@File : test_gpt_v_generator.py +""" +from pathlib import Path + +import pytest + +from metagpt import logs +from metagpt.const import METAGPT_ROOT +from metagpt.tools.libs.gpt_v_generator import GPTvGenerator + + +@pytest.fixture +def mock_webpage_filename_with_styles_and_scripts(mocker): + mock_data = """```html\n\n +\n\n```\n +```css\n/* styles.css */\n```\n +```javascript\n// scripts.js\n```\n""" + mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", return_value=mock_data) + return mocker + + +@pytest.fixture +def mock_webpage_filename_with_style_and_script(mocker): + mock_data = """```html\n\n +\n\n```\n +```css\n/* style.css */\n```\n +```javascript\n// script.js\n```\n""" + mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", return_value=mock_data) + return mocker + + +@pytest.fixture +def mock_image_layout(mocker): + image_layout = "The layout information of the sketch image is ..." + mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", return_value=image_layout) + return mocker + + +@pytest.fixture +def image_path(): + return f"{METAGPT_ROOT}/docs/resources/workspace/content_rec_sys/resources/competitive_analysis.png" + + +@pytest.mark.asyncio +async def test_generate_webpages(mock_webpage_filename_with_styles_and_scripts, image_path): + generator = GPTvGenerator() + rsp = await generator.generate_webpages(image_path=image_path) + logs.logger.info(rsp) + assert "html" in rsp + assert "css" in rsp + assert "javascript" in rsp + + +@pytest.mark.asyncio +async def test_save_webpages_with_styles_and_scripts(mock_webpage_filename_with_styles_and_scripts, image_path): + generator = GPTvGenerator() + webpages = await generator.generate_webpages(image_path) + webpages_dir = generator.save_webpages(image_path=image_path, webpages=webpages) + logs.logger.info(webpages_dir) + assert webpages_dir.exists() + + +@pytest.mark.asyncio +async def test_save_webpages_with_style_and_script(mock_webpage_filename_with_style_and_script, image_path): + generator = GPTvGenerator() + webpages = await generator.generate_webpages(image_path) + webpages_dir = generator.save_webpages(image_path=image_path, webpages=webpages) + logs.logger.info(webpages_dir) + assert webpages_dir.exists() + + +@pytest.mark.asyncio +async def test_analyze_layout(mock_image_layout, image_path): + layout = await GPTvGenerator().analyze_layout(Path(image_path)) + logs.logger.info(layout) + assert layout + + +if __name__ == "__main__": + pytest.main([__file__, "-s"]) diff --git a/tests/metagpt/tools/libs/test_sd_engine.py b/tests/metagpt/tools/libs/test_sd_engine.py new file mode 100644 index 000000000..e2c46e72a --- /dev/null +++ b/tests/metagpt/tools/libs/test_sd_engine.py @@ -0,0 +1,61 @@ +# -*- coding: utf-8 -*- +# @Date : 1/10/2024 10:07 PM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : +import base64 +import io +import json + +import pytest +from PIL import Image, ImageDraw + +from metagpt.tools.libs.sd_engine import SDEngine + + +def generate_mock_image_data(): + # 创建一个简单的图片对象 + image = Image.new("RGB", (100, 100), color="white") + draw = ImageDraw.Draw(image) + draw.text((10, 10), "Mock Image", fill="black") + + # 将图片转换为二进制数据 + with io.BytesIO() as buffer: + image.save(buffer, format="PNG") + image_binary = buffer.getvalue() + + # 对图片二进制数据进行 base64 编码 + image_base64 = base64.b64encode(image_binary).decode("utf-8") + + return image_base64 + + +def test_sd_tools(mocker): + mock_response = mocker.MagicMock() + mock_response.json.return_value = {"images": [generate_mock_image_data()]} + mocker.patch("requests.Session.post", return_value=mock_response) + + engine = SDEngine(sd_url="http://example_localhost:7860") + prompt = "1boy, hansom" + engine.construct_payload(prompt) + engine.simple_run_t2i(engine.payload) + + +def test_sd_construct_payload(): + engine = SDEngine(sd_url="http://example_localhost:7860") + prompt = "1boy, hansom" + engine.construct_payload(prompt) + assert "negative_prompt" in engine.payload + + +@pytest.mark.asyncio +async def test_sd_asyn_t2i(mocker): + mock_post = mocker.patch("aiohttp.ClientSession.post") + mock_response = mocker.AsyncMock() + mock_response.read.return_value = json.dumps({"images": [generate_mock_image_data()]}) + mock_post.return_value.__aenter__.return_value = mock_response + + engine = SDEngine(sd_url="http://example_localhost:7860") + prompt = "1boy, hansom" + engine.construct_payload(prompt) + await engine.run_t2i([engine.payload]) + assert "negative_prompt" in engine.payload diff --git a/tests/metagpt/tools/libs/test_web_scraping.py b/tests/metagpt/tools/libs/test_web_scraping.py new file mode 100644 index 000000000..c11960e68 --- /dev/null +++ b/tests/metagpt/tools/libs/test_web_scraping.py @@ -0,0 +1,23 @@ +import pytest + +from metagpt.tools.libs.web_scraping import scrape_web_playwright + + +@pytest.mark.asyncio +async def test_scrape_web_playwright(): + test_url = "https://www.deepwisdom.ai" + + result = await scrape_web_playwright(test_url) + + # Assert that the result is a dictionary + assert isinstance(result, dict) + + # Assert that the result contains 'inner_text' and 'html' keys + assert "inner_text" in result + assert "html" in result + + # Assert startswith and endswith + assert not result["inner_text"].startswith(" ") + assert not result["inner_text"].endswith(" ") + assert not result["html"].startswith(" ") + assert not result["html"].endswith(" ") diff --git a/tests/metagpt/tools/test_azure_tts.py b/tests/metagpt/tools/test_azure_tts.py index 74d23e439..f72b5663b 100644 --- a/tests/metagpt/tools/test_azure_tts.py +++ b/tests/metagpt/tools/test_azure_tts.py @@ -28,10 +28,10 @@ async def test_azure_tts(mocker): mocker.patch.object(Path, "exists", return_value=True) # Prerequisites - assert config.AZURE_TTS_SUBSCRIPTION_KEY and config.AZURE_TTS_SUBSCRIPTION_KEY != "YOUR_API_KEY" - assert config.AZURE_TTS_REGION + assert config.azure_tts_subscription_key and config.azure_tts_subscription_key != "YOUR_API_KEY" + assert config.azure_tts_region - azure_tts = AzureTTS(subscription_key=config.AZURE_TTS_SUBSCRIPTION_KEY, region=config.AZURE_TTS_REGION) + azure_tts = AzureTTS(subscription_key=config.azure_tts_subscription_key, region=config.azure_tts_region) text = """ 女儿看见父亲走了进来,问道: diff --git a/tests/metagpt/tools/test_iflytek_tts.py b/tests/metagpt/tools/test_iflytek_tts.py index 8e4c0cf54..c51f62b8e 100644 --- a/tests/metagpt/tools/test_iflytek_tts.py +++ b/tests/metagpt/tools/test_iflytek_tts.py @@ -15,8 +15,8 @@ from metagpt.tools.iflytek_tts import IFlyTekTTS, oas3_iflytek_tts async def test_iflytek_tts(mocker): # mock config = Config.default() - config.AZURE_TTS_SUBSCRIPTION_KEY = None - config.AZURE_TTS_REGION = None + config.azure_tts_subscription_key = None + config.azure_tts_region = None mocker.patch.object(IFlyTekTTS, "synthesize_speech", return_value=None) mock_data = mocker.AsyncMock() mock_data.read.return_value = b"mock iflytek" @@ -24,15 +24,15 @@ async def test_iflytek_tts(mocker): mock_reader.return_value.__aenter__.return_value = mock_data # Prerequisites - assert config.IFLYTEK_APP_ID - assert config.IFLYTEK_API_KEY - assert config.IFLYTEK_API_SECRET + assert config.iflytek_app_id + assert config.iflytek_api_key + assert config.iflytek_api_secret result = await oas3_iflytek_tts( text="你好,hello", - app_id=config.IFLYTEK_APP_ID, - api_key=config.IFLYTEK_API_KEY, - api_secret=config.IFLYTEK_API_SECRET, + app_id=config.iflytek_app_id, + api_key=config.iflytek_api_key, + api_secret=config.iflytek_api_secret, ) assert result diff --git a/tests/metagpt/tools/test_metagpt_oas3_api_svc.py b/tests/metagpt/tools/test_metagpt_oas3_api_svc.py index 3cf5e515b..5be139106 100644 --- a/tests/metagpt/tools/test_metagpt_oas3_api_svc.py +++ b/tests/metagpt/tools/test_metagpt_oas3_api_svc.py @@ -12,14 +12,12 @@ from pathlib import Path import pytest import requests -from metagpt.context import CONTEXT - @pytest.mark.asyncio -async def test_oas2_svc(): +async def test_oas2_svc(context): workdir = Path(__file__).parent.parent.parent.parent script_pathname = workdir / "metagpt/tools/metagpt_oas3_api_svc.py" - env = CONTEXT.new_environ() + env = context.new_environ() env["PYTHONPATH"] = str(workdir) + ":" + env.get("PYTHONPATH", "") process = subprocess.Popen(["python", str(script_pathname)], cwd=str(workdir), env=env) await asyncio.sleep(5) diff --git a/tests/metagpt/tools/test_metagpt_text_to_image.py b/tests/metagpt/tools/test_metagpt_text_to_image.py index 0dcad20d2..d3797a460 100644 --- a/tests/metagpt/tools/test_metagpt_text_to_image.py +++ b/tests/metagpt/tools/test_metagpt_text_to_image.py @@ -24,7 +24,7 @@ async def test_draw(mocker): mock_post.return_value.__aenter__.return_value = mock_response # Prerequisites - assert config.METAGPT_TEXT_TO_IMAGE_MODEL_URL + assert config.metagpt_tti_url binary_data = await oas3_metagpt_text_to_image("Panda emoji") assert binary_data diff --git a/tests/metagpt/tools/test_openai_text_to_embedding.py b/tests/metagpt/tools/test_openai_text_to_embedding.py index 047206d48..81b3895c3 100644 --- a/tests/metagpt/tools/test_openai_text_to_embedding.py +++ b/tests/metagpt/tools/test_openai_text_to_embedding.py @@ -10,7 +10,7 @@ from pathlib import Path import pytest -from metagpt.config2 import config +from metagpt.config2 import Config from metagpt.tools.openai_text_to_embedding import oas3_openai_text_to_embedding from metagpt.utils.common import aread @@ -18,6 +18,7 @@ from metagpt.utils.common import aread @pytest.mark.asyncio async def test_embedding(mocker): # mock + config = Config.default() mock_post = mocker.patch("aiohttp.ClientSession.post") mock_response = mocker.AsyncMock() mock_response.status = 200 diff --git a/tests/metagpt/tools/test_openapi_v3_hello.py b/tests/metagpt/tools/test_openapi_v3_hello.py index daa5d21c6..f49b8412a 100644 --- a/tests/metagpt/tools/test_openapi_v3_hello.py +++ b/tests/metagpt/tools/test_openapi_v3_hello.py @@ -12,14 +12,12 @@ from pathlib import Path import pytest import requests -from metagpt.context import CONTEXT - @pytest.mark.asyncio -async def test_hello(): +async def test_hello(context): workdir = Path(__file__).parent.parent.parent.parent script_pathname = workdir / "metagpt/tools/openapi_v3_hello.py" - env = CONTEXT.new_environ() + env = context.new_environ() env["PYTHONPATH"] = str(workdir) + ":" + env.get("PYTHONPATH", "") process = subprocess.Popen(["python", str(script_pathname)], cwd=workdir, env=env) await asyncio.sleep(5) diff --git a/tests/metagpt/tools/test_search_engine.py b/tests/metagpt/tools/test_search_engine.py index 966f53a38..a1f03ef7b 100644 --- a/tests/metagpt/tools/test_search_engine.py +++ b/tests/metagpt/tools/test_search_engine.py @@ -12,6 +12,7 @@ from typing import Callable import pytest from metagpt.config2 import config +from metagpt.configs.search_config import SearchConfig from metagpt.logs import logger from metagpt.tools import SearchEngineType from metagpt.tools.search_engine import SearchEngine @@ -49,27 +50,34 @@ async def test_search_engine( search_engine_mocker, ): # Prerequisites - search_engine_config = {} + search_engine_config = {"engine": search_engine_type, "run_func": run_func} if search_engine_type is SearchEngineType.SERPAPI_GOOGLE: assert config.search - search_engine_config["serpapi_api_key"] = "mock-serpapi-key" + search_engine_config["api_key"] = "mock-serpapi-key" elif search_engine_type is SearchEngineType.DIRECT_GOOGLE: assert config.search - search_engine_config["google_api_key"] = "mock-google-key" - search_engine_config["google_cse_id"] = "mock-google-cse" + search_engine_config["api_key"] = "mock-google-key" + search_engine_config["cse_id"] = "mock-google-cse" elif search_engine_type is SearchEngineType.SERPER_GOOGLE: assert config.search - search_engine_config["serper_api_key"] = "mock-serper-key" + search_engine_config["api_key"] = "mock-serper-key" - search_engine = SearchEngine(search_engine_type, run_func, **search_engine_config) - rsp = await search_engine.run("metagpt", max_results, as_string) - logger.info(rsp) - if as_string: - assert isinstance(rsp, str) - else: - assert isinstance(rsp, list) - assert len(rsp) <= max_results + async def test(search_engine): + rsp = await search_engine.run("metagpt", max_results, as_string) + logger.info(rsp) + if as_string: + assert isinstance(rsp, str) + else: + assert isinstance(rsp, list) + assert len(rsp) <= max_results + + await test(SearchEngine(**search_engine_config)) + search_engine_config["api_type"] = search_engine_config.pop("engine") + if run_func: + await test(SearchEngine.from_search_func(run_func)) + search_engine_config["search_func"] = search_engine_config.pop("run_func") + await test(SearchEngine.from_search_config(SearchConfig(**search_engine_config))) if __name__ == "__main__": diff --git a/tests/metagpt/tools/test_tool_convert.py b/tests/metagpt/tools/test_tool_convert.py new file mode 100644 index 000000000..8f26a211c --- /dev/null +++ b/tests/metagpt/tools/test_tool_convert.py @@ -0,0 +1,175 @@ +import pandas as pd + +from metagpt.tools.tool_convert import convert_code_to_tool_schema, docstring_to_schema + + +def test_docstring_to_schema(): + docstring = """ + Some test desc. + + Args: + features (list): Columns to be processed. + strategy (str, optional): The imputation strategy, notice 'mean' and 'median' can only be + used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'. + fill_value (int, optional): Fill_value is used to replace all occurrences of missing_values. + Defaults to None. + Returns: + pd.DataFrame: The transformed DataFrame. + """ + expected = { + "description": "Some test desc.", + "parameters": { + "properties": { + "features": {"type": "list", "description": "Columns to be processed."}, + "strategy": { + "type": "str", + "description": "The imputation strategy, notice 'mean' and 'median' can only be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'.", + "default": "'mean'", + "enum": ["'mean'", "'median'", "'most_frequent'", "'constant'"], + }, + "fill_value": { + "type": "int", + "description": "Fill_value is used to replace all occurrences of missing_values. Defaults to None.", + "default": "None", + }, + }, + "required": ["features"], + }, + "returns": [{"type": "pd.DataFrame", "description": "The transformed DataFrame."}], + } + schema = docstring_to_schema(docstring) + assert schema == expected + + +class DummyClass: + """ + Completing missing values with simple strategies. + """ + + def __init__(self, features: list, strategy: str = "mean", fill_value=None): + """ + Initialize self. + + Args: + features (list): Columns to be processed. + strategy (str, optional): The imputation strategy, notice 'mean' and 'median' can only + be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'. + fill_value (int, optional): Fill_value is used to replace all occurrences of missing_values. + Defaults to None. + """ + pass + + def fit(self, df: pd.DataFrame): + """ + Fit the FillMissingValue model. + + Args: + df (pd.DataFrame): The input DataFrame. + """ + pass + + def transform(self, df: pd.DataFrame) -> pd.DataFrame: + """ + Transform the input DataFrame with the fitted model. + + Args: + df (pd.DataFrame): The input DataFrame. + + Returns: + pd.DataFrame: The transformed DataFrame. + """ + pass + + +def dummy_fn(df: pd.DataFrame) -> dict: + """ + Analyzes a DataFrame and categorizes its columns based on data types. + + Args: + df (pd.DataFrame): The DataFrame to be analyzed. + + Returns: + dict: A dictionary with four keys ('Category', 'Numeric', 'Datetime', 'Others'). + Each key corresponds to a list of column names belonging to that category. + """ + pass + + +async def dummy_async_fn(df: pd.DataFrame) -> dict: + """ + A dummy async function for test + + Args: + df (pd.DataFrame): test args. + + Returns: + dict: test returns. + """ + pass + + +def test_convert_code_to_tool_schema_class(): + expected = { + "type": "class", + "description": "Completing missing values with simple strategies.", + "methods": { + "__init__": { + "type": "function", + "description": "Initialize self.", + "parameters": { + "properties": { + "features": {"type": "list", "description": "Columns to be processed."}, + "strategy": { + "type": "str", + "description": "The imputation strategy, notice 'mean' and 'median' can only be used for numeric features. Enum: ['mean', 'median', 'most_frequent', 'constant']. Defaults to 'mean'.", + "default": "'mean'", + "enum": ["'mean'", "'median'", "'most_frequent'", "'constant'"], + }, + "fill_value": { + "type": "int", + "description": "Fill_value is used to replace all occurrences of missing_values. Defaults to None.", + "default": "None", + }, + }, + "required": ["features"], + }, + }, + "fit": { + "type": "function", + "description": "Fit the FillMissingValue model.", + "parameters": { + "properties": {"df": {"type": "pd.DataFrame", "description": "The input DataFrame."}}, + "required": ["df"], + }, + }, + "transform": { + "type": "function", + "description": "Transform the input DataFrame with the fitted model.", + "parameters": { + "properties": {"df": {"type": "pd.DataFrame", "description": "The input DataFrame."}}, + "required": ["df"], + }, + "returns": [{"type": "pd.DataFrame", "description": "The transformed DataFrame."}], + }, + }, + } + schema = convert_code_to_tool_schema(DummyClass) + assert schema == expected + + +def test_convert_code_to_tool_schema_function(): + expected = { + "type": "function", + "description": "Analyzes a DataFrame and categorizes its columns based on data types.", + "parameters": { + "properties": {"df": {"type": "pd.DataFrame", "description": "The DataFrame to be analyzed."}}, + "required": ["df"], + }, + } + schema = convert_code_to_tool_schema(dummy_fn) + assert schema == expected + + +def test_convert_code_to_tool_schema_async_function(): + schema = convert_code_to_tool_schema(dummy_async_fn) + assert schema.get("type") == "async_function" diff --git a/tests/metagpt/tools/test_tool_registry.py b/tests/metagpt/tools/test_tool_registry.py new file mode 100644 index 000000000..2fd487fb7 --- /dev/null +++ b/tests/metagpt/tools/test_tool_registry.py @@ -0,0 +1,102 @@ +import pytest + +from metagpt.tools.tool_registry import ToolRegistry +from metagpt.tools.tool_type import ToolType + + +@pytest.fixture +def tool_registry(): + return ToolRegistry() + + +@pytest.fixture +def tool_registry_full(): + return ToolRegistry(tool_types=ToolType) + + +# Test Initialization +def test_initialization(tool_registry): + assert isinstance(tool_registry, ToolRegistry) + assert tool_registry.tools == {} + assert tool_registry.tool_types == {} + assert tool_registry.tools_by_types == {} + + +# Test Initialization with tool types +def test_initialize_with_tool_types(tool_registry_full): + assert isinstance(tool_registry_full, ToolRegistry) + assert tool_registry_full.tools == {} + assert tool_registry_full.tools_by_types == {} + assert "data_preprocess" in tool_registry_full.tool_types + + +class TestClassTool: + """test class""" + + def test_class_fn(self): + """test class fn""" + pass + + +def test_fn(): + """test function""" + pass + + +# Test Tool Registration Class +def test_register_tool_class(tool_registry): + tool_registry.register_tool("TestClassTool", "/path/to/tool", tool_source_object=TestClassTool) + assert "TestClassTool" in tool_registry.tools + + +# Test Tool Registration Function +def test_register_tool_fn(tool_registry): + tool_registry.register_tool("test_fn", "/path/to/tool", tool_source_object=test_fn) + assert "test_fn" in tool_registry.tools + + +# Test Tool Existence Checks +def test_has_tool(tool_registry): + tool_registry.register_tool("TestClassTool", "/path/to/tool", tool_source_object=TestClassTool) + assert tool_registry.has_tool("TestClassTool") + assert not tool_registry.has_tool("NonexistentTool") + + +# Test Tool Retrieval +def test_get_tool(tool_registry): + tool_registry.register_tool("TestClassTool", "/path/to/tool", tool_source_object=TestClassTool) + tool = tool_registry.get_tool("TestClassTool") + assert tool is not None + assert tool.name == "TestClassTool" + assert tool.path == "/path/to/tool" + assert "description" in tool.schemas + + +# Similar tests for has_tool_type, get_tool_type, get_tools_by_type +def test_has_tool_type(tool_registry_full): + assert tool_registry_full.has_tool_type("data_preprocess") + assert not tool_registry_full.has_tool_type("NonexistentType") + + +def test_get_tool_type(tool_registry_full): + retrieved_type = tool_registry_full.get_tool_type("data_preprocess") + assert retrieved_type is not None + assert retrieved_type.name == "data_preprocess" + + +def test_get_tools_by_type(tool_registry): + tool_type_name = "TestType" + tool_name = "TestTool" + tool_path = "/path/to/tool" + + tool_registry.register_tool(tool_name, tool_path, tool_type=tool_type_name, tool_source_object=TestClassTool) + + tools_by_type = tool_registry.get_tools_by_type(tool_type_name) + assert tools_by_type is not None + assert tool_name in tools_by_type + + +# Test case for when the tool type does not exist +def test_get_tools_by_nonexistent_type(tool_registry): + tools_by_type = tool_registry.get_tools_by_type("NonexistentType") + assert not tools_by_type diff --git a/tests/metagpt/tools/test_ut_writer.py b/tests/metagpt/tools/test_ut_writer.py index 29b6572c2..3cc7e86bb 100644 --- a/tests/metagpt/tools/test_ut_writer.py +++ b/tests/metagpt/tools/test_ut_writer.py @@ -8,6 +8,17 @@ from pathlib import Path import pytest +from openai.resources.chat.completions import AsyncCompletions +from openai.types import CompletionUsage +from openai.types.chat.chat_completion import ( + ChatCompletion, + ChatCompletionMessage, + Choice, +) +from openai.types.chat.chat_completion_message_tool_call import ( + ChatCompletionMessageToolCall, + Function, +) from metagpt.config2 import config from metagpt.const import API_QUESTIONS_PATH, UT_PY_PATH @@ -16,7 +27,43 @@ from metagpt.tools.ut_writer import YFT_PROMPT_PREFIX, UTGenerator class TestUTWriter: @pytest.mark.asyncio - async def test_api_to_ut_sample(self): + async def test_api_to_ut_sample(self, mocker): + async def mock_create(*args, **kwargs): + return ChatCompletion( + id="chatcmpl-8n5fAd21w2J1IIFkI4qxWlNfM7QRC", + choices=[ + Choice( + finish_reason="stop", + index=0, + logprobs=None, + message=ChatCompletionMessage( + content=None, + role="assistant", + function_call=None, + tool_calls=[ + ChatCompletionMessageToolCall( + id="call_EjjmIY7GMspHu3r9mx8gPA2k", + function=Function( + arguments='{"code":"import string\\nimport random\\n\\ndef random_string' + "(length=10):\\n return ''.join(random.choice(string.ascii_" + 'lowercase) for i in range(length))"}', + name="execute", + ), + type="function", + ) + ], + ), + ) + ], + created=1706710532, + model="gpt-3.5-turbo-1106", + object="chat.completion", + system_fingerprint="fp_04f9a1eebf", + usage=CompletionUsage(completion_tokens=35, prompt_tokens=1982, total_tokens=2017), + ) + + mocker.patch.object(AsyncCompletions, "create", mock_create) + # Prerequisites swagger_file = Path(__file__).parent / "../../data/ut_writer/yft_swaggerApi.json" assert swagger_file.exists() diff --git a/tests/metagpt/tools/test_web_browser_engine_playwright.py b/tests/metagpt/tools/test_web_browser_engine_playwright.py index 0e838a2f8..f35848cf4 100644 --- a/tests/metagpt/tools/test_web_browser_engine_playwright.py +++ b/tests/metagpt/tools/test_web_browser_engine_playwright.py @@ -3,7 +3,6 @@ import pytest -from metagpt.config2 import config from metagpt.tools import web_browser_engine_playwright from metagpt.utils.parse_html import WebPage @@ -19,26 +18,22 @@ from metagpt.utils.parse_html import WebPage ids=["chromium-normal", "firefox-normal", "webkit-normal"], ) async def test_scrape_web_page(browser_type, use_proxy, kwagrs, url, urls, proxy, capfd): - global_proxy = config.proxy - try: - if use_proxy: - server, proxy_url = await proxy() - config.proxy = proxy_url - browser = web_browser_engine_playwright.PlaywrightWrapper(browser_type=browser_type, **kwagrs) - result = await browser.run(url) - assert isinstance(result, WebPage) - assert "MetaGPT" in result.inner_text + proxy_url = None + if use_proxy: + server, proxy_url = await proxy() + browser = web_browser_engine_playwright.PlaywrightWrapper(browser_type=browser_type, proxy=proxy_url, **kwagrs) + result = await browser.run(url) + assert isinstance(result, WebPage) + assert "MetaGPT" in result.inner_text - if urls: - results = await browser.run(url, *urls) - assert isinstance(results, list) - assert len(results) == len(urls) + 1 - assert all(("MetaGPT" in i.inner_text) for i in results) - if use_proxy: - server.close() - assert "Proxy:" in capfd.readouterr().out - finally: - config.proxy = global_proxy + if urls: + results = await browser.run(url, *urls) + assert isinstance(results, list) + assert len(results) == len(urls) + 1 + assert all(("MetaGPT" in i.inner_text) for i in results) + if use_proxy: + server.close() + assert "Proxy:" in capfd.readouterr().out if __name__ == "__main__": diff --git a/tests/metagpt/tools/test_web_browser_engine_selenium.py b/tests/metagpt/tools/test_web_browser_engine_selenium.py index e38905b85..a88a5d0f4 100644 --- a/tests/metagpt/tools/test_web_browser_engine_selenium.py +++ b/tests/metagpt/tools/test_web_browser_engine_selenium.py @@ -1,9 +1,9 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- +import browsers import pytest -from metagpt.config2 import config from metagpt.tools import web_browser_engine_selenium from metagpt.utils.parse_html import WebPage @@ -12,36 +12,49 @@ from metagpt.utils.parse_html import WebPage @pytest.mark.parametrize( "browser_type, use_proxy, url, urls", [ - ("chrome", True, "https://deepwisdom.ai", ("https://deepwisdom.ai",)), - ("firefox", False, "https://deepwisdom.ai", ("https://deepwisdom.ai",)), - ("edge", False, "https://deepwisdom.ai", ("https://deepwisdom.ai",)), + pytest.param( + "chrome", + True, + "https://deepwisdom.ai", + ("https://deepwisdom.ai",), + marks=pytest.mark.skipif(not browsers.get("chrome"), reason="chrome browser not found"), + ), + pytest.param( + "firefox", + False, + "https://deepwisdom.ai", + ("https://deepwisdom.ai",), + marks=pytest.mark.skipif(not browsers.get("firefox"), reason="firefox browser not found"), + ), + pytest.param( + "edge", + False, + "https://deepwisdom.ai", + ("https://deepwisdom.ai",), + marks=pytest.mark.skipif(not browsers.get("msedge"), reason="edge browser not found"), + ), ], ids=["chrome-normal", "firefox-normal", "edge-normal"], ) async def test_scrape_web_page(browser_type, use_proxy, url, urls, proxy, capfd): # Prerequisites # firefox, chrome, Microsoft Edge + proxy_url = None + if use_proxy: + server, proxy_url = await proxy() + browser = web_browser_engine_selenium.SeleniumWrapper(browser_type=browser_type, proxy=proxy_url) + result = await browser.run(url) + assert isinstance(result, WebPage) + assert "MetaGPT" in result.inner_text - global_proxy = config.proxy - try: - if use_proxy: - server, proxy_url = await proxy() - config.proxy = proxy_url - browser = web_browser_engine_selenium.SeleniumWrapper(browser_type=browser_type) - result = await browser.run(url) - assert isinstance(result, WebPage) - assert "MetaGPT" in result.inner_text - - if urls: - results = await browser.run(url, *urls) - assert isinstance(results, list) - assert len(results) == len(urls) + 1 - assert all(("MetaGPT" in i.inner_text) for i in results) - if use_proxy: - server.close() - assert "Proxy:" in capfd.readouterr().out - finally: - config.proxy = global_proxy + if urls: + results = await browser.run(url, *urls) + assert isinstance(results, list) + assert len(results) == len(urls) + 1 + assert all(("MetaGPT" in i.inner_text) for i in results) + if use_proxy: + server.close() + assert "Proxy:" in capfd.readouterr().out if __name__ == "__main__": diff --git a/tests/metagpt/utils/test_mermaid.py b/tests/metagpt/utils/test_mermaid.py index 89847994d..7367463dc 100644 --- a/tests/metagpt/utils/test_mermaid.py +++ b/tests/metagpt/utils/test_mermaid.py @@ -8,20 +8,19 @@ import pytest -from metagpt.context import CONTEXT from metagpt.utils.common import check_cmd_exists from metagpt.utils.mermaid import MMC1, mermaid_to_file @pytest.mark.asyncio @pytest.mark.parametrize("engine", ["nodejs", "ink"]) # TODO: playwright and pyppeteer -async def test_mermaid(engine, mermaid_mocker): +async def test_mermaid(engine, context, mermaid_mocker): # nodejs prerequisites: npm install -g @mermaid-js/mermaid-cli # ink prerequisites: connected to internet # playwright prerequisites: playwright install --with-deps chromium assert check_cmd_exists("npm") == 0 - save_to = CONTEXT.git_repo.workdir / f"{engine}/1" + save_to = context.git_repo.workdir / f"{engine}/1" await mermaid_to_file(engine, MMC1, save_to) # ink does not support pdf diff --git a/tests/metagpt/utils/test_redis.py b/tests/metagpt/utils/test_redis.py index 748c44f54..6fd4250a6 100644 --- a/tests/metagpt/utils/test_redis.py +++ b/tests/metagpt/utils/test_redis.py @@ -9,23 +9,25 @@ from unittest.mock import AsyncMock import pytest -from metagpt.config2 import Config from metagpt.utils.redis import Redis -async def async_mock_from_url(*args, **kwargs): - mock_client = AsyncMock() - mock_client.set.return_value = None - mock_client.get.side_effect = [b"test", b""] - return mock_client - - @pytest.mark.asyncio async def test_redis(mocker): - redis = Config.default().redis - mocker.patch("aioredis.from_url", return_value=async_mock_from_url()) + async def async_mock_from_url(*args, **kwargs): + mock_client = AsyncMock() + mock_client.set.return_value = None + mock_client.get.return_value = b"test" + return mock_client - conn = Redis(redis) + mocker.patch("aioredis.from_url", return_value=async_mock_from_url()) + mock_config = mocker.Mock() + mock_config.to_url.return_value = "http://mock.com" + mock_config.username = "mockusername" + mock_config.password = "mockpwd" + mock_config.db = "0" + + conn = Redis(mock_config) await conn.set("test", "test", timeout_sec=0) assert await conn.get("test") == b"test" await conn.close() diff --git a/tests/metagpt/utils/test_repair_llm_raw_output.py b/tests/metagpt/utils/test_repair_llm_raw_output.py index bd6169d71..e28423b91 100644 --- a/tests/metagpt/utils/test_repair_llm_raw_output.py +++ b/tests/metagpt/utils/test_repair_llm_raw_output.py @@ -128,6 +128,45 @@ def test_repair_json_format(): output = repair_llm_raw_output(output=raw_output, req_keys=[None], repair_type=RepairType.JSON) assert output == target_output + raw_output = """ +{ + "Language": "en_us", # define language + "Programming Language": "Python" +} +""" + target_output = """{ + "Language": "en_us", + "Programming Language": "Python" +}""" + output = repair_llm_raw_output(output=raw_output, req_keys=[None], repair_type=RepairType.JSON) + assert output == target_output + + raw_output = """ +{ + "Language": "en_us", // define language + "Programming Language": "Python" # define code language +} +""" + target_output = """{ + "Language": "en_us", + "Programming Language": "Python" +}""" + output = repair_llm_raw_output(output=raw_output, req_keys=[None], repair_type=RepairType.JSON) + assert output == target_output + + raw_output = """ + { + "Language": "#en_us#", // define language + "Programming Language": "//Python # Code // Language//" # define code language + } + """ + target_output = """{ + "Language": "#en_us#", + "Programming Language": "//Python # Code // Language//" + }""" + output = repair_llm_raw_output(output=raw_output, req_keys=[None], repair_type=RepairType.JSON) + assert output == target_output + def test_repair_invalid_json(): from metagpt.utils.repair_llm_raw_output import repair_invalid_json @@ -204,6 +243,25 @@ def test_retry_parse_json_text(): output = retry_parse_json_text(output=invalid_json_text) assert output == target_json + invalid_json_text = '''{ + "Data structures and interfaces": """ + class UI: + - game_engine: GameEngine + + __init__(engine: GameEngine) -> None + + display_board() -> None + + display_score() -> None + + prompt_move() -> str + + reset_game() -> None + """ + "Anything UNCLEAR": "no" +}''' + target_json = { + "Data structures and interfaces": "\n class UI:\n - game_engine: GameEngine\n + __init__(engine: GameEngine) -> None\n + display_board() -> None\n + display_score() -> None\n + prompt_move() -> str\n + reset_game() -> None\n ", + "Anything UNCLEAR": "no", + } + output = retry_parse_json_text(output=invalid_json_text) + assert output == target_json + def test_extract_content_from_output(): """ diff --git a/tests/metagpt/utils/test_s3.py b/tests/metagpt/utils/test_s3.py index 4dc3b1e42..b26ebe94d 100644 --- a/tests/metagpt/utils/test_s3.py +++ b/tests/metagpt/utils/test_s3.py @@ -8,8 +8,8 @@ import uuid from pathlib import Path +import aioboto3 import aiofiles -import mock import pytest from metagpt.config2 import Config @@ -18,21 +18,18 @@ from metagpt.utils.s3 import S3 @pytest.mark.asyncio -@mock.patch("aioboto3.Session") -async def test_s3(mock_session_class): +async def test_s3(mocker): # Set up the mock response data = await aread(__file__, "utf-8") - mock_session_object = mock.Mock() - reader_mock = mock.AsyncMock() + reader_mock = mocker.AsyncMock() reader_mock.read.side_effect = [data.encode("utf-8"), b"", data.encode("utf-8")] - type(reader_mock).url = mock.PropertyMock(return_value="https://mock") - mock_client = mock.AsyncMock() + type(reader_mock).url = mocker.PropertyMock(return_value="https://mock") + mock_client = mocker.AsyncMock() mock_client.put_object.return_value = None mock_client.get_object.return_value = {"Body": reader_mock} mock_client.__aenter__.return_value = mock_client mock_client.__aexit__.return_value = None - mock_session_object.client.return_value = mock_client - mock_session_class.return_value = mock_session_object + mocker.patch.object(aioboto3.Session, "client", return_value=mock_client) # Prerequisites s3 = Config.default().s3 @@ -55,7 +52,7 @@ async def test_s3(mock_session_class): # Mock session env s3.access_key = "ABC" - type(reader_mock).url = mock.PropertyMock(return_value="") + type(reader_mock).url = mocker.PropertyMock(return_value="") try: conn = S3(s3) res = await conn.cache("ABC", ".bak", "script") diff --git a/tests/metagpt/utils/test_save_code.py b/tests/metagpt/utils/test_save_code.py new file mode 100644 index 000000000..35ad84baf --- /dev/null +++ b/tests/metagpt/utils/test_save_code.py @@ -0,0 +1,44 @@ +# -*- coding: utf-8 -*- +# @Date : 12/12/2023 4:17 PM +# @Author : stellahong (stellahong@fuzhi.ai) +# @Desc : + +import nbformat +import pytest + +from metagpt.actions.ci.execute_nb_code import ExecuteNbCode +from metagpt.utils.common import read_json_file +from metagpt.utils.save_code import DATA_PATH, save_code_file + + +def test_save_code_file_python(): + save_code_file("example", "print('Hello, World!')") + file_path = DATA_PATH / "output" / "example" / "code.py" + assert file_path.exists(), f"File does not exist: {file_path}" + content = file_path.read_text() + assert "print('Hello, World!')" in content, "File content does not match" + + +def test_save_code_file_json(): + save_code_file("example_json", "print('Hello, JSON!')", file_format="json") + file_path = DATA_PATH / "output" / "example_json" / "code.json" + data = read_json_file(file_path) + assert "code" in data, "JSON key 'code' is missing" + assert data["code"] == "print('Hello, JSON!')", "JSON content does not match" + + +@pytest.mark.asyncio +async def test_save_code_file_notebook(): + code = "print('Hello, World!')" + executor = ExecuteNbCode() + await executor.run(code) + # Save as a Notebook file + save_code_file("example_nb", executor.nb, file_format="ipynb") + file_path = DATA_PATH / "output" / "example_nb" / "code.ipynb" + assert file_path.exists(), f"Notebook file does not exist: {file_path}" + + # Additional checks specific to notebook format + notebook = nbformat.read(file_path, as_version=4) + assert len(notebook.cells) > 0, "Notebook should have at least one cell" + first_cell_source = notebook.cells[0].source + assert "print" in first_cell_source, "Notebook cell content does not match" diff --git a/tests/metagpt/utils/test_text.py b/tests/metagpt/utils/test_text.py index 7003c7767..c9a9753be 100644 --- a/tests/metagpt/utils/test_text.py +++ b/tests/metagpt/utils/test_text.py @@ -42,6 +42,7 @@ def test_reduce_message_length(msgs, model_name, system_text, reserved, expected (" ".join("Hello World." for _ in range(1000)), "Prompt: {}", "gpt-3.5-turbo-16k", "System", 3000, 1), (" ".join("Hello World." for _ in range(4000)), "Prompt: {}", "gpt-4", "System", 2000, 2), (" ".join("Hello World." for _ in range(8000)), "Prompt: {}", "gpt-4-32k", "System", 4000, 1), + (" ".join("Hello World" for _ in range(8000)), "Prompt: {}", "gpt-3.5-turbo", "System", 1000, 8), ], ) def test_generate_prompt_chunk(text, prompt_template, model_name, system_text, reserved, expected): diff --git a/tests/mock/mock_aiohttp.py b/tests/mock/mock_aiohttp.py index 805c84b0c..ca98ed8b5 100644 --- a/tests/mock/mock_aiohttp.py +++ b/tests/mock/mock_aiohttp.py @@ -14,7 +14,8 @@ class MockAioResponse: def __init__(self, session, method, url, **kwargs) -> None: fn = self.check_funcs.get((method, url)) - self.key = f"{self.name}-{method}-{url}-{fn(kwargs) if fn else json.dumps(kwargs, sort_keys=True)}" + _kwargs = {k: v for k, v in kwargs.items() if k != "proxy"} + self.key = f"{self.name}-{method}-{url}-{fn(kwargs) if fn else json.dumps(_kwargs, sort_keys=True)}" self.mng = self.response = None if self.key not in self.rsp_cache: self.mng = origin_request(session, method, url, **kwargs) @@ -52,3 +53,7 @@ class MockAioResponse: data = await self.response.content.read() self.rsp_cache[self.key] = str(data) return data + + def raise_for_status(self): + if self.response: + self.response.raise_for_status() diff --git a/tests/mock/mock_llm.py b/tests/mock/mock_llm.py index bef380c83..50e75dabf 100644 --- a/tests/mock/mock_llm.py +++ b/tests/mock/mock_llm.py @@ -1,13 +1,22 @@ -from typing import Optional +import json +from typing import Optional, Union from metagpt.config2 import config +from metagpt.configs.llm_config import LLMType from metagpt.logs import log_llm_stream, logger +from metagpt.provider.azure_openai_api import AzureOpenAILLM from metagpt.provider.openai_api import OpenAILLM +from metagpt.schema import Message + +OriginalLLM = OpenAILLM if config.llm.api_type == LLMType.OPENAI else AzureOpenAILLM -class MockLLM(OpenAILLM): +class MockLLM(OriginalLLM): def __init__(self, allow_open_api_call): - super().__init__(config.get_openai_llm()) + original_llm_config = ( + config.get_openai_llm() if config.llm.api_type == LLMType.OPENAI else config.get_azure_llm() + ) + super().__init__(original_llm_config) self.allow_open_api_call = allow_open_api_call self.rsp_cache: dict = {} self.rsp_candidates: list[dict] = [] # a test can have multiple calls with the same llm, thus a list @@ -35,6 +44,7 @@ class MockLLM(OpenAILLM): msg: str, system_msgs: Optional[list[str]] = None, format_msgs: Optional[list[dict[str, str]]] = None, + images: Optional[Union[str, list[str]]] = None, timeout=3, stream=True, ): @@ -47,7 +57,7 @@ class MockLLM(OpenAILLM): message = [] if format_msgs: message.extend(format_msgs) - message.append(self._user_msg(msg)) + message.append(self._user_msg(msg, images=images)) rsp = await self.acompletion_text(message, stream=stream, timeout=timeout) return rsp @@ -61,11 +71,20 @@ class MockLLM(OpenAILLM): context.append(self._assistant_msg(rsp_text)) return self._extract_assistant_rsp(context) + async def original_aask_code(self, messages: Union[str, Message, list[dict]], **kwargs) -> dict: + """ + A copy of metagpt.provider.openai_api.OpenAILLM.aask_code, we can't use super().aask because it will be mocked. + Since openai_api.OpenAILLM.aask_code is different from base_llm.BaseLLM.aask_code, we use the former. + """ + rsp = await self._achat_completion_function(messages, **kwargs) + return self.get_choice_function_arguments(rsp) + async def aask( self, msg: str, system_msgs: Optional[list[str]] = None, format_msgs: Optional[list[dict[str, str]]] = None, + images: Optional[Union[str, list[str]]] = None, timeout=3, stream=True, ) -> str: @@ -73,7 +92,7 @@ class MockLLM(OpenAILLM): if system_msgs: joined_system_msg = "#MSG_SEP#".join(system_msgs) + "#SYSTEM_MSG_END#" msg_key = joined_system_msg + msg_key - rsp = await self._mock_rsp(msg_key, self.original_aask, msg, system_msgs, format_msgs, timeout, stream) + rsp = await self._mock_rsp(msg_key, self.original_aask, msg, system_msgs, format_msgs, images, timeout, stream) return rsp async def aask_batch(self, msgs: list, timeout=3) -> str: @@ -81,6 +100,12 @@ class MockLLM(OpenAILLM): rsp = await self._mock_rsp(msg_key, self.original_aask_batch, msgs, timeout) return rsp + async def aask_code(self, messages: Union[str, Message, list[dict]], **kwargs) -> dict: + messages = self._process_message(messages) + msg_key = json.dumps(messages, ensure_ascii=False) + rsp = await self._mock_rsp(msg_key, self.original_aask_code, messages, **kwargs) + return rsp + async def _mock_rsp(self, msg_key, ask_func, *args, **kwargs): if msg_key not in self.rsp_cache: if not self.allow_open_api_call: diff --git a/tests/spark.yaml b/tests/spark.yaml new file mode 100644 index 000000000..a5bbd98bd --- /dev/null +++ b/tests/spark.yaml @@ -0,0 +1,7 @@ +llm: + api_type: "spark" + app_id: "xxx" + api_key: "xxx" + api_secret: "xxx" + domain: "generalv2" + base_url: "wss://spark-api.xf-yun.com/v3.1/chat" \ No newline at end of file