From 41d6d2be1d26518593c30ce8af9b9ddc865eeded Mon Sep 17 00:00:00 2001 From: better629 Date: Sun, 18 Feb 2024 20:02:42 +0800 Subject: [PATCH] update dashscope stream output --- metagpt/provider/dashscope_api.py | 2 ++ requirements.txt | 4 ++-- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/metagpt/provider/dashscope_api.py b/metagpt/provider/dashscope_api.py index 58031e452..c02a302e4 100644 --- a/metagpt/provider/dashscope_api.py +++ b/metagpt/provider/dashscope_api.py @@ -192,6 +192,8 @@ class DashScopeLLM(BaseLLM): if self.config.temperature > 0: # different model has default temperature. only set when it"s specified. kwargs["temperature"] = self.config.temperature + if stream: + kwargs["incremental_output"] = True return kwargs def _check_response(self, resp: GenerationResponse): diff --git a/requirements.txt b/requirements.txt index cfac35717..9759248bb 100644 --- a/requirements.txt +++ b/requirements.txt @@ -27,7 +27,7 @@ python_docx==0.8.11 PyYAML==6.0.1 # sentence_transformers==2.2.2 setuptools==65.6.3 -tenacity==8.2.2 +tenacity==8.2.3 tiktoken==0.5.2 tqdm==4.65.0 #unstructured[local-inference] @@ -68,4 +68,4 @@ anytree ipywidgets==8.1.1 Pillow qianfan==0.3.1 -dashscope==1.14.1 +dashscope==1.14.1 \ No newline at end of file