diff --git a/metagpt/ext/android_assistant/README.md b/metagpt/ext/android_assistant/README.md index 7dc310afe..610f7f895 100644 --- a/metagpt/ext/android_assistant/README.md +++ b/metagpt/ext/android_assistant/README.md @@ -28,7 +28,9 @@ #### Learning Based on Human Demonstration ``` After running this command, you will first see a screenshot of an Android screen that has been marked at various interactive locations, as shown in the figure below: -##### TODO Add Image + + + After remembering the location where you want to operate, a request similar to the one below will be output in the terminal. Reply to it and thereby direct the Android assistant to learn your demonstration action: ```bash diff --git a/metagpt/ext/android_assistant/README_CN.md b/metagpt/ext/android_assistant/README_CN.md index ae29574cb..46bc381d7 100644 --- a/metagpt/ext/android_assistant/README_CN.md +++ b/metagpt/ext/android_assistant/README_CN.md @@ -30,7 +30,9 @@ #### 基于人类演示的学习 ``` 在运行这一指令后,你将首先看到一个在各个可交互的位置进行了标记的安卓屏幕的截图,如下图: -###### TODO Add Image + + + 在记住你要操作的位置之后,终端中将会输出与下面类似的要求,回复它,进而指挥安卓助理学习你的演示行为: ```bash diff --git a/metagpt/ext/android_assistant/actions/manual_record.py b/metagpt/ext/android_assistant/actions/manual_record.py index fa9235da4..d0f7a6c13 100644 --- a/metagpt/ext/android_assistant/actions/manual_record.py +++ b/metagpt/ext/android_assistant/actions/manual_record.py @@ -73,6 +73,7 @@ class ManualRecord(Action): screenshot_labeled_path = Path(self.screenshot_after_path).joinpath(f"{step}_labeled.png") labeled_img = draw_bbox_multi(screenshot_path, screenshot_labeled_path, elem_list) + cv2.namedWindow('image', cv2.WINDOW_NORMAL) cv2.imshow("image", labeled_img) cv2.waitKey(0) cv2.destroyAllWindows()