From d7b4e84cda27fbb60d0811b9747960a19ab7a13f Mon Sep 17 00:00:00 2001 From: Liu An Date: Tue, 12 Aug 2025 18:05:52 +0800 Subject: [PATCH] Refa: Update LLM stream response type to Generator (#9420) ### What problem does this PR solve? Change return type of _generate_streamly from str to Generator[str, None, None] to properly type hint streaming responses. ### Type of change - [x] Refactoring --- agent/component/llm.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/agent/component/llm.py b/agent/component/llm.py index 532bedc86..2d2e831bf 100644 --- a/agent/component/llm.py +++ b/agent/component/llm.py @@ -17,7 +17,7 @@ import json import logging import os import re -from typing import Any +from typing import Any, Generator import json_repair from copy import deepcopy @@ -154,7 +154,7 @@ class LLM(ComponentBase): return self.chat_mdl.chat(msg[0]["content"], msg[1:], self._param.gen_conf(), **kwargs) return self.chat_mdl.chat(msg[0]["content"], msg[1:], self._param.gen_conf(), images=self.imgs, **kwargs) - def _generate_streamly(self, msg:list[dict], **kwargs) -> str: + def _generate_streamly(self, msg:list[dict], **kwargs) -> Generator[str, None, None]: ans = "" last_idx = 0 endswith_think = False