From c0a7cc2102c39f6cbb439cae27fc801496781e67 Mon Sep 17 00:00:00 2001 From: foxplaying <166147707+foxplaying@users.noreply.github.com> Date: Fri, 17 Oct 2025 23:14:33 +0800 Subject: [PATCH] Update gemini_client.py --- src/llm_models/model_client/gemini_client.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/llm_models/model_client/gemini_client.py b/src/llm_models/model_client/gemini_client.py index 02889373..87ca8697 100644 --- a/src/llm_models/model_client/gemini_client.py +++ b/src/llm_models/model_client/gemini_client.py @@ -271,7 +271,7 @@ def _build_stream_api_resp( f" 可能会对回复内容造成影响,建议修改模型 {warn_target} 配置!" ) else: - logger.warning(f"⚠ Gemini 响应因达到 {warn_target} 限制被截断,\n请修改模型 {warn_target} 配置!") + logger.warning(f"⚠ Gemini 响应因达到 {warn_target} 限制被截断,\n 请修改模型 {warn_target} 配置!") if not resp.content and not resp.tool_calls: if not getattr(resp, "reasoning_content", None): @@ -422,7 +422,7 @@ def _default_normal_response_parser( f" 可能会对回复内容造成影响,建议修改模型 {warn_target} 配置!" ) else: - logger.warning(f"⚠ Gemini 响应因达到 {warn_target} 限制被截断,\n请修改模型 {warn_target} 配置!") + logger.warning(f"⚠ Gemini 响应因达到 {warn_target} 限制被截断,\n 请修改模型 {warn_target} 配置!") return api_response, _usage_record except Exception as e: