From f776ffa05330d344e33eb46d902e2a9b0aae5c33 Mon Sep 17 00:00:00 2001 From: XBC_D2O Date: Thu, 13 Mar 2025 22:52:07 +0800 Subject: [PATCH] =?UTF-8?q?=E8=B0=83=E6=95=B4=E6=B8=A9=E5=BA=A6=E5=8F=82?= =?UTF-8?q?=E6=95=B0=E8=87=B30.6=EF=BC=8C=E5=B9=B6=E6=9B=B4=E6=96=B0?= =?UTF-8?q?=E4=B8=BB=E9=A2=98=E8=AF=86=E5=88=AB=E6=8F=90=E7=A4=BA=E4=BF=A1?= =?UTF-8?q?=E6=81=AF=E4=BB=A5=E5=A2=9E=E5=BC=BA=E7=94=A8=E6=88=B7=E4=BD=93?= =?UTF-8?q?=E9=AA=8C?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/llm_generator.py | 2 +- src/plugins/chat/topic_identifier.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/plugins/chat/llm_generator.py b/src/plugins/chat/llm_generator.py index 46dc34e9..0e7992f0 100644 --- a/src/plugins/chat/llm_generator.py +++ b/src/plugins/chat/llm_generator.py @@ -21,7 +21,7 @@ class ResponseGenerator: def __init__(self): self.model_r1 = LLM_request( model=global_config.llm_reasoning, - temperature=0.7, + temperature=0.6, max_tokens=1000, stream=True, ) diff --git a/src/plugins/chat/topic_identifier.py b/src/plugins/chat/topic_identifier.py index a0c5bae3..bf55430c 100644 --- a/src/plugins/chat/topic_identifier.py +++ b/src/plugins/chat/topic_identifier.py @@ -20,7 +20,7 @@ class TopicIdentifier: prompt = f"""判断这条消息的主题,如果没有明显主题请回复"无主题",要求: 1. 主题通常2-4个字,必须简短,要求精准概括,不要太具体。 2. 建议给出多个主题,之间用英文逗号分割。只输出主题本身就好,不要有前后缀。 - +请注意:如果没有明显主题请只回复"无主题" 消息内容:{text}""" # 使用 LLM_request 类进行请求