Merge branch 'debug' of https://github.com/EightOrange/MaiMBot into debug

# Please enter a commit message to explain why this merge is necessary,
# especially if it merges an updated upstream into a topic branch.
#
# Lines starting with '#' will be ignored, and an empty message aborts
# the commit.
pull/220/head
八分橘子 2025-03-11 20:39:40 +08:00
commit c8a36ac08c
7 changed files with 93 additions and 74 deletions

10
bot.py
View File

@ -51,15 +51,15 @@ def init_env():
with open(".env", "w") as f: with open(".env", "w") as f:
f.write("ENVIRONMENT=prod") f.write("ENVIRONMENT=prod")
# 检测.env.prod文件是否存在 # 检测.env.prod文件是否存在
if not os.path.exists(".env.prod"): if not os.path.exists(".env.prod"):
logger.error("检测到.env.prod文件不存在") logger.error("检测到.env.prod文件不存在")
shutil.copy("template.env", "./.env.prod") shutil.copy("template.env", "./.env.prod")
# 检测.env.dev文件是否存在不存在的话直接复制生产环境配置 # 检测.env.dev文件是否存在不存在的话直接复制生产环境配置
if not os.path.exists(".env.dev"): if not os.path.exists(".env.dev"):
logger.error("检测到.env.dev文件不存在") logger.error("检测到.env.dev文件不存在")
shutil.copy(".env.prod", "./.env.dev") shutil.copy("template.env", "./.env.dev")
# 首先加载基础环境变量.env # 首先加载基础环境变量.env
if os.path.exists(".env"): if os.path.exists(".env"):

View File

@ -56,6 +56,7 @@ class ChatBot:
group_name = None group_name = None
# 白名单设定由nontbot侧完成 # 白名单设定由nontbot侧完成
# 消息过滤涉及到config有待更新
if event.group_id: if event.group_id:
if event.group_id not in global_config.talk_allowed_groups: if event.group_id not in global_config.talk_allowed_groups:
return return

View File

@ -1,4 +1,5 @@
import os import os
import sys
from dataclasses import dataclass, field from dataclasses import dataclass, field
from typing import Dict, List, Optional from typing import Dict, List, Optional
@ -67,6 +68,7 @@ class BotConfig:
enable_advance_output: bool = False # 是否启用高级输出 enable_advance_output: bool = False # 是否启用高级输出
enable_kuuki_read: bool = True # 是否启用读空气功能 enable_kuuki_read: bool = True # 是否启用读空气功能
enable_debug_output: bool = False # 是否启用调试输出
mood_update_interval: float = 1.0 # 情绪更新间隔 单位秒 mood_update_interval: float = 1.0 # 情绪更新间隔 单位秒
mood_decay_rate: float = 0.95 # 情绪衰减率 mood_decay_rate: float = 0.95 # 情绪衰减率
@ -325,6 +327,7 @@ class BotConfig:
others_config = parent["others"] others_config = parent["others"]
config.enable_advance_output = others_config.get("enable_advance_output", config.enable_advance_output) config.enable_advance_output = others_config.get("enable_advance_output", config.enable_advance_output)
config.enable_kuuki_read = others_config.get("enable_kuuki_read", config.enable_kuuki_read) config.enable_kuuki_read = others_config.get("enable_kuuki_read", config.enable_kuuki_read)
config.enable_debug_output = others_config.get("enable_debug_output", config.enable_debug_output)
# 版本表达式:>=1.0.0,<2.0.0 # 版本表达式:>=1.0.0,<2.0.0
# 允许字段func: method, support: str, notice: str, necessary: bool # 允许字段func: method, support: str, notice: str, necessary: bool
@ -419,4 +422,8 @@ global_config = BotConfig.load_config(config_path=bot_config_path)
if not global_config.enable_advance_output: if not global_config.enable_advance_output:
logger.remove() logger.remove()
pass
# 调试输出功能
if global_config.enable_debug_output:
logger.remove()
logger.add(sys.stdout, level="DEBUG")

View File

@ -231,7 +231,8 @@ class CQCode:
group_info=group_info, group_info=group_info,
) )
content_seg = Seg( content_seg = Seg(
type="seglist", data=message_obj.message_segment ) type="seglist", data=[message_obj.message_segment]
)
else: else:
content_seg = Seg(type="text", data="[空消息]") content_seg = Seg(type="text", data="[空消息]")
else: else:
@ -256,7 +257,7 @@ class CQCode:
group_info=group_info, group_info=group_info,
) )
content_seg = Seg( content_seg = Seg(
type="seglist", data=message_obj.message_segment type="seglist", data=[message_obj.message_segment]
) )
else: else:
content_seg = Seg(type="text", data="[空消息]") content_seg = Seg(type="text", data="[空消息]")
@ -281,11 +282,12 @@ class CQCode:
if self.reply_message.sender.user_id: if self.reply_message.sender.user_id:
message_obj = MessageRecvCQ( message_obj = MessageRecvCQ(
user_info=UserInfo(user_id=self.reply_message.sender.user_id,user_nickname=self.reply_message.sender.get("nickname",None)), user_info=UserInfo(user_id=self.reply_message.sender.user_id,user_nickname=self.reply_message.sender.nickname),
message_id=self.reply_message.message_id, message_id=self.reply_message.message_id,
raw_message=str(self.reply_message.message), raw_message=str(self.reply_message.message),
group_info=GroupInfo(group_id=self.reply_message.group_id), group_info=GroupInfo(group_id=self.reply_message.group_id),
) )
segments = [] segments = []
if message_obj.message_info.user_info.user_id == global_config.BOT_QQ: if message_obj.message_info.user_info.user_id == global_config.BOT_QQ:
@ -302,7 +304,7 @@ class CQCode:
) )
) )
segments.append(Seg(type="seglist", data=message_obj.message_segment)) segments.append(Seg(type="seglist", data=[message_obj.message_segment]))
segments.append(Seg(type="text", data="]")) segments.append(Seg(type="text", data="]"))
return segments return segments
else: else:

View File

@ -14,9 +14,52 @@ urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
#这个类是消息数据类,用于存储和管理消息数据。 #这个类是消息数据类,用于存储和管理消息数据。
#它定义了消息的属性包括群组ID、用户ID、消息ID、原始消息内容、纯文本内容和时间戳。 #它定义了消息的属性包括群组ID、用户ID、消息ID、原始消息内容、纯文本内容和时间戳。
#它还定义了两个辅助属性keywords用于提取消息的关键词is_plain_text用于判断消息是否为纯文本。 #它还定义了两个辅助属性keywords用于提取消息的关键词is_plain_text用于判断消息是否为纯文本。
@dataclass @dataclass
class MessageRecv(MessageBase): class Message(MessageBase):
chat_stream: ChatStream=None
reply: Optional['Message'] = None
detailed_plain_text: str = ""
processed_plain_text: str = ""
def __init__(
self,
message_id: str,
time: int,
chat_stream: ChatStream,
user_info: UserInfo,
message_segment: Optional[Seg] = None,
reply: Optional['MessageRecv'] = None,
detailed_plain_text: str = "",
processed_plain_text: str = "",
):
# 构造基础消息信息
message_info = BaseMessageInfo(
platform=chat_stream.platform,
message_id=message_id,
time=time,
group_info=chat_stream.group_info,
user_info=user_info
)
# 调用父类初始化
super().__init__(
message_info=message_info,
message_segment=message_segment,
raw_message=None
)
self.chat_stream = chat_stream
# 文本处理相关属性
self.processed_plain_text = processed_plain_text
self.detailed_plain_text = detailed_plain_text
# 回复消息
self.reply = reply
@dataclass
class MessageRecv(Message):
"""接收消息类用于处理从MessageCQ序列化的消息""" """接收消息类用于处理从MessageCQ序列化的消息"""
def __init__(self, message_dict: Dict): def __init__(self, message_dict: Dict):
@ -25,20 +68,16 @@ class MessageRecv(MessageBase):
Args: Args:
message_dict: MessageCQ序列化后的字典 message_dict: MessageCQ序列化后的字典
""" """
message_info = BaseMessageInfo.from_dict(message_dict.get('message_info', {})) self.message_info = BaseMessageInfo.from_dict(message_dict.get('message_info', {}))
message_segment = Seg.from_dict(message_dict.get('message_segment', {})) self.message_segment = Seg.from_dict(message_dict.get('message_segment', {}))
raw_message = message_dict.get('raw_message') self.raw_message = message_dict.get('raw_message')
super().__init__(
message_info=message_info,
message_segment=message_segment,
raw_message=raw_message
)
# 处理消息内容 # 处理消息内容
self.processed_plain_text = "" # 初始化为空字符串 self.processed_plain_text = "" # 初始化为空字符串
self.detailed_plain_text = "" # 初始化为空字符串 self.detailed_plain_text = "" # 初始化为空字符串
self.is_emoji=False self.is_emoji=False
def update_chat_stream(self,chat_stream:ChatStream): def update_chat_stream(self,chat_stream:ChatStream):
self.chat_stream=chat_stream self.chat_stream=chat_stream
@ -110,48 +149,6 @@ class MessageRecv(MessageBase):
) )
return f"[{time_str}] {name}: {self.processed_plain_text}\n" return f"[{time_str}] {name}: {self.processed_plain_text}\n"
@dataclass
class Message(MessageBase):
chat_stream: ChatStream=None
reply: Optional['Message'] = None
detailed_plain_text: str = ""
processed_plain_text: str = ""
def __init__(
self,
message_id: str,
time: int,
chat_stream: ChatStream,
user_info: UserInfo,
message_segment: Optional[Seg] = None,
reply: Optional['MessageRecv'] = None,
detailed_plain_text: str = "",
processed_plain_text: str = "",
):
# 构造基础消息信息
message_info = BaseMessageInfo(
platform=chat_stream.platform,
message_id=message_id,
time=time,
group_info=chat_stream.group_info,
user_info=user_info
)
# 调用父类初始化
super().__init__(
message_info=message_info,
message_segment=message_segment,
raw_message=None
)
self.chat_stream = chat_stream
# 文本处理相关属性
self.processed_plain_text = detailed_plain_text
self.detailed_plain_text = processed_plain_text
# 回复消息
self.reply = reply
@dataclass @dataclass
class MessageProcessBase(Message): class MessageProcessBase(Message):

View File

@ -192,13 +192,11 @@ class LLM_request:
logger.warning(f"检测到403错误模型从 {old_model_name} 降级为 {self.model_name}") logger.warning(f"检测到403错误模型从 {old_model_name} 降级为 {self.model_name}")
# 对全局配置进行更新 # 对全局配置进行更新
if hasattr(global_config, 'llm_normal') and global_config.llm_normal.get( if global_config.llm_normal.get('name') == old_model_name:
'name') == old_model_name:
global_config.llm_normal['name'] = self.model_name global_config.llm_normal['name'] = self.model_name
logger.warning(f"将全局配置中的 llm_normal 模型临时降级至{self.model_name}") logger.warning(f"将全局配置中的 llm_normal 模型临时降级至{self.model_name}")
if hasattr(global_config, 'llm_reasoning') and global_config.llm_reasoning.get( if global_config.llm_reasoning.get('name') == old_model_name:
'name') == old_model_name:
global_config.llm_reasoning['name'] = self.model_name global_config.llm_reasoning['name'] = self.model_name
logger.warning(f"将全局配置中的 llm_reasoning 模型临时降级至{self.model_name}") logger.warning(f"将全局配置中的 llm_reasoning 模型临时降级至{self.model_name}")
@ -216,6 +214,7 @@ class LLM_request:
# 将流式输出转化为非流式输出 # 将流式输出转化为非流式输出
if stream_mode: if stream_mode:
flag_delta_content_finished = False
accumulated_content = "" accumulated_content = ""
async for line_bytes in response.content: async for line_bytes in response.content:
line = line_bytes.decode("utf-8").strip() line = line_bytes.decode("utf-8").strip()
@ -227,13 +226,25 @@ class LLM_request:
break break
try: try:
chunk = json.loads(data_str) chunk = json.loads(data_str)
delta = chunk["choices"][0]["delta"] if flag_delta_content_finished:
delta_content = delta.get("content") usage = chunk.get("usage", None) # 获取tokn用量
if delta_content is None: else:
delta_content = "" delta = chunk["choices"][0]["delta"]
accumulated_content += delta_content delta_content = delta.get("content")
if delta_content is None:
delta_content = ""
accumulated_content += delta_content
# 检测流式输出文本是否结束
finish_reason = chunk["choices"][0]["finish_reason"]
if finish_reason == "stop":
usage = chunk.get("usage", None)
if usage:
break
# 部分平台在文本输出结束前不会返回token用量此时需要再获取一次chunk
flag_delta_content_finished = True
except Exception: except Exception:
logger.exception("解析流式输出错") logger.exception("解析流式输出错")
content = accumulated_content content = accumulated_content
reasoning_content = "" reasoning_content = ""
think_match = re.search(r'<think>(.*?)</think>', content, re.DOTALL) think_match = re.search(r'<think>(.*?)</think>', content, re.DOTALL)
@ -242,7 +253,7 @@ class LLM_request:
content = re.sub(r'<think>.*?</think>', '', content, flags=re.DOTALL).strip() content = re.sub(r'<think>.*?</think>', '', content, flags=re.DOTALL).strip()
# 构造一个伪result以便调用自定义响应处理器或默认处理器 # 构造一个伪result以便调用自定义响应处理器或默认处理器
result = { result = {
"choices": [{"message": {"content": content, "reasoning_content": reasoning_content}}]} "choices": [{"message": {"content": content, "reasoning_content": reasoning_content}}], "usage": usage}
return response_handler(result) if response_handler else self._default_response_handler( return response_handler(result) if response_handler else self._default_response_handler(
result, user_id, request_type, endpoint) result, user_id, request_type, endpoint)
else: else:

View File

@ -100,6 +100,7 @@ word_replace_rate=0.006 # 整词替换概率
[others] [others]
enable_advance_output = true # 是否启用高级输出 enable_advance_output = true # 是否启用高级输出
enable_kuuki_read = true # 是否启用读空气功能 enable_kuuki_read = true # 是否启用读空气功能
enable_debug_output = false # 是否启用调试输出
[groups] [groups]
talk_allowed = [ talk_allowed = [