refactor(core): 提升类型安全性并添加配置空值检查

此提交在核心模块中引入了多项改进,以增强类型安全性和健壮性,主要通过为类型提示添加 `cast` 并在访问 `global_config` 属性前进行空值检查实现。

主要改动包括:
- **类型安全**:在 `message_handler.py`、`unified_manager.py` 和 `napcat_adapter` 的消息处理器中使用 `typing.cast` 来解决类型不一致问题并提高静态分析的准确性。
- **配置空值检查**:在 `message_handler.py` 和 `message_processor.py` 中添加对 `global_config` 及其嵌套属性的检查,以防止在应用启动或配置加载过程中出现 `NoneType` 错误。
- **内存管理提示**:优化了 `unified_manager.py` 中内存判断器的提示,使其在获取长期记忆时更加保守,从而提升简单交互的性能。
- **Napcat 适配器**:新增了视频处理的配置选项以及回复行为。同时改进了消息解析逻辑的鲁棒性。- **消息处理器**:重构了 `_process_message_segments` 及相关函数,移除了未使用的 `message_info` 参数,从而简化了函数签名。
This commit is contained in:
tt-P607
2025-11-28 10:15:53 +08:00
parent 883e391010
commit abfcf56941
6 changed files with 119 additions and 94 deletions

View File

@@ -43,6 +43,7 @@ from src.config.config import global_config
from src.mood.mood_manager import mood_manager
from src.plugin_system.base import BaseCommand, EventType
from src.plugin_system.core import component_registry, event_manager, global_announcement_manager
from typing import cast
if TYPE_CHECKING:
from src.chat.message_receive.chat_stream import ChatStream
@@ -55,23 +56,25 @@ PROJECT_ROOT = os.path.abspath(os.path.join(os.path.dirname(__file__), "../.."))
def _check_ban_words(text: str, chat: "ChatStream", userinfo) -> bool:
"""检查消息是否包含过滤词"""
for word in global_config.message_receive.ban_words:
if word in text:
chat_name = chat.group_info.group_name if chat.group_info else "私聊"
logger.info(f"[{chat_name}]{userinfo.user_nickname}:{text}")
logger.info(f"[过滤词识别]消息中含有{word}filtered")
return True
if global_config and global_config.message_receive:
for word in global_config.message_receive.ban_words:
if word in text:
chat_name = chat.group_info.group_name if chat.group_info else "私聊"
logger.info(f"[{chat_name}]{userinfo.user_nickname}:{text}")
logger.info(f"[过滤词识别]消息中含有{word}filtered")
return True
return False
def _check_ban_regex(text: str, chat: "ChatStream", userinfo) -> bool:
"""检查消息是否匹配过滤正则表达式"""
for pattern in global_config.message_receive.ban_msgs_regex:
if re.search(pattern, text):
chat_name = chat.group_info.group_name if chat.group_info else "私聊"
logger.info(f"[{chat_name}]{userinfo.user_nickname}:{text}")
logger.info(f"[正则表达式过滤]消息匹配到{pattern}filtered")
return True
if global_config and global_config.message_receive:
for pattern in global_config.message_receive.ban_msgs_regex:
if re.search(pattern, text):
chat_name = chat.group_info.group_name if chat.group_info else "私聊"
logger.info(f"[{chat_name}]{userinfo.user_nickname}:{text}")
logger.info(f"[正则表达式过滤]消息匹配到{pattern}filtered")
return True
return False
@@ -281,7 +284,7 @@ class MessageHandler:
chat = await get_chat_manager().get_or_create_stream(
platform=platform,
user_info=DatabaseUserInfo.from_dict(user_info) if user_info else None, # type: ignore
group_info=DatabaseGroupInfo.from_dict(group_info) if group_info else None,
group_info=DatabaseGroupInfo.from_dict(cast(dict, group_info)) if group_info else None,
)
# 将消息信封转换为 DatabaseMessages
@@ -431,7 +434,7 @@ class MessageHandler:
chat = await get_chat_manager().get_or_create_stream(
platform=platform,
user_info=DatabaseUserInfo.from_dict(user_info) if user_info else None, # type: ignore
group_info=DatabaseGroupInfo.from_dict(group_info) if group_info else None,
group_info=DatabaseGroupInfo.from_dict(cast(dict, group_info)) if group_info else None,
)
# 将消息信封转换为 DatabaseMessages
@@ -535,7 +538,9 @@ class MessageHandler:
text = message.processed_plain_text or ""
# 获取配置的命令前缀
prefixes = global_config.command.command_prefixes
prefixes = []
if global_config and global_config.command:
prefixes = global_config.command.command_prefixes
# 检查是否以任何前缀开头
matched_prefix = None
@@ -707,7 +712,7 @@ class MessageHandler:
# 检查是否需要处理消息
should_process_in_manager = True
if group_info and str(group_info.group_id) in global_config.message_receive.mute_group_list:
if group_info and global_config and global_config.message_receive and str(group_info.group_id) in global_config.message_receive.mute_group_list:
is_image_or_emoji = message.is_picid or message.is_emoji
if not message.is_mentioned and not is_image_or_emoji:
logger.debug(
@@ -731,7 +736,7 @@ class MessageHandler:
# 情绪系统更新
try:
if global_config.mood.enable_mood:
if global_config and global_config.mood and global_config.mood.enable_mood:
interest_rate = message.interest_value or 0.0
logger.debug(f"开始更新情绪状态,兴趣度: {interest_rate:.2f}")

View File

@@ -56,7 +56,7 @@ async def process_message_from_dict(message_dict: MessageEnvelope, stream_id: st
}
# 异步处理消息段,生成纯文本
processed_plain_text = await _process_message_segments(message_segment, processing_state, message_info)
processed_plain_text = await _process_message_segments(message_segment, processing_state)
# 解析 notice 信息
is_notify = False
@@ -155,15 +155,13 @@ async def process_message_from_dict(message_dict: MessageEnvelope, stream_id: st
async def _process_message_segments(
segment: SegPayload | list[SegPayload],
state: dict,
message_info: MessageInfoPayload
state: dict
) -> str:
"""递归处理消息段,转换为文字描述
Args:
segment: 要处理的消息段TypedDict 或列表)
state: 处理状态字典(用于记录消息类型标记)
message_info: 消息基础信息TypedDict 格式)
Returns:
str: 处理后的文本
@@ -172,7 +170,7 @@ async def _process_message_segments(
if isinstance(segment, list):
segments_text = []
for seg in segment:
processed = await _process_message_segments(seg, state, message_info)
processed = await _process_message_segments(seg, state)
if processed:
segments_text.append(processed)
return " ".join(segments_text)
@@ -186,28 +184,26 @@ async def _process_message_segments(
if seg_type == "seglist" and isinstance(seg_data, list):
segments_text = []
for sub_seg in seg_data:
processed = await _process_message_segments(sub_seg, state, message_info)
processed = await _process_message_segments(sub_seg, state)
if processed:
segments_text.append(processed)
return " ".join(segments_text)
# 处理其他类型
return await _process_single_segment(segment, state, message_info)
return await _process_single_segment(segment, state)
return ""
async def _process_single_segment(
segment: SegPayload,
state: dict,
message_info: MessageInfoPayload
state: dict
) -> str:
"""处理单个消息段
Args:
segment: 消息段TypedDict 格式)
state: 处理状态字典
message_info: 消息基础信息TypedDict 格式)
Returns:
str: 处理后的文本
@@ -234,7 +230,6 @@ async def _process_single_segment(
return f"@{seg_data}" if isinstance(seg_data, str) else "@未知用户"
elif seg_type == "image":
# 如果是base64图片数据
if isinstance(seg_data, str):
state["has_picid"] = True
state["is_picid"] = True
@@ -247,27 +242,17 @@ async def _process_single_segment(
state["has_emoji"] = True
state["is_emoji"] = True
if isinstance(seg_data, str):
return await get_image_manager().get_emoji_description(seg_data)
image_manager = get_image_manager()
return await image_manager.get_emoji_description(seg_data)
return "[发了一个表情包,网卡了加载不出来]"
elif seg_type == "voice":
state["is_voice"] = True
# 检查消息是否由机器人自己发送
user_info = message_info.get("user_info", {})
user_id_str = str(user_info.get("user_id", ""))
if user_id_str == str(global_config.bot.qq_account):
logger.info(f"检测到机器人自身发送的语音消息 (User ID: {user_id_str}),尝试从缓存获取文本。")
if isinstance(seg_data, str):
cached_text = consume_self_voice_text(seg_data)
if cached_text:
logger.info(f"成功从缓存中获取语音文本: '{cached_text[:70]}...'")
return f"[语音:{cached_text}]"
else:
logger.warning("机器人自身语音消息缓存未命中,将回退到标准语音识别。")
# 标准语音识别流程
# 检查是否是自己发送的语音
if isinstance(seg_data, str):
cached_text = consume_self_voice_text(seg_data)
if cached_text:
return f"[语音:{cached_text}]"
return await get_voice_text(seg_data)
return "[发了一段语音,网卡了加载不出来]"
@@ -299,7 +284,7 @@ async def _process_single_segment(
logger.warning("⚠️ Rust视频处理模块不可用跳过视频分析")
return "[视频]"
if global_config.video_analysis.enable:
if global_config and global_config.video_analysis and global_config.video_analysis.enable:
logger.info("已启用视频识别,开始识别")
if isinstance(seg_data, dict):
try:
@@ -317,8 +302,9 @@ async def _process_single_segment(
# 使用video analyzer分析视频
video_analyzer = get_video_analyzer()
prompt = global_config.video_analysis.batch_analysis_prompt if global_config and global_config.video_analysis else ""
result = await video_analyzer.analyze_video_from_bytes(
video_bytes, filename, prompt=global_config.video_analysis.batch_analysis_prompt
video_bytes, filename, prompt=prompt
)
logger.info(f"视频分析结果: {result}")

View File

@@ -13,9 +13,10 @@ import asyncio
import time
from datetime import datetime
from pathlib import Path
from typing import Any
from typing import Any, cast
from src.common.logger import get_logger
from src.config.config import TaskConfig
from src.memory_graph.manager import MemoryManager
from src.memory_graph.long_term_manager import LongTermMemoryManager
from src.memory_graph.models import JudgeDecision, MemoryBlock, ShortTermMemory
@@ -83,7 +84,7 @@ class UnifiedMemoryManager:
self.long_term_manager: LongTermMemoryManager
# 底层 MemoryManager长期记忆
self.memory_manager: MemoryManager = memory_manager
self.memory_manager: MemoryManager = cast(MemoryManager, memory_manager)
# 配置参数存储(用于初始化)
self._config = {
@@ -330,7 +331,11 @@ class UnifiedMemoryManager:
"""
prompt = f"""你是一个记忆检索评估专家。请判断检索到的记忆是否足以回答用户的问题
prompt = f"""你是一个记忆检索评估专家。你的任务是判断当前检索到的“感知记忆”(即时对话)和“短期记忆”(结构化信息)是否足以支撑一次有深度、有上下文的回复
**核心原则:**
- **不要轻易检索长期记忆!** 只有在当前对话需要深入探讨、回忆过去复杂事件或需要特定背景知识时,才认为记忆不足。
- **闲聊、简单问候、表情互动或无特定主题的对话,现有记忆通常是充足的。** 频繁检索长期记忆会拖慢响应速度。
**用户查询:**
{query}
@@ -341,27 +346,36 @@ class UnifiedMemoryManager:
**检索到的短期记忆(结构化信息,自然语言描述):**
{short_term_desc or '(无)'}
**任务要求**
1. 判断这些记忆是否足以回答用户的问题
2. 如果不充足,分析缺少哪些方面的信息
3. 生成额外需要检索的 query用于在长期记忆中检索
**评估指南**
1. **分析用户意图**:用户是在闲聊,还是在讨论一个需要深入挖掘的话题?
2. **检查现有记忆**:当前的感知和短期记忆是否已经包含了足够的信息来回应用户的查询?
- 对于闲聊(如“你好”、“哈哈”、“[表情]”),现有记忆总是充足的 (`"is_sufficient": true`)。
- 对于需要回忆具体细节、深入探讨个人经历或专业知识的查询,如果现有记忆中没有相关信息,则可能不充足。
3. **决策**
- 如果记忆充足,设置 `"is_sufficient": true`。
- 如果确实需要更多信息才能进行有意义的对话,设置 `"is_sufficient": false`,并提供具体的补充查询。
**输出格式JSON**
```json
{{
"is_sufficient": true/false,
"confidence": 0.85,
"reasoning": "判断理由",
"reasoning": "在这里解释你的判断理由。例如:‘用户只是在打招呼,现有记忆已足够’或‘用户问到了一个具体的历史事件,需要检索长期记忆’。",
"missing_aspects": ["缺失的信息1", "缺失的信息2"],
"additional_queries": ["补充query1", "补充query2"]
}}
```
请输出JSON"""
严格按照上述原则进行判断,并输出JSON"""
# 调用记忆裁判模型
model_set = (
model_config.model_task_config.memory_judge
if model_config and model_config.model_task_config
else TaskConfig(model_name="deepseek/deepseek-v2", provider="deepseek")
)
llm = LLMRequest(
model_set=model_config.model_task_config.memory_judge,
model_set=model_set,
request_type="unified_memory.judge",
)

View File

@@ -63,7 +63,7 @@ class CoreActionsPlugin(BasePlugin):
"""返回插件包含的组件列表"""
# --- 根据配置注册组件 ---
components: ClassVar = []
components = []
# 注册 reply 动作
if self.get_config("components.enable_reply", True):

View File

@@ -317,6 +317,9 @@ class NapcatAdapterPlugin(BasePlugin):
"ignore_non_self_poke": ConfigField(type=bool, default=False, description="是否忽略不是针对自己的戳一戳消息"),
"poke_debounce_seconds": ConfigField(type=float, default=2.0, description="戳一戳防抖时间(秒)"),
"enable_emoji_like": ConfigField(type=bool, default=True, description="是否启用群聊表情回复处理"),
"enable_reply_at": ConfigField(type=bool, default=True, description="是否在回复时自动@原消息发送者"),
"reply_at_rate": ConfigField(type=float, default=0.5, description="回复时@的概率0.0-1.0"),
"enable_video_processing": ConfigField(type=bool, default=True, description="是否启用视频消息处理(下载和解析)"),
},
}

View File

@@ -5,7 +5,7 @@ from __future__ import annotations
import base64
import time
from pathlib import Path
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, cast
import uuid
from mofox_wire import MessageBuilder
@@ -214,6 +214,9 @@ class MessageHandler:
case RealMessageType.record:
return await self._handle_record_message(segment)
case RealMessageType.video:
if not config_api.get_plugin_config(self.plugin_config, "features.enable_video_processing", True):
logger.debug("视频消息处理已禁用,跳过")
return None
return await self._handle_video_message(segment)
case RealMessageType.rps:
return await self._handle_rps_message(segment)
@@ -331,10 +334,13 @@ class MessageHandler:
{"type": seg.get("type", "text"), "data": seg.get("data", "")} for seg in reply_segments
] or [{"type": "text", "data": "[无法获取被引用的消息]"}]
return {
"type": "seglist",
"data": [{"type": "text", "data": prefix_text}, *brief_segments, {"type": "text", "data": suffix_text}],
}
return cast(
SegPayload,
{
"type": "seglist",
"data": [{"type": "text", "data": prefix_text}, *brief_segments, {"type": "text", "data": suffix_text}],
},
)
async def _handle_record_message(self, segment: dict) -> SegPayload | None:
"""处理语音消息"""
@@ -380,14 +386,17 @@ class MessageHandler:
video_base64 = base64.b64encode(video_data).decode("utf-8")
logger.debug(f"视频文件大小: {len(video_data) / (1024 * 1024):.2f} MB")
return {
"type": "video",
"data": {
"base64": video_base64,
"filename": Path(file_path).name,
"size_mb": len(video_data) / (1024 * 1024),
return cast(
SegPayload,
{
"type": "video",
"data": {
"base64": video_base64,
"filename": Path(file_path).name,
"size_mb": len(video_data) / (1024 * 1024),
},
},
}
)
elif video_url:
# URL下载处理
from ..video_handler import get_video_downloader
@@ -401,15 +410,18 @@ class MessageHandler:
video_base64 = base64.b64encode(download_result["data"]).decode("utf-8")
logger.debug(f"视频下载成功,大小: {len(download_result['data']) / (1024 * 1024):.2f} MB")
return {
"type": "video",
"data": {
"base64": video_base64,
"filename": download_result.get("filename", "video.mp4"),
"size_mb": len(download_result["data"]) / (1024 * 1024),
"url": video_url,
return cast(
SegPayload,
{
"type": "video",
"data": {
"base64": video_base64,
"filename": download_result.get("filename", "video.mp4"),
"size_mb": len(download_result["data"]) / (1024 * 1024),
"url": video_url,
},
},
}
)
else:
logger.warning("既没有有效的本地文件路径也没有有效的视频URL")
return None
@@ -454,34 +466,39 @@ class MessageHandler:
processed_message = handled_message
forward_hint = {"type": "text", "data": "这是一条转发消息:\n"}
return {"type": "seglist", "data": [forward_hint, processed_message]}
return cast(SegPayload, {"type": "seglist", "data": [forward_hint, processed_message]})
async def _recursive_parse_image_seg(self, seg_data: SegPayload, to_image: bool) -> SegPayload:
# sourcery skip: merge-else-if-into-elif
if seg_data.get("type") == "seglist":
new_seg_list = []
for i_seg in seg_data.get("data", []):
parsed_seg = await self._recursive_parse_image_seg(i_seg, to_image)
new_seg_list.append(parsed_seg)
if isinstance(i_seg, dict): # 确保是字典类型
parsed_seg = await self._recursive_parse_image_seg(i_seg, to_image)
new_seg_list.append(parsed_seg)
return {"type": "seglist", "data": new_seg_list}
if to_image:
if seg_data.get("type") == "image":
image_url = seg_data.get("data")
try:
encoded_image = await get_image_base64(image_url)
except Exception as e:
logger.error(f"图片处理失败: {str(e)}")
return {"type": "text", "data": "[图片]"}
return {"type": "image", "data": encoded_image}
if isinstance(image_url, str):
try:
encoded_image = await get_image_base64(image_url)
except Exception as e:
logger.error(f"图片处理失败: {str(e)}")
return {"type": "text", "data": "[图片]"}
return {"type": "image", "data": encoded_image}
return {"type": "text", "data": "[图片]"}
if seg_data.get("type") == "emoji":
image_url = seg_data.get("data")
try:
encoded_image = await get_image_base64(image_url)
except Exception as e:
logger.error(f"图片处理失败: {str(e)}")
return {"type": "text", "data": "[表情包]"}
return {"type": "emoji", "data": encoded_image}
if isinstance(image_url, str):
try:
encoded_image = await get_image_base64(image_url)
except Exception as e:
logger.error(f"图片处理失败: {str(e)}")
return {"type": "text", "data": "[表情包]"}
return {"type": "emoji", "data": encoded_image}
return {"type": "text", "data": "[表情包]"}
logger.debug(f"不处理类型: {seg_data.get('type')}")
return seg_data
@@ -595,7 +612,7 @@ class MessageHandler:
"id": file_id,
}
return {"type": "file", "data": file_data}
return cast(SegPayload, {"type": "file", "data": file_data})
async def _handle_json_message(self, segment: dict) -> SegPayload | None:
"""
@@ -623,7 +640,7 @@ class MessageHandler:
# 从回声消息中提取文件信息
file_info = self._extract_file_info_from_echo(nested_data)
if file_info:
return {"type": "file", "data": file_info}
return cast(SegPayload, {"type": "file", "data": file_info})
# 检查是否是QQ小程序分享消息
if "app" in nested_data and "com.tencent.miniapp" in str(nested_data.get("app", "")):