改各种小问题

This commit is contained in:
春河晴
2025-04-16 17:37:28 +09:00
parent a0b1b1f8d8
commit dc2cf843e5
36 changed files with 114 additions and 107 deletions

View File

@@ -1,6 +1,6 @@
from typing import Tuple
from src.common.logger import get_module_logger
from ..models.utils_model import LLM_request
from ..models.utils_model import LLMRequest
from ..config.config import global_config
from .chat_observer import ChatObserver
from .pfc_utils import get_items_from_json
@@ -23,7 +23,7 @@ class ActionPlanner:
"""行动规划器"""
def __init__(self, stream_id: str):
self.llm = LLM_request(
self.llm = LLMRequest(
model=global_config.llm_normal,
temperature=global_config.llm_normal["temp"],
max_tokens=1000,

View File

@@ -4,7 +4,7 @@ from ..chat.chat_stream import ChatStream
from ..chat.message import Message
from ..message.message_base import Seg
from src.plugins.chat.message import MessageSending, MessageSet
from src.plugins.chat.message_sender import message_manager
from src.plugins.chat.messagesender import message_manager
logger = get_module_logger("message_sender")

View File

@@ -120,6 +120,10 @@ class ObservationInfo:
# #spec
# meta_plan_trigger: bool = False
def __init__(self):
self.last_message_id = None
self.chat_observer = None
def __post_init__(self):
"""初始化后创建handler"""
self.chat_observer = None
@@ -129,7 +133,7 @@ class ObservationInfo:
"""绑定到指定的chat_observer
Args:
stream_id: 聊天流ID
chat_observer: 要绑定的ChatObserver实例
"""
self.chat_observer = chat_observer
self.chat_observer.notification_manager.register_handler(
@@ -171,7 +175,8 @@ class ObservationInfo:
self.last_bot_speak_time = message["time"]
else:
self.last_user_speak_time = message["time"]
self.active_users.add(user_info.user_id)
if user_info.user_id is not None:
self.active_users.add(str(user_info.user_id))
self.new_messages_count += 1
self.unprocessed_messages.append(message)
@@ -227,7 +232,7 @@ class ObservationInfo:
"""清空未处理消息列表"""
# 将未处理消息添加到历史记录中
for message in self.unprocessed_messages:
self.chat_history.append(message)
self.chat_history.append(message) # TODO NEED FIX TYPE???
# 清空未处理消息列表
self.has_unread_messages = False
self.unprocessed_messages.clear()

View File

@@ -8,7 +8,7 @@ from src.common.logger import get_module_logger
from ..chat.chat_stream import ChatStream
from ..message.message_base import UserInfo, Seg
from ..chat.message import Message
from ..models.utils_model import LLM_request
from ..models.utils_model import LLMRequest
from ..config.config import global_config
from src.plugins.chat.message import MessageSending
from ..message.api import global_api
@@ -30,7 +30,7 @@ class GoalAnalyzer:
"""对话目标分析器"""
def __init__(self, stream_id: str):
self.llm = LLM_request(
self.llm = LLMRequest(
model=global_config.llm_normal, temperature=0.7, max_tokens=1000, request_type="conversation_goal"
)
@@ -350,7 +350,7 @@ class DirectMessageSender:
# logger.info(f"发送消息到{end_point}")
# logger.info(message_json)
try:
await global_api.send_message_REST(end_point, message_json)
await global_api.send_message_rest(end_point, message_json)
except Exception as e:
logger.error(f"REST方式发送失败出现错误: {str(e)}")
logger.info("尝试使用ws发送")

View File

@@ -1,7 +1,7 @@
from typing import List, Tuple
from src.common.logger import get_module_logger
from src.plugins.memory_system.Hippocampus import HippocampusManager
from ..models.utils_model import LLM_request
from ..models.utils_model import LLMRequest
from ..config.config import global_config
from ..chat.message import Message
@@ -12,7 +12,7 @@ class KnowledgeFetcher:
"""知识调取器"""
def __init__(self):
self.llm = LLM_request(
self.llm = LLMRequest(
model=global_config.llm_normal,
temperature=global_config.llm_normal["temp"],
max_tokens=1000,

View File

@@ -2,7 +2,7 @@ import json
import datetime
from typing import Tuple
from src.common.logger import get_module_logger
from ..models.utils_model import LLM_request
from ..models.utils_model import LLMRequest
from ..config.config import global_config
from .chat_observer import ChatObserver
from ..message.message_base import UserInfo
@@ -14,7 +14,7 @@ class ReplyChecker:
"""回复检查器"""
def __init__(self, stream_id: str):
self.llm = LLM_request(
self.llm = LLMRequest(
model=global_config.llm_normal, temperature=0.7, max_tokens=1000, request_type="reply_check"
)
self.name = global_config.BOT_NICKNAME

View File

@@ -1,6 +1,6 @@
from typing import Tuple
from src.common.logger import get_module_logger
from ..models.utils_model import LLM_request
from ..models.utils_model import LLMRequest
from ..config.config import global_config
from .chat_observer import ChatObserver
from .reply_checker import ReplyChecker
@@ -15,7 +15,7 @@ class ReplyGenerator:
"""回复生成器"""
def __init__(self, stream_id: str):
self.llm = LLM_request(
self.llm = LLMRequest(
model=global_config.llm_normal,
temperature=global_config.llm_normal["temp"],
max_tokens=300,