From 2335ec6577679c6d28c723b318c40f89d33fc9c4 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=A2=A8=E6=A2=93=E6=9F=92?= <1787882683@qq.com> Date: Fri, 25 Jul 2025 16:44:55 +0800 Subject: [PATCH] =?UTF-8?q?fix:=20=E5=85=BC=E5=AE=B9=E6=96=B0=E6=97=A7?= =?UTF-8?q?=E6=A0=BC=E5=BC=8F=E7=9A=84=E6=A8=A1=E5=9E=8B=E5=90=8D=E7=A7=B0?= =?UTF-8?q?=E8=8E=B7=E5=8F=96=EF=BC=8C=E4=BF=9D=E7=95=99provider=E5=85=BC?= =?UTF-8?q?=E5=AE=B9=E6=80=A7?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/chat/replyer/default_generator.py | 8 ++++++-- src/llm_models/utils_model.py | 3 ++- 2 files changed, 8 insertions(+), 3 deletions(-) diff --git a/src/chat/replyer/default_generator.py b/src/chat/replyer/default_generator.py index 9d75671c6..bb2aa34e3 100644 --- a/src/chat/replyer/default_generator.py +++ b/src/chat/replyer/default_generator.py @@ -213,8 +213,10 @@ class DefaultReplyer: with Timer("LLM生成", {}): # 内部计时器,可选保留 # 加权随机选择一个模型配置 selected_model_config = self._select_weighted_model_config() + # 兼容新旧格式的模型名称获取 + model_display_name = selected_model_config.get('model_name', selected_model_config.get('name', 'N/A')) logger.info( - f"使用模型生成回复: {selected_model_config.get('name', 'N/A')} (选中概率: {selected_model_config.get('weight', 1.0)})" + f"使用模型生成回复: {model_display_name} (选中概率: {selected_model_config.get('weight', 1.0)})" ) express_model = LLMRequest( @@ -277,8 +279,10 @@ class DefaultReplyer: with Timer("LLM生成", {}): # 内部计时器,可选保留 # 加权随机选择一个模型配置 selected_model_config = self._select_weighted_model_config() + # 兼容新旧格式的模型名称获取 + model_display_name = selected_model_config.get('model_name', selected_model_config.get('name', 'N/A')) logger.info( - f"使用模型重写回复: {selected_model_config.get('name', 'N/A')} (选中概率: {selected_model_config.get('weight', 1.0)})" + f"使用模型重写回复: {model_display_name} (选中概率: {selected_model_config.get('weight', 1.0)})" ) express_model = LLMRequest( diff --git a/src/llm_models/utils_model.py b/src/llm_models/utils_model.py index 461d4a890..0e79b63b1 100644 --- a/src/llm_models/utils_model.py +++ b/src/llm_models/utils_model.py @@ -126,7 +126,8 @@ class LLMRequest: # 兼容新旧模型配置格式 # 新格式使用 model_name,旧格式使用 name self.model_name: str = model.get("model_name", model.get("name", "")) - self.provider = model.get("provider", "") + # 在新架构中,provider信息从model_config.toml自动获取,不需要在这里设置 + self.provider = model.get("provider", "") # 保留兼容性,但在新架构中不使用 # 从全局配置中获取任务配置 self.request_type = kwargs.pop("request_type", "default")