diff --git a/script/run_thingking.bat b/script/run_thingking.bat index a134da6fe..0806e46ed 100644 --- a/script/run_thingking.bat +++ b/script/run_thingking.bat @@ -1,5 +1,5 @@ -call conda activate niuniu -cd src\gui -start /b python reasoning_gui.py +@REM call conda activate niuniu +cd ../src\gui +start /b ../../venv/scripts/python.exe reasoning_gui.py exit diff --git a/src/gui/reasoning_gui.py b/src/gui/reasoning_gui.py index a93d80afd..43f692d58 100644 --- a/src/gui/reasoning_gui.py +++ b/src/gui/reasoning_gui.py @@ -6,6 +6,8 @@ import time from datetime import datetime from typing import Dict, List from typing import Optional +sys.path.insert(0, sys.path[0]+"/../") +sys.path.insert(0, sys.path[0]+"/../") from src.common.logger import get_module_logger import customtkinter as ctk diff --git a/src/plugins/memory_system/memory_manual_build.py b/src/plugins/memory_system/memory_manual_build.py index b575f455e..7e392668f 100644 --- a/src/plugins/memory_system/memory_manual_build.py +++ b/src/plugins/memory_system/memory_manual_build.py @@ -10,6 +10,12 @@ from pathlib import Path import matplotlib.pyplot as plt import networkx as nx from dotenv import load_dotenv +sys.path.insert(0, sys.path[0]+"/../") +sys.path.insert(0, sys.path[0]+"/../") +sys.path.insert(0, sys.path[0]+"/../") +sys.path.insert(0, sys.path[0]+"/../") +sys.path.insert(0, sys.path[0]+"/../") +from src.common.logger import get_module_logger import jieba # from chat.config import global_config diff --git a/src/plugins/models/utils_model.py b/src/plugins/models/utils_model.py index 975bcaf7b..5ad69ff25 100644 --- a/src/plugins/models/utils_model.py +++ b/src/plugins/models/utils_model.py @@ -274,6 +274,7 @@ class LLM_request: raise RuntimeError(f"请求被拒绝: {error_code_mapping.get(response.status)}") response.raise_for_status() + reasoning_content = "" # 将流式输出转化为非流式输出 if stream_mode: @@ -303,6 +304,8 @@ class LLM_request: accumulated_content += delta_content # 检测流式输出文本是否结束 finish_reason = chunk["choices"][0].get("finish_reason") + if delta.get("reasoning_content", None): + reasoning_content += delta["reasoning_content"] if finish_reason == "stop": chunk_usage = chunk.get("usage", None) if chunk_usage: @@ -314,7 +317,6 @@ class LLM_request: except Exception as e: logger.exception(f"解析流式输出错误: {str(e)}") content = accumulated_content - reasoning_content = "" think_match = re.search(r"(.*?)", content, re.DOTALL) if think_match: reasoning_content = think_match.group(1).strip()