From 7a3e47cfc64e5e0976c373b644a043487c86c502 Mon Sep 17 00:00:00 2001 From: tcmofashi Date: Mon, 17 Mar 2025 18:57:33 +0800 Subject: [PATCH 01/21] pull from upstream main-fix --- MaiLauncher.bat | 1264 +++++++++++++++++++++++------------------------ 1 file changed, 632 insertions(+), 632 deletions(-) diff --git a/MaiLauncher.bat b/MaiLauncher.bat index 7d33946b3..3a43d68ac 100644 --- a/MaiLauncher.bat +++ b/MaiLauncher.bat @@ -1,632 +1,632 @@ -@echo off -@setlocal enabledelayedexpansion -@chcp 936 - -@REM 设置版本号 -set "VERSION=1.0" - -title 麦麦Bot控制台 v%VERSION% - -@REM 设置Python和Git环境变量 -set "_root=%~dp0" -set "_root=%_root:~0,-1%" -cd "%_root%" - - -:search_python -cls -if exist "%_root%\python" ( - set "PYTHON_HOME=%_root%\python" -) else if exist "%_root%\venv" ( - call "%_root%\venv\Scripts\activate.bat" - set "PYTHON_HOME=%_root%\venv\Scripts" -) else ( - echo 正在自动查找Python解释器... - - where python >nul 2>&1 - if %errorlevel% equ 0 ( - for /f "delims=" %%i in ('where python') do ( - echo %%i | findstr /i /c:"!LocalAppData!\Microsoft\WindowsApps\python.exe" >nul - if errorlevel 1 ( - echo 找到Python解释器:%%i - set "py_path=%%i" - goto :validate_python - ) - ) - ) - set "search_paths=%ProgramFiles%\Git*;!LocalAppData!\Programs\Python\Python*" - for /d %%d in (!search_paths!) do ( - if exist "%%d\python.exe" ( - set "py_path=%%d\python.exe" - goto :validate_python - ) - ) - echo 没有找到Python解释器,要安装吗? - set /p pyinstall_confirm="继续?(Y/n): " - if /i "!pyinstall_confirm!"=="Y" ( - cls - echo 正在安装Python... - winget install --id Python.Python.3.13 -e --accept-package-agreements --accept-source-agreements - if %errorlevel% neq 0 ( - echo 安装失败,请手动安装Python - start https://www.python.org/downloads/ - exit /b - ) - echo 安装完成,正在验证Python... - goto search_python - - ) else ( - echo 取消安装Python,按任意键退出... - pause >nul - exit /b - ) - - echo 错误:未找到可用的Python解释器! - exit /b 1 - - :validate_python - "!py_path!" --version >nul 2>&1 - if %errorlevel% neq 0 ( - echo 无效的Python解释器:%py_path% - exit /b 1 - ) - - :: 提取安装目录 - for %%i in ("%py_path%") do set "PYTHON_HOME=%%~dpi" - set "PYTHON_HOME=%PYTHON_HOME:~0,-1%" -) -if not exist "%PYTHON_HOME%\python.exe" ( - echo Python路径验证失败:%PYTHON_HOME% - echo 请检查Python安装路径中是否有python.exe文件 - exit /b 1 -) -echo 成功设置Python路径:%PYTHON_HOME% - - - -:search_git -cls -if exist "%_root%\tools\git\bin" ( - set "GIT_HOME=%_root%\tools\git\bin" -) else ( - echo 正在自动查找Git... - - where git >nul 2>&1 - if %errorlevel% equ 0 ( - for /f "delims=" %%i in ('where git') do ( - set "git_path=%%i" - goto :validate_git - ) - ) - echo 正在扫描常见安装路径... - set "search_paths=!ProgramFiles!\Git\cmd" - for /f "tokens=*" %%d in ("!search_paths!") do ( - if exist "%%d\git.exe" ( - set "git_path=%%d\git.exe" - goto :validate_git - ) - ) - echo 没有找到Git,要安装吗? - set /p confirm="继续?(Y/N): " - if /i "!confirm!"=="Y" ( - cls - echo 正在安装Git... - set "custom_url=https://ghfast.top/https://github.com/git-for-windows/git/releases/download/v2.48.1.windows.1/Git-2.48.1-64-bit.exe" - - set "download_path=%TEMP%\Git-Installer.exe" - - echo 正在下载Git安装包... - curl -L -o "!download_path!" "!custom_url!" - - if exist "!download_path!" ( - echo 下载成功,开始安装Git... - start /wait "" "!download_path!" /SILENT /NORESTART - ) else ( - echo 下载失败,请手动安装Git - start https://git-scm.com/download/win - exit /b - ) - - del "!download_path!" - echo 临时文件已清理。 - - echo 安装完成,正在验证Git... - where git >nul 2>&1 - if %errorlevel% equ 0 ( - for /f "delims=" %%i in ('where git') do ( - set "git_path=%%i" - goto :validate_git - ) - goto :search_git - - ) else ( - echo 安装完成,但未找到Git,请手动安装Git - start https://git-scm.com/download/win - exit /b - ) - - ) else ( - echo 取消安装Git,按任意键退出... - pause >nul - exit /b - ) - - echo 错误:未找到可用的Git! - exit /b 1 - - :validate_git - "%git_path%" --version >nul 2>&1 - if %errorlevel% neq 0 ( - echo 无效的Git:%git_path% - exit /b 1 - ) - - :: 提取安装目录 - for %%i in ("%git_path%") do set "GIT_HOME=%%~dpi" - set "GIT_HOME=%GIT_HOME:~0,-1%" -) - -:search_mongodb -cls -sc query | findstr /i "MongoDB" >nul -if !errorlevel! neq 0 ( - echo MongoDB服务未运行,是否尝试运行服务? - set /p confirm="是否启动?(Y/N): " - if /i "!confirm!"=="Y" ( - echo 正在尝试启动MongoDB服务... - powershell -Command "Start-Process -Verb RunAs cmd -ArgumentList '/c net start MongoDB'" - echo 正在等待MongoDB服务启动... - echo 按下任意键跳过等待... - timeout /t 30 >nul - sc query | findstr /i "MongoDB" >nul - if !errorlevel! neq 0 ( - echo MongoDB服务启动失败,可能是没有安装,要安装吗? - set /p install_confirm="继续安装?(Y/N): " - if /i "!install_confirm!"=="Y" ( - echo 正在安装MongoDB... - winget install --id MongoDB.Server -e --accept-package-agreements --accept-source-agreements - echo 安装完成,正在启动MongoDB服务... - net start MongoDB - if !errorlevel! neq 0 ( - echo 启动MongoDB服务失败,请手动启动 - exit /b - ) else ( - echo MongoDB服务已成功启动 - ) - ) else ( - echo 取消安装MongoDB,按任意键退出... - pause >nul - exit /b - ) - ) - ) else ( - echo "警告:MongoDB服务未运行,将导致MaiMBot无法访问数据库!" - ) -) else ( - echo MongoDB服务已运行 -) - -@REM set "GIT_HOME=%_root%\tools\git\bin" -set "PATH=%PYTHON_HOME%;%GIT_HOME%;%PATH%" - -:install_maim -if not exist "!_root!\bot.py" ( - cls - echo 你似乎没有安装麦麦Bot,要安装在当前目录吗? - set /p confirm="继续?(Y/N): " - if /i "!confirm!"=="Y" ( - echo 要使用Git代理下载吗? - set /p proxy_confirm="继续?(Y/N): " - if /i "!proxy_confirm!"=="Y" ( - echo 正在安装麦麦Bot... - git clone https://ghfast.top/https://github.com/SengokuCola/MaiMBot - ) else ( - echo 正在安装麦麦Bot... - git clone https://github.com/SengokuCola/MaiMBot - ) - xcopy /E /H /I MaiMBot . >nul 2>&1 - rmdir /s /q MaiMBot - git checkout main-fix - - echo 安装完成,正在安装依赖... - python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple - python -m pip install virtualenv - python -m virtualenv venv - call venv\Scripts\activate.bat - python -m pip install -r requirements.txt - - echo 安装完成,要编辑配置文件吗? - set /p edit_confirm="继续?(Y/N): " - if /i "!edit_confirm!"=="Y" ( - goto config_menu - ) else ( - echo 取消编辑配置文件,按任意键返回主菜单... - ) - ) -) - - -@REM git获取当前分支名并保存在变量里 -for /f "delims=" %%b in ('git symbolic-ref --short HEAD 2^>nul') do ( - set "BRANCH=%%b" -) - -@REM 根据不同分支名给分支名字符串使用不同颜色 -echo 分支名: %BRANCH% -if "!BRANCH!"=="main" ( - set "BRANCH_COLOR=" -) else if "!BRANCH!"=="main-fix" ( - set "BRANCH_COLOR=" -@REM ) else if "%BRANCH%"=="stable-dev" ( -@REM set "BRANCH_COLOR=" -) else ( - set "BRANCH_COLOR=" -) - -@REM endlocal & set "BRANCH_COLOR=%BRANCH_COLOR%" - -:check_is_venv -echo 正在检查虚拟环境状态... -if exist "%_root%\config\no_venv" ( - echo 检测到no_venv,跳过虚拟环境检查 - goto menu -) - -:: 环境检测 -if defined VIRTUAL_ENV ( - goto menu -) - -echo ===================================== -echo 虚拟环境检测警告: -echo 当前使用系统Python路径:!PYTHON_HOME! -echo 未检测到激活的虚拟环境! - -:env_interaction -echo ===================================== -echo 请选择操作: -echo 1 - 创建并激活Venv虚拟环境 -echo 2 - 创建/激活Conda虚拟环境 -echo 3 - 临时跳过本次检查 -echo 4 - 永久跳过虚拟环境检查 -set /p choice="请输入选项(1-4): " - -if "!choice!" = "4" ( - echo 要永久跳过虚拟环境检查吗? - set /p no_venv_confirm="继续?(Y/N): ....." - if /i "!no_venv_confirm!"=="Y" ( - echo 1 > "%_root%\config\no_venv" - echo 已创建no_venv文件 - pause >nul - goto menu - ) else ( - echo 取消跳过虚拟环境检查,按任意键返回... - pause >nul - goto env_interaction - ) -) - -if "!choice!" = "3"( - echo 警告:使用系统环境可能导致依赖冲突! - timeout /t 2 >nul - goto menu -) - -if "!choice!" = "2" goto handle_conda -if "!choice!" = "1" goto handle_venv - -echo 无效的输入,请输入1-4之间的数字 -timeout /t 2 >nul -goto env_interaction - -:handle_venv -python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple -echo 正在初始化Venv环境... -python -m pip install virtualenv || ( - echo 安装环境失败,错误码:!errorlevel! - pause - goto env_interaction -) -echo 创建虚拟环境到:venv - python -m virtualenv venv || ( - echo 环境创建失败,错误码:!errorlevel! - pause - goto env_interaction -) - -call venv\Scripts\activate.bat -echo 已激活Venv环境 -echo 要安装依赖吗? -set /p install_confirm="继续?(Y/N): " -if /i "!install_confirm!"=="Y" ( - goto update_dependencies -) -goto menu - -:handle_conda -where conda >nul 2>&1 || ( - echo 未检测到conda,可能原因: - echo 1. 未安装Miniconda - echo 2. conda配置异常 - timeout /t 10 >nul - goto env_interaction -) - -:conda_menu -echo 请选择Conda操作: -echo 1 - 创建新环境 -echo 2 - 激活已有环境 -echo 3 - 返回上级菜单 -set /p choice="请输入选项(1-3): " - -if "!choice!"=="3" goto env_interaction -if "!choice!"=="2" goto activate_conda -if "!choice!"=="1" goto create_conda - -:create_conda -set /p "CONDA_ENV=请输入新环境名称:" -if "!CONDA_ENV!"=="" ( - echo 环境名称不能为空! - goto create_conda -) -conda create -n !CONDA_ENV! python=3.13 -y || ( - echo 环境创建失败,错误码:!errorlevel! - pause - goto conda_menu -) -goto activate_conda - -:activate_conda -set /p "CONDA_ENV=请输入要激活的环境名称:" -conda activate !CONDA_ENV! || ( - echo 激活失败,可能原因: - echo 1. 环境不存在 - echo 2. conda配置异常 - pause - goto conda_menu -) -echo 成功激活conda环境:!CONDA_ENV! -echo 要安装依赖吗? -set /p install_confirm="继续?(Y/N): " -if /i "!install_confirm!"=="Y" ( - goto update_dependencies -) -:menu -@chcp 936 -cls -echo 麦麦Bot控制台 v%VERSION% 当前分支: %BRANCH_COLOR%%BRANCH% -echo 当前Python环境: !PYTHON_HOME! -echo ====================== -echo 1. 更新并启动麦麦Bot (默认) -echo 2. 直接启动麦麦Bot -echo 3. 启动麦麦配置界面 -echo 4. 打开麦麦神奇工具箱 -echo 5. 退出 -echo ====================== - -set /p choice="请输入选项数字 (1-5)并按下回车以选择: " - -if "!choice!"=="" set choice=1 - -if "!choice!"=="1" goto update_and_start -if "!choice!"=="2" goto start_bot -if "!choice!"=="3" goto config_menu -if "!choice!"=="4" goto tools_menu -if "!choice!"=="5" exit /b - -echo 无效的输入,请输入1-5之间的数字 -timeout /t 2 >nul -goto menu - -:config_menu -@chcp 936 -cls -if not exist config/bot_config.toml ( - copy /Y "template\bot_config_template.toml" "config\bot_config.toml" - -) -if not exist .env.prod ( - copy /Y "template\.env.prod" ".env.prod" -) - -start python webui.py - -goto menu - - -:tools_menu -@chcp 936 -cls -echo 麦麦时尚工具箱 当前分支: %BRANCH_COLOR%%BRANCH% -echo ====================== -echo 1. 更新依赖 -echo 2. 切换分支 -echo 3. 重置当前分支 -echo 4. 更新配置文件 -echo 5. 学习新的知识库 -echo 6. 打开知识库文件夹 -echo 7. 返回主菜单 -echo ====================== - -set /p choice="请输入选项数字: " -if "!choice!"=="1" goto update_dependencies -if "!choice!"=="2" goto switch_branch -if "!choice!"=="3" goto reset_branch -if "!choice!"=="4" goto update_config -if "!choice!"=="5" goto learn_new_knowledge -if "!choice!"=="6" goto open_knowledge_folder -if "!choice!"=="7" goto menu - -echo 无效的输入,请输入1-6之间的数字 -timeout /t 2 >nul -goto tools_menu - -:update_dependencies -cls -echo 正在更新依赖... -python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple -python.exe -m pip install -r requirements.txt - -echo 依赖更新完成,按任意键返回工具箱菜单... -pause -goto tools_menu - -:switch_branch -cls -echo 正在切换分支... -echo 当前分支: %BRANCH% -@REM echo 可用分支: main, debug, stable-dev -echo 1. 切换到main -echo 2. 切换到main-fix -echo 请输入要切换到的分支: -set /p branch_name="分支名: " -if "%branch_name%"=="" set branch_name=main -if "%branch_name%"=="main" ( - set "BRANCH_COLOR=" -) else if "%branch_name%"=="main-fix" ( - set "BRANCH_COLOR=" -@REM ) else if "%branch_name%"=="stable-dev" ( -@REM set "BRANCH_COLOR=" -) else if "%branch_name%"=="1" ( - set "BRANCH_COLOR=" - set "branch_name=main" -) else if "%branch_name%"=="2" ( - set "BRANCH_COLOR=" - set "branch_name=main-fix" -) else ( - echo 无效的分支名, 请重新输入 - timeout /t 2 >nul - goto switch_branch -) - -echo 正在切换到分支 %branch_name%... -git checkout %branch_name% -echo 分支切换完成,当前分支: %BRANCH_COLOR%%branch_name% -set "BRANCH=%branch_name%" -echo 按任意键返回工具箱菜单... -pause >nul -goto tools_menu - - -:reset_branch -cls -echo 正在重置当前分支... -echo 当前分支: !BRANCH! -echo 确认要重置当前分支吗? -set /p confirm="继续?(Y/N): " -if /i "!confirm!"=="Y" ( - echo 正在重置当前分支... - git reset --hard !BRANCH! - echo 分支重置完成,按任意键返回工具箱菜单... -) else ( - echo 取消重置当前分支,按任意键返回工具箱菜单... -) -pause >nul -goto tools_menu - - -:update_config -cls -echo 正在更新配置文件... -echo 请确保已备份重要数据,继续将修改当前配置文件。 -echo 继续请按Y,取消请按任意键... -set /p confirm="继续?(Y/N): " -if /i "!confirm!"=="Y" ( - echo 正在更新配置文件... - python.exe config\auto_update.py - echo 配置文件更新完成,按任意键返回工具箱菜单... -) else ( - echo 取消更新配置文件,按任意键返回工具箱菜单... -) -pause >nul -goto tools_menu - -:learn_new_knowledge -cls -echo 正在学习新的知识库... -echo 请确保已备份重要数据,继续将修改当前知识库。 -echo 继续请按Y,取消请按任意键... -set /p confirm="继续?(Y/N): " -if /i "!confirm!"=="Y" ( - echo 正在学习新的知识库... - python.exe src\plugins\zhishi\knowledge_library.py - echo 学习完成,按任意键返回工具箱菜单... -) else ( - echo 取消学习新的知识库,按任意键返回工具箱菜单... -) -pause >nul -goto tools_menu - -:open_knowledge_folder -cls -echo 正在打开知识库文件夹... -if exist data\raw_info ( - start explorer data\raw_info -) else ( - echo 知识库文件夹不存在! - echo 正在创建文件夹... - mkdir data\raw_info - timeout /t 2 >nul -) -goto tools_menu - - -:update_and_start -cls -:retry_git_pull -git pull > temp.log 2>&1 -findstr /C:"detected dubious ownership" temp.log >nul -if %errorlevel% equ 0 ( - echo 检测到仓库权限问题,正在自动修复... - git config --global --add safe.directory "%cd%" - echo 已添加例外,正在重试git pull... - del temp.log - goto retry_git_pull -) -del temp.log -echo 正在更新依赖... -python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple -python -m pip install -r requirements.txt && cls - -echo 当前代理设置: -echo HTTP_PROXY=%HTTP_PROXY% -echo HTTPS_PROXY=%HTTPS_PROXY% - -echo Disable Proxy... -set HTTP_PROXY= -set HTTPS_PROXY= -set no_proxy=0.0.0.0/32 - -REM chcp 65001 -python bot.py -echo. -echo Bot已停止运行,按任意键返回主菜单... -pause >nul -goto menu - -:start_bot -cls -echo 正在更新依赖... -python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple -python -m pip install -r requirements.txt && cls - -echo 当前代理设置: -echo HTTP_PROXY=%HTTP_PROXY% -echo HTTPS_PROXY=%HTTPS_PROXY% - -echo Disable Proxy... -set HTTP_PROXY= -set HTTPS_PROXY= -set no_proxy=0.0.0.0/32 - -REM chcp 65001 -python bot.py -echo. -echo Bot已停止运行,按任意键返回主菜单... -pause >nul -goto menu - - -:open_dir -start explorer "%cd%" -goto menu +@echo off +@setlocal enabledelayedexpansion +@chcp 936 + +@REM 设置版本号 +set "VERSION=1.0" + +title 麦麦Bot控制台 v%VERSION% + +@REM 设置Python和Git环境变量 +set "_root=%~dp0" +set "_root=%_root:~0,-1%" +cd "%_root%" + + +:search_python +cls +if exist "%_root%\python" ( + set "PYTHON_HOME=%_root%\python" +) else if exist "%_root%\venv" ( + call "%_root%\venv\Scripts\activate.bat" + set "PYTHON_HOME=%_root%\venv\Scripts" +) else ( + echo 正在自动查找Python解释器... + + where python >nul 2>&1 + if %errorlevel% equ 0 ( + for /f "delims=" %%i in ('where python') do ( + echo %%i | findstr /i /c:"!LocalAppData!\Microsoft\WindowsApps\python.exe" >nul + if errorlevel 1 ( + echo 找到Python解释器:%%i + set "py_path=%%i" + goto :validate_python + ) + ) + ) + set "search_paths=%ProgramFiles%\Git*;!LocalAppData!\Programs\Python\Python*" + for /d %%d in (!search_paths!) do ( + if exist "%%d\python.exe" ( + set "py_path=%%d\python.exe" + goto :validate_python + ) + ) + echo 没有找到Python解释器,要安装吗? + set /p pyinstall_confirm="继续?(Y/n): " + if /i "!pyinstall_confirm!"=="Y" ( + cls + echo 正在安装Python... + winget install --id Python.Python.3.13 -e --accept-package-agreements --accept-source-agreements + if %errorlevel% neq 0 ( + echo 安装失败,请手动安装Python + start https://www.python.org/downloads/ + exit /b + ) + echo 安装完成,正在验证Python... + goto search_python + + ) else ( + echo 取消安装Python,按任意键退出... + pause >nul + exit /b + ) + + echo 错误:未找到可用的Python解释器! + exit /b 1 + + :validate_python + "!py_path!" --version >nul 2>&1 + if %errorlevel% neq 0 ( + echo 无效的Python解释器:%py_path% + exit /b 1 + ) + + :: 提取安装目录 + for %%i in ("%py_path%") do set "PYTHON_HOME=%%~dpi" + set "PYTHON_HOME=%PYTHON_HOME:~0,-1%" +) +if not exist "%PYTHON_HOME%\python.exe" ( + echo Python路径验证失败:%PYTHON_HOME% + echo 请检查Python安装路径中是否有python.exe文件 + exit /b 1 +) +echo 成功设置Python路径:%PYTHON_HOME% + + + +:search_git +cls +if exist "%_root%\tools\git\bin" ( + set "GIT_HOME=%_root%\tools\git\bin" +) else ( + echo 正在自动查找Git... + + where git >nul 2>&1 + if %errorlevel% equ 0 ( + for /f "delims=" %%i in ('where git') do ( + set "git_path=%%i" + goto :validate_git + ) + ) + echo 正在扫描常见安装路径... + set "search_paths=!ProgramFiles!\Git\cmd" + for /f "tokens=*" %%d in ("!search_paths!") do ( + if exist "%%d\git.exe" ( + set "git_path=%%d\git.exe" + goto :validate_git + ) + ) + echo 没有找到Git,要安装吗? + set /p confirm="继续?(Y/N): " + if /i "!confirm!"=="Y" ( + cls + echo 正在安装Git... + set "custom_url=https://ghfast.top/https://github.com/git-for-windows/git/releases/download/v2.48.1.windows.1/Git-2.48.1-64-bit.exe" + + set "download_path=%TEMP%\Git-Installer.exe" + + echo 正在下载Git安装包... + curl -L -o "!download_path!" "!custom_url!" + + if exist "!download_path!" ( + echo 下载成功,开始安装Git... + start /wait "" "!download_path!" /SILENT /NORESTART + ) else ( + echo 下载失败,请手动安装Git + start https://git-scm.com/download/win + exit /b + ) + + del "!download_path!" + echo 临时文件已清理。 + + echo 安装完成,正在验证Git... + where git >nul 2>&1 + if %errorlevel% equ 0 ( + for /f "delims=" %%i in ('where git') do ( + set "git_path=%%i" + goto :validate_git + ) + goto :search_git + + ) else ( + echo 安装完成,但未找到Git,请手动安装Git + start https://git-scm.com/download/win + exit /b + ) + + ) else ( + echo 取消安装Git,按任意键退出... + pause >nul + exit /b + ) + + echo 错误:未找到可用的Git! + exit /b 1 + + :validate_git + "%git_path%" --version >nul 2>&1 + if %errorlevel% neq 0 ( + echo 无效的Git:%git_path% + exit /b 1 + ) + + :: 提取安装目录 + for %%i in ("%git_path%") do set "GIT_HOME=%%~dpi" + set "GIT_HOME=%GIT_HOME:~0,-1%" +) + +:search_mongodb +cls +sc query | findstr /i "MongoDB" >nul +if !errorlevel! neq 0 ( + echo MongoDB服务未运行,是否尝试运行服务? + set /p confirm="是否启动?(Y/N): " + if /i "!confirm!"=="Y" ( + echo 正在尝试启动MongoDB服务... + powershell -Command "Start-Process -Verb RunAs cmd -ArgumentList '/c net start MongoDB'" + echo 正在等待MongoDB服务启动... + echo 按下任意键跳过等待... + timeout /t 30 >nul + sc query | findstr /i "MongoDB" >nul + if !errorlevel! neq 0 ( + echo MongoDB服务启动失败,可能是没有安装,要安装吗? + set /p install_confirm="继续安装?(Y/N): " + if /i "!install_confirm!"=="Y" ( + echo 正在安装MongoDB... + winget install --id MongoDB.Server -e --accept-package-agreements --accept-source-agreements + echo 安装完成,正在启动MongoDB服务... + net start MongoDB + if !errorlevel! neq 0 ( + echo 启动MongoDB服务失败,请手动启动 + exit /b + ) else ( + echo MongoDB服务已成功启动 + ) + ) else ( + echo 取消安装MongoDB,按任意键退出... + pause >nul + exit /b + ) + ) + ) else ( + echo "警告:MongoDB服务未运行,将导致MaiMBot无法访问数据库!" + ) +) else ( + echo MongoDB服务已运行 +) + +@REM set "GIT_HOME=%_root%\tools\git\bin" +set "PATH=%PYTHON_HOME%;%GIT_HOME%;%PATH%" + +:install_maim +if not exist "!_root!\bot.py" ( + cls + echo 你似乎没有安装麦麦Bot,要安装在当前目录吗? + set /p confirm="继续?(Y/N): " + if /i "!confirm!"=="Y" ( + echo 要使用Git代理下载吗? + set /p proxy_confirm="继续?(Y/N): " + if /i "!proxy_confirm!"=="Y" ( + echo 正在安装麦麦Bot... + git clone https://ghfast.top/https://github.com/SengokuCola/MaiMBot + ) else ( + echo 正在安装麦麦Bot... + git clone https://github.com/SengokuCola/MaiMBot + ) + xcopy /E /H /I MaiMBot . >nul 2>&1 + rmdir /s /q MaiMBot + git checkout main-fix + + echo 安装完成,正在安装依赖... + python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple + python -m pip install virtualenv + python -m virtualenv venv + call venv\Scripts\activate.bat + python -m pip install -r requirements.txt + + echo 安装完成,要编辑配置文件吗? + set /p edit_confirm="继续?(Y/N): " + if /i "!edit_confirm!"=="Y" ( + goto config_menu + ) else ( + echo 取消编辑配置文件,按任意键返回主菜单... + ) + ) +) + + +@REM git获取当前分支名并保存在变量里 +for /f "delims=" %%b in ('git symbolic-ref --short HEAD 2^>nul') do ( + set "BRANCH=%%b" +) + +@REM 根据不同分支名给分支名字符串使用不同颜色 +echo 分支名: %BRANCH% +if "!BRANCH!"=="main" ( + set "BRANCH_COLOR=" +) else if "!BRANCH!"=="main-fix" ( + set "BRANCH_COLOR=" +@REM ) else if "%BRANCH%"=="stable-dev" ( +@REM set "BRANCH_COLOR=" +) else ( + set "BRANCH_COLOR=" +) + +@REM endlocal & set "BRANCH_COLOR=%BRANCH_COLOR%" + +:check_is_venv +echo 正在检查虚拟环境状态... +if exist "%_root%\config\no_venv" ( + echo 检测到no_venv,跳过虚拟环境检查 + goto menu +) + +:: 环境检测 +if defined VIRTUAL_ENV ( + goto menu +) + +echo ===================================== +echo 虚拟环境检测警告: +echo 当前使用系统Python路径:!PYTHON_HOME! +echo 未检测到激活的虚拟环境! + +:env_interaction +echo ===================================== +echo 请选择操作: +echo 1 - 创建并激活Venv虚拟环境 +echo 2 - 创建/激活Conda虚拟环境 +echo 3 - 临时跳过本次检查 +echo 4 - 永久跳过虚拟环境检查 +set /p choice="请输入选项(1-4): " + +if "!choice!" = "4" ( + echo 要永久跳过虚拟环境检查吗? + set /p no_venv_confirm="继续?(Y/N): ....." + if /i "!no_venv_confirm!"=="Y" ( + echo 1 > "%_root%\config\no_venv" + echo 已创建no_venv文件 + pause >nul + goto menu + ) else ( + echo 取消跳过虚拟环境检查,按任意键返回... + pause >nul + goto env_interaction + ) +) + +if "!choice!" = "3"( + echo 警告:使用系统环境可能导致依赖冲突! + timeout /t 2 >nul + goto menu +) + +if "!choice!" = "2" goto handle_conda +if "!choice!" = "1" goto handle_venv + +echo 无效的输入,请输入1-4之间的数字 +timeout /t 2 >nul +goto env_interaction + +:handle_venv +python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple +echo 正在初始化Venv环境... +python -m pip install virtualenv || ( + echo 安装环境失败,错误码:!errorlevel! + pause + goto env_interaction +) +echo 创建虚拟环境到:venv + python -m virtualenv venv || ( + echo 环境创建失败,错误码:!errorlevel! + pause + goto env_interaction +) + +call venv\Scripts\activate.bat +echo 已激活Venv环境 +echo 要安装依赖吗? +set /p install_confirm="继续?(Y/N): " +if /i "!install_confirm!"=="Y" ( + goto update_dependencies +) +goto menu + +:handle_conda +where conda >nul 2>&1 || ( + echo 未检测到conda,可能原因: + echo 1. 未安装Miniconda + echo 2. conda配置异常 + timeout /t 10 >nul + goto env_interaction +) + +:conda_menu +echo 请选择Conda操作: +echo 1 - 创建新环境 +echo 2 - 激活已有环境 +echo 3 - 返回上级菜单 +set /p choice="请输入选项(1-3): " + +if "!choice!"=="3" goto env_interaction +if "!choice!"=="2" goto activate_conda +if "!choice!"=="1" goto create_conda + +:create_conda +set /p "CONDA_ENV=请输入新环境名称:" +if "!CONDA_ENV!"=="" ( + echo 环境名称不能为空! + goto create_conda +) +conda create -n !CONDA_ENV! python=3.13 -y || ( + echo 环境创建失败,错误码:!errorlevel! + pause + goto conda_menu +) +goto activate_conda + +:activate_conda +set /p "CONDA_ENV=请输入要激活的环境名称:" +conda activate !CONDA_ENV! || ( + echo 激活失败,可能原因: + echo 1. 环境不存在 + echo 2. conda配置异常 + pause + goto conda_menu +) +echo 成功激活conda环境:!CONDA_ENV! +echo 要安装依赖吗? +set /p install_confirm="继续?(Y/N): " +if /i "!install_confirm!"=="Y" ( + goto update_dependencies +) +:menu +@chcp 936 +cls +echo 麦麦Bot控制台 v%VERSION% 当前分支: %BRANCH_COLOR%%BRANCH% +echo 当前Python环境: !PYTHON_HOME! +echo ====================== +echo 1. 更新并启动麦麦Bot (默认) +echo 2. 直接启动麦麦Bot +echo 3. 启动麦麦配置界面 +echo 4. 打开麦麦神奇工具箱 +echo 5. 退出 +echo ====================== + +set /p choice="请输入选项数字 (1-5)并按下回车以选择: " + +if "!choice!"=="" set choice=1 + +if "!choice!"=="1" goto update_and_start +if "!choice!"=="2" goto start_bot +if "!choice!"=="3" goto config_menu +if "!choice!"=="4" goto tools_menu +if "!choice!"=="5" exit /b + +echo 无效的输入,请输入1-5之间的数字 +timeout /t 2 >nul +goto menu + +:config_menu +@chcp 936 +cls +if not exist config/bot_config.toml ( + copy /Y "template\bot_config_template.toml" "config\bot_config.toml" + +) +if not exist .env.prod ( + copy /Y "template\.env.prod" ".env.prod" +) + +start python webui.py + +goto menu + + +:tools_menu +@chcp 936 +cls +echo 麦麦时尚工具箱 当前分支: %BRANCH_COLOR%%BRANCH% +echo ====================== +echo 1. 更新依赖 +echo 2. 切换分支 +echo 3. 重置当前分支 +echo 4. 更新配置文件 +echo 5. 学习新的知识库 +echo 6. 打开知识库文件夹 +echo 7. 返回主菜单 +echo ====================== + +set /p choice="请输入选项数字: " +if "!choice!"=="1" goto update_dependencies +if "!choice!"=="2" goto switch_branch +if "!choice!"=="3" goto reset_branch +if "!choice!"=="4" goto update_config +if "!choice!"=="5" goto learn_new_knowledge +if "!choice!"=="6" goto open_knowledge_folder +if "!choice!"=="7" goto menu + +echo 无效的输入,请输入1-6之间的数字 +timeout /t 2 >nul +goto tools_menu + +:update_dependencies +cls +echo 正在更新依赖... +python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple +python.exe -m pip install -r requirements.txt + +echo 依赖更新完成,按任意键返回工具箱菜单... +pause +goto tools_menu + +:switch_branch +cls +echo 正在切换分支... +echo 当前分支: %BRANCH% +@REM echo 可用分支: main, debug, stable-dev +echo 1. 切换到main +echo 2. 切换到main-fix +echo 请输入要切换到的分支: +set /p branch_name="分支名: " +if "%branch_name%"=="" set branch_name=main +if "%branch_name%"=="main" ( + set "BRANCH_COLOR=" +) else if "%branch_name%"=="main-fix" ( + set "BRANCH_COLOR=" +@REM ) else if "%branch_name%"=="stable-dev" ( +@REM set "BRANCH_COLOR=" +) else if "%branch_name%"=="1" ( + set "BRANCH_COLOR=" + set "branch_name=main" +) else if "%branch_name%"=="2" ( + set "BRANCH_COLOR=" + set "branch_name=main-fix" +) else ( + echo 无效的分支名, 请重新输入 + timeout /t 2 >nul + goto switch_branch +) + +echo 正在切换到分支 %branch_name%... +git checkout %branch_name% +echo 分支切换完成,当前分支: %BRANCH_COLOR%%branch_name% +set "BRANCH=%branch_name%" +echo 按任意键返回工具箱菜单... +pause >nul +goto tools_menu + + +:reset_branch +cls +echo 正在重置当前分支... +echo 当前分支: !BRANCH! +echo 确认要重置当前分支吗? +set /p confirm="继续?(Y/N): " +if /i "!confirm!"=="Y" ( + echo 正在重置当前分支... + git reset --hard !BRANCH! + echo 分支重置完成,按任意键返回工具箱菜单... +) else ( + echo 取消重置当前分支,按任意键返回工具箱菜单... +) +pause >nul +goto tools_menu + + +:update_config +cls +echo 正在更新配置文件... +echo 请确保已备份重要数据,继续将修改当前配置文件。 +echo 继续请按Y,取消请按任意键... +set /p confirm="继续?(Y/N): " +if /i "!confirm!"=="Y" ( + echo 正在更新配置文件... + python.exe config\auto_update.py + echo 配置文件更新完成,按任意键返回工具箱菜单... +) else ( + echo 取消更新配置文件,按任意键返回工具箱菜单... +) +pause >nul +goto tools_menu + +:learn_new_knowledge +cls +echo 正在学习新的知识库... +echo 请确保已备份重要数据,继续将修改当前知识库。 +echo 继续请按Y,取消请按任意键... +set /p confirm="继续?(Y/N): " +if /i "!confirm!"=="Y" ( + echo 正在学习新的知识库... + python.exe src\plugins\zhishi\knowledge_library.py + echo 学习完成,按任意键返回工具箱菜单... +) else ( + echo 取消学习新的知识库,按任意键返回工具箱菜单... +) +pause >nul +goto tools_menu + +:open_knowledge_folder +cls +echo 正在打开知识库文件夹... +if exist data\raw_info ( + start explorer data\raw_info +) else ( + echo 知识库文件夹不存在! + echo 正在创建文件夹... + mkdir data\raw_info + timeout /t 2 >nul +) +goto tools_menu + + +:update_and_start +cls +:retry_git_pull +git pull > temp.log 2>&1 +findstr /C:"detected dubious ownership" temp.log >nul +if %errorlevel% equ 0 ( + echo 检测到仓库权限问题,正在自动修复... + git config --global --add safe.directory "%cd%" + echo 已添加例外,正在重试git pull... + del temp.log + goto retry_git_pull +) +del temp.log +echo 正在更新依赖... +python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple +python -m pip install -r requirements.txt && cls + +echo 当前代理设置: +echo HTTP_PROXY=%HTTP_PROXY% +echo HTTPS_PROXY=%HTTPS_PROXY% + +echo Disable Proxy... +set HTTP_PROXY= +set HTTPS_PROXY= +set no_proxy=0.0.0.0/32 + +REM chcp 65001 +python bot.py +echo. +echo Bot已停止运行,按任意键返回主菜单... +pause >nul +goto menu + +:start_bot +cls +echo 正在更新依赖... +python -m pip config set global.index-url https://mirrors.aliyun.com/pypi/simple +python -m pip install -r requirements.txt && cls + +echo 当前代理设置: +echo HTTP_PROXY=%HTTP_PROXY% +echo HTTPS_PROXY=%HTTPS_PROXY% + +echo Disable Proxy... +set HTTP_PROXY= +set HTTPS_PROXY= +set no_proxy=0.0.0.0/32 + +REM chcp 65001 +python bot.py +echo. +echo Bot已停止运行,按任意键返回主菜单... +pause >nul +goto menu + + +:open_dir +start explorer "%cd%" +goto menu From e9bd3196ba1f6f393d6eb97ebd6d6df5c27cb388 Mon Sep 17 00:00:00 2001 From: UnCLAS-Prommer Date: Thu, 20 Mar 2025 16:47:50 +0800 Subject: [PATCH 02/21] =?UTF-8?q?=E6=AD=A3=E7=A1=AE=E4=BF=9D=E5=AD=98?= =?UTF-8?q?=E6=A8=A1=E5=9E=8B=E5=90=8D=E7=A7=B0=E5=88=B0Database?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/llm_generator.py | 5 +++-- src/plugins/models/utils_model.py | 2 +- 2 files changed, 4 insertions(+), 3 deletions(-) diff --git a/src/plugins/chat/llm_generator.py b/src/plugins/chat/llm_generator.py index bcd0b9e87..73cd12ed7 100644 --- a/src/plugins/chat/llm_generator.py +++ b/src/plugins/chat/llm_generator.py @@ -37,6 +37,7 @@ class ResponseGenerator: self.model_r1_distill = LLM_request(model=global_config.llm_reasoning_minor, temperature=0.7, max_tokens=3000) self.model_v25 = LLM_request(model=global_config.llm_normal_minor, temperature=0.7, max_tokens=3000) self.current_model_type = "r1" # 默认使用 R1 + self.current_model_name = "unknown model" async def generate_response(self, message: MessageThinking) -> Optional[Union[str, List[str]]]: """根据当前模型类型选择对应的生成函数""" @@ -107,7 +108,7 @@ class ResponseGenerator: # 生成回复 try: - content, reasoning_content = await model.generate_response(prompt) + content, reasoning_content, self.current_model_name = await model.generate_response(prompt) except Exception: logger.exception("生成回复时出错") return None @@ -144,7 +145,7 @@ class ResponseGenerator: "chat_id": message.chat_stream.stream_id, "user": sender_name, "message": message.processed_plain_text, - "model": self.current_model_type, + "model": self.current_model_name, # 'reasoning_check': reasoning_content_check, # 'response_check': content_check, "reasoning": reasoning_content, diff --git a/src/plugins/models/utils_model.py b/src/plugins/models/utils_model.py index d915b3759..ba85a1bd2 100644 --- a/src/plugins/models/utils_model.py +++ b/src/plugins/models/utils_model.py @@ -526,7 +526,7 @@ class LLM_request: """根据输入的提示生成模型的异步响应""" content, reasoning_content = await self._execute_request(endpoint="/chat/completions", prompt=prompt) - return content, reasoning_content + return content, reasoning_content, self.model_name async def generate_response_for_image(self, prompt: str, image_base64: str, image_format: str) -> Tuple[str, str]: """根据输入的提示和图片生成模型的异步响应""" From a9730575792db81b95afd97a105156f9df147ced Mon Sep 17 00:00:00 2001 From: UnCLAS-Prommer Date: Thu, 20 Mar 2025 17:08:53 +0800 Subject: [PATCH 03/21] =?UTF-8?q?=E5=AE=8C=E6=88=90=E5=85=B6=E4=BB=96?= =?UTF-8?q?=E9=83=A8=E5=88=86=E5=AF=B9=E8=BF=94=E5=9B=9E=E5=80=BC=E5=A4=84?= =?UTF-8?q?=E7=90=86=E7=9A=84=E4=BF=AE=E6=94=B9?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/llm_generator.py | 6 +++--- src/plugins/chat/topic_identifier.py | 2 +- src/plugins/models/utils_model.py | 2 +- src/plugins/schedule/schedule_generator.py | 2 +- 4 files changed, 6 insertions(+), 6 deletions(-) diff --git a/src/plugins/chat/llm_generator.py b/src/plugins/chat/llm_generator.py index 73cd12ed7..80daa250b 100644 --- a/src/plugins/chat/llm_generator.py +++ b/src/plugins/chat/llm_generator.py @@ -175,7 +175,7 @@ class ResponseGenerator: """ # 调用模型生成结果 - result, _ = await self.model_v25.generate_response(prompt) + result, _, _ = await self.model_v25.generate_response(prompt) result = result.strip() # 解析模型输出的结果 @@ -216,7 +216,7 @@ class InitiativeMessageGenerate: topic_select_prompt, dots_for_select, prompt_template = prompt_builder._build_initiative_prompt_select( message.group_id ) - content_select, reasoning = self.model_v3.generate_response(topic_select_prompt) + content_select, reasoning, _ = self.model_v3.generate_response(topic_select_prompt) logger.debug(f"{content_select} {reasoning}") topics_list = [dot[0] for dot in dots_for_select] if content_select: @@ -227,7 +227,7 @@ class InitiativeMessageGenerate: else: return None prompt_check, memory = prompt_builder._build_initiative_prompt_check(select_dot[1], prompt_template) - content_check, reasoning_check = self.model_v3.generate_response(prompt_check) + content_check, reasoning_check, _ = self.model_v3.generate_response(prompt_check) logger.info(f"{content_check} {reasoning_check}") if "yes" not in content_check.lower(): return None diff --git a/src/plugins/chat/topic_identifier.py b/src/plugins/chat/topic_identifier.py index c87c37155..6e11bc9d7 100644 --- a/src/plugins/chat/topic_identifier.py +++ b/src/plugins/chat/topic_identifier.py @@ -33,7 +33,7 @@ class TopicIdentifier: 消息内容:{text}""" # 使用 LLM_request 类进行请求 - topic, _ = await self.llm_topic_judge.generate_response(prompt) + topic, _, _ = await self.llm_topic_judge.generate_response(prompt) if not topic: logger.error("LLM API 返回为空") diff --git a/src/plugins/models/utils_model.py b/src/plugins/models/utils_model.py index ba85a1bd2..91e43fd4f 100644 --- a/src/plugins/models/utils_model.py +++ b/src/plugins/models/utils_model.py @@ -522,7 +522,7 @@ class LLM_request: return {"Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json"} # 防止小朋友们截图自己的key - async def generate_response(self, prompt: str) -> Tuple[str, str]: + async def generate_response(self, prompt: str) -> Tuple[str, str, str]: """根据输入的提示生成模型的异步响应""" content, reasoning_content = await self._execute_request(endpoint="/chat/completions", prompt=prompt) diff --git a/src/plugins/schedule/schedule_generator.py b/src/plugins/schedule/schedule_generator.py index fe9f77b90..11db6664d 100644 --- a/src/plugins/schedule/schedule_generator.py +++ b/src/plugins/schedule/schedule_generator.py @@ -73,7 +73,7 @@ class ScheduleGenerator: ) try: - schedule_text, _ = await self.llm_scheduler.generate_response(prompt) + schedule_text, _, _ = await self.llm_scheduler.generate_response(prompt) db.schedule.insert_one({"date": date_str, "schedule": schedule_text}) self.enable_output = True except Exception as e: From 3cda0fa74555b1238726f2d42e5b1c9db902257b Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Thu, 20 Mar 2025 19:38:10 +0800 Subject: [PATCH 04/21] =?UTF-8?q?WebUI=E5=A2=9E=E5=8A=A0=E5=9B=9E=E5=A4=8D?= =?UTF-8?q?=E6=84=8F=E6=84=BF=E6=A8=A1=E5=BC=8F=E9=80=89=E6=8B=A9=E5=8A=9F?= =?UTF-8?q?=E8=83=BD?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- webui.py | 205 ++++++++++++++++++++++++++++++++----------------------- 1 file changed, 118 insertions(+), 87 deletions(-) diff --git a/webui.py b/webui.py index 86215b745..ca87613f5 100644 --- a/webui.py +++ b/webui.py @@ -66,6 +66,16 @@ else: HAVE_ONLINE_STATUS_VERSION = version.parse("0.0.9") +#定义意愿模式可选项 +WILLING_MODE_CHOICES = [ + "classical", + "dynamic", + "custom", +] + + + + #添加WebUI配置文件版本 WEBUI_VERSION = version.parse("0.0.9") @@ -321,19 +331,19 @@ def format_list_to_str(lst): # env保存函数 def save_trigger( - server_address, - server_port, - final_result_list, - t_mongodb_host, - t_mongodb_port, - t_mongodb_database_name, - t_console_log_level, - t_file_log_level, - t_default_console_log_level, - t_default_file_log_level, - t_api_provider, - t_api_base_url, - t_api_key, + server_address, + server_port, + final_result_list, + t_mongodb_host, + t_mongodb_port, + t_mongodb_database_name, + t_console_log_level, + t_file_log_level, + t_default_console_log_level, + t_default_file_log_level, + t_api_provider, + t_api_base_url, + t_api_key, ): final_result_lists = format_list_to_str(final_result_list) env_config_data["env_HOST"] = server_address @@ -402,12 +412,12 @@ def save_bot_config(t_qqbot_qq, t_nickname, t_nickname_final_result): # 监听滑块的值变化,确保总和不超过 1,并显示警告 def adjust_personality_greater_probabilities( - t_personality_1_probability, t_personality_2_probability, t_personality_3_probability + t_personality_1_probability, t_personality_2_probability, t_personality_3_probability ): total = ( - Decimal(str(t_personality_1_probability)) - + Decimal(str(t_personality_2_probability)) - + Decimal(str(t_personality_3_probability)) + Decimal(str(t_personality_1_probability)) + + Decimal(str(t_personality_2_probability)) + + Decimal(str(t_personality_3_probability)) ) if total > Decimal("1.0"): warning_message = ( @@ -418,12 +428,12 @@ def adjust_personality_greater_probabilities( def adjust_personality_less_probabilities( - t_personality_1_probability, t_personality_2_probability, t_personality_3_probability + t_personality_1_probability, t_personality_2_probability, t_personality_3_probability ): total = ( - Decimal(str(t_personality_1_probability)) - + Decimal(str(t_personality_2_probability)) - + Decimal(str(t_personality_3_probability)) + Decimal(str(t_personality_1_probability)) + + Decimal(str(t_personality_2_probability)) + + Decimal(str(t_personality_3_probability)) ) if total < Decimal("1.0"): warning_message = ( @@ -435,7 +445,7 @@ def adjust_personality_less_probabilities( def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) ) if total > Decimal("1.0"): warning_message = ( @@ -447,7 +457,7 @@ def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probabil def adjust_model_less_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) ) if total < Decimal("1.0"): warning_message = ( @@ -460,13 +470,13 @@ def adjust_model_less_probabilities(t_model_1_probability, t_model_2_probability # ============================================== # 人格保存函数 def save_personality_config( - t_prompt_personality_1, - t_prompt_personality_2, - t_prompt_personality_3, - t_prompt_schedule, - t_personality_1_probability, - t_personality_2_probability, - t_personality_3_probability, + t_prompt_personality_1, + t_prompt_personality_2, + t_prompt_personality_3, + t_prompt_schedule, + t_personality_1_probability, + t_personality_2_probability, + t_personality_3_probability, ): # 保存人格提示词 config_data["personality"]["prompt_personality"][0] = t_prompt_personality_1 @@ -487,20 +497,20 @@ def save_personality_config( def save_message_and_emoji_config( - t_min_text_length, - t_max_context_size, - t_emoji_chance, - t_thinking_timeout, - t_response_willing_amplifier, - t_response_interested_rate_amplifier, - t_down_frequency_rate, - t_ban_words_final_result, - t_ban_msgs_regex_final_result, - t_check_interval, - t_register_interval, - t_auto_save, - t_enable_check, - t_check_prompt, + t_min_text_length, + t_max_context_size, + t_emoji_chance, + t_thinking_timeout, + t_response_willing_amplifier, + t_response_interested_rate_amplifier, + t_down_frequency_rate, + t_ban_words_final_result, + t_ban_msgs_regex_final_result, + t_check_interval, + t_register_interval, + t_auto_save, + t_enable_check, + t_check_prompt, ): config_data["message"]["min_text_length"] = t_min_text_length config_data["message"]["max_context_size"] = t_max_context_size @@ -522,27 +532,30 @@ def save_message_and_emoji_config( def save_response_model_config( - t_model_r1_probability, - t_model_r2_probability, - t_model_r3_probability, - t_max_response_length, - t_model1_name, - t_model1_provider, - t_model1_pri_in, - t_model1_pri_out, - t_model2_name, - t_model2_provider, - t_model3_name, - t_model3_provider, - t_emotion_model_name, - t_emotion_model_provider, - t_topic_judge_model_name, - t_topic_judge_model_provider, - t_summary_by_topic_model_name, - t_summary_by_topic_model_provider, - t_vlm_model_name, - t_vlm_model_provider, + t_willing_mode, + t_model_r1_probability, + t_model_r2_probability, + t_model_r3_probability, + t_max_response_length, + t_model1_name, + t_model1_provider, + t_model1_pri_in, + t_model1_pri_out, + t_model2_name, + t_model2_provider, + t_model3_name, + t_model3_provider, + t_emotion_model_name, + t_emotion_model_provider, + t_topic_judge_model_name, + t_topic_judge_model_provider, + t_summary_by_topic_model_name, + t_summary_by_topic_model_provider, + t_vlm_model_name, + t_vlm_model_provider, ): + if PARSED_CONFIG_VERSION >= version.parse("0.0.10"): + config_data["willing"]["willing_mode"] = t_willing_mode config_data["response"]["model_r1_probability"] = t_model_r1_probability config_data["response"]["model_v3_probability"] = t_model_r2_probability config_data["response"]["model_r1_distill_probability"] = t_model_r3_probability @@ -569,15 +582,15 @@ def save_response_model_config( def save_memory_mood_config( - t_build_memory_interval, - t_memory_compress_rate, - t_forget_memory_interval, - t_memory_forget_time, - t_memory_forget_percentage, - t_memory_ban_words_final_result, - t_mood_update_interval, - t_mood_decay_rate, - t_mood_intensity_factor, + t_build_memory_interval, + t_memory_compress_rate, + t_forget_memory_interval, + t_memory_forget_time, + t_memory_forget_percentage, + t_memory_ban_words_final_result, + t_mood_update_interval, + t_mood_decay_rate, + t_mood_intensity_factor, ): config_data["memory"]["build_memory_interval"] = t_build_memory_interval config_data["memory"]["memory_compress_rate"] = t_memory_compress_rate @@ -594,17 +607,17 @@ def save_memory_mood_config( def save_other_config( - t_keywords_reaction_enabled, - t_enable_advance_output, - t_enable_kuuki_read, - t_enable_debug_output, - t_enable_friend_chat, - t_chinese_typo_enabled, - t_error_rate, - t_min_freq, - t_tone_error_rate, - t_word_replace_rate, - t_remote_status, + t_keywords_reaction_enabled, + t_enable_advance_output, + t_enable_kuuki_read, + t_enable_debug_output, + t_enable_friend_chat, + t_chinese_typo_enabled, + t_error_rate, + t_min_freq, + t_tone_error_rate, + t_word_replace_rate, + t_remote_status, ): config_data["keywords_reaction"]["enable"] = t_keywords_reaction_enabled config_data["others"]["enable_advance_output"] = t_enable_advance_output @@ -624,9 +637,9 @@ def save_other_config( def save_group_config( - t_talk_allowed_final_result, - t_talk_frequency_down_final_result, - t_ban_user_id_final_result, + t_talk_allowed_final_result, + t_talk_frequency_down_final_result, + t_ban_user_id_final_result, ): config_data["groups"]["talk_allowed"] = t_talk_allowed_final_result config_data["groups"]["talk_frequency_down"] = t_talk_frequency_down_final_result @@ -1182,6 +1195,23 @@ with gr.Blocks(title="MaimBot配置文件编辑") as app: with gr.Column(scale=3): with gr.Row(): gr.Markdown("""### 回复设置""") + if PARSED_CONFIG_VERSION >= version.parse("0.0.10"): + with gr.Row(): + gr.Markdown("""#### 回复意愿模式""") + with gr.Row(): + gr.Markdown("""回复意愿模式说明:\n + classical为经典回复意愿管理器\n + dynamic为动态意愿管理器\n + custom为自定义意愿管理器 + """) + with gr.Row(): + willing_mode = gr.Dropdown( + choices=WILLING_MODE_CHOICES, + value=config_data["willing"]["willing_mode"], + label="回复意愿模式" + ) + else: + willing_mode = gr.Textbox(visible=False,value="disabled") with gr.Row(): model_r1_probability = gr.Slider( minimum=0, @@ -1355,6 +1385,7 @@ with gr.Blocks(title="MaimBot配置文件编辑") as app: save_model_btn.click( save_response_model_config, inputs=[ + willing_mode, model_r1_probability, model_r2_probability, model_r3_probability, From c0400aeb41621a0b69d77f8792435389b99016b5 Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Thu, 20 Mar 2025 19:40:31 +0800 Subject: [PATCH 05/21] =?UTF-8?q?=E8=BF=87Ruff=E6=A3=80=E6=B5=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- webui.py | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/webui.py b/webui.py index ca87613f5..b598df7c0 100644 --- a/webui.py +++ b/webui.py @@ -445,7 +445,9 @@ def adjust_personality_less_probabilities( def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + + Decimal(str(t_model_2_probability)) + + Decimal(str(t_model_3_probability)) ) if total > Decimal("1.0"): warning_message = ( @@ -457,7 +459,9 @@ def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probabil def adjust_model_less_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + + Decimal(str(t_model_2_probability)) + + Decimal(str(t_model_3_probability)) ) if total < Decimal("1.0"): warning_message = ( From 7c5cdb82bc475449a02ffedbf91bf5e23ccbcef9 Mon Sep 17 00:00:00 2001 From: enKl03b Date: Thu, 20 Mar 2025 23:10:33 +0800 Subject: [PATCH 06/21] =?UTF-8?q?=E6=95=B4=E7=90=86doc=E6=96=87=E4=BB=B6?= =?UTF-8?q?=E5=A4=B9=EF=BC=8C=E6=B7=BB=E5=8A=A0macos=E6=95=99=E7=A8=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- README.md | 2 +- docs/fast_q_a.md | 8 +- docs/linux_deploy_guide_for_beginners.md | 220 +++++--------------- docs/manual_deploy_linux.md | 48 +++-- docs/manual_deploy_macos.md | 201 ++++++++++++++++++ docs/{ => pic}/API_KEY.png | Bin docs/{ => pic}/MONGO_DB_0.png | Bin docs/{ => pic}/MONGO_DB_1.png | Bin docs/{ => pic}/MONGO_DB_2.png | Bin docs/pic/MongoDB_Ubuntu_guide.png | Bin 0 -> 14733 bytes docs/pic/QQ_Download_guide_Linux.png | Bin 0 -> 37847 bytes docs/pic/linux_beginner_downloadguide.png | Bin 0 -> 10333 bytes docs/{ => pic}/synology_.env.prod.png | Bin docs/{ => pic}/synology_create_project.png | Bin docs/{ => pic}/synology_docker-compose.png | Bin docs/{ => pic}/synology_how_to_download.png | Bin docs/{ => pic}/video.png | Bin docs/synology_deploy.md | 8 +- 18 files changed, 297 insertions(+), 190 deletions(-) create mode 100644 docs/manual_deploy_macos.md rename docs/{ => pic}/API_KEY.png (100%) rename docs/{ => pic}/MONGO_DB_0.png (100%) rename docs/{ => pic}/MONGO_DB_1.png (100%) rename docs/{ => pic}/MONGO_DB_2.png (100%) create mode 100644 docs/pic/MongoDB_Ubuntu_guide.png create mode 100644 docs/pic/QQ_Download_guide_Linux.png create mode 100644 docs/pic/linux_beginner_downloadguide.png rename docs/{ => pic}/synology_.env.prod.png (100%) rename docs/{ => pic}/synology_create_project.png (100%) rename docs/{ => pic}/synology_docker-compose.png (100%) rename docs/{ => pic}/synology_how_to_download.png (100%) rename docs/{ => pic}/video.png (100%) diff --git a/README.md b/README.md index 8dea5bc15..b016a2569 100644 --- a/README.md +++ b/README.md @@ -101,7 +101,7 @@
- 麦麦演示视频 + 麦麦演示视频
👆 点击观看麦麦演示视频 👆 diff --git a/docs/fast_q_a.md b/docs/fast_q_a.md index 1f015565d..92800bad2 100644 --- a/docs/fast_q_a.md +++ b/docs/fast_q_a.md @@ -10,7 +10,7 @@ - 为什么显示:"缺失必要的API KEY" ❓ - + >你需要在 [Silicon Flow Api](https://cloud.siliconflow.cn/account/ak) 网站上注册一个账号,然后点击这个链接打开API KEY获取页面。 > @@ -41,19 +41,19 @@ >打开你的MongoDB Compass软件,你会在左上角看到这样的一个界面: > -> +> > >
> >点击 "CONNECT" 之后,点击展开 MegBot 标签栏 > -> +> > >
> >点进 "emoji" 再点击 "DELETE" 删掉所有条目,如图所示 > -> +> > >
> diff --git a/docs/linux_deploy_guide_for_beginners.md b/docs/linux_deploy_guide_for_beginners.md index 04601923f..ece0a3334 100644 --- a/docs/linux_deploy_guide_for_beginners.md +++ b/docs/linux_deploy_guide_for_beginners.md @@ -1,48 +1,51 @@ # 面向纯新手的Linux服务器麦麦部署指南 -## 你得先有一个服务器 -为了能使麦麦在你的电脑关机之后还能运行,你需要一台不间断开机的主机,也就是我们常说的服务器。 +## 事前准备 +为了能使麦麦不间断的运行,你需要一台一直开着的主机。 +### 如果你想购买服务器 华为云、阿里云、腾讯云等等都是在国内可以选择的选择。 -你可以去租一台最低配置的就足敷需要了,按月租大概十几块钱就能租到了。 +租一台最低配置的就足敷需要了,按月租大概十几块钱就能租到了。 -我们假设你已经租好了一台Linux架构的云服务器。我用的是阿里云ubuntu24.04,其他的原理相似。 +### 如果你不想购买服务器 +你可以准备一台可以一直开着的电脑/主机,只需要保证能够正常访问互联网即可 + +我们假设你已经有了一台Linux架构的服务器。举例使用的是Ubuntu24.04,其他的原理相似。 ## 0.我们就从零开始吧 ### 网络问题 -为访问github相关界面,推荐去下一款加速器,新手可以试试watttoolkit。 +为访问Github相关界面,推荐去下一款加速器,新手可以试试[Watt Toolkit](https://gitee.com/rmbgame/SteamTools/releases/latest)。 ### 安装包下载 #### MongoDB +进入[MongoDB下载页](https://www.mongodb.com/try/download/community-kubernetes-operator),并选择版本 -对于ubuntu24.04 x86来说是这个: +以Ubuntu24.04 x86为例,保持如图所示选项,点击`Download`即可,如果是其他系统,请在`Platform`中自行选择: -https://repo.mongodb.org/apt/ubuntu/dists/noble/mongodb-org/8.0/multiverse/binary-amd64/mongodb-org-server_8.0.5_amd64.deb +![](./pic/MongoDB_Ubuntu_guide.png) -如果不是就在这里自行选择对应版本 -https://www.mongodb.com/try/download/community-kubernetes-operator +不想使用上述方式?你也可以参考[官方文档](https://www.mongodb.com/zh-cn/docs/manual/administration/install-on-linux/#std-label-install-mdb-community-edition-linux)进行安装,进入后选择自己的系统版本即可 -#### Napcat - -在这里选择对应版本。 - -https://github.com/NapNeko/NapCatQQ/releases/tag/v4.6.7 - -对于ubuntu24.04 x86来说是这个: - -https://dldir1.qq.com/qqfile/qq/QQNT/ee4bd910/linuxqq_3.2.16-32793_amd64.deb +#### QQ(可选)/Napcat +*如果你使用Napcat的脚本安装,可以忽略此步* +访问https://github.com/NapNeko/NapCatQQ/releases/latest +在图中所示区域可以找到QQ的下载链接,选择对应版本下载即可 +从这里下载,可以保证你下载到的QQ版本兼容最新版Napcat +![](./pic/QQ_Download_guide_Linux.png) +如果你不想使用Napcat的脚本安装,还需参考[Napcat-Linux手动安装](https://www.napcat.wiki/guide/boot/Shell-Linux-SemiAuto) #### 麦麦 -https://github.com/SengokuCola/MaiMBot/archive/refs/tags/0.5.8-alpha.zip - -下载这个官方压缩包。 +先打开https://github.com/MaiM-with-u/MaiBot/releases +往下滑找到这个 +![下载指引](./pic/linux_beginner_downloadguide.png "") +下载箭头所指这个压缩包。 ### 路径 @@ -53,10 +56,10 @@ https://github.com/SengokuCola/MaiMBot/archive/refs/tags/0.5.8-alpha.zip ``` moi └─ mai - ├─ linuxqq_3.2.16-32793_amd64.deb - ├─ mongodb-org-server_8.0.5_amd64.deb + ├─ linuxqq_3.2.16-32793_amd64.deb # linuxqq安装包 + ├─ mongodb-org-server_8.0.5_amd64.deb # MongoDB的安装包 └─ bot - └─ MaiMBot-0.5.8-alpha.zip + └─ MaiMBot-0.5.8-alpha.zip # 麦麦的压缩包 ``` ### 网络 @@ -69,7 +72,7 @@ moi ## 2. Python的安装 -- 导入 Python 的稳定版 PPA: +- 导入 Python 的稳定版 PPA(Ubuntu需执行此步,Debian可忽略): ```bash sudo add-apt-repository ppa:deadsnakes/ppa @@ -92,6 +95,11 @@ sudo apt install python3.12 ```bash python3.12 --version ``` +- (可选)更新替代方案,设置 python3.12 为默认的 python3 版本: +```bash +sudo update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1 +sudo update-alternatives --config python3 +``` - 在「终端」中,执行以下命令安装 pip: @@ -141,23 +149,17 @@ systemctl status mongod #通过这条指令检查运行状态 sudo systemctl enable mongod ``` -## 5.napcat的安装 +## 5.Napcat的安装 ``` bash +# 该脚本适用于支持Ubuntu 20+/Debian 10+/Centos9 curl -o napcat.sh https://nclatest.znin.net/NapNeko/NapCat-Installer/main/script/install.sh && sudo bash napcat.sh ``` - -上面的不行试试下面的 - -``` bash -dpkg -i linuxqq_3.2.16-32793_amd64.deb -apt-get install -f -dpkg -i linuxqq_3.2.16-32793_amd64.deb -``` +执行后,脚本会自动帮你部署好QQ及Napcat 成功的标志是输入``` napcat ```出来炫酷的彩虹色界面 -## 6.napcat的运行 +## 6.Napcat的运行 此时你就可以根据提示在```napcat```里面登录你的QQ号了。 @@ -170,6 +172,13 @@ napcat status #检查运行状态 ```http://<你服务器的公网IP>:6099/webui?token=napcat``` +如果你部署在自己的电脑上: +```http://127.0.0.1:6099/webui?token=napcat``` + +> [!WARNING] +> 如果你的麦麦部署在公网,请**务必**修改Napcat的默认密码 + + 第一次是这个,后续改了密码之后token就会对应修改。你也可以使用```napcat log <你的QQ号>```来查看webui地址。把里面的```127.0.0.1```改成<你服务器的公网IP>即可。 登录上之后在网络配置界面添加websocket客户端,名称随便输一个,url改成`ws://127.0.0.1:8080/onebot/v11/ws`保存之后点启用,就大功告成了。 @@ -178,7 +187,7 @@ napcat status #检查运行状态 ### step 1 安装解压软件 -``` +```bash sudo apt-get install unzip ``` @@ -229,138 +238,11 @@ bot 你可以注册一个硅基流动的账号,通过邀请码注册有14块钱的免费额度:https://cloud.siliconflow.cn/i/7Yld7cfg。 -#### 在.env.prod中定义API凭证: +#### 修改配置文件 +请参考 +- [🎀 新手配置指南](./installation_cute.md) - 通俗易懂的配置教程,适合初次使用的猫娘 +- [⚙️ 标准配置指南](./installation_standard.md) - 简明专业的配置说明,适合有经验的用户 -``` -# API凭证配置 -SILICONFLOW_KEY=your_key # 硅基流动API密钥 -SILICONFLOW_BASE_URL=https://api.siliconflow.cn/v1/ # 硅基流动API地址 - -DEEP_SEEK_KEY=your_key # DeepSeek API密钥 -DEEP_SEEK_BASE_URL=https://api.deepseek.com/v1 # DeepSeek API地址 - -CHAT_ANY_WHERE_KEY=your_key # ChatAnyWhere API密钥 -CHAT_ANY_WHERE_BASE_URL=https://api.chatanywhere.tech/v1 # ChatAnyWhere API地址 -``` - -#### 在bot_config.toml中引用API凭证: - -``` -[model.llm_reasoning] -name = "Pro/deepseek-ai/DeepSeek-R1" -base_url = "SILICONFLOW_BASE_URL" # 引用.env.prod中定义的地址 -key = "SILICONFLOW_KEY" # 引用.env.prod中定义的密钥 -``` - -如需切换到其他API服务,只需修改引用: - -``` -[model.llm_reasoning] -name = "Pro/deepseek-ai/DeepSeek-R1" -base_url = "DEEP_SEEK_BASE_URL" # 切换为DeepSeek服务 -key = "DEEP_SEEK_KEY" # 使用DeepSeek密钥 -``` - -#### 配置文件详解 - -##### 环境配置文件 (.env.prod) - -``` -# API配置 -SILICONFLOW_KEY=your_key -SILICONFLOW_BASE_URL=https://api.siliconflow.cn/v1/ -DEEP_SEEK_KEY=your_key -DEEP_SEEK_BASE_URL=https://api.deepseek.com/v1 -CHAT_ANY_WHERE_KEY=your_key -CHAT_ANY_WHERE_BASE_URL=https://api.chatanywhere.tech/v1 - -# 服务配置 -HOST=127.0.0.1 # 如果使用Docker部署,需要改成0.0.0.0,否则QQ消息无法传入 -PORT=8080 - -# 数据库配置 -MONGODB_HOST=127.0.0.1 # 如果使用Docker部署,需要改成数据库容器的名字,默认是mongodb -MONGODB_PORT=27017 -DATABASE_NAME=MegBot -MONGODB_USERNAME = "" # 数据库用户名 -MONGODB_PASSWORD = "" # 数据库密码 -MONGODB_AUTH_SOURCE = "" # 认证数据库 - -# 插件配置 -PLUGINS=["src2.plugins.chat"] -``` - -##### 机器人配置文件 (bot_config.toml) - -``` -[bot] -qq = "机器人QQ号" # 必填 -nickname = "麦麦" # 机器人昵称(你希望机器人怎么称呼它自己) - -[personality] -prompt_personality = [ - "曾经是一个学习地质的女大学生,现在学习心理学和脑科学,你会刷贴吧", - "是一个女大学生,你有黑色头发,你会刷小红书" -] -prompt_schedule = "一个曾经学习地质,现在学习心理学和脑科学的女大学生,喜欢刷qq,贴吧,知乎和小红书" - -[message] -min_text_length = 2 # 最小回复长度 -max_context_size = 15 # 上下文记忆条数 -emoji_chance = 0.2 # 表情使用概率 -ban_words = [] # 禁用词列表 - -[emoji] -auto_save = true # 自动保存表情 -enable_check = false # 启用表情审核 -check_prompt = "符合公序良俗" - -[groups] -talk_allowed = [] # 允许对话的群号 -talk_frequency_down = [] # 降低回复频率的群号 -ban_user_id = [] # 禁止回复的用户QQ号 - -[others] -enable_advance_output = true # 启用详细日志 -enable_kuuki_read = true # 启用场景理解 - -# 模型配置 -[model.llm_reasoning] # 推理模型 -name = "Pro/deepseek-ai/DeepSeek-R1" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - -[model.llm_reasoning_minor] # 轻量推理模型 -name = "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - -[model.llm_normal] # 对话模型 -name = "Pro/deepseek-ai/DeepSeek-V3" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - -[model.llm_normal_minor] # 备用对话模型 -name = "deepseek-ai/DeepSeek-V2.5" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - -[model.vlm] # 图像识别模型 -name = "deepseek-ai/deepseek-vl2" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - -[model.embedding] # 文本向量模型 -name = "BAAI/bge-m3" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" - - -[topic.llm_topic] -name = "Pro/deepseek-ai/DeepSeek-V3" -base_url = "SILICONFLOW_BASE_URL" -key = "SILICONFLOW_KEY" -``` **step # 6** 运行 @@ -438,7 +320,7 @@ sudo systemctl enable bot.service # 启动bot服务 sudo systemctl status bot.service # 检查bot服务状态 ``` -``` -python bot.py +```python +python bot.py # 运行麦麦 ``` diff --git a/docs/manual_deploy_linux.md b/docs/manual_deploy_linux.md index a5c91d6e2..653284bf5 100644 --- a/docs/manual_deploy_linux.md +++ b/docs/manual_deploy_linux.md @@ -6,7 +6,7 @@ - QQ小号(QQ框架的使用可能导致qq被风控,严重(小概率)可能会导致账号封禁,强烈不推荐使用大号) - 可用的大模型API - 一个AI助手,网上随便搜一家打开来用都行,可以帮你解决一些不懂的问题 -- 以下内容假设你对Linux系统有一定的了解,如果觉得难以理解,请直接用Windows系统部署[Windows系统部署指南](./manual_deploy_windows.md) +- 以下内容假设你对Linux系统有一定的了解,如果觉得难以理解,请直接用Windows系统部署[Windows系统部署指南](./manual_deploy_windows.md)或[使用Windows一键包部署](https://github.com/MaiM-with-u/MaiBot/releases/tag/EasyInstall-windows) ## 你需要知道什么? @@ -24,6 +24,9 @@ --- +## 一键部署 +请下载并运行项目根目录中的run.sh并按照提示安装,部署完成后请参照后续配置指南进行配置 + ## 环境配置 ### 1️⃣ **确认Python版本** @@ -36,17 +39,26 @@ python --version python3 --version ``` -如果版本低于3.9,请更新Python版本。 +如果版本低于3.9,请更新Python版本,目前建议使用python3.12 ```bash -# Ubuntu/Debian +# Debian sudo apt update -sudo apt install python3.9 -# 如执行了这一步,建议在执行时将python3指向python3.9 -# 更新替代方案,设置 python3.9 为默认的 python3 版本: -sudo update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.9 1 +sudo apt install python3.12 +# Ubuntu +sudo add-apt-repository ppa:deadsnakes/ppa +sudo apt update +sudo apt install python3.12 + +# 执行完以上命令后,建议在执行时将python3指向python3.12 +# 更新替代方案,设置 python3.12 为默认的 python3 版本: +sudo update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1 sudo update-alternatives --config python3 ``` +建议再执行以下命令,使后续运行命令中的`python3`等同于`python` +```bash +sudo apt install python-is-python3 +``` ### 2️⃣ **创建虚拟环境** @@ -73,7 +85,7 @@ pip install -r requirements.txt ### 3️⃣ **安装并启动MongoDB** -- 安装与启动:Debian参考[官方文档](https://docs.mongodb.com/manual/tutorial/install-mongodb-on-debian/),Ubuntu参考[官方文档](https://docs.mongodb.com/manual/tutorial/install-mongodb-on-ubuntu/) +- 安装与启动:请参考[官方文档](https://www.mongodb.com/zh-cn/docs/manual/administration/install-on-linux/#std-label-install-mdb-community-edition-linux),进入后选择自己的系统版本即可 - 默认连接本地27017端口 --- @@ -82,7 +94,11 @@ pip install -r requirements.txt ### 4️⃣ **安装NapCat框架** -- 参考[NapCat官方文档](https://www.napcat.wiki/guide/boot/Shell#napcat-installer-linux%E4%B8%80%E9%94%AE%E4%BD%BF%E7%94%A8%E8%84%9A%E6%9C%AC-%E6%94%AF%E6%8C%81ubuntu-20-debian-10-centos9)安装 +- 执行NapCat的Linux一键使用脚本(支持Ubuntu 20+/Debian 10+/Centos9) +```bash +curl -o napcat.sh https://nclatest.znin.net/NapNeko/NapCat-Installer/main/script/install.sh && sudo bash napcat.sh +``` +- 如果你不想使用Napcat的脚本安装,可参考[Napcat-Linux手动安装](https://www.napcat.wiki/guide/boot/Shell-Linux-SemiAuto) - 使用QQ小号登录,添加反向WS地址: `ws://127.0.0.1:8080/onebot/v11/ws` @@ -91,9 +107,17 @@ pip install -r requirements.txt ## 配置文件设置 ### 5️⃣ **配置文件设置,让麦麦Bot正常工作** - -- 修改环境配置文件:`.env.prod` -- 修改机器人配置文件:`bot_config.toml` +可先运行一次 +```bash +# 在项目目录下操作 +nb run +# 或 +python3 bot.py +``` +之后你就可以找到`.env.prod`和`bot_config.toml`这两个文件了 +关于文件内容的配置请参考: +- [🎀 新手配置指南](./installation_cute.md) - 通俗易懂的配置教程,适合初次使用的猫娘 +- [⚙️ 标准配置指南](./installation_standard.md) - 简明专业的配置说明,适合有经验的用户 --- diff --git a/docs/manual_deploy_macos.md b/docs/manual_deploy_macos.md new file mode 100644 index 000000000..00e2686b3 --- /dev/null +++ b/docs/manual_deploy_macos.md @@ -0,0 +1,201 @@ +# 📦 macOS系统手动部署MaiMbot麦麦指南 + +## 准备工作 + +- 一台搭载了macOS系统的设备(macOS 12.0 或以上) +- QQ小号(QQ框架的使用可能导致qq被风控,严重(小概率)可能会导致账号封禁,强烈不推荐使用大号) +- Homebrew包管理器 + - 如未安装,你可以在https://github.com/Homebrew/brew/releases/latest 找到.pkg格式的安装包 +- 可用的大模型API +- 一个AI助手,网上随便搜一家打开来用都行,可以帮你解决一些不懂的问题 +- 以下内容假设你对macOS系统有一定的了解,如果觉得难以理解,请直接用Windows系统部署[Windows系统部署指南](./manual_deploy_windows.md)或[使用Windows一键包部署](https://github.com/MaiM-with-u/MaiBot/releases/tag/EasyInstall-windows) +- 终端应用(iTerm2等) + +--- + +## 环境配置 + +### 1️⃣ **Python环境配置** + +```bash +# 检查Python版本(macOS自带python可能为2.7) +python3 --version + +# 通过Homebrew安装Python +brew install python@3.12 + +# 设置环境变量(如使用zsh) +echo 'export PATH="/usr/local/opt/python@3.12/bin:$PATH"' >> ~/.zshrc +source ~/.zshrc + +# 验证安装 +python3 --version # 应显示3.12.x +pip3 --version # 应关联3.12版本 +``` + +### 2️⃣ **创建虚拟环境** + +```bash +# 方法1:使用venv(推荐) +python3 -m venv maimbot-venv +source maimbot-venv/bin/activate # 激活虚拟环境 + +# 方法2:使用conda +brew install --cask miniconda +conda create -n maimbot python=3.9 +conda activate maimbot # 激活虚拟环境 + +# 安装项目依赖 +# 请确保已经进入虚拟环境再执行 +pip install -r requirements.txt +``` + +--- + +## 数据库配置 + +### 3️⃣ **安装MongoDB** + +请参考[官方文档](https://www.mongodb.com/zh-cn/docs/manual/tutorial/install-mongodb-on-os-x/#install-mongodb-community-edition) + +--- + +## NapCat + +### 4️⃣ **安装与配置Napcat** +- 安装 +可以使用Napcat官方提供的[macOS安装工具](https://github.com/NapNeko/NapCat-Mac-Installer/releases/) +由于权限问题,补丁过程需要手动替换 package.json,请注意备份原文件~ +- 配置 +使用QQ小号登录,添加反向WS地址: `ws://127.0.0.1:8080/onebot/v11/ws` + +--- + +## 配置文件设置 + +### 5️⃣ **生成配置文件** +可先运行一次 +```bash +# 在项目目录下操作 +nb run +# 或 +python3 bot.py +``` + +之后你就可以找到`.env.prod`和`bot_config.toml`这两个文件了 + +关于文件内容的配置请参考: +- [🎀 新手配置指南](./installation_cute.md) - 通俗易懂的配置教程,适合初次使用的猫娘 +- [⚙️ 标准配置指南](./installation_standard.md) - 简明专业的配置说明,适合有经验的用户 + + +--- + +## 启动机器人 + +### 6️⃣ **启动麦麦机器人** + +```bash +# 在项目目录下操作 +nb run +# 或 +python3 bot.py +``` + +## 启动管理 + +### 7️⃣ **通过launchd管理服务** + +创建plist文件: + +```bash +nano ~/Library/LaunchAgents/com.maimbot.plist +``` + +内容示例(需替换实际路径): + +```xml + + + + + Label + com.maimbot + + ProgramArguments + + /path/to/maimbot-venv/bin/python + /path/to/MaiMbot/bot.py + + + WorkingDirectory + /path/to/MaiMbot + + StandardOutPath + /tmp/maimbot.log + StandardErrorPath + /tmp/maimbot.err + + RunAtLoad + + KeepAlive + + + +``` + +加载服务: + +```bash +launchctl load ~/Library/LaunchAgents/com.maimbot.plist +launchctl start com.maimbot +``` + +查看日志: + +```bash +tail -f /tmp/maimbot.log +``` + +--- + +## 常见问题处理 + +1. **权限问题** +```bash +# 遇到文件权限错误时 +chmod -R 755 ~/Documents/MaiMbot +``` + +2. **Python模块缺失** +```bash +# 确保在虚拟环境中 +source maimbot-venv/bin/activate # 或 conda 激活 +pip install --force-reinstall -r requirements.txt +``` + +3. **MongoDB连接失败** +```bash +# 检查服务状态 +brew services list +# 重置数据库权限 +mongosh --eval "db.adminCommand({setFeatureCompatibilityVersion: '5.0'})" +``` + +--- + +## 系统优化建议 + +1. **关闭App Nap** +```bash +# 防止系统休眠NapCat进程 +defaults write NSGlobalDomain NSAppSleepDisabled -bool YES +``` + +2. **电源管理设置** +```bash +# 防止睡眠影响机器人运行 +sudo systemsetup -setcomputersleep Never +``` + +--- diff --git a/docs/API_KEY.png b/docs/pic/API_KEY.png similarity index 100% rename from docs/API_KEY.png rename to docs/pic/API_KEY.png diff --git a/docs/MONGO_DB_0.png b/docs/pic/MONGO_DB_0.png similarity index 100% rename from docs/MONGO_DB_0.png rename to docs/pic/MONGO_DB_0.png diff --git a/docs/MONGO_DB_1.png b/docs/pic/MONGO_DB_1.png similarity index 100% rename from docs/MONGO_DB_1.png rename to docs/pic/MONGO_DB_1.png diff --git a/docs/MONGO_DB_2.png b/docs/pic/MONGO_DB_2.png similarity index 100% rename from docs/MONGO_DB_2.png rename to docs/pic/MONGO_DB_2.png diff --git a/docs/pic/MongoDB_Ubuntu_guide.png b/docs/pic/MongoDB_Ubuntu_guide.png new file mode 100644 index 0000000000000000000000000000000000000000..abd47c2834a27794808fb7a82d7c0164a92c1e90 GIT binary patch literal 14733 zcmd^mXH-y9myRe?br$VkcEuX`<4;+w0)qyt$O36cJHHIK_$41#nDm|4Q45 zi0Dcy;g7h(F8@6dkq>6pl;dI$4_3mfHrG)JfB>`p3UL zzs$~Z@%r^UEPEY7UaQG2OuLfT`|hP-kuCd44X& zy@WXsx!&m^0Zr_^f3p#m{pB43CyeC$T@3!jY;Akr3_5u8@i&#%uS0H_^i$mimu`Nr z)pT-#CeqBPz$X(t?7oMy{sDJQ-)nP+n!S$hADUuVF)_BTCpiZ?PyIf8Vtxbqv`0+` z^zpO3x|}dP^oV(;N&#GybSr=Q*V2D&wrs?b5Y>l?vx(f=k;_oRgvq4isZ9Q)*ja!- z_G-gQcn-KidCWW&?K;KWQ_6QeF0o)j{PaegV=A1me&_G@1gSJ&gxYeF^ThABC+dIe zkLQpp^r;%y6uuvnhHq{@S^mJk$;00OI{V!T#68-mp)IJ`punVgdP#a&!``zh)oX{6 zx)xSc-b<$foz|2H-q*!B6joF;_2eq8*6o^6#_v!w8i*k%QRML97*Vufp@DBWy=xo? zuyAO1yb$|6m_;rR8CJWP(j0Xxp#oDdC)|V4ASy4L4*Q31{ifVd!N&$?cg57iFs*K< z3j%YIa&`HaSYD;{i&fVpYtf)-(CzOR@z<$5!&XH zvlCc|s-2bL87e3fA8Y=lBD}p;^^HMTr_+gPTjV?i4JJ&!2?- zHyySU`h|;A2C56$Ni(17fEU*$R`OV3d!Y)w&S`|ua-ubArol7_Jmgsx#W3-A}*HvS&FbGB^{axYc|QQ!y>l7G^4#g*|;`2Dld3qb#Z4AhGu`#nVH{{!o+}=Ey6cy3QK({|HNAm$MNUAWZdUb?$l4-g{>y4;rGB zut$WbfBWsf-XQKTS-Chw$ZIVpF^U=VRaS9u^9*gAgOU;v-8s6p;^miY$`_M!sp)Nb z`{u@ueC*PCnq89PLDK05g;7jP1sens_cL`6koVE{lRTA^qCM(0Kj!pzr z{Y5a_!@;rw>r-p)x69*Kr3kl3UU;0uI%Bma6D{|G<`G(}2!BxdEe`r+!7f*7CYz8!Zh_l8;M&BJq$ zBH}0J*`MW9mkp{U!j;uftV{(bG9^zNuD)9)rBNMf>k2&#Cnh4QSGAO{ga>L~U1^=n zUtn~JeBumg+4N}M&2`V#G;c_u!zgRqc1(5?cq_-=CD3r9AD{VL&Xn0X`}K-+4vSn> zkT{=+ch&w24(Zz*0ox#E9&^(8enzV=VA>V@4bxz`M%KLZhhEhz*Juol$ zcGAz@W98+Rpd1B|-aUmINu2C2x%T9YFMhl)fMRiVQmMBWT-jzFA z%BlG$x_dnQm!C{asC+FWgn~pq>=HufsFDd)|FmNP+BQegX+LOfd3+BRn ze|MH#@JCp9vB%3dc0oq1A_t;GMD{BN9vE)@aKZA32X}ex7YBV=^_R4}KT*Apg<(78}K8&Es=Ub2@%-oo2Wf+k%Kq-J1_Dj3YbH zhwYz~BW1t(SP|VU-ZgkPE!IYD4EfnQsdFFg@hLtd4hJbN`+m|LCm(BBeJ1>AH*Tha zSmb)l<|sb43~U7LH`*E{d8rabJEpk3vqd`mK?Z++Vm^7AEk zDw^vS6NC4TadTf`!@hA{zAq_Ji|YJ22)@LFJ_T}DYF09ys+GLy=kd>)z-~?2GS`3w zjk{(MRVu-&BD6Gj&)J@zO~4C7c9quB^($GFhn>ukkSg@OpEoMnB|b=%It?xiX1Bd; zYV>XLBFiu^$7WP>zX5nsQ12lKLy}U)noTa&atpaXRkP!Jnb1MsVfXClZiR}i$FEIb|zQv06 zF4RHvJtWp;v`o}2-iw8dpxmcQj;~z#$xv0B{Yy)@a_rErz`NuiChU;872Y*Sj%HPu z-HqLQt^NM|k%dl0Urn&Wo6c_%=nF3v=3+S`w>$QIs6MHQ&Qj%b*qg1S1V@tBPuY3* zRgyJ~I=6JWpkEw%c)_&J(lmlT)jo_!=9Uz3$Yln|ivjFrcI`r02=kQ~{(hLx6m6!} z>~NMXV*?c1sM3n!YQJ1~0Xyn<%sbnoU5JFt*XG>x4ZZoS0eap#eBHX^Z3gYk0OL2i zvJ=1hBYvIoUku44;o3`-yHdewz>1cg*@~2WeF=2fhq{OC`;1d(#%A-iA!KRfAnjdE z0qYmd)j8~JxQ)>wq&_OLMxV4xRz1;WT=9&%w@i`dR`tGDd9@42yPMAF9+5O{l*Zc7 zoW{UzXMPrq$6ooUqGedA4xM3iI0+P4DQuYLX4dCNThzf_KTqQKP-L(dtx}Dw$|+Cw zz$0iN)cEriWjg@|o2v{{-PY2tnxDjU#gXtb4P?$ej_G5#ZmjznSXEH4{sVr4KR!WX z^%F;xKkU=ulkD9ZvR$y1gadI*Ee&f`r_7H=sh?4J$ZTnWaj$MYgo5-)bT7S!fTSh! z?*E$*`rn1#|L*U;(Fuz-!L5J%^v`vN5*B35S;j;;UQy*^f(JNiQ4Rd3W^C;5yu=v! z00ScO%OU%}ujz|;#B4$UAw-Ptm8(PkF~1if#y6;`@3&3tkpSf2cOfxpbb7kj?1h_p zwX4}I!;%R(SX`nt@5i&nbkxhx{e#cA_*+RRKX9@|;Bj*gT^kqWUi26^OlNrRvGU_- zkF47k0izBuf(xJ}`q!>7zvIzIBHyq@W#$x=4e)sA+h$oAl(#~~X=JZ0`1Kt!Bk^-= z?~WxzwPvlah?t~u+G$P?3=$x*@ZnSfz`8b92muywbw$@9(>scQ4FO~>^ghs}Om;C0 zB7p#wL0Pv~lD>+ii!8~Vl$QoiBBIGpj}TWV8x&l;rxZ5a*17)DpEcM2L_S7ZnSG9q zZFP5(Nk>*ssXEht^;#EbcFM+)fmgQ`N!iV3-^uSk|3qt^qn@7#c$pi5B&wg1vqWIbsLsn$e?K zvgu-&I)LJ%-n+F&A&t_J>NObcQS@ZG6hb#AeirgEi~!01Bqk~)5h4-i`sz<=sm~&; zWU{3HLd@@HQq6#T*lH^QYZmzO|DZKoLleHJ9Z%D|D>Xt&C(*Y@&T(j+KtN^1pqvffUAWu zqaC)|Z(utJ&D}=C+T7x}SO0p85^FzsV?OY%A4py$U;2|bczhI_mgyj;?H>z|r<_(E zVbUCv5-mA3p!0&Zy1kzfYXIaEFHYVYsUTRYDxT5?}qHn2Sm{*848v@;5xy|^E*lwGyv@4d@JN&~~ zKSDm3c~~0S9WGc)YvJM#AKv5A)5sn&hIQ2ESWaO=_Mo<_7vA)#(%@hQUT2J#nD)f& zG@9YV7j~;Ah5D--4Qf&ufW0(E%|Y82xBk#RU|gkKM07q)1emRSbUZni26@v$H4@GQ z|E2MQLtj+t>r1b(MuCfS0UheXysg$-_ka~Mid~WF)jrj^w8f7}&u@%50TTu8_bi5R z>n&3(F7sM>qO7qy#&s;FYe%NTJ$bd$tJShUO`0i!gl+T(cRh!l3^fy2OlPw*$;t`< zGTmq|-}DYA(3k)Xv$Bk&eo7*uiO@&ke9>i`PgqAA^OoB~_PWm5)habEO@3b%pwRL# z-?I{w*rdSQK<)KD^Zr@2aQN!mPtU!5;mH2)kUGHf!hvAUS+x>4qSG<91Y_VH4J(E# z2qez<*W;#%h`tCEuy>kK>qy0;EtR&MehWJqin-+PUPbJYm?c7NiYNk*BP8b8DLT%1 z=O4%*9{0CuxiMKjXi!8FJHE7x&z@wohqrGv-Ik&&v= zHm`hWLz`3`Wdah1=7XvfcKFLqu_YSwH#Bj)}}kOc4=rzQcQ~ zh)b?;n&Ikx3Vh>$6nAhdQ-xS!(Yplo;VqALA-tLHtk?r@MRd+;o}@W0prcXBpbzGe z%}-%&50Z4U%)h|(aXSwOwb&T-WKjUD&Ijm=;KmC?a_ACFARV*$ZF{GNHPXDoCab(Q zVlskf4Jy(7ysbXovD(VjV>E)6J$q*{Mx1Gerq`2)G*&r6alq1TGKf!cR|w~ zDGhwmGfxI_xca%P&{7Q3Lt#WhfY;mS`bU&@ zvr46UxNRvi*(A;XQ;`npRdk76fmMy*pX1_sg>J!dIXP-)Y<=wYL|HId3muiYDO0wk zfOp{!XrAS0rnnPkeE-*lBfz7c3Li@cH2Ob6m~!lU#<$cvrnp^%Ol$)WACPI}8SIWq zW#us7EW1zL5*^Hoonkj^4dHSN(J%awWYeqePc0m>9itC)YrAwWQkNUFcs$e9HJiOr zxgr)F$;w$(^-})zt~=T}-tC!V>1a-0pSTo70K{K>-FSJ$8DB?CWMBG-L`+It6zph* zsyx-ZmSrTSR0ppL)ke+Dl#$m-4L=A}b>qvV4Y`tOE`(Q|AJ~oK7aO0p$ZcR8N#*f9 zy?SP+N8q7VLDFYb6~}KgEhKoyl{F97-V$Bry5d#$_Kem4PddQJMj&TPv8mt>(bHQp?x+Vnb9v z&AQPl3KP2opbi+&D3w_Y=^mE2%~2l@nCzJi?G)ZfNh#DlrMmRXjb9?|a3<*TDoiC< zfy0|sX#Qr@XC5hGHN00l?^}SOLoZZBW=Gawi%cm?sdj3IWB%aByGl#mm zSfYQW(Bg^?5#66+D!T$UELF3TB;{?it<41abi?If$m+ty0e+o^I;CA37drP;o4JuM z$IK#YgM;*k-pfJ~ByK@WXz7tyeg6}~4y})~u~Z){dCiOD&R%rjDU4@=%R_Y@6EAh6sD#EnIBjc!f{RaLcbZlk!Fe1c`|I$sU2C5?eJm$A{QHIuWBZ1SJg z*h`H~vSC!OWa(!Rl_7&`;T`RDYpm)9-r$0Wp?FS;6M2=B=n^Wm6{m}Y^tRPU-8feT z_^kmKRzFaP_EmFaroh=0$jr`fdIsj#d(E;ajZM4~6v5L(GSA9V*hb(5rP&GrgehG6 zf4OP9%$PcBEigGJsn!T5sC3Tt85L|j1Qm+_8Q-!0`lrUqiZEHf)Uf|0A^5L9s{e^k z{Ga*x>Z|zh#~?cr`CX}G({2C9{O-Mc{O#-42;)thKUmb3bG^NmwzitkOc4%U5k55K zZpC?mDbCHe?8ZXA?rM8_wR*n}QFN0e_#o+);Yx^c>}L;oD{weKnJMa|L%)*M;QE&$ zT>%%sX`+8=Jr4R12=||-aj^tQ?@EZ9l6yU}-$9AtsWC3qkMt+O;=zDNh;P)(Nk7W$ zqXDK@U53_J&J`b%5%AfUOMy7o5C0IZf!d#-4^B$f1OrMl(-p23`d8Gf|1pKXBS%p1 zXR1M+g8;S&6#%!_{{h1R^c1c;qi(}6FNm{M8#=bcuc%>!cfI|A6kQLg(oIvFL8Pil zA2?GU&2sRkh?ce%trk^ON&^-CyZI0|$|?L5gBY>f)YHJHC5^t-zUF-LX7%d-pap*z zo&I~lflye9;O9>fNXop(x3~5EyO=%!4uTs*c}q1HR1a~t7R($ai!S0j840Z7?wQR; zB|M}^WCbb6G?I=kQ3W=@?kfoT{~m_ansUUDd) zlfUKkG0fKeIf2h)w|8A)oZ0?z+!&TB0|xx|6`SAxrNWl5qMI`DcAw79O+Y?nO_y`a zfx)G6I6`BVi4BA3oetO0q|qo~md0@BqGj>Xy?1(h1)eE*Pdmw}^)hTBPQB^aC4Vt# zbpWsrdi);MbZ}&b8?2<(|Ne;8+#f#&aD#FRBmCvxpb`*M_&dk!qBo=E>q#45d*lvl z!IAPO%L#nq(9L2sJLO|5_eR2%RXOdRkKP-ft;Xcr?kAt3(_0kQRJIaN#%wg`}}fAFPJ-~d2U6iLiZ zep4O-@6bP5WJ5VkdR_s=8BRcT|B}gQd~k}C%kf{E8{p5cS_(4w+Q8-7Z#&hazBH(X z)+L2{8;Rm{yJr~9Bh99VOJ}T=qU{JAb3v7_6=k;QdjL!pA!||UMyd?^@&6@fpM7g zB$}^~6PUW>`N7*O?0@ANt{3q@4IuhI(nzYmQj`-&8u%8U(K?*e`;y^z`@xf>ddX%@ zT0}M%7737iL|z*!@kbUpP#geqKAt7ZZDDAu=@PnW9hD9De1igP3r$A%>-Q7iEbhe!wk=Z5DShX4f1M+>%EoRHa4ERgX`en8D)8F_ z#MAR2Edz)Gt-1MM{henV0CY=Ka>C5br7T7x@#KICz<>Tx*MJ}T3-g4#@u29f{_lbqoISKizr+VB1d5Q0}o%&9gupC{z_&Wh# z0?O;_f;GR-TB1dDkM#q>eecf58g(-R$s0Yx*(_(;9_(J_*jnwq z07K{x_lO6qLdsXv5Bwc|g2%_zPd$>#->JeMF^>V!U1CbGVfLOw*S(lhOLl04_BZN< z!2|gAU-TuPnBI8KgSKbo{j0)ZHgtSP{_q&h&$7{7vZn7hL5N&bN{0dFAg>^I_fL>i zyxoh=NKu&INo5xcyHjyZ=_!yT%$Mt5mF#-xDFUvOq4c4iG8@SGYsBLoU~)N$!rBd(ab^o3$Xek{2h; z#ghT+y)+>{q-^EYbWE<7y;0NOSV!9^aW;OXs#X(EOObKQ>(B8PpB+y;L8W-V$!C`Z zSl?Fk=pn_WajPFKGuMaG+>o7DKAD@hOAKX>YPeNJpB{4hCfKk6-+Ge$MRtB<@d<>4 zFtGlw)jZJU|4QxiA6EZ|#%-U@K5I?L_QEp*6VzsEBk#<>G7**EF#4aH=qmodX;Hc3 ze~JJ884dn_byV<=RIZy$pHgb#)-%vqL8q#h45xH%YJZ8rl}n#8BhqI@U{AfWR5SC+ zK2s{GOs4zGGsR~=R}MI8$ExV@qN^Z_>kmfLo)t{iU)1o`&uW!z9>dFM6qXbTQh`YA zdt%jkQZTqKyt{O69mzOU(WL5>C3Jqx^Ra{KS;SC_!qc^bQ5A+Y>V@Nn$bf5IxA^r) zj;^kkD2E37dwJ52DQhJm2P_Yl`^d1NznQk zL@yUx6PF+Z_N`!r8?m+CZ7HU~th&dRV>a2jEEXsiEvtcaDbMw-@O@O?u`u@^g&{Bs z;s5(XQ_L3NTB6s?sJx=wlOSU7j?g>1c+#sZ$N=XV_ZjsVh;|9gnsdCOo_giGl2b7% z5MNoMXKJX4Y)9sDfKE<;Wv&?qdd%H+fUjLvyd|RZB1wF-Pb{bh#RGNfze{6tq)y!) zHfEZ|*8>!hflp+6EkbjZgLL%|t)dMOl?p+HR=p@5xBPm|GCO!qaOKNuYs5rrF9SVV z9JOh)t$Nld)K6@+<6=jDth}8er-A9_KF;DzU%Ch^m5wLnicO5nwutK3*s`H9=l>%T zHkvXpHO2ANX7}yQwsU}aknQ=k-WGV&V=b97j;FdJGtu_(|ZvEVG!Nv@Jir%oWy2Vbx z{XQ74S}7bX{vg?&SeqZS1-9JOlNaX*%~P03s(#%NbgEYbPbpW+rlil8;qT~O;oIds z!+6Luo$KzdOp_Oo%`m!#fXAm}h1c3_uAv6$Vl8~8GlE(N5 zC+Gsd9f4J}sCb}NI?m5$h7%hrdY#Y(MNuN9Ddi)xGw9p)g$Ja#X~CeLFLd*p7Ui6z z@9pB;)DPZ;!vmT89wAsiuc!6@;(FEx`I>s3vBd6ClE`LMN`nR-%exOd>)*S^yzF}> zowhk`Bi_^-HYWIRFzTH^P)vh&3My?u1%9MmzjBZlGMg7IVG=tQtI(A^&T98XztwAY zyeGaajV1{~G|6dY5Z|Cz*S#Obu5#THvh3=SU=|e3kA96BIelX#tye7S(_lAre+~Vh z&omI%H}FGOVg~jc-rS+g&}d{O`OEv%!B6PcT5al$mz|+yt}~F!IRBM+h{bqS2sUQ-e9x}P zvY3odx=_7dQINZWmnPfUG(}cVd0Pw$y*wZ7a*xwuAHK`W@+Rql(u<*nA>bmZ$$4CIRs1f* zS&X*0(7i`KH7R)h?j~50KP!ya^8Zj4j zEVG6028akm>Q;-E=~g$U6WT!pE*EYSTHVx;dAi|S<`NniUHsZ`E+y*3Y3txJ#ckT8 z5?>eTok7LP2^Myq)+VC(DErilujmI>9R>GAY(&6sv}Q;shku+ud-D0bBEzjZ*R$Cl zrbpv`tFOK#(ZAh8U`ajg-c_Hoee_##RH6eX7|0}e$H-^9*7eX~!TB`%>@rmQtjV_& zwdq>|4{`eCylk)#Vd`4nHGLRtB;T8P}3dtWafK;Ci>2g7h{cW^iY#eUQAZqVu{{%*;8o3W;QJDlyjNf zee;EaKAm&|AL!;P;?QMWqP<3$?U696y)7$$CBD_ABg$^g2YT|r$>5J_Trqq5rVr{# z6YBhjIN2yPrQ2f({VY}10zIqN*v9A$bJpd^>|zK1jAv_@s0SPPD+WK(IoJF)w zn;3ZXm#kd1y1LRZz!pjwc|Ce;iow-4;`^NfDF&E;g5&k4+&jW@%}R zbIm$IzhOs)6N@OF9g?m!C6p~spRVW8}vOkRHrr@Qe{4;tuySlz!-%D)b^{sta z@f&iq8*E^QCeMN$=P3Ae-A%lbM}jKIBmVQA%i)-o6=7B-oWnMk4(_OBa%h3I9-rZ+ zffSs=9x9*< zs(2Z6;h38din4fFhhbx342zZG7}tS0##JpxO5;m~7FUqwZD2{6J$w;0qDz7*bB(Tl zR3#7r*hG#OdN%5sIXZIkV!Ye8h)spK`3@ztc}_RePP&}CQaZmdBOT6GMTAcv{X-;4 zO>TnI6rsg8Si400=kb@#dWP;Zkc3~g2}4mdmq%r1c{OdF%6hVChoeM3gp`vKid*!n zl;|L3xs4eFt!wa)7CXL3t)k#_W{ZY4hre`FseMyMu#WoII2mmwN*Lt_XvpIlY%Gi4 z%*o<6$6<^}oF+{ItE zbxUR?-*G6Ea2T>isk?=AN2{0G+%@YArZ_8p;_=i>XO2M&PI}j@=i6E47QgBOD`LC3 zPS{X60WBDekqPaOR>!X@kq_VPmr-4qqab031VIhnw#acTp~S9G?z6Nh?0?~PmIsSI z0HdvIZE_hjgFz44Pa#Dl&p6m!HG<>+`CSSo*o~PclOkdS*Hb{Krr^s1YP>LV;(AGzL#)S*O zRx};Ylqq%`Fb#kcn-}*@ys*f4-XBTLlP6Lwl4z ztJ0wJqat}!Y`4|5&Nu5$mh=Q%XudYK{>AM=VurdG2MO$8_EP_ddrc`OsY^Aby77i^ zAjz6|+o!#%>P9E|*?X6Ces73M>q|cKm3h`#!UV;L#9z*&(7uD&+}Z*pgz`}l#ZGBF%OPQ(@L z#kFnjm&aJsE|=M7x(OGG#P$;qT1=Fexco2IUWVK{FA9lk}Iz&vxR}ke_mxVs_9XHE7G{{7L?DD;LHnYVALhQ?}g;T~U zuREI|jj)5;e2LdqL}aY1@uPOMv}`R*HgEG7QU>(TyiSJ-!kIsW*$lK##wNJtjl8tY zk9GbbEs{(u#VWvPzfWExj!!%7n<>aL)H;o*(Ob=T(@G^K-_v!4KBnEN_RxFZYblVV z=hDmO+{pgy5B6zLxOd4a%I`~M=MK;Xv{C&fu-VC9dUR}7%tOp+l=gw@H z&&kIYxRY<+>w0a3Est;%%d(G*No0?p5MtAN=s5)2=|EgIb(*T$%&>MJolD%(q1}V> zo=Zm0k3_+4&P(4(?`bRhy_5nTcaiZy(!D*L>A48qd#hTRi(I)S9$Hj^g*#NY5Z8~G zANLP4(jEmY^faR-SNaT59x=XGSB_=f?)JEmqcexd@tDQF*qD{ff!vv+b-dw)k->%> z0n347%D%_5BR3HT3n?1LAf?5M?dXRlMK>YJ!kY2e#JMT6Vc!s}bc?cX$BRcb=(qY; z1^-yvYP{4dwA~aLvleIDl2Fevve-Ii8GU8jI4`VkKl{ow)z{#6$A#6WKA>IbLZaZ;ax--~Z@cQJ0zVSTT`RqoxrQjJ) zZL^q@!Q75h+&adKjpOl1(n=Rqq z5TawIX`a#MZ6GSC^NOdw2Q4ya>T98Olik?VwHrsLpzh)D5>b=(y`zBOyqOZx8CJr# zHrvA)Dx3eot;(l!?*>O-N2{zd)#q2q;x#2_S#J^uC4XqmPZ9-Yv zOhF_KHLFXkbNJ7&ow6EE`|(2;bK>!*CoVH8sIcBc!y0~VWDVdxSM9YzPa+mRmQK>$@leM0ZMnL zC+q%0sMLr@ZS`Bz(4fivFNt>$pTlDGO)PPkj6v^o%pRJP7m-%e4ium4=U4nas(WcbADKvuf>|OA~$; zi%teTqI$cQrOfH94PO4j{!Y?Y$%3&N#F<_hmQ8!9;-M%jl$QCY6jb*g%Jhm3XxA5* z%{sKjb5YViedP<+v~_*)$u6XSB;vm8KK|f)S`Vno3KR3~`Q^+a3?3*;C%+8;)~pO; z6tgTBC3hf{E$eRqMm~}2CNZWF7dkYz^c;#Trb|xFFXCayQAGO?%8hj^l1i0jfeFQv~x2TRu(!P`~+g>JSb*&$3XBo)#wwwI%;U2WjY%DRSWf- z%Laj~>k>a3+61q(>ia#SN+uDX)y~j2^>RL43htBd>;?DLm%!cI;+Ue=XTEZQs6_~T z=aGD$&w|(MUB0v$daWT+bp0(Hc`1tGNpk6 zVygd0nl9_JmE@Vvq;KuZAj`0#`dkB9Bm*FtDxrDG3Xuz*7*%QS_S{zvcG8+ixw}?`-1Bj<&dkg$kvFkLutToMFVY`TeQqbEa{{ zaB3Hz+3*9UVdx^I`fdvgtLOC*!RGJG`ChI-- z-VhPZaX|%o(tu!jUzz87^Xy&5sqV{SePQv>3eH|x%(2otA6sH(JMV#>0}vqF?Or(g z-zb}s-$8WD^lkk8W@rEXrWvTB|68B%{Cl7NY@Csl2VRoYW=y^Y?n@;1Lg{&_^t*un E29!Ydd;kCd literal 0 HcmV?d00001 diff --git a/docs/pic/QQ_Download_guide_Linux.png b/docs/pic/QQ_Download_guide_Linux.png new file mode 100644 index 0000000000000000000000000000000000000000..1d47e9d27f92c6d751fdc14e96dfd1bbe0d355fb GIT binary patch literal 37847 zcmd?RcTkhx_wVb+E{c3CbO99<0R`zzh$6*^fRuzzL|O=42)(F?DAEy-Pz3~%0HK8* zP!Nz3dI%5#BGMs93B8;r_&sONoS8Fof49usx%Ur-`H9d-p5O=XPM*H2 z`~5aSVD^rM<7ce+2Nl-22tIsZX~~|{XHER;e7a_sP3dip`h&j9f&RB* z!iwIN_fATXV=IZQ7f@%8|KN(gpQ3Q%W#Jg8Q(kg9MP4%L>znkbokbbh)Qv59yPG3Z z^0)JJq8l0*_kUonjS1Cc3VDi$aMC;(54%5XX8gc?M*$z=&YGVW(?#eA#$`V^ z;Vw)i<~(M`BAyQ1@?s3c5wjs*d3b^__Ih@0tlq{^n~r7YC@x4=Ehd`DcjXZW(;X5b z2$@)Y)TeF6{{HPz%}c05vxi|R+#7SGLn^)z+=$olUhCCQ2Q?8@kuoUn^qS6&UAMx~w@oj%Se zBoS*pRJ>8$8Hv&L)HZUWv=NuXJDm%^U1?2(jW!nIlpqC#mtsL8y6n^p|9{X~t-SwY zNfsCSX+94^Oo&X=wHcmygvkyyjjT1D7Mgma9K(-o*q%`Ho>WAA$v51 z^;h+rNLE+!Tw@{rhET?Wgx!5&JIf<^$zRQOC*4OVTnfpk$@X-bQlFz+odhXgy9;6W zMdHeiN9ZJHTeu>q`EVlL#~s&#wPv*`_vD}7c%7%WwXGSmew=EX)qpT%J)7He9|7TX zih7FmGmVq)bMzef{_sm1S0c7>8#(83Cno=2Vg7z-|Bu`kJgz)_TSm#h>5SR=H9m{| zwJloWMU?QG3c>G{grIRe$m*Q+LX|Okft7?9qjy z3!SCZgBLpQmv`w|J(RfLO~l8Cy(z~%Z5ckry%Gn;-ev3&9tpyz?Wf( z_IRo;`eqcTWx$PMw>VDsg^jMVW(_F92ogKt6BbdwfQjNE2i}H(A2LZ!NF284Yy351 z&xNK8pV`t=L|WBN5M$h5wi}f!JnBAN2W!B~PwN>&OF5>bcSHT4 zWJafSC;#0%{2u-ph+$f!IJg|`9O`Zf#LVrS1jk-Swy*un^vXUpg;48xHXRl^+TUI< zBz)y6Sav4zm!fo8dUjkW-rX-N(sVxcUdFIA1=(F(oSk`MSWXNEEf1G6nE%G->{h@S z(w}BG+9MNfyKnnGu1|5#b*|hFS4rNtC^e+-Z*`i~J?K+^q^F96M9s|d(4?b3!AcWw z;TZfC)SdApWc7E%*55B^1`RU$LCQ3pZ(2Q{(T zJc{$AK)qhO3n5$w6;=Ue{%<=w47;hfz<27q_xkvHWo7U)WkdaedAedCdNeJ4pYfMD z*D&?km4L|sW$(boTqv)okCc{)Mv2CZN3&}$d8DT5eJ0vuMZEO>&7(wLwDIl2xABY} zpwo6Vtvb>^!U^AepOem_yyFCTdLsD(4ZCFIpmHuV3V!7)_&@3F@R1|nu$jHZddMo+ z3M{bp)?!roZuxG9#iFS2@g>>V&5~AIs3XL@0a9gJ1}Up)FT_jSD{H z37AXDY~t5IjlTS`jsxZjH0@}}layd^oOH84i%vx?8e>D+ZtaR_X{WHXA9PE(|Kgk7 zteJ<9Hx%fLWWnvCd5RdD#0@lV4Cipe{n`6PJ&L>9%f82Q)a0A>_A=HAg;DLK+?Y`Dj;cIPs8=&z7ibqmn zhz@0y@N8SEWbe6mq4-sHy6;BN-Xop8q6Yuk%#BQQJ8t?4i^y8P=j!T3=l=G1EWNPa z=tH`w=cs5C$fVQ)PyHqE%}OmPfIqowO$!%I>G|6D{Uuyp!WG9S#EtQ$mhW-b$rjC zQU>c8yE5XulT(M-|LN-MLhu~DYf`ZWx_0nRQM&Mtc7({exsxJa)@O01OiTThugFQv zb{^TwMb~|{+7?}8zJPzLeyCwMoA+7P^twNu-*-;uyn5px8bp3dDY1ufSsD~KIAQ#p zD$b)I*t%f#!PoJykcas+d?AQF^GB|*M-I;m&ElYZ9M{G9nY*^MMQooE* z+}NhHM+cm8GB!x%E;QBMOL2V+ZMGV#YU0H`VeFoR7&^doTQ%{=vlgXpk^=fGfXO5G zIKV>U@)b!FueIihgmYjzLa7(MvEhkoQOpWftjJr(-x=PrBdg)SV=-~`5fF5yY9$G^ znRrQh>(zVI?n@sQE%2PI(FE_OyjIwYx1J;?@=i;N7MsJA$<8rG-z?4JGth61UiND} zpt76aweUJ8c{HYnXPt-O_4VA5U}qAw=ZJUI3nC&>`P?kC!*zKjB1a|D9ydZNJHZJAxHyCGCv zT?DttjK^$ShNNweOB9UM!7h|?Pcd6*;(uSWD)`Y%=n;lNJ5wW*^aRIj>v6xIQJry* zl1Id#?$>&OHjSV!JmpKc*CAPz59|#S>5=%{Y|u!pN{_8ZxB}n zH+W~lPhUV9vtBhVwGn)2YFQggC^P|WZHcWl3z5;Ev;}Jj@9FPUtsO0J!#Vu13akpx zcA||+`xSUiFNYN{wyBK7HH+@DQ;(RTw9BhOEtlbW@-@?+K^#;|^1&$P4>cpDc*=Le z`2)0ccw@2uL8&~_&f^dM28xgmRpjELk8+DhO(*(gRrFL|^R~-xQ?M_gm4)SMjP_)@ z7{VvHBYbh6bzN%BK5I<5L_zIRPtKe6t=M$L_^N-_3c_gwR+_j-NzD=7dX)tZ#R+f0 zA|!t4@KN^-ZPc4&R&v$@;I!+0Z=UnK6?Ew6R*sF#+k0rmz2gL<)euwuOj&=JvydGw zDXe&oP$g1=K9BLpM0)PBQ1C0EVV)@SHh%8AM|=m|K!OD zI=Wk=Vw(TCx+Q8RJ(fE9-S<-|@o}%8!x$++tIM9MTgvF1uO~^&Ae!nu7VYV7n zFJb$r?E2QeDb~QQz-p*Q9eHXbBGw~nCI9H}VqZe7a3EsrM~;cKk-1`)W%o zGQU(a)A#^E4dmDlN^akiKVj?jtEbxPom)aKiwuHL)#`U(g9OvDMIPw1PtQ+`efWN` ziRodM8J@&Py^WNtX&wzo?tgZC`yKu4*B}O2NEw`GqkEYfSELOBsdm(X-#?TIB!p-gWOMs`l@=^!Eq>hAOu^kq zi$bLP{<2mb@KB|;a$ARF^lksHuic(e_t>tdI>m;aY-!ei8}uAteU{lb`ycxb}^E;5=kw0tU}zwLaD; zUF`TOu*yCHbKsT}mT%f9fCp=4mDc8LEl+6Cg5~MlpN+*JGs#>9+#L-Ivyv9%gWKOv z^f3>Y`;@$<(FL|zCPm0hWwFDc-vhT8gIP60&9Gm(uJG(#UFMEQjX2fOQG&bGczho8 zBt*Wct)?|)aH`v4XeycDKiFJkOI6T*n6<$cG4qqAP9Z!Zx}kcKd^fP;tfm|v9LbsW zTjgogr?}vOyFajdTne@&Q@CyP6uXzKm2<9&Ekt}V3gqz^o7VqkCl}F>vp2PT9fL(- z&TJUY6Z$x-ExVnnORX!N?*@+wws<+%d|`zuIlEj8sO2b_ODp(!+1+-0%WD|(^#by4 zG{X~@6C`j;QO#ASqoB@wV1+EuZf+@JEe-Om?V zqefN%mq13Yj4EZGpFcPjY%pVnC~@&KIF%)P3?E_1b5|ACTBYM7%Q+!7M!5zwFv1O` zR)Lz3R1#F5>yJ!K%v-}nfL|gGo)vc6MTgbdN1i3jIQDb~T(DZ0JJuI2+8C>ya^6oPLPJS4ksGz92aql_4tKW;)$DlNpT0QwB_x!$Py^2K?zZ?7AsUNzj`@Pnz zIR?xOdFbbt#NeE@QCj(owVZLqIjM%;BsS?vKZv@Bx86Xax_XOS-GQP*!l%*Inx_kC z=@kbMw6oWp4R*@Svn3yDv9=C<|7`hIen-wTgZG8>tth>x}lY*zevL^(ssiy7|MYDE_##|y} zlrjT8e`~9;seD~=&hSsdxAGmhEgW7(^%fu?+gsTPE8pW#i}X!0Aq`}qdYo0f*S701AVGsZRN7QSy;{%MJs}y&{^4#J!V>|g=CJ5 z_ig${sweH({#aCW%uKikGVR-cQAXpCwCI}>g%CGZcq*>Gx9^WSAfj5!qnv}_RT%cp z{s1dS%04KI=%k2ucfv~ws5c_{sDYGcV7va2e?MgQ_{EI6COh*9@O!&>X@^VACnc6S z(yUpj&n+Ymo&|idHgoJ4k7+#99BCn!1T)8=+FTGLK{(yTC8wArV8J8*6rnOnEsYA! ztjS8k9}7nH|M7BZ`QY0-z+AD$tuH;kkM{lJk_g{>`KrgcY#;l@mYf4L_eeE)=Yc*R zAvaUY_uOy3hL`n=(TT3-BX+Cz-Z^%+kvBA40)EwhF)FQt0j|GWp)D(mVfuiHlf9dh zRoL!}_rhaxj?VMs?uJ&>MpUuWo}s3m>e9iA{GBjA@nToNMdkwoUW+;WlHT=`wIB5tomwmeZ zy?Q2A>K1H_lah&n)7=QD0>5-$gInnV#PR~utELB5*)=`Ct|a<)d$#5V*uQZrxS=+@ zA6@3lI2C`T_?$-REuLDZ;ijx*c{5q>wdT?Gl;=D<=cos5rLVUzcB+)C%sPr1R_o|lJ+H1Bh>7}8h2CH#* zv00Zo>l42u5qm!!a}*it&=oR397?BfzpjWhfTtVYI@&p4tyJNhL%S=uEZ-%$-6I`iD!>{Z#`)@}-7ipv_}@4b<~M z!T0BGDyAid-7NILZ-=In=PNt&_GCrY7GdZ58~X~Pg;Je|GWo;tHlJQap^ zaqr3&UPOYN3ZNu0g5dUSOqT+Fwduwi#!>L+ZQY*lKJGg+JqRa9sn=wVm$xsg03PCos=A;704$;58wiA5$j#<`LNT8Ry-nIF?c^jI$BCLgAeK~@h z?`DI7dK=6y1?dsdwTXY_ki{Zpzml|;WW0ayP2PFhYQpAHDP@ru#8NG9m78Fqht`MC+jUb-RDnorSVWmFVK;yXhu^G_Nu~N&ABF~Ecl*||UV&f5 z@&--tj-=bIx)@p3@1pmgJ>R1WZ0>jk-MmxWm2a1((zY3hb-_w?)_y!3hMb{o1&!6O$a=I@|LAF@cFIS&Ishus8t5nrVe(-Ix zS8$1JA9x}1EE%}?e8fdsahMIO?ad>}6+1$fr+x7E^*N!1`J&mw#+%5~Je0+t$UbhB3t=v(@gphK3bJkzaxo`WOqA(jx{2k|`=O^0jJoHeRmv7eG zGRcT27t7&hAKocVOXc~%`#$jfJ=qC-r4(#2In5y0FJ^LZT(pk^tC{`Q0W#fgO?Q*f@jGZk0~ka4dQ4Hf>iI*s|lI+~Hhg5hPnowsu_Fs*n|f1yYv7$@9wl zlQOM9oK21!5cPuhCUrKOT8kzcv-F;2hlL5SGD8!tBZiYQc2EK#+P5;tx#CvXKP8!n z+f=F~k3i7w*4&Nu0rFIdXD`Cc`_4g|9L;0r`v}Xz)kNHnQ}xK>na1rAqTcT@XRD5x z!J_pyU%R|JkeH=ET2oWh9)h1RhnS0vk~+{-sFa`)gqRk{=2u;3KL96Jl1q!N-Upho z>$*N(`6)9$zlU!Mvf%BG% z*`#dDyE)9SF~KEQI~IdxADsWly~LuSkp2m!3d8Ro*b>lx2KOtOd^)_Dd9LDvKK&l! zX7i=ag%oUJX+jso@>zX(WwVMl_fx<|RChITzy=CJ zIj`z5B`?Fu3A|_O;ffrvLK`SluZpDNkG$cgCKs)u9uODz)mt zIM;L+`j7A~jP^vbTiK6X*~$(?pwmcl2*umm=(wZ^2nq1$v%8WbY-6fG%ywX@g&U2J$NFu%loQ^nT6 zH;36Z;@WLfM2GND){&jaw7^Q}=lgXz>7=Ufrdyl+-!EZ59CjS#y{yIE^N2|5;XZ|VKMb%b@y;hjPKj5N?K2|+=GLc7leW$Yn0TZ zd<>iP$13D+dTu?j8kySTO~>0_x8DOe_b?XT`!|5+N-ff2;lHF+L% zlzeH105p}K?bWjRs`=BCw$n*d(kZffS`cjb1kEV( z)KZXs(rT@*-1p{dI{F2=t=dLfGQ0+o{~USgdk3D=O~eC6TW|9X$_650MPk!YVm?+< zjYXD@oqYya$g(;$rbz8;WUT#TfQ0JEVA*pM~*(e zCi6{uH_?neO5Ba6VMeC@U1YWh-_HgcB3UokhOG_84U9b3$@h%eP23lLOEsxfc*!Z8 zB6x-aiwMCg5%lj0>i13gQq+*#u=sgNJPmiVMQ(A^Y2s%8nf1gmx)&>R3vYT$WmkH8 zTMN`-CdN1&%nEjTsAFOPh>*Fa-91vQ zYHE`Glls2S`%~P3Y70B+QoaD{b|aJCoruwiPreOGXPtVU*$k!UURA!zRjR%&Od0I1 z6pr+>1)JiVH?OB}wmTx?f0a(`u7FHLWVkC{zUS+!FexQE*N&OW4HkbNcN9h`?cH8? z-7{_%6d7E}S>0xgS)BRmOn@wnh zY%*CywZPSrsL?lsW}_S?qzx?04+sYXR#Af-#P}%Fe*$aS#;haA(#nZFg){nZMat~s zPz5YC@`di}!t-+LU2eaua4yT3%ws^X)_Cc-7FdK&w%WQmLP!Gp;#=jiPecy~Qc`Xs$mP-;4nKUv6Kn#_95A=S9u90&ONIcoK5 zxy>_aN#3#P{o&+alj|Fe{(Iwu$mE)I$*w=cc|#WFJXzsl#Vn!<3l{%Q=stJ>sug|= z^j=QgK44*sVDo<4RZrfNLS)9odp?R^h*@lj%{;N4W=Ph@nU)XUlsDND$?_{)WqYWJ zue{Y{^#<5!UUAK(HeZrxe(SsY?04F=;W)ouF3CSW$G=$z2?EU$)Ii1lG|LP>Q_;(o z(P8{-n#H?Mt_$Ofc6ENudunNBwnZPklJkp4i2%o2rL68YWc4`6`^6g%c?6B>+Zw#( zZ>iQA0vu1KUWW<r4&QERL36Q!GGf7o^e(xM${arkLEoRn0H8V&>~ou+n~gffNBY z8$IN10C$`TJ^QBP@KRnzxcw7l&Zn#5WCw*0JDp8LHU~d}{84YtXF;HjGb+{dEjhGusG2pyi65bbELaalXG%rux)26{q8al04l^- z)-K4|<@Bh6-sb?nMXgyo&tuYcWi@q2{D8w-&sR>cLg5*L)xe7-#An%QKZ7#Y2;C;I z!Ip^W+&S2t0f0T&*t@N|5vQsH`f!FU+X+ct0A20q4ab_qnGd4Ek~iZOO4>q&u2=Ju<-I%_>}^tT1A0=-ps>L?Bxuz1#sF03A6{ zHd^|Y+?HbJpP&-C@#D8QEaJ@T@^hcC?^|lShW3*@_nnMdq5?dXqNeSEc(gzcR_9G`%&|fr~2h(6tJhLGsa(xT*$72oEk*c9=;Wk`ak` z;Hb~u@)s2=R5kdEPbdLre!ZSUI^(}WP!KJm#Rao1n>uA+;9w?U#BX;QM#Sr@&Ghx! zC|h)ia9e!9a|V+|vly%itH3g7xgTUuyp>MGV;ct&huu6wsT8*`fBf05BU&RPB4SpnRmo1Bsc0ETyvwU zO0K-PapI%#(>b%dhge3RR zvNGxa7V(j45{9Eo>cpWj1GXDSU9#T(z-7%no*V&aC0v`ba^GoUPa$6Lye^S39Iiw2 z@fm2P(93Ryq8iLGFd~BDW56^5Ka2X(_JcVOyX7SLhBLmk_M!u2a24nbR)Zv#mhTGj zZEY-hS?HQ*&a>I4P$QZbf7I#L{m~FQ`MrV!<%3RTM>Ao#K2DwntQW+pWQ{Q_D3 zxUS<_S(9P3Ak2ZFP(j8%tZI7)-Fv-qBKwga(!*S=ZufRNdz zG4Ga%r4$=ftI0GzG05Or6ebs!eU0t;VidSa?^(W8Jxyax&JRN}mQ=5qK0z`Nvq^M< zyxqBf_~4}ZFBw8e8EH-VE!%DhYy!_`_-SGXRtkE%(=Z1`n*CHodN^-FSty)D;3|;c zh;_Qklfs+KLe>#oW_Zmuu_&Y!55NwBkCDpT{8rLMC9f**#UsvYn$zQX+n-0}pf;l& z7rGRnX#YK`1LKKbC(-2BIScXax?Vz(&$QO9-ad6N$q27CL*qv!cZWEyNhcKzTi}T^ z)w!xp1$tAJTw4ZcOB*XzAEj@<@4vsSW9XS->?BQ6Z*q_1j3iukiNJa^^G>#%f7F&V zUVQuGf9SCHvxE$LoSE&|Ov*y;BENgd)vR7}z`F(Yp@urG#P~rSrlfrq=)lC>9mzaw!aqiq78U42wnXR(3KAc>5x-RzQDKbP&<_A{yxZc-wl0_O z%?M}w*=GZzKTtJUZK$b*C0B9bgd}Srx*er$R0- zxV+Im?C+PF<|PY*BIb5Ap0a3>=-hDK^?}DWzaN>87x`Ct$u|pcGp+nSD~n#uf0)Ct zq4;RDo=t#N{`CSP?Syz1;&KN|9yZXtOJypLYZ5ZlVh6<%uIC-25OW33UywK)+jZX? za75QNwaWmD>2xtw_u+5D(qT{D_@lDJheYbJIYc1koF(my_wz2vsW%2sZIrKEnl}&@ z0cAwy&{S-1NAtp8h!qRZ!CvD_k%P>a5j1xF@ zW6V+uV#OpmoR(RCw~|e8os0Cjr-c%kka_^_73xj%Z{UFiR!N{<4=ah~QCA247n}Ov zu5>I_twa+`+RKF-35;0B5wj^@ zA^77|Eux-w9cNhw#zx~7mx|lKQZQdfj;@%^Fa4$apjdm-Zz^yH+##FLcgSNF0(%Idq-rc^!nKi0nB*Tgrer; z{E`J@85|MSA+Sl=;LIYc0s6H_!IK-(-kD>IKKOS_z~=Yl;{y7b23Yt7y|^>jMF>Sz z(eH8pfUr#@Hl}OgJhzY1vh_nw_qO;UGfwV_p}sKN<$=n(Tz(%#6f?e$$5S#G52y@h z;30;hr7j-Y>wW7n^2X7+v-8TTo>`0hU0}~@h48!HPOSlLA3KWUrmXY|v&~KrMaxsg zj;BhFulKL}O*Ft4ilICbwv-4M+Y`&9{D*~ltUd3R3br`7*4EaJ*IMVd@1o3|KJ`)F zvO25RXP~1s)PJbc7G`^2!R--W^2AM^s6$&SCHCO!6N3OLJ}NR5K!dHqch)9#>^g1& zMZRpeSURb-DkGS*M~&Mvz?RD2hH2B}#|FMX3m@u8u}=2S^~S!Ix5PUswj(TcYC3N| z8&Qt|ra)uw8J# z@Y=Kd%~EKG0qeY`w388`K2j+6{~L^#(cJqcbl>h2Ab?ZuG@9X*oQOyeu+M=y$_&eh zURg35;xjGZszP_#oY|riPiW%3X$Dj;&A0BlLwc)&>l25o&A0D=J^Uy^f#b~-`! zitop)1^YvExNd_L_V+y(&!xH*T-IIXuJ7h9D9=gJgVf9EJ|yOzf~9Bmrt`|AeFT(v zp+@2jK=I!f#P<##+7Q4^XjI_mn^2O`UYpljc+IW%Zc{OF5)mkhn4X zmhyptzqv)ZeH)yLHIU;6aY-O;ae)=X=NOU7BQMWp8VaKlD4J~W{bO5gD||hEIKpQ- z2nR{MiP&G=xTT@jvSc_vo5&r3gO5TQ+F%>XH4- zQ>%Ta%4g6m7iUF7nHT;{u~_4f}$J%-DXW0p;*jntlMm+`T@yH~-f2yY^%g!GGY-qO$x z$#Un)fkx@?lFn5V%kPzo(+^KmKl9fftChk>b)<3d=w%J^?btYz2?@Z-aJncy~%l#TjyyNS>D#}x)byWa~nXRC&p9Dt?S7qrg3P#v1)}td?{-?2vg>4WkHON#Mtx~2v zh5em7!$ZPCcKJ7hgdQr?n(OTC{Cg)&Trc28h>TJuDvFP^=67IL!SeWSX;+dq&u)DL z#=PdqsruUcV_SMbvvd1gv3(kgcQhvJE`$jP-F^CPX8$_kpi$Qp(r-E+DM(*%pg|ce zddYADG#~Uu5AR7u!R<&)i?=@&Z1spI+VxvhfblI%Qj4SG)!$N(P{C1BY0zHE$KJp$ zZFrvn(PHut4L>r-fRcNW1HHRTrUpC1Yz1agkV`aKUL7ZpqjudT-JMtbVesf?BDQ-s znHxX(hm+@^(Q_|*m*Q*PBUZMbt26gGbUF|5kZ?XtE;CgOBzQCA(6jwux?~{5_X*3Y z7ECSR-Of8G`d0l7mAd}*%#vu4Oo3ure-yX+d;&LBvv(pYl#5ki>VK-wB`Ka#LQZ0- zgTvd`R&QrMbV7V-eV4W$PK-OZxEz|obHGS09VhEHZ^muCfl2;XEz63%|C)i43L zKg^B6*L6}cn;AX~uHEp_0>F4vT6Tl}y18qQfc|cO;GQ`J3W|C$LR9*=IhwGZ!FjmG z6i8)oK8?qRKJH6ihPOxmn}GIGL`UotkspADk4dLaWNp3kJIsoC`2iy4m5Pk~d@rD+ zrp9`BZ859Pko#c1_RD*oxPKV?%GZ!+V7b0f>vOJ0mxTdx`v>Xtjz~NR9`{U(!wbc(~^FffWN)Y^BzJcuF zu!=AyT>F1=P2_>@?Q$R+GfMwD8TDU=Eqo`HQN}|gnK~i7j z-T=O^Cro1EKvwjMiz$;ggDd#1%*aRlTJ_&SlmC&Ucbg%t#l=G^p0J4ewh*2I!j+T$ zTh0Ga4X$Fc6)e@sH5b4hwH9|BwoYnsUjQbbFxsQ=Fc!_&nFyKLnIR5R_1$7%CMO1J zG2-y)FL|*Q?+p_lkS@VFIA9e%zh7OMH2j-qviZTWq}&ZRjSpg<_n#PTxm=D{KBjcR zD@7#QhssEOv5L13{0v&Q6W;cT?FVAfQ8-VWB^S_+4v1w&a+e^`mcuL)ZMD)oHD=_6 zt+UezipbgDqRMPQ5K;jA%3rI#L-(KTg{2y6otjbbn`@ea?$U1?w50=!PV{c;M2*z0SXxaR*u%Ejyf#=<=b_?g2E{=PT z+fH**lQd?{uW`Br!CzYFb2qP8_WdOol;!u|Td% zwJPlM>mkpRc_evBRxH0A;6bYaQCFbU;qC}5+ zltfhW^U0FjUm(WZs_bQ$cKx2Rd|y3AU!UR&(qmD_LK$YkBr6jQM2Z#`n-ksqnuv2R zxD2H~1SWYwe`T_k#sD}_lbM#tm3;1P_5Q$T^(vu?$YI$>6*?%abi3W0r#nlta~S?l z6O$73o}O{w?B@6$6OUZxk8~UOOl2s2q>F8)E=};Ei(*1$Vdx&pXk7!FL7OjKaogF3 zeo97*^QTahOzi|F9#KnXlz%lgAPOW!(nG$i&#eMcm>|qe0j4lDAja}F?@^)V<-9KJ zg=%E92YkbmL~}IoON2c7WLma0MPf6CLdgIvPU zSW*=h`SKfFh&{V1Enzv!!s+*V@hpvoC)X1N)RaeX-F*2i7CiUHEml>$0iRVcrMt z0Fa#K+K;HY8(Y$gNvtUPje2dq(ZaTR)6ZkZB+UbOfcrPIfO$<$1;Pevz{HZwtrvx? zS}e``DNe1}N1Yf;caJOZ{4h9yg|;`?62ATsFJ|{;t3voOTF(~i>I>bk9xvoo*qq$Z zxLiD1(U7~=9^y6k_R`WE-~P7zo;EOC6NaQ|>LP+yMoTStF`clClF8CH9gq>O-r7wp z^EP`s9HjqQBe0)CF&xN96SLn>+*6?D>X@nZRf^7i>7rLbrIfr0EMS}@V;%p^)s(lW z;w$WZ0|d*}pD6EeKF^XY%GoT6I&lY;gBul?9jw8WggyZ)tVgB=?aLw(Khh^#QMbYS z_SMsxfwLVJPDQ}oFsV6qO_^E3yMC|VhV+6f6p1Z}To;p3qx53@2`+@_8SQ1c7FBYl zBl?0w;|0l?1w#f2(;%@#pXlgcqg-YNQ-^?%Y?dAb!bfekpg&o^dvyMZ*2XyuG-TMq z2@l$cE*D3YM+H3t}F-Wll=qUPTj^vq`c2`>jHhQjVbI zqeQNG@;BK)D;n9pn{I*Z6krf*nO$lZ0qCWN@9b;+0r;}k)awgzO3GfmR#rDAT9`_A ze7;xHko2(QD|DqUI<6-H{WU}k30CvGzLD`f8vibP4>rWw%q$MAXd+yOGZn}Z z;NHUztr%rPF-HWe10OHX321K_Q{dhgZANaSr-&%FTe_vcZXD#;1x z!X||cfd}pDHuW?B8vW-KI{_WNE1Z(pP{0W7&kyS2H6zb>=U@FLmcq?^rI=#a(a+?N z0G!Dfao|CK%~ys_x7f>_eTP&?>vbKx#1H%x=k+ zBc^^NnQX)B3LS(;~C&97^K z5^2uGo}~%gaWZOiqXu!L&2b>KC9W(vGRN@V?XZBT-&EO^1p$NYbMz_UIl8psBwA=o zwnCMxnIJgKL!stgv(%ioYHVR2a_Dy;4jw~jFimWzxNuU1+)q{FHq{zZe2^1ewhk+I?AAsxFNlF#MScapPd@2ynEWEV zAjX5+3E6Mm{?Y);Z_?p5k9UBXE#}8k%J>Fx8=lU(i=FBr=QRWlp0Nhb3*C9LdoqXo zt`!B7GfR2^?s2Qp+X?Ly9Rwi4Pmnd=J?|LF2!P?@l;g4JY`cAV_XwDPTz|B*7o@%cqG0l{?^=V6yOi}2HY zE>PZrj8E3KWhvIQgJ)Lj&6gWD&c>wNexeDDw>!2=2r54Pwe&LdRgg78F1^D%Nu9iycl?rP@ z2J7{aV*3QBuh+vw0TK`&&rK^B&`B%kUZ@XHbsv+vF7z;@Db7k{$VE*8fR|dPZ{c01 z&A%jEN79+VOp$5!2GuvfJEUjaVpq7cxOL6(WTOrSNQXxo_A}hyM2oYi(eb(4DQaq~ zgq*CfM{RA%x4OU&HIy^NWx-4pvmnzOiwNDC>^gp?wL2S~0<9q9kZ;Pd=dnID0ukiy zeCZ_bNLG4zFGyRBD1u~gWUa2)^5WOKuUHmed^Wveuw zq)4izmuiKONyd9#0L}CCuQWu~ml%F1cow_OL7bW9ce0YYK_blwf^Y+we9GwF(4dSv ze2@1jOFfNNBO=YKKth@3*)Zuf6hFpqYl;3NDj%>;2l`g(YvM;IMx>|gy3Tb?P51eo z%U;{xaSD4h=a(wZT_9jT@xga4l(pGPE9-jL3_*Hom2dx%$QXXqu;@heHLVU;iBAK=JZR;_InRcxBiXO^HXhIdP^hjb1fk z+w9YCtkO*geyY7M6{T^{Go6bTp~r3_#`S~Hx50}Cb)$nhbar3&B~EnLBdvnlB%bx6 zS-t`nz|Je~>iyoRC)F#|QChr0q>s+%@|f9#Bm3U%!b!c=%s=uciyu#wCPy}7jC;DT zaNAV@E->q&|M%0Zk!`lusZl*6%j3EsGFqf4P*FMO1(0$d+fN=+)1>4Vh~}@OVdBnx zwQ6Ay@Y(dIrN4c!AK@DmtI$iK3uTqi07BrsqIU6Si&0nG`hl*4!WM#L6h_(?dvV_F z`{_(O1>_%=9g;_$esf2~`r>Z&5>|cCeV6A71-HS)712^G8JpeL(lN~(p)eEAXa1|Y z0rgMWUL(nO`%V_h0uH$oWV8QOw_FO^nva3_t+|hzgGc3P6~09niwnZZKv&Luwx}Y! z|BusrX|yi8hLZe=wcZMk`h)$}T-QMNT_0~mIYSv~;bk^o%_8hm06#mNzN%bAZMORT zq`=#=cNREuyzN;hTL-X> zkY!_zlNq|xqwByqJ2lL$2%U#aLvG)&y-6gKoSnZxfJ1!>*B0*CsTK_n94CvpH*pBkVt zj}5}cE5^QIH=UFy#LU`f+uk`i1spL1O-NTv>5iuRB6dVBc3g7-XhvDt{hD*LQ zTh@;t+|;n!d}0t2=WEmxZYm3HdfW!Jd1BJDP#(sCogwkG8(O|zPFWX!6}z}OwKF+V zK`4QWXcgC|UB)&fN|ODquMWSRn~R5h0@#hdU_Nbe?92{6Fiz^umoNoZ#_ve1nE@hz z$b;2PyLqVJ_8(8kK9y-tbf{9!uVVO8gh^GUitN2=^@l-|7v7j6=nKVundV#*)j>BG zwC7lP3G#v-#VXJJ$wRd7r?k9`7DY3D){2;7A^_l|0^{q{ap9vi6GkfMNvA|Sm<5KvGMkWLa> zf`}9oIs~NK&_|>wMWjoIR9fgoL_m-tB?Jfo1!)OIO6Y;PgTLpL^`3L)omn$$&6@wH z$(_4g-@U)r-k-f^9Q+9XzYeqDmT?9uMfac(idHti;35UUz3jHvA7S||#wDH&pEnR&gAwZ-UqfB<`#lI8us8 z$$Qbt`TnWp_=B{;mXfI{7pAMQ8syos?uwrZvU0I9ygs);x73zp{Qg8^lS?$pm4&Kj z9NXI7BPek{=OMizxcXgPTpz1P14!&okz9u`HDVixQ3Y@bv~_z-8Sp{>K}3|nScZfD z#!{eUHFglMua6VCQ8!fkED{x*W4uzuXRAF0_}vmZcW8U<71zt474RTKA-z5 z(`xS)0_+SyKfv@vl~UHa0Rtj3D>k&d{<1~7{+d(Qx#Hm}4)uQ2yqlU>>muximf-<4 zP~D}?=V5`#4?~mHR73~YxVCC6L%9Jd-7IRm@f><1+Pi4^SjS@3Zu0P-XJ|!r7d}fM z`vzYIx+c)%5}>9rNdDnav|w#BJcEjo6tGdaJi~di6%LeMe;Q;ti$uE1B$?Yeg7I+W zfPvr-8v!mWmV=S~`{waTUt4S3AW)>A&9n*)&6IwvgOb6kc_s{JL^JzJB$5@t-;~_G zmqm6>Cr7P%+oisOuGvOAr0e$ZrmDkVn~y)&+z zWV`B0e;j0KfOy5I>!T00KrukksdgYxbxNhfS0m=*DVesDOz{*aU(|37(7Mxn`tFLF zcdh6uzANe#*Cg9U(gM6prQ7MN|8~VI9n`fUYAV)*SBPLfY)H?FL8H9 z+Pc;4#7Cds3M}hfmx$2ljHSMG8ecFq{RUTTtvLr+S&|Gi$N82~;JE^ByZbMvnO5m- zqpqyd@?ALAUG^LhF$e>V0#BCl+|P4u8%}MoCymxe?i$$y4*Qj0>Mb6db;n8Emwnz-Ak<4 zKTGZL+5s-hOHS}mV)@-34+ER)Biot+g84a#N$_c-lV8kEoqu&5xZjQ+>ZX}|zUeg+ z#aTUW&1JRRCLJ5m3tZRLx2#)mP1+1L@OWq^V*{stFiMbfzMT2wB)822yGvW1Oa(x@d1L-sy3@rrnimoiPf?6%VlsCSR;Zj~O&*Lhrz1 zSS?o9Gpm3aWjPRp5@hjnh`FJ_HhaQP#$dy4H{;tTB|L1FwH9p;M?Y zo%PddFxMJRAcu9F=nG1X3VzI+(d1VhYQaOuU=tF{jCzYa1>Cm&-{0Jl&%6t~e?NgH z$zUh)*!Hws8wGaPaBeR<+uFm6Exs0o>iQg6Rjz+>qVIb}SkLBY`E+SE_R^{}dTnO) zQqN4RlhAR~&IEh?J(i0eVstInQnuzdMBVLW(#^U_>F9%#FIytT-FrJvS#7rgu}fPw zS~OC+iBq~eDzl+2M(8+Mfx_(;__r6-Ed_RRlvSWz??Tbk(Z?YzuX(`GZai|DmQU5^ zC$CqsMbOWxOuqYMsyGB6w8?5J+li`-LQ{GyqMuw>BRE?-8BW<>54)qqt{v2omeDf1jY`I{*HKjiHtpos=YR^~ zx%9@{Tn2v1u##s<>pBO$AEZ@;1~w;$dY8R*h4?YUykqB$V$7AAs*yyl4^$X+0yPuR4j4lh{GYD6B?|Nn3QDru_RM7Ml7N*O+gAE05tC^-B4$3s9*#p*Dula zf6js(zv`O4d%Y;pi!|;*2BPl9yo~63i0Qyu2_Rg|Z}+^V%*%7HD@r-N&At+_>jzPk z)w(kZL)*G3)O)yFWzoW7QPo63K3&0ohhwOs)n>u=J*}w7Q}IKb1$8si7Lr9ha|OKx z&T{P_B?X6O$xRtR>^{}_atikQ1UmF?9&m$c$@2_hj8>8A4e$c}ujsczoR)KfNwAf6 z>E-j>hW({sMaa@cCvQ~x+-a`&*NhqU+j8gJ;K8@v!FEn$sn} zmkuC`)?8rMgj{z9U8t9IRhk0ED$|=s@>sBybZHWFI~?MU`{)`bv+(_7^`T~ZXfuf>wpy@XVG_fxErgE%`VK&20RX9T7ziO4(~;b2Rw*8I?3v`61(2S<33dC zP4o>q`A+*+@)_ef(g_{mPFj^qtw!e*MM<>4*{!VW7f9WT$q{9!P>nei!u6HBHXGlT zNf7S}sMootw2C}WnE_LY$%;c)whsfw^C_7Qx5dd(e7%R;>@bYWhOV=3-1g}5?%@xS z{m}-{hsu6HeW2`g)5k~VHdtw`1~EL^vXthyqbdR^(AGI2pmH~5eQVZd2@ zQ}9Y3RHxJ(m9^txN>lYI&DD*WH*+h?_DH7jJ@hGpts5n2JgUH}{l+6~DVeE~J>Nek zT2*}V$Ey2}_F*ZP5s!nbX(w)H?=&iI6dc@~9CR>9UbM171V4pM@T7x$e)8F;e6%F` zdvvQN5r?rZoy%x}oYzc!U(OZY<-#2eEyL+@cM|>n?IlhD!a{pG4=hBd!5QX8jM$XOW42{WviZ3djM&`;YWNiq?5R}^#yr;cU67oBM;@WdYdBs z6b|@eB?w#8DjSmfX}i%#A7dU<_+`i z&y#raX4T)W!$jF?tm5nMz}E^7ioHiYxhOiN4zaS$W>>V&FFYNjU)ccxo0BPmKwB2h zEd_s6EJ7(-R)_6Si(jtkasM$D=C3J|)ykKET0hgOpHYGB9*F6{rCmh# z==RMwA$6KY8cNpn1A*E!c}+j{`_?nq8v6(`q5x$uCTkN`KBrjskQMT^i+6p{rh}46 zS1?O<;K%JW{Uzzot#?FaO>WPzwMFb3!$0t!4xClCGKOSKB}f8+Scj~$aUR#WmX#gtA_)Axu?sI^!g>tP&kw4G;A~S&M~#;x|gqGmHW++QzrK3 za?fmG7_sS4WN_;8dW_nW57blCF0--u_b%YGk!wCV7}|8u?i?l8qPu3){&c=G*{KBD z;GCuJB;;ba{cD{PnLiJ0UTuDOz4LWt__>iYsWS?tVmt1Xq?;{Bxui*ZM{Mb5lQ@YV z!oT=asL>T&Otf~1a+n{v&v~LKM|PfZceM@AyBVTraYhGdg)IJZCT)hFUm?(^f*0%} ze?Lj%9FH2;gY?4$x>LCq>2%g*>`y+kbAiXZ1r7np(Uc{yd$~0mwHlt=?FgY_ZHF6A ziG8=NoKF?{g+3;9q+r2yG|yMpQpK+=UEZ*qGv@NE9l}Vg`BsCS?sK0l=pi})y;`D2 zB!#*gvYGoNztp3tdL`kKhwf3kjZEw55$uU#J;nM(H%@mC1?vbgOd!7vbyGKgWdmQ1 zmK1ofT3$jG2X4F?aOJpA5|jk~`Gd;Qx4kas9TR7Oc<)CZ1We^K$7fnwoU58dWZwxX z&PN)q%7wCaFZIl4@mBTzS)%rvFIuPLEt z%9iiyDtzP#0C6l}Cl)VVCAEaht*xr%JB!S+yC4)ocmjU^)G6Ju0HO%nk4*%Z@1>{a zX*q9PR|jvOVH0*V*VF<`_zyp?0)@znH&7#IuJ{`~d}x6~IbK{FT?J3o3j%$m-P_6% zebb7EP`MBb7_5=y)QqCP_HTXS6Znx+S=5h(SAC$iM_OwPKJy>!`AhypAO&!#bWJ^0 zOta|j6>xgMN*Zt(vyE7n(V}oYJl7xdH2IKTc>WBowp%*s^)FEKlO|0Ap=)9?rgz}F zC3fW}UJ5%ze~&nN%V8$MN@GAuH{6AT_tHxLmV(>Yi!2NVa?+Q~v-^5`hf9so$W9)= zgNx?B)G0anP%clyEw(Rz{FE+Vc822`_3G|jhl4Q(&m!LszC*~>-?WYwZ@Dcs$ql~} zWEGoTlg0n~*;50fX6JK~*WSELx=R)BfD^yeFzzYe%2vH&G?O=)j{SUL=04Szw6iQ% zyS?>g3ZO#00bKTO!DxNum1@Hnpz7u58bZ5#sEK36V$qc*ID%0wBbbpm%jq8jiE$R? z!@T%M$>ixk$8|$uU?!OAt&VNWZjnT~L=R$SDc0l3wFln&<@BC9fLjcU+LGc+uY3fA zS5J1@Vl%1EwOR;lV0kSq5j`BXb~T@guz0w=e)2+X-}E!&$YkROR~4`THMmwpuK2ZY zFRCNrJV*5w=_GfL%i5}I0qMo7guDHt+Q4SUd@ilRo`hJaUvuU>xuy0ZQs8KYL2_RT zEvzu)2(RpL1?lxxKWnpU%0&7v11G;S;4;aIV5gPpWq-Da)89Bgd#wV%lOWU4huWN> zweOV~KdiSRr4wz}hWdfZ?evDKtG*XP1@y^I`V4+N6hx^p~S@nFE))s90P))y2T&W|@z%S+!5dOk88 z`dp5R$#Ee%LP-*r>$mqaJ|^c~+JQ$w9C}o=JZ67h@I$&R@EL8inc@oPzaf$b6|Of? zYX@lo$d@L|@_jM^&nu0t%&M>djt%%;2KUelS~FXV=IrhQAbS|FIFxZg}zw0edSr#m+Pmjf}j@++JFH z2Qbbez!iM`$SJ|jyHA5UtQi0}MUrGc>aW+uoE4Vbg7}bMCWiwYI*qI_;D_IKHUt$G zMRWz{do=3Zu8l0xlG#(=ZAN{-IuLfuNnO%lMm@IMyNR_)Hcwui|LE3_O`--pN>Fih zDdANZ{&ugqcBd|>@C`=UTDjpb``_6X+#_j`;JB#t?b|VDx7*|c##D2x*0;S3EX+H3e8so$mYJ+OsZ4wPYD z7Yw-MY&9;uhWuu&lXEk4rZKDM?{3p^pa#Pkw;oVeG`ub65LC31lxXTm9Mi>>QX8cp zw~8F!Dl7=kjPzA?ytc73?TOk|m?w(e@$o2Uu3maYT$&NyywiBf{!rf-02Ta{T1DM` zZDs9>(GA~j38?ja@U~gaBhO9nyG{#dO;&mw;qOS@;c6=*sh~TkS64uD=d&})kCDHB z;=L0!3!aN>m`#cJfZYhwA!WbQ7QR+fz9p^;G=i3*aQ7xwkH4-y5E!_Shau(& zUp~%L@S5{p-<(q!-S&KhUz21hlS}dNj!o>FWyL_x+C4^`I}e1NaB7E?5$dypeUsLvKtl9axnAi?UlKR;iL1PGlQVxVLA3R*qM3FD`wQ z2%29-#{eCw&`Gq+$@9z5KjpbxK!@{5M@eOgvywzl+c7Mv_C2i1WxT<(3^*r1j!PB> za2$f_*qO&+$K&8)CN}oMl2#mRa<;`l1!`$YBrt&w22qhcpVQUiEKiLmO;Nwf6 zgZ-Rjy5(}n`+mACzM#FUwo(JhWWiu+gy-EgJ;_#9MH>RB523c&UuC)g7Kf5{O*OoZsVAMXE3%4YGLseh#`$=$p=Q|SXf=)V{i9Eo z-P_b<6Fh4jcC9^Y(wj%`EOnJEU!aEibs|TpfpaJ(aGrXAo|xG%013m}UVQMJt6cL_ z1G8G%p3n3VQ)|{vEY}q`viN3x{mTF2z0zawVdQlm?&1v4*<-JBG zh?)j0jD=iW+F2Bxm%rRp-+#R`^vaxT!)-$EaOjdZ5dU5W+Ao1dM1zaLV|5=`Mc5Tr zg@7_7`9Jh-kEYDbJ=*?s5kPPX-}CSPZ&dc7DO=7vYl>ugJ)yObaNtv8dhk$z#@33x zA1fyCu=H}$sdjrl+0rQPBe#rP1@C)&OrUTYHcw4ZM&-Sd0R&dsn*)HZ1ILWODAtDT#nRf`SN-g^b1c%k`OXPrM4 zkpIuZwZGh%LRXj5qR?-5f+n-(?*NCrPip^vMk+@E&Gjt4*+}z$|C?+E zHkm}X_5K^j`31g{p!(@LCuZv57~?=HWnM80U&Zi0p6W@EbFY9vbwP}uQv8oAZ>L}I zCOq=)V59l%+Vd^O+uSKTgt}inII>`J`~w)>@VsL@ot-{jg&?{FnM85VPHcKPXn?eo zSEF4iixhF9+eX5)IF!0s^YUJ`KHH(ceB#H*i>z4!8uuv$uoBJXd9afnq0(Hr-QWCq z2^-BjGTW$ebA1g-tj{$~SMlQ`ISo~B_ShbK`?;FcFomCvF+IMy<#oJnCr;felyG8a z>r(47Q9;aH-BOUU%FEy;R@qyn;iD+!28$4`du{PnWW*ZBMA^Y*g~zQ~e0@f=$4n{y z0k;dbW!^i#MyC+AE%toOn|O69xL}P8rj*rJy!hujE`I{tiLhuTP(Q2UKne<6PqIE& zhw1SEv}U6>8_2u#mMf!!R^|A2v0wn)(qC&>;=DG=xo_04{|VhlY`8+-RkFpW!@yo~ zshe*D?@73(%Q?NdpfnxT=1|4NR4&0>ZI!f?)n$!TRQ9J1?v}0YQYf_RT*l4o{#B@< zV_V@PZk@f1V2yqq6cK@Qha?=}`wivdgWTx1$uNH|kcQpA)4N%*=*sVQBioGpm%U%!0`D@nJ151jDiAf-o?&iz zX6aW&AAHB?>yd&%|IN(WVDmcsg44FUw%cph1`C)*(39*!b+-V6^m-H_L__+q$*Vp7 z_iKjX_nN?}L$=bXhyW7?3+=LC^osHOql%MkrA)Sr za%Y}z7WyzP9HxdHeoTsp3)E~u%#*eV$}{r1iRA`L(CeAeqxr^-&s4brHYlK_B^Uzj+T&COQ4VT7 zV(c`Ml-r*^j0}vUJr??i@zI;3)M2_zLpn4_PX$K``jZxD>s612pv*G$cfUABa*$b~ zvk=@BdL%iJY1`h*O?c;HP+7eD;BOza<<@S_@R1sFd+(1|&FIkr$FQAH`@Zd^a@00d z7_^>sx(kuJkf9SFJXi1{zArzze0tnI2G4tWH{ZfLy~g@iuUM*z7%Vkj)YhjJw)r_S z#_;9JV%osw-otKCds-oQkWE-Oy%f^KmplS-8pvuB))fcW?~n1ho4);ul)Ch`iZ7{; z3(71ur+<~VFZCmxp@AmgW#IInt732S+IQcHiWxDRc^Uu-z|F^x2lLT;#B@Od{k*zP zh9M7?YI0+Dw`@~j9!6>Xp_KIM3G?Ym8(mjxhV5>FK+U99PhNNJU01vDne@*fS7cPL zIsSdFSr5^lZ_~ISh_mB+Bv~)D9@p0b39q5Uk1y#9?goZ|Vo^r#O*Sr=@EA|Ik3#YS zEmO!VEzj+A_rucHo(wudBpVr^0CUqH=YnM^HgO4g((o(6zRu-_ZLSsO;>&9n~*wM@KHQ_TiBi6e4ERQbz->h5&r5hbBw8!a)M($Z`N%d6oPr_)|qw?gaxk9 zCeW9-+5#@xs)wO|jk-1r{*-bEn($@3^T{k|t~KsC8N7(}n*pVoZmo*8KM6QTP%%cI z-&vJ%n2rPOG;;ZjU+8Hi?Pf+B>efkj^tx{BpJ~K_3WZlUwbpQQTN|KaR37M@TJsuKGe6DvwkMER?Umdv}XkHKT;8T^aKyE z*efRbzRsU6%)iVxcEr*3OllYFmYF`LJ`zrRXL7(5SbM*_w3nb2pGQ}xuG zDAx6f7GK!1j$a;tL~d6HqH2NqIv8y|+A%m6sW^R}*cR-9@k)CP_4iA6CDNoEZh(vi z$>}|U=gSeh^{(KpZnLF{uHM42fe%d&gM^Q3lV)GnLsl6&q|M&b{EM`H{|>gQ9rtvo z%#>Ade--4_s3%xUcxKa|FHT}+U5cM+gef{q7{vbOGLJ*9ADQ_oF-m$`?psL?Uwf1j zwe2);Y%Oo4+dV?6`aP0R4Kcl`9HM9x+P_`4x`5~rO+VWC74Oawz06nPu^$()=sbuo=pOYnq3!z1)AXx{W{j|WMIuO zvFOPgvENn?#h3@~;4DQoj!$o%s%*R9VX6~~^tr;EjSD>HsJsKnNs2d49SO$-ONT;R z%7dV;`efQb9$9=@8Ym&V@95V)opsByZ1F`NpVA}NNu{JY24`Ms+n5JYH$Nf|r0t(* zORK+ES$@a|FT0pGL#;HQ=1^y_+eIet#K!mmFVLhH=hfH=!U`} z^H@*-@9&yJQ-`x-^L+HkwC1sdzLwgkZ2R+~$kA`3{foeU)LO!E5O6(e>ifanp9CP* z29MULNPcltW)n{y@7LJP$TkuQl%fxk=CAu*Rn@d+X_}R8`UMMaDRAX+-5IzwveL`-Be0|3%xOn0yf3&)I z9|?A6E(zJe0YTr2?lhhx+7W9NEDyV^kt5_0Sh?+(p%HS+s%~)2%CzDdDP5R1~{VJx;CF-I@PWM5DHQ|Gygp2xc{xOWsBLfYc#b}p! z$mU8a)^^IDexf>Jp^os*2uuF>x(j(z7w45OMT%r*n?+hjy6RAK_LKED&6h@DxRjO% z&wQ0=S8!*$p3a*2_C`#xu zOzju^cx52g^Ua4Tbl*^bUfrciKA8u9Zw3ux`$JGCDCxs*z|tB@{jMb%S4#n zWvoV=PdRg}swe)ivLCoBTtnr`o7>}dT%g1a;`B0|(>&j{g^<w53h;NDBoK1KL8$^f`8_<=>?Jp|BXL!Yz3X`6yAqwSGvf!@o}$*4F3=a z?h{r&9<>qM`NdPQ2$EM8iIQ}qF;fhA5M}p zy9gS_86o`jQR+rQ0J7p1j>VXuzKQs#^629bB+_I{VD0nBPgYBL-Hj;AJ-!AqVGr%2 z7!AN70D1y;3U?I#rZNp6Mwpw;!TI6=iqQS2YA|hSpz}KUNtOY-P8ZeNKXbX{HATwp z{Ib;m#52WqFExskEajl!ZOh4c(?+!8Rr$>ut3y)4KQlDGr@IBoY~Jwr%^UnPD>4cN z9Q~7T@v5prun*yrtAWJ8k&WZ$V2bu<5y!7?omVW{C)5gE$}eW?U?0?S1YSCBqQD+( zTs@T-{nsTa`)88Is63RbOXxDdXK=?|W|2sS(8t!Qe!>u5Sb7yaY63}mC$Hr%oKwa% zw$XRC+@K+)ZGgS0cY^Xrb&^F>sC>zl{H=VH4hr?whynS6v>bmigr<}be^+IGdyYT@ z;D&q;mYOL+W%r%MYz)U+CfH^^406Y2M^p5KjhpcMA4LcpQBh}p5 zXq&^6V6@^m@oDr@iLO z)TBhEQ6g~U%H`ia$Uy4)iDk?<`X_qF(u7fT*o>NRH$%7pGkF|!^`fmWmenfsDTn^~ zuZ}TXr+6A82w(u9HN|BFyY-R==Xy&TG z2lGTwzT9%So0~L{^<&lZz(lvfYZJd(po)Q?5@zNRf$A)v#z~B0QWZ$E%qgY}pbW81 zWIbScC^o00?tNd0!~9UrHZE0lxOOu4q0#H43-KiY>7$WH^$}+bOi`Nk(PzUC9rRA( z@-2Ie1bF{_Hc3*MA9*1ZOwo6HpS*VPOB{L-H%E^+ zX=#dxsO^&q8xxt975}^LCO}G{yTBZU3(V_wVx;SBcnykVUu5*#9F%xquPQ$oC$aTq zs6scWb0|$Kyn=GTo}WZLKv^7rZ`8(kED0A^Y#+*<)h2dufU4U-Ja!<@eb>O+Zb8O# zE{OMDjsxg!?H_*8_KbLi+qTxgs6k&!%U4#9a78xk_O0dos?JJ~1#T4@dDLZTY2pZ8 zHl9yO49WtfY(xhb-5jSGA_XZkz~fiAE2YTD4j5pePW;L%McKR#%aWJM+LzX4`0{KH zXHxnk8YMop(Y#z<%v6k9h384Ae#aX69ifI0eCtQB?YfY=A?5)n9Ox_rYVAVAC&?CF zBFv2-KET1q8`PkmD?yYyjKj#+N=Y=(PYdcvYM!h}x23Cs+u<>Ng+LeaV)r=d z9M6o`ymr@(c;O9$W|b$fO9Qz8)eCEBANykF4}GaRoc1BrH)2Tf_KuK-5Rd!;(R80b zE9-a28JmXLGX@(Zh~Kf+ITZzYS+O)+jQZdJUoQo=`aSQlJMHR0a(*&*sL_Au?6 zCCl+VXSPd8=I8Biu-S7(bZR1f6NyGY#*@Mz{R74=zLPwS{mMP^i_tJ(xiIH#xb9y_ zhc~|{YQy&#}N? zP}W;>m;Yy;k#>6M4Xmc9eEqrL5av^M27Mj1;(wXfTa$gV!n(1UDOLg?d}7WPi$7_? zPsONsb{v~wSfDzfJs)d==t=?|?6}RlgAV?*zvtzl{AG+t<|KNYp@)i;0EktCsAGw- zBXJVe)qQ9Aqn$6P_FCnjTw~m_n^zM%;&nw2FVzXar^Z84WnqkF*&7L>YC28zB9-N` z1g&j5`-NoQo58}z+rGIL!=#f<6x$?940g_x*+R9pH&oX`3yyFypYHgKKkuG;`n+qJ zL(tU+%ep~q@WX@Mc=9t5kIE}79?}oJNkHTh?)|H~I_K%Xx~nmzeyMYQy|^momi2JQ zO}!nS0O%GxXn-1IlKKBfmBYvmcgRM@r%buv#z*>_uBRcb%g1oyAL7mj&J&H+KnRHQ zNG&Z+2#l~G9dj6Z-+VoKE{Hk5EI0c63B!i=SPh^W*>_26|L;g1JZ4`X2LRu?+Q(Suv6%(_6rphDIU8Pi3?;nrowU_`J8(WUp6^$#A^ama0Y3>%V zHZ$$d;T}11!x{b0rRi6~V$zHXikH+)&3sZnYT_!{waH@S{gmPXP>j{6$IJGe>dY-! zcrV1^cJ!IJBnkVi$F-J;v6_643ourC@k#g>pfMcT-V`a+!DjJq>@t5aPU9$yrf8O4 z0)j{ch&jw;(kWl*yLX5~w$vQ_r%2uB2x!L}1jk_(vd`+Y@#+gbXmSJ8D_zbyne-F> zqX%sEAdtEzJ1U)dR&&zVB~qFd@16Ey(fdTPWv-4g)C>bWO-1JX5Kla2=Jii!Rik)H+thh{{ZWKaG0rNoo!#Fzx_>og2L!rFm&C3Lf z|0IDCN9`w&F#2@n%%0@_n;2HTphBzTQ*S)vLz7UNfMoOI_*MN1Bt%CcT%~d1SCj$9 zQ~*F%|6Gw0^ zaKCayQS9nPUR*9%;p&H8>|2aJR$BfeezJTb`3y%*l<~Dhdv>RP3F8ZKMHf?i6^%TI z4u|Va(4|qR7bzXC!U!Z0deES$Hy#H=EG(`!_jR_gNPw9a=kbo($K$%d`vI}d-qE`8 zyo{7uDcSnr&oV}MbJe-38d{feVnRGibbO_Z@W!|Gp9!u~SRUBz;HOV?^5ryB_jK}g zv#=hE_!RuWwWl@o#zcEz7&!>l$>Cca9icLDR14~4R>A^)eYVnu<3YY6X+mEMfbqZm z8^d{cFg3+!5i&7x`1JfJ^ewk-$MA9)=rg{ksL6U@$0xwBG1%^a!8aMJwq(3yf7r?RQw(f-o zE^?FY#*U8@G`Zj<-}%Nbjo)5-^g`^QOGZIkh2$&g`Wtq4-$3zf7~cYkDfQ@&^c0kD z`IhgM6WS;x`K}|_x@R}U@OsH;fLY12f57sN5;{%$hk@KJQ(&}j{MBDWY&hBH1Znw` zqx=KKOt>QH)i|2SIz3Xy6CmGxqGy&r+!M9$Jhs8oRcDs2`V9V?uIgOhxz6HK0>WD3 zHpSA_KCtaz$$T1s@jA!t)AboRGxX5(Vr{Uv#BRP5H>JdBWOV7vklO)5w*wT%D5wqb zzUI_2?~}%&WIlP*Bz#5z0K8RdW^^Cy?}R?M_q!Kj<{GhhECMVu#Oz6$ZB}|3jXY;% zX|Kgk;f-Z)id4Jzypy}XDhj29CJkP52pL!32sH1&_Q^oW)Cd*%6mTnxVP})dQ)dt%s*fsIzz+{k8ujDMhi#GU19XUx(-l5; z*Vq~C5EsYg{@aS(fe+sQiHj4E7Dcl9F1cf(q$0(mCK@7M?~0~2V+`S8E`8rkjc!kDQGM#GQ4&z_$!!#|5!`yMo)g(w|6kM#Ol>Zh1? zsdhfaCm#cg@+>$AKqWnKETR0-^2OQqBE)+iF^3B^*Teo{v(VcvOiWN3<47-Bew*yLndOn{{IYy#@ zwO4qUK=`HZ@Y1g!xO&hS_{2{g!(19_mcxj&BN1Ro(O_YoP+46zK1sPR<4Mq9(H`<} zstX_#|1%N(-z1aA@!-*tx74k7Kqhtu(lU>@au3K2I~(JG z{QsFsd~+x+*s4XmEIn(3FT7m4YjCkDv!OsYl^S+0UYvfOMPc>#BboDDTD*WE4z_AJ zF65g-LOww{6v8yTQ0`JtC(N&|;#xB6^hojjBblaGhnTWJHq-h2eP7yX#oCc;wmUxF z=dV;0)R`W3AxY`{rm%FRvaRfunSSRn2ZhS&4~OE^IzcUZd(-UM3+@l=IC7hobe#J; zIB`Epv0gHr*Q)k&zK}z^Y}KowE(Uf^CHt!icx6vnK>?=&fYU$Cb(&u<{vFbWlYkmS zC%&&}LjV+7VOz)V-H6r?@{NHStfTCF1L(j@(kkLwHV;+MJN{?Bwy*sbh?v;W&5G2B z`j>Rb27ccQAWq)T@F{7xnnI#jq0KDy94;p5>DBrhd{(v1JezuaXJ_m64URUuuG-v% z+4))D9CTPMqzg1U+omCmF?}+~PzPnSD_rK0apwoKJ%Z?kh1X4w-N&BMRs&F9xkmH_ zhz7?V5*7cIai$Y=Sv8~kzEuAGY|;dUqSThE0{%8CAL>}T zPOhm=OuHWUz*zTZjdg0gUKSFg(p<=z!yYXj*MI+PoV1&!)!M_G^Rij^303z9l)hm; zkbZ7qpxWTO2h+|R;eDzw5Soy5y5k`-s%bWY13dw{Z=mq-Q0qsyc|Mp%J939iL-U^D zSPD?U(-nckda^uqnLX)I@zmSP^Gaqw3bs@!5$ND(O}NKy4jBFON?R)L#gR7nZrAHp zj}+Vw$HY3X4xEkPyIfzRY0c?kn^y) zKHR>RE7pW>Jc?n}8|{-i1SF;kRTV>SkL(OB?5BWiUZ#bIZ2-Ld5_G_xt5jrDtP|?^ zOnlTb^00x)u>Xf!-^rWJmM^}#v6jks*#hIG`w7p+X4-sF^1eLaEW0S(xOTEbvK<+S z>C!P2;1OPkG)H4kPt~d{+nnKYX!kV0n*-Tt+Mn$75eXmyE8#l>4bivGHeL7}c;^kWWBc*RZn|T`VhE)2GXrCNT zjd~plI3VTE#jbucYeCsMI_Be}CiOwV*cHz*_V%S{{GS(&c>mDj64+#8&NLs+>S8;b z-p#E(G;P6RjaAqHJ}GPg@`rs6z3>Nxb`jSzf+`y|D5tG1xO*zIaetXg4@Pe;MEfY` z9o(y}@euX~>TpUA8~o6_#e$-Jif>QlpPL5lq1oL&Y_DMP9(cJ6uD$^K z+~DobtoCt}ggFZXhN*92iC3!ZPi^90=b%LkXbOmh2Raq23@z9fOslUvJYk-UQiIOL zEPdlAwC0IWh?bbB6khQUX1vy{MD!-1Qs&n3fp1^GfuYN95Xj?K12x@WMYEjO09CYh z!*&+U8Nf3eOmQGd8BdplIl3qd;CygfdrAE~P%t9bLK+Z{RAOaLj;I$@4x81uFP}gD zyFQk6a;JdJ>{B0JDmS&<*pHem74OPv<4eHcH@WS}-bSxA8J)`u*di$U8woD>ev@-C zB$#zC^w9ofp@qUJ6t1v@k_A^hxM5;oElJ!M^ zU%qGQIyy7hvtr_8^K1r1`$c$n2l&Ssn^vRKWnQ#z$L@-rsC;G*#NS-DKLG;#Y>Fe1 z^de;rU|GvI$_s(cPDz+`&{k|(0Hb%&sdA(Qh{yHKJAl8u|BO|qxbJTLKe~IwXEdj$ z5y)fYpJx$%nXoYFWk(-ii(bMy+>={a!fR1CTsEdZ=8(Dz zWQ;1XW%A&uRdh4zX&Df3n;R=Z-jV0H>{Y$5Ij-nD{Q$ValR}#P^fu@e8bX*#C z5Gf;^2=nOBk!hK(O5Uv1#*p9uwi)sm<=GBO{&h}7POA{h?hYiiX4B!xccC3T(XU@c z$QA;*uBpSyy2+Hq?ER)Pmqy*7eS&=5Q%cOI|v*812LW ze*ZQjUc2d=$xn>Ac9g0QZ*=ZvA+8!3;BQc4e{bjv1e&}ycS93587z1PETG0Kk6CrL z!6c{&BUEJ}mQBFJtNgv>&0HlwPM51X>UZWbYrsQ}!t0nAhz4sxNeTUzdQ12&jxR+l zzUn|=3Y3~3A8HnNMzI;l&9whZwpZ@=q@~f&{<3aznI+wWeb{!-Xz}ISU7fP_Q4+s^ zy*L>insf1IwCuRMO~T~$`Jz&(CdJM`?@lqls21REwu?HN){ac`^4QC4r2>?lC)*1< zO8IEFpvIj2wWqVjkK;W3bNBQ*|4KmMTQ7dVg9FAm9G6h4&M)g1&CiU-lz?9^T=6#q zPS=RlouzLb%Sw(3b8bCpCH4l|3XR@258Hrtn&boFx7i*bsz8T8RY1WKix6HmUj%pq z3-eVjY8S>upD!I8bGJL~onYHr#o3F*UBj70&#v);QyT@_a9s}EFR^(I<6R=C9YphpAAw>9U{LO)I|fvK89VaI?EIo z(_s(b&L_IWQG~hgi&d^n|66k*%>3eSn&KbMgkrO&-u^=EEd`2NUV=!Hl$iW9oGNlv zkpc%S$+ncUJ%taZtbm;h@qBib#_UKV14Ouh9}eW@_@%xMaJpGxpg3h?@N+)PL(bgF zXU$R}WHv7|SfcVIK-GCu2sJUBhc%fOjxjI51{Gu}&cg5Wpc3TRIuLD_$II5)*HY!Vn-5vfQ)^$$oy1M@{V^_i~*G#})=G~k=IY$f`T;_~G;u~u`c$vK}Q z|6KQ@0H3|@wQ^rSA232d2L;e_5?M-G!Oar{EC9rQ3?SxH>OsP5tdR+G^C$hJSZgVG zk@QWV{zM6%b$#c#o*tL2+99=t5$~aT&*RB`i5FQDBmw*eFd1H~07%ZoLaXF1;IvQD z2NIpJwmHR1Y&2&%bO-?TUs72$K}QTfsY)6(EMyMSX&zAKF|`2xta($%i`Xct8Wt_` zL&D!2$k&OCQaz+HD|C4>ZWEl?thfR!t&uGdGO`rYSEyTi;Unvjv{?#J112H>;QJ}@ z#u0uLkkOgPARYosED39f++Zb!L{8mztKLW8M}3HbC}rM&ZBP{Yn~4~nF?>#+nIY$!ZRR{x>#cBcNi-x>k+&~U`pM>97 ze!1nEb42r!Suy_|*|$6gGE954F;m%EQVX?ft6m$8T^DP+oB~I;rPk`ogE!Q=mXilC zU5ne)pt*rhsao^!b9KwjtrqSQ6JC0(~YiOzG{c^eABi}6&yiv z;()hyc6;qs$sk$T1J4B)=N7qAHLDG2*1il$U~fR5i&l**omc!lB~#pBHeqdWI5^tu z9g*SF{~(8VlBQq1l5>t)J;2v#COCQ`g?wVO4!2-aLp&RAxZe*|hkLcMxW>^vogC3{ z!?>#%Fw1v8L{(Ig@c=jWJ3Z>6v(l@@AMn2)lTCjYBdid2lrKbWWJ7L2rLP4JeS#B zzPwGT|0&Pf|kf`rjqaSUnBztbKU-PF$ z)aS(u-F#n{D|YlRJr?On`31a}_Jtp6Wui6b{4{o>L91C&X;I@vT#<}1V5{K?Q#JGB z;B3i-sOSp_yPE)g$>l{_{NPILg>2=RWor^9XxS}~v zSG&yT-Ai#Y6v%7`9Nu!Zo63hbe0E^IvLJ4!p^)cv*~cik;?Hq|&BLjW`;>VbH>xI+ zNU9yki;{BImQ;FwxwzxOf$cK^iglNC$(Dh~Up5n} z4Td9~U0e=NZ_8Jwau0PpC;q)?^)nD@LQRQoDp)O1gt~RK^ajDa4MLiyyc9O|h1;1N zCK`SBF9Uf#v_@v#ijh7R`Lo3R%Tkxd1lg$bUXv}`!X#M~k~=n#<-UZR23r$CFNCHA zN!QPBF?t`*loN*ffk%27ox=Yn%0mY*W}m<>vU0(<=1V<;_Jp)ZWsOaIQ1aCGM#1{O zi3AO?(Rp4y;L}e#Ow3xu-Eh_VJL# z`n@PZzy=FQt5>GmY7sWGHg73xj*`>jJm^mvwvQw zKuzB%RD)@^9Xw{^dGve3H+t_o*B#?Jz+~*o@$$VkB+okZo0jFX`TtwT^4}ihfBTC~ a?ykf`LDv&oZUc+bKHWQpxAB_ypZy;oHV->w!SQGy^CB%=4;jc5_k61}(4iQbJCf&?Q<^b#RN^xiw8n!IC8i3Q-Il*{QUNs1#y?bEJ>=xYnF%%$?nB15 zh-_YX4l2}9?p%N+y(%UQ=4&Yb(6{PmBTrA|jfI*sm01H-7?@1yc!i@PZw`<8hj7>6 z3K{v^<6(-z?H5T6GYXQMWA|Bz{%u<~_8sfnHWK3hbbR&y)RFs9LVp+>Mtl{FfrW*Q zf0v+J`Y}dTHa&Afzo^m8bjy`r+QH$-s=Z+mpR3vHmk=jJ2&L?;!kl2pi`v~?|Fgfl z>w5p!y%C1%!>vKXO!)4=%O@S1V39-`q>>lCOTTISGHgjQEf~jF;8}` zKM+daJv_m#CIgDsqgt3Em4~#4|qrRF$=dNtxh) z0PQTUn>munv%eLE*U`}M)Bx2|7q74ZtZIvGPQ61s+a(=C_wpkMyQasWzqf1z zp^NlT>Gs~fAhxzK8onC^dL<`D-&T>dLIcdjtUshI+vy8?ymlwExXE`9uWc3kZG9jV zOfeHT(||o|o!jU`rha)qyeVYjjd_fRbMTFziDPSPOHf3JD*e_hxRty%_a%m0EBMYW z=VqKtWn{2plBpR5xa?Rl#_;aL2W^&v%y|Zh8dt5iU1`$$um|ZuC{ASl~bcBv-#J1uHqnxN3n zw#+kUkFJD7NP<_Peg@QUXtN>1&y{mylEQJ*LR{U0`YZjNt0PvzUqX^}ez{Dk9GBa6 zN|ZI4MdSH5PzBJm>77XW(zy3erwkUgrFO{^!|9W;J8lh{8!rIJn@q=s_+7}nUC)Im zF54?>&Cgj(+|}@|vv!kXe8zPKoFr_YYfIE1G^6W@(dTAEhO0PHTj4rx8hZ11&$fLO zJDMNdIx4g5VyQv#dCv?Snqz!&^s{9a2s#f+5pZDkmTJtBPM(aw83sq+A(QRU>CZQn zSWOj5P(UV*zD+%z-UfL+INuW4ufNzH13T83>h8^{IJNmR*TIbY=j)T~=IP$k|ENR~ z&m5NRr&O8IdCz)EtG6De23$~12A9fid$%g6^iX8!*Jjw5@O)v~rnr2g7VoMR!M#R$ z4=ZJF+=$_P%-?a453Ch3Xd_bkR04N@2h0fm-V}Xzrr_kaO`hK)GguI`fXZH>D&wtBwc4QM>%d}i4$2M+ z`Zb_|k)9m?U3EwGC;N$bA ztC6XzX4u!Sn!o$(xItwrS$J*pzVt$bn*oRiEEv29BzSr3j>l;=g4tgE?e_Az>E-=g z_S*7;0^T{ZkDtPm9h@#;B%ypKZ47 z5B&A`;v$*g>%i~Kt;F1ij~~k#Y-kdg&Cdrpw<{r9&nPC$oSr> zW?9uBTrVhET^Z08##>%feeegSdSlt8oO2@(Df4?o6_(dYYbWxBz1rlFT+8C`!sp)| z?8S`yx+hm;_=zPy_D6<1JJAU|Ja>|$Q}nESyJ(lnXwVL`SX~YeIQ|N;aCwE%kwfwJ zLKVF(v7NOV&Qln;7$avFeDwDb79K7y>x}GYxV*Be*tAILLPbrsSRWIWT2}>9KWL)0ZV+sovf5_b;=VymeQ=`>o0DbR=1Iy0B_%QA z<2oSs*qYi}cFx<=Rth2d&G}>25(PhIrlwvNy{x#&p3D5X?VKiK_EqK%b~!CAXD_c% zC@)7Yu6u(wuIi?bhFoP14i32ug@uSL%ZZ(txj8jMLz>zoinR1Jt?+*!g6ocTgPBpN zt0w5FYD|oijKOUXa?5Y$ymk4d5Ui@(^+)2}8@g>{pkw9zhxmLx|9{a@{kL}4bW9AT z=6Qrz zolmqRMqha7=?N%1{*bR3^KbVVlArneOQ;r5i<#_RpGA-U$@N9{PaZ6HTRlEY$_Q6r*Ri$ivF;| zLOjGTjQz@3a}Z*j9dG)|cJnsP$5rh-x88>WTE5a+htkgl;Ppg6$_qb#RE3xsp*!cl zl_jDRm$mkzrdqdy4$9k#;QiwFkc`Q_e6KU;0;;>&x&yv@L%56PIRHtIu=BUFd-w0L z0X@S0rX?&>KCBExNI0^yYS|_URHtLxQO7;M@T~gQwryhY#p(0xZxMB@hc|FT;?Anm z`Sv?N;E^m*taNj5FTKgLR?wP$j($dl`H9xj)+DZJ*LpiIcuE~HI%f}HzcJ8&NGx{| z{$Hc@d@JI6_koOU3vw9f>ERU{WKJqnNa1mtp{j1~S;f5%q#8PjANLOEzO-!#FAscd zO-&?c>i$Y2;m7^hftU^6fb&SgD2$u)iPfSTs7kSgYMQ3hyg`T)bLtJwRuI|M9<58` z@GZmOuIVPo^5G11g?dG>?I%^*R#uGlEOY93y0#qf+#8yGB*mVD&xW{OgV*M1UT#a{ zDkFMxGO&xO4Ja~ z#_Qds7ayRLn~R^HIw0b-IB6JFI9HuAKeO{RPOxQa3rK@*LS6msQ{RUrCed5b)4HG?n33vC9^>24O|D^6=K>F7 zibHJ(|LQcU=&p(6*E^`oAMcpDdnq(4E!k^0S)6(Z7zS$He~YA&bft-*mI$>HM;O2; z(&y3DX8bb#2)b6VqxfO>LUu#aaWxLsMNSL+<(Pkz*y*Bhown!QP5UhFal#blW&XX9 z_XPLCP*DB>R9EEkqh#a?#oMaN_cZjOZ9LT?iLX&poULr^?#3ZqZc93rKMv3{Z6ePb z>MNWISS9XPqCP&1Lrz+GnIx;X&rr!v;!OoTKck(98D_HEHiK}b`k9y;nO~E*JmU0> zp6^Q9l2;pG6xCUz^*BhHVJX+DSaKSgj5zgHUTU~F<@KNb38pB8uLG^^jn@1pF44A= zLaVg+w87EUBQ{+u8?VcIvR8Cu-L7dFPwLFro07UEOlU+y zK7oS74GU}EExIQZdP0JbqFrlF`MEBEK~#hUPt1zXv|Dm+Op_rjFCIFLHZ!XG6I^CsJfV>-Jl;dPdUnn2 z=6hr1@RNoU0Fly)Kg2WUdg~F&K9?klSC>VzgRqv0?Z6iy4Kz~R{nj*|JZgxlnPIIl z!@HTH5xcQ#B&7JfQGG~8Gy6%oe)l8IBYmPnXVsN<6#Jt@70d6grQBL?)+Y14H%-;k zpye{Jbn8=o#lQtA%0AnxQm|xZ>$b#%@>q?69fh>Z-V|iLz*Wz%%BVQ=ZPcr~>b87q zm%nsasI@heQ(-2KC+#*XmZhjw`L1?}@L^jGwV|utyrN;&~r}?wtEJ z532n=*C4x8#Q^`GR!8+_PkW2;!@OqdjARqWWZT7n3=Ok@7hPjlQKt?;&m8oywwML- z8}Ci21`0wlRJ!Wd*Dl0`R)wh*FA^q9hxrn0>5|-JOltcos>5R%AD7iGu&UvO`(Aui z3`Dc5KHtLW{-P0MIQGdyx_6N|C^IoDN5FWCbHZo^iGkXP;#a+;vh@=gb59B1`SPwe z__1za?UAICEU)B{k_K)Ggaq4PJzYVzU30)?pOtQa12>XxtNxe7#&(2O{2MdQej?)g zRgYAeTfIiNW|;p)C`&*V_hlwN7j67PCUS$sc(K8lVDa%CRf)g5FZN z@t2pI@cIlRTBe`0(u(1ETfMH5|r|Ea8XC55DmUG|ZYT*UbR@$q*SgIe6 zLSE>!|H5?fYrNycM!}HbU)DF!{likLNh!{L;Zo-T$=(plaDE4G)L}umNA2WbU9R52 za!iI^-4xS=?!2GjA7V>>mXyo_hSo#kO#Ex^DE_42pV=rIm$^2d=EppS-L{T#?Z@TBx zAnM)2IME~R@qD30U(W5$1(QF(O!KXFh^eUQW33K61%-(^d90{%TWKO1>?OWBueweN zQw`>ca>_wRjaEX}o78#P?H*u%%eQvtR?Q<57kVGU5_Rm?o4pmq72A=Kq~Ze>56tPZ zwNqmDSOpJV_2rvSI)+C>p7Jg~JY=%UAj+2Srmf>&;Bm0?x(IJd?VTEKGz8~Aa18*M z^?>Vw?2TtM$t$`kA;TIpC1*#c$(K{vf#7|Y`grll@# z_v5Qe>9e(HW7$o{ZzrJA-gx9TfvsIICi4{M&kR(ZbaeaCG6^o-IXza?K0fPNx~|D3 zx&KblWYpC}Q}gAX#O|Hd?>xI_*4Y!LtuKpI|tmkG`+yy7dfLcldwA9c!EYA0X==Yme34~iHIIBXtS%+({4hby!d0A zzo-h0q$&MXiJ%3|QRSgKN5osj9Uc+m-i6Abopt<88B}i_$?H8J4Y02B{i{-Cb*mTY z=&FIh@qFs;aW*IO6}_o%igcR(KJ(90;fBbnrRsWZyR>zc4ig$%(FcOh zV?QvNYdI<^)%v>8<&Kiwy?fA$cUJhD#oPI_2g0p))s^(JuH^Bh^;zLqwa}=Kj{wsK z#pq1rg-h|e7~{;N&^G+ESjW|LOwIQYF@GgtjIK&5;F((ZbAuhssqb~}X8IH{;R%Iy z4*3;efe0Ng6o3LIu70$Ef*C1yF5_Tr3Wu-U0jZfvi)*IrlOG-`t=bP5*YItG49*U%D z>dq?_F;SyqUP#1MXFIktsM2rU+nL#Ps)>U~RF_#`3rK0ZjoIoFvg$;?a36@~C%V$~b%Nb_xvz$2a}{_s^y)fybYWpD`I9CX6)w%ES> z_HPFdUUT1VJk=qt)h^{cx$~}g-d6hE)XDkE(1BpxQJeph2vgu;M(5tUO6jQhS-D8q z_R7<`#s`f%ufr!CMUmwMlh|(SI=F$m!mZl;jY+ zW%viqX!h6lLLC83-00&u{&@bD7_*Ou)e^nRdOoY>X}xSd7-^FSeLu&`7y7Sd2S)uP zFKeeia>$)pqsi%mWT(wP4Z+1yd6sLWW}Ed|zeEx$Ed+9)Uf0{XZb&NNvs6IpA?@#>(prI{2o?oEL3_o@=sVBLRX!v+uBohVFvN;Evm?jU7;!BK~Ukubx>wla~x>? zTAY^op=UDW=?m$G#A(Gx63Z|1+h+(tN+|sp2y==x`+WU7kTX}l->3z;6O2{hza z<`1N5*ep$j_)KzISM5_DX%=k!v^$OM4vQroo}@8jfi1&|@qNj@~&A zhNapLfFb`UgWl{=nhmcH=)!;n(Q9}t>CVK6^wS%205RKC*lUo(uEetgpujJwhm*09 z$F@`)u-e^~`pemPAE!3p(43cG_B#i!#@JNqiIce+X}(YvYekLY7u<$`S%me-TE`A8 ziVt$)J)RdYjz#2b0v#wRz;ddiOsr znX)P^9S_L*=iYV^4Ky|!{lG3o4pz9c)w4}OD9^my`E0BOYHL&YYnjz%BID4f>0>uz zVZ3`$=9BWIYe##)*%ZsSz@QXkWv9;;SAcL};&30%w+qpT00$lS13j{Q2uwM62>=YT zzN1)lP6eLOB=Ew|&TLFCi1O@hmG2ySPatkQ6n6;UoHtO=(}*6N;lr_%QnuK8kOnm9 zC~$BSoulrGEFKH1KL?9P$2^Hd2yWH2l^Bz&|8y#nI_7r89TN+C{_u4*W0tY<<^_pJ z!zF-=Aa-)Nlxf$7aIo2?&1)il%w@ED#NmFv_!I(OV~NdD?zX9p`sJdhtVxKYFzlN7 z>N>2I3O*^!+ZDb~uMh3?l&FC8{b@$}x=|G66IrY^TA?l#O{{J7aMg#Rg;jpN`>4f$ znl3YKAkQBfd~AUnq}?aiq2_T(0Sh^L1v&vE$wsrDH~_qC_RYS(GcCKB@b93T89=g#*43CHwsd^1SsQCeHmc1A2H!v*OLRmU8r z3t8ej$ZV)=HUrW1lr3q#RI{`~p8$4k33pr0l#z~DBX7tW$tTC05+Cdy=Mtz+hU!TY z!_;{~WIfy}g=0em66`1JP^h;;Pyx}f8#GI9cU+k~gMQxpn57ctDM3yv(d|%pJn9vj zP1)#zE7PQ_xp)_ep~_N>Bj4RbM25EQI1bfWH51S60lL0|9aedQ8oWU^5TblLAUy*T z^TW*z$lgO5qzWUxS2W#!7EbG=?30Kna;tSbtZpK39QK4$%`3g`ZTzD2o*1#C$ldhU z(B1dQw)fMVzPEH9I)3kzze3jw0Wj;@?DzmLt4S{yqL0UUag?{qeA-tXSdo|ENWKx{bwBU8`v z?y%-|Snap8R5a`lPd<8!q1nlKCXTK=@=M{9nO(ue+%mc|E5C##qDgW0?A=IeEFmj8 z?8+i^*zc|i`7%=j<-T_BxT;8Ib^#s_nTckl{Ed!kAz0aQT;$+8+qI z+N8XNm?7egVqnM%Co^q`-S#japQT6Vf+yx$oYy5(mEPR1YQGN-9)0I`41(p3-^ogbYaex4uNWkdp z_?8U?m_)Y?IRyfE<KMO{!t<~DG?~eF-bPZOBr}!}@$12pA#~@QNRZ#gWXXhw(f+ zW@NCQQ7j+pk})(M8l7!S2Xf?4-S4tlCdm@dUKo;m&aBv|lBGDDW37TuMQ}B4i;ICB zZpFV@sxnkMJ+5{(Xv>S^@)W`C`s}5~5$5zRftCLWO8$#@7Kv2Ew{#=o69cuGnfHRt zB4b@_v+j3Q%YVGeF}ArjhF!ubP^WFXzA}6T5%ZlXa5&>! z8b*&V_;T|CMY;ik15Un0uLk2h9~oiZ{vH6{dNSQ;XFT$Aeswa4M$nZ4u{z8m4loGL zpP+CzJ90gTj+o7;>+)aDN#IKxhbvcj6mYMud(CEZt}Y~5q`i%QbP~$cZ46o|qV3^M za9&$RCTnUCmo_}P|*5R2KOtj`SOZepvHH*k9Z;I%SwX2mb z`$`%%63V$j-6^CKnkC+qZtiSP3&c+g@bM=+TC4G*{LUsi_2aqzGk@2U znW7G>2;af9Wca!`Y?5cj`*V}=(I=(YPyA^3_f*pH!G=id=j5JrNWLkO@MA28JsNJ zwK!=X>O85El;%}5D3640AnF0r|4mUiqA$dTNVvIFo02yK9k0!8eF8e10Sk3HdYQbp zeO;!uYxwvdl5#7R-C2nlT8$C&Q)YtS!s!1~C;Xv~5LVp2KdDe^&zZIL&`gj8UbZg`s6H z2Z!&w`vPRt6i)x8T`4mPass5K1!{!GprybZQylByfwDFRQO2+*uXRayd0!pWiEhj<6?VcM;F$X)7%R!$4ciuV7G zRNiX2|D}Ciz`piDF7g=-uVa|Je0Tc)^ntd2 Date: Thu, 20 Mar 2025 23:17:51 +0800 Subject: [PATCH 07/21] Update README.md --- README.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/README.md b/README.md index b016a2569..3ff2548d7 100644 --- a/README.md +++ b/README.md @@ -149,6 +149,8 @@ MaiMBot是一个开源项目,我们非常欢迎你的参与。你的贡献, - [📦 Linux 手动部署指南 ](docs/manual_deploy_linux.md) +- [📦 macOS 手动部署指南 ](docs/manual_deploy_macos.md) + 如果你不知道Docker是什么,建议寻找相关教程或使用手动部署 **(现在不建议使用docker,更新慢,可能不适配)** - [🐳 Docker部署指南](docs/docker_deploy.md) From b7d7a9b2db00707fc151bd74c2b5984ce7cbbf2b Mon Sep 17 00:00:00 2001 From: tcmofashi Date: Fri, 21 Mar 2025 00:30:47 +0800 Subject: [PATCH 08/21] =?UTF-8?q?fix:=20=E4=BF=AE=E5=A4=8D=E4=BA=86filenam?= =?UTF-8?q?e=E9=87=8D=E5=A4=8D?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/emoji_manager.py | 34 ++++++++++++++++++++++++++++++- 1 file changed, 33 insertions(+), 1 deletion(-) diff --git a/src/plugins/chat/emoji_manager.py b/src/plugins/chat/emoji_manager.py index b1056a0ec..e3a6b77af 100644 --- a/src/plugins/chat/emoji_manager.py +++ b/src/plugins/chat/emoji_manager.py @@ -242,7 +242,33 @@ class EmojiManager: image_hash = hashlib.md5(image_bytes).hexdigest() image_format = Image.open(io.BytesIO(image_bytes)).format.lower() # 检查是否已经注册过 - existing_emoji = db["emoji"].find_one({"hash": image_hash}) + existing_emoji_by_path = db["emoji"].find_one({"filename": filename}) + existing_emoji_by_hash = db["emoji"].find_one({"hash": image_hash}) + if existing_emoji_by_path and existing_emoji_by_hash: + if existing_emoji_by_path["_id"] != existing_emoji_by_hash["_id"]: + logger.error(f"[错误] 表情包已存在但记录不一致: {filename}") + db.emoji.delete_one({"_id": existing_emoji_by_path["_id"]}) + db.emoji.update_one( + {"_id": existing_emoji_by_hash["_id"]}, {"$set": {"path": image_path, "filename": filename}} + ) + existing_emoji_by_hash["path"] = image_path + existing_emoji_by_hash["filename"] = filename + existing_emoji = existing_emoji_by_hash + elif existing_emoji_by_hash: + logger.error(f"[错误] 表情包hash已存在但path不存在: {filename}") + db.emoji.update_one( + {"_id": existing_emoji_by_hash["_id"]}, {"$set": {"path": image_path, "filename": filename}} + ) + existing_emoji_by_hash["path"] = image_path + existing_emoji_by_hash["filename"] = filename + existing_emoji = existing_emoji_by_hash + elif existing_emoji_by_path: + logger.error(f"[错误] 表情包path已存在但hash不存在: {filename}") + db.emoji.delete_one({"_id": existing_emoji_by_path["_id"]}) + existing_emoji = None + else: + existing_emoji = None + description = None if existing_emoji: @@ -366,6 +392,12 @@ class EmojiManager: logger.warning(f"[检查] 发现缺失记录(缺少hash字段),ID: {emoji.get('_id', 'unknown')}") hash = hashlib.md5(open(emoji["path"], "rb").read()).hexdigest() db.emoji.update_one({"_id": emoji["_id"]}, {"$set": {"hash": hash}}) + else: + file_hash = hashlib.md5(open(emoji["path"], "rb").read()).hexdigest() + if emoji["hash"] != file_hash: + logger.warning(f"[检查] 表情包文件hash不匹配,ID: {emoji.get('_id', 'unknown')}") + db.emoji.delete_one({"_id": emoji["_id"]}) + removed_count += 1 except Exception as item_error: logger.error(f"[错误] 处理表情包记录时出错: {str(item_error)}") From dd1a4cd731cff4746a79ded0097b59a2b538c780 Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Fri, 21 Mar 2025 13:31:54 +0800 Subject: [PATCH 09/21] =?UTF-8?q?=E8=BF=87Ruff=E6=A3=80=E6=B5=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/personality/big5_test.py | 6 ++--- src/plugins/personality/combined_test.py | 13 ++++++----- src/plugins/personality/questionnaire.py | 29 ++++++++++++++++++------ src/plugins/personality/renqingziji.py | 23 +++++++++++-------- src/plugins/personality/scene.py | 18 +++++++++------ 5 files changed, 56 insertions(+), 33 deletions(-) diff --git a/src/plugins/personality/big5_test.py b/src/plugins/personality/big5_test.py index 80114ec36..e77dfbc4f 100644 --- a/src/plugins/personality/big5_test.py +++ b/src/plugins/personality/big5_test.py @@ -4,9 +4,10 @@ # from .questionnaire import PERSONALITY_QUESTIONS, FACTOR_DESCRIPTIONS import os +import random import sys from pathlib import Path -import random +from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS,FACTOR_DESCRIPTIONS current_dir = Path(__file__).resolve().parent project_root = current_dir.parent.parent.parent @@ -15,9 +16,6 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.scene import get_scene_by_factor,get_all_scenes,PERSONALITY_SCENES -from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS,FACTOR_DESCRIPTIONS -from src.plugins.personality.offline_llm import LLMModel diff --git a/src/plugins/personality/combined_test.py b/src/plugins/personality/combined_test.py index a842847fb..2aaca4266 100644 --- a/src/plugins/personality/combined_test.py +++ b/src/plugins/personality/combined_test.py @@ -1,11 +1,14 @@ -from typing import Dict, List import json import os -from pathlib import Path +import random import sys from datetime import datetime -import random +from pathlib import Path +from typing import Dict from scipy import stats # 添加scipy导入用于t检验 +from src.plugins.personality.big5_test import BigFiveTest +from src.plugins.personality.renqingziji import PersonalityEvaluator_direct +from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS, PERSONALITY_QUESTIONS current_dir = Path(__file__).resolve().parent project_root = current_dir.parent.parent.parent @@ -14,9 +17,7 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.big5_test import BigFiveTest -from src.plugins.personality.renqingziji import PersonalityEvaluator_direct -from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS, PERSONALITY_QUESTIONS + class CombinedPersonalityTest: def __init__(self): diff --git a/src/plugins/personality/questionnaire.py b/src/plugins/personality/questionnaire.py index 4afff1185..c6d1de068 100644 --- a/src/plugins/personality/questionnaire.py +++ b/src/plugins/personality/questionnaire.py @@ -1,4 +1,5 @@ -# 人格测试问卷题目 王孟成, 戴晓阳, & 姚树桥. (2011). 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. 中国临床心理学杂志, 19(04), Article 04. +# 人格测试问卷题目 王孟成, 戴晓阳, & 姚树桥. (2011). 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. 中国临床心理学杂志, +# 19(04), Article 04. # 王孟成, 戴晓阳, & 姚树桥. (2010). 中国大五人格问卷的初步编制Ⅰ:理论框架与信度分析. 中国临床心理学杂志, 18(05), Article 05. PERSONALITY_QUESTIONS = [ @@ -23,7 +24,11 @@ PERSONALITY_QUESTIONS = [ {"id": 16, "content": "我是个倾尽全力做事的人", "factor": "严谨性", "reverse_scoring": False}, # 宜人性维度 (F3) - {"id": 17, "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈),我仍然相信人性总的来说是善良的", "factor": "宜人性", "reverse_scoring": False}, + {"id": 17, + "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈),我仍然相信人性总的来说是善良的", + "factor": "宜人性", + "reverse_scoring": False + }, {"id": 18, "content": "我觉得大部分人基本上是心怀善意的", "factor": "宜人性", "reverse_scoring": False}, {"id": 19, "content": "虽然社会上有骗子,但我觉得大部分人还是可信的", "factor": "宜人性", "reverse_scoring": False}, {"id": 20, "content": "我不太关心别人是否受到不公正的待遇", "factor": "宜人性", "reverse_scoring": True}, @@ -56,7 +61,9 @@ PERSONALITY_QUESTIONS = [ # 因子维度说明 FACTOR_DESCRIPTIONS = { "外向性": { - "description": "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性,包括对社交活动的兴趣、对人群的态度、社交互动中的主动程度以及在群体中的影响力。高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我,并往往在群体中发挥领导作用;低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。", + "description": "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性,包括对社交活动的兴趣、对人 \ + 群的态度、社交互动中的主动程度以及在群体中的影响力。高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我,并往往在群体中发挥领导 \ + 作用;低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。", "trait_words": ["热情", "活力", "社交", "主动"], "subfactors": { "合群性": "个体愿意与他人聚在一起,即接近人群的倾向;高分表现乐群、好交际,低分表现封闭、独处", @@ -66,7 +73,9 @@ FACTOR_DESCRIPTIONS = { } }, "神经质": { - "description": "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、挫折和日常生活挑战时的情绪稳定性和适应能力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度,以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波动较大;低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。", + "description": "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、挫折和日常生活挑战时的情绪稳定性和适应能 \ + 力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度,以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波 \ + 动较大;低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。", "trait_words": ["稳定", "沉着", "从容", "坚韧"], "subfactors": { "焦虑": "个体体验焦虑感的个体差异;高分表现坐立不安,低分表现平静", @@ -77,7 +86,9 @@ FACTOR_DESCRIPTIONS = { } }, "严谨性": { - "description": "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、学习等目标性活动中的自我约束和行为管理能力。它涉及到个体的责任感、自律性、计划性、条理性以及完成任务的态度。高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的努力精神;低分者则可能表现出随意性强、缺乏规划、做事马虎或易放弃的特点。", + "description": "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、学习等目标性活动中的自我约束和行为管理能 \ + 力。它涉及到个体的责任感、自律性、计划性、条理性以及完成任务的态度。高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的 \ + 努力精神;低分者则可能表现出随意性强、缺乏规划、做事马虎或易放弃的特点。", "trait_words": ["负责", "自律", "条理", "勤奋"], "subfactors": { "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;高分表现有责任心、负责任,低分表现推卸责任、逃避处罚", @@ -88,7 +99,9 @@ FACTOR_DESCRIPTIONS = { } }, "开放性": { - "description": "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。这个维度体现了个体在认知和体验方面的广度、深度和灵活性。它包括对艺术的欣赏能力、对知识的求知欲、想象力的丰富程度,以及对冒险和创新的态度。高分者往往具有丰富的想象力、广泛的兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、传统,喜欢熟悉和常规的事物。", + "description": "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。这个维度体现了个体在认知和体验方面的 \ + 广度、深度和灵活性。它包括对艺术的欣赏能力、对知识的求知欲、想象力的丰富程度,以及对冒险和创新的态度。高分者往往具有丰富的想象力、广泛的 \ + 兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、传统,喜欢熟悉和常规的事物。", "trait_words": ["创新", "好奇", "艺术", "冒险"], "subfactors": { "幻想": "个体富于幻想和想象的水平;高分表现想象力丰富,低分表现想象力匮乏", @@ -99,7 +112,9 @@ FACTOR_DESCRIPTIONS = { } }, "宜人性": { - "description": "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。这个维度主要关注个体与他人互动时的态度和行为特征,包括对他人的信任程度、同理心水平、助人意愿以及在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人建立和谐关系;低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑他人感受。", + "description": "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。这个维度主要关注个体与他人互动时的态度和行为特 \ + 征,包括对他人的信任程度、同理心水平、助人意愿以及在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人 \ + 建立和谐关系;低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑他人感受。", "trait_words": ["友善", "同理", "信任", "合作"], "subfactors": { "信任": "个体对他人和/或他人言论的相信程度;高分表现信任他人,低分表现怀疑", diff --git a/src/plugins/personality/renqingziji.py b/src/plugins/personality/renqingziji.py index b3a3e267e..5431f4e68 100644 --- a/src/plugins/personality/renqingziji.py +++ b/src/plugins/personality/renqingziji.py @@ -1,17 +1,25 @@ ''' -The definition of artificial personality in this paper follows the dispositional para-digm and adapts a definition of personality developed for humans [17]: +The definition of artificial personality in this paper follows the dispositional para-digm and adapts a definition of +personality developed for humans [17]: Personality for a human is the "whole and organisation of relatively stable tendencies and patterns of experience and -behaviour within one person (distinguishing it from other persons)". This definition is modified for artificial personality: -Artificial personality describes the relatively stable tendencies and patterns of behav-iour of an AI-based machine that +behaviour within one person (distinguishing it from other persons)". +This definition is modified for artificial personality: +Artificial personality describes the relatively stable tendencies +and patterns of behav-iour of an AI-based machine that can be designed by developers and designers via different modalities, such as language, creating the impression of individuality of a humanized social agent when users interact with the machine.''' -from typing import Dict, List import json import os -from pathlib import Path -from dotenv import load_dotenv import sys +from pathlib import Path +from typing import Dict, List + +from dotenv import load_dotenv + +from src.plugins.personality.offline_llm import LLMModel +from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS +from src.plugins.personality.scene import get_scene_by_factor, PERSONALITY_SCENES ''' 第一种方案:基于情景评估的人格测定 @@ -23,9 +31,6 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.scene import get_scene_by_factor,get_all_scenes,PERSONALITY_SCENES -from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS,FACTOR_DESCRIPTIONS -from src.plugins.personality.offline_llm import LLMModel # 加载环境变量 if env_path.exists(): diff --git a/src/plugins/personality/scene.py b/src/plugins/personality/scene.py index 936b07a3e..9bf3b4ec1 100644 --- a/src/plugins/personality/scene.py +++ b/src/plugins/personality/scene.py @@ -1,4 +1,4 @@ -from typing import Dict, List +from typing import Dict PERSONALITY_SCENES = { "外向性": { @@ -44,11 +44,12 @@ PERSONALITY_SCENES = { "神经质": { "场景1": { - "scenario": """你正在准备一个重要的项目演示,这关系到你的晋升机会。就在演示前30分钟,你收到了主管发来的消息: - + "scenario": """你正在准备一个重要的项目演示,这关系到你的晋升机会。就在演示前30分钟 +,你收到了主管发来的消息: 主管:「临时有个变动,CEO也会来听你的演示。他对这个项目特别感兴趣。」 -正当你准备回复时,主管又发来一条:「对了,能不能把演示时间压缩到15分钟?CEO下午还有其他安排。你之前准备的是30分钟的版本对吧?」""", +正当你准备回复时,主管又发来一条:「对了,能不能把演示时间压缩到15分钟?CEO下午还有其他安排。 +你之前准备的是30分钟的版本对吧?」""", "explanation": "这个场景通过突发的压力情境,观察个体在面对计划外变化时的情绪反应和调节能力。" }, "场景2": { @@ -142,9 +143,11 @@ PERSONALITY_SCENES = { "场景1": { "scenario": """周末下午,你的好友小美兴致勃勃地给你打电话: -小美:「我刚发现一个特别有意思的沉浸式艺术展!不是传统那种挂画的展览,而是把整个空间都变成了艺术品。观众要穿特制的服装,还要带上VR眼镜,好像还有AI实时互动!」 +小美:「我刚发现一个特别有意思的沉浸式艺术展!不是传统那种挂画的展览,而是把整个空间都变成了艺术品。观众要穿特制的服装, +还要带上VR眼镜,好像还有AI实时互动!」 -小美继续说:「虽然票价不便宜,但听说体验很独特。网上评价两极分化,有人说是前所未有的艺术革新,也有人说是哗众取宠。要不要周末一起去体验一下?」""", +小美继续说:「虽然票价不便宜,但听说体验很独特。网上评价两极分化,有人说是前所未有的艺术革新, +也有人说是哗众取宠。要不要周末一起去体验一下?」""", "explanation": "这个场景通过新型艺术体验,反映个体对创新事物的接受程度和尝试意愿。" }, "场景2": { @@ -158,7 +161,8 @@ PERSONALITY_SCENES = { "场景3": { "scenario": """在社交媒体上,你看到一个朋友分享了一种新的生活方式: -「最近我在尝试'数字游牧'生活,就是一边远程工作一边环游世界。没有固定住所,住青旅或短租,认识来自世界各地的朋友。虽然有时会很不稳定,但这种自由的生活方式真的很棒!」 +「最近我在尝试'数字游牧'生活,就是一边远程工作一边环游世界。没有固定住所,住青旅或短租,认识来自世界各地的朋友。 +虽然有时会很不稳定,但这种自由的生活方式真的很棒!」 评论区里争论不断,有人向往这种生活,也有人觉得太冒险。""", "explanation": "通过另类生活方式,观察个体对非传统选择的态度。" From 94ba8e0927c576328256633b3256aa10817bd0d4 Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Fri, 21 Mar 2025 13:36:09 +0800 Subject: [PATCH 10/21] =?UTF-8?q?=E8=BF=87Ruff=E6=A3=80=E6=B5=8B2?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/personality/combined_test.py | 2 +- src/plugins/personality/questionnaire.py | 165 +++++++++++++++-------- 2 files changed, 108 insertions(+), 59 deletions(-) diff --git a/src/plugins/personality/combined_test.py b/src/plugins/personality/combined_test.py index 2aaca4266..96ca3736a 100644 --- a/src/plugins/personality/combined_test.py +++ b/src/plugins/personality/combined_test.py @@ -245,7 +245,7 @@ class CombinedPersonalityTest: # 对所有维度进行整体t检验 t_stat, p_value = stats.ttest_rel(questionnaire_values, scenario_values) - print(f"\n整体统计分析:") + print("\n整体统计分析:") print(f"平均差异: {mean_diff:.3f}") print(f"差异标准差: {std_diff:.3f}") print(f"效应量(Cohen's d): {cohens_d:.3f}") diff --git a/src/plugins/personality/questionnaire.py b/src/plugins/personality/questionnaire.py index c6d1de068..0366b1c27 100644 --- a/src/plugins/personality/questionnaire.py +++ b/src/plugins/personality/questionnaire.py @@ -1,6 +1,7 @@ -# 人格测试问卷题目 王孟成, 戴晓阳, & 姚树桥. (2011). 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. 中国临床心理学杂志, -# 19(04), Article 04. -# 王孟成, 戴晓阳, & 姚树桥. (2010). 中国大五人格问卷的初步编制Ⅰ:理论框架与信度分析. 中国临床心理学杂志, 18(05), Article 05. +# 人格测试问卷题目 王孟成, 戴晓阳, & 姚树桥. (2011). 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. +# 中国临床心理学杂志, 19(04), Article 04. +# 王孟成, 戴晓阳, & 姚树桥. (2010). 中国大五人格问卷的初步编制Ⅰ:理论框架与信度分析. +# 中国临床心理学杂志, 18(05), Article 05. PERSONALITY_QUESTIONS = [ # 神经质维度 (F1) @@ -8,62 +9,97 @@ PERSONALITY_QUESTIONS = [ {"id": 2, "content": "我常感到害怕", "factor": "神经质", "reverse_scoring": False}, {"id": 3, "content": "有时我觉得自己一无是处", "factor": "神经质", "reverse_scoring": False}, {"id": 4, "content": "我很少感到忧郁或沮丧", "factor": "神经质", "reverse_scoring": True}, - {"id": 5, "content": "别人一句漫不经心的话,我常会联系在自己身上", "factor": "神经质", "reverse_scoring": False}, - {"id": 6, "content": "在面对压力时,我有种快要崩溃的感觉", "factor": "神经质", "reverse_scoring": False}, - {"id": 7, "content": "我常担忧一些无关紧要的事情", "factor": "神经质", "reverse_scoring": False}, - {"id": 8, "content": "我常常感到内心不踏实", "factor": "神经质", "reverse_scoring": False}, + {"id": 5, "content": "别人一句漫不经心的话,我常会联系在自己身上", + "factor": "神经质", "reverse_scoring": False}, + {"id": 6, "content": "在面对压力时,我有种快要崩溃的感觉", + "factor": "神经质", "reverse_scoring": False}, + {"id": 7, "content": "我常担忧一些无关紧要的事情", + "factor": "神经质", "reverse_scoring": False}, + {"id": 8, "content": "我常常感到内心不踏实", + "factor": "神经质", "reverse_scoring": False}, # 严谨性维度 (F2) - {"id": 9, "content": "在工作上,我常只求能应付过去便可", "factor": "严谨性", "reverse_scoring": True}, - {"id": 10, "content": "一旦确定了目标,我会坚持努力地实现它", "factor": "严谨性", "reverse_scoring": False}, - {"id": 11, "content": "我常常是仔细考虑之后才做出决定", "factor": "严谨性", "reverse_scoring": False}, - {"id": 12, "content": "别人认为我是个慎重的人", "factor": "严谨性", "reverse_scoring": False}, - {"id": 13, "content": "做事讲究逻辑和条理是我的一个特点", "factor": "严谨性", "reverse_scoring": False}, - {"id": 14, "content": "我喜欢一开头就把事情计划好", "factor": "严谨性", "reverse_scoring": False}, - {"id": 15, "content": "我工作或学习很勤奋", "factor": "严谨性", "reverse_scoring": False}, - {"id": 16, "content": "我是个倾尽全力做事的人", "factor": "严谨性", "reverse_scoring": False}, + {"id": 9, "content": "在工作上,我常只求能应付过去便可", + "factor": "严谨性", "reverse_scoring": True}, + {"id": 10, "content": "一旦确定了目标,我会坚持努力地实现它", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 11, "content": "我常常是仔细考虑之后才做出决定", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 12, "content": "别人认为我是个慎重的人", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 13, "content": "做事讲究逻辑和条理是我的一个特点", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 14, "content": "我喜欢一开头就把事情计划好", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 15, "content": "我工作或学习很勤奋", + "factor": "严谨性", "reverse_scoring": False}, + {"id": 16, "content": "我是个倾尽全力做事的人", + "factor": "严谨性", "reverse_scoring": False}, # 宜人性维度 (F3) - {"id": 17, - "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈),我仍然相信人性总的来说是善良的", - "factor": "宜人性", - "reverse_scoring": False - }, - {"id": 18, "content": "我觉得大部分人基本上是心怀善意的", "factor": "宜人性", "reverse_scoring": False}, - {"id": 19, "content": "虽然社会上有骗子,但我觉得大部分人还是可信的", "factor": "宜人性", "reverse_scoring": False}, - {"id": 20, "content": "我不太关心别人是否受到不公正的待遇", "factor": "宜人性", "reverse_scoring": True}, - {"id": 21, "content": "我时常觉得别人的痛苦与我无关", "factor": "宜人性", "reverse_scoring": True}, - {"id": 22, "content": "我常为那些遭遇不幸的人感到难过", "factor": "宜人性", "reverse_scoring": False}, - {"id": 23, "content": "我是那种只照顾好自己,不替别人担忧的人", "factor": "宜人性", "reverse_scoring": True}, - {"id": 24, "content": "当别人向我诉说不幸时,我常感到难过", "factor": "宜人性", "reverse_scoring": False}, + {"id": 17, "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈)," + "我仍然相信人性总的来说是善良的", "factor": "宜人性", "reverse_scoring": False}, + {"id": 18, "content": "我觉得大部分人基本上是心怀善意的", + "factor": "宜人性", "reverse_scoring": False}, + {"id": 19, "content": "虽然社会上有骗子,但我觉得大部分人还是可信的", + "factor": "宜人性", "reverse_scoring": False}, + {"id": 20, "content": "我不太关心别人是否受到不公正的待遇", + "factor": "宜人性", "reverse_scoring": True}, + {"id": 21, "content": "我时常觉得别人的痛苦与我无关", + "factor": "宜人性", "reverse_scoring": True}, + {"id": 22, "content": "我常为那些遭遇不幸的人感到难过", + "factor": "宜人性", "reverse_scoring": False}, + {"id": 23, "content": "我是那种只照顾好自己,不替别人担忧的人", + "factor": "宜人性", "reverse_scoring": True}, + {"id": 24, "content": "当别人向我诉说不幸时,我常感到难过", + "factor": "宜人性", "reverse_scoring": False}, # 开放性维度 (F4) - {"id": 25, "content": "我的想象力相当丰富", "factor": "开放性", "reverse_scoring": False}, - {"id": 26, "content": "我头脑中经常充满生动的画面", "factor": "开放性", "reverse_scoring": False}, - {"id": 27, "content": "我对许多事情有着很强的好奇心", "factor": "开放性", "reverse_scoring": False}, - {"id": 28, "content": "我喜欢冒险", "factor": "开放性", "reverse_scoring": False}, - {"id": 29, "content": "我是个勇于冒险,突破常规的人", "factor": "开放性", "reverse_scoring": False}, - {"id": 30, "content": "我身上具有别人没有的冒险精神", "factor": "开放性", "reverse_scoring": False}, - {"id": 31, "content": "我渴望学习一些新东西,即使它们与我的日常生活无关", "factor": "开放性", "reverse_scoring": False}, - {"id": 32, "content": "我很愿意也很容易接受那些新事物、新观点、新想法", "factor": "开放性", "reverse_scoring": False}, + {"id": 25, "content": "我的想象力相当丰富", + "factor": "开放性", "reverse_scoring": False}, + {"id": 26, "content": "我头脑中经常充满生动的画面", + "factor": "开放性", "reverse_scoring": False}, + {"id": 27, "content": "我对许多事情有着很强的好奇心", + "factor": "开放性", "reverse_scoring": False}, + {"id": 28, "content": "我喜欢冒险", + "factor": "开放性", "reverse_scoring": False}, + {"id": 29, "content": "我是个勇于冒险,突破常规的人", + "factor": "开放性", "reverse_scoring": False}, + {"id": 30, "content": "我身上具有别人没有的冒险精神", + "factor": "开放性", "reverse_scoring": False}, + {"id": 31, "content": "我渴望学习一些新东西,即使它们与我的日常生活无关", + "factor": "开放性", "reverse_scoring": False}, + {"id": 32, "content": "我很愿意也很容易接受那些新事物、新观点、新想法", + "factor": "开放性", "reverse_scoring": False}, # 外向性维度 (F5) - {"id": 33, "content": "我喜欢参加社交与娱乐聚会", "factor": "外向性", "reverse_scoring": False}, - {"id": 34, "content": "我对人多的聚会感到乏味", "factor": "外向性", "reverse_scoring": True}, - {"id": 35, "content": "我尽量避免参加人多的聚会和嘈杂的环境", "factor": "外向性", "reverse_scoring": True}, - {"id": 36, "content": "在热闹的聚会上,我常常表现主动并尽情玩耍", "factor": "外向性", "reverse_scoring": False}, - {"id": 37, "content": "有我在的场合一般不会冷场", "factor": "外向性", "reverse_scoring": False}, - {"id": 38, "content": "我希望成为领导者而不是被领导者", "factor": "外向性", "reverse_scoring": False}, - {"id": 39, "content": "在一个团体中,我希望处于领导地位", "factor": "外向性", "reverse_scoring": False}, - {"id": 40, "content": "别人多认为我是一个热情和友好的人", "factor": "外向性", "reverse_scoring": False} + {"id": 33, "content": "我喜欢参加社交与娱乐聚会", + "factor": "外向性", "reverse_scoring": False}, + {"id": 34, "content": "我对人多的聚会感到乏味", + "factor": "外向性", "reverse_scoring": True}, + {"id": 35, "content": "我尽量避免参加人多的聚会和嘈杂的环境", + "factor": "外向性", "reverse_scoring": True}, + {"id": 36, "content": "在热闹的聚会上,我常常表现主动并尽情玩耍", + "factor": "外向性", "reverse_scoring": False}, + {"id": 37, "content": "有我在的场合一般不会冷场", + "factor": "外向性", "reverse_scoring": False}, + {"id": 38, "content": "我希望成为领导者而不是被领导者", + "factor": "外向性", "reverse_scoring": False}, + {"id": 39, "content": "在一个团体中,我希望处于领导地位", + "factor": "外向性", "reverse_scoring": False}, + {"id": 40, "content": "别人多认为我是一个热情和友好的人", + "factor": "外向性", "reverse_scoring": False} ] # 因子维度说明 FACTOR_DESCRIPTIONS = { "外向性": { - "description": "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性,包括对社交活动的兴趣、对人 \ - 群的态度、社交互动中的主动程度以及在群体中的影响力。高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我,并往往在群体中发挥领导 \ - 作用;低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。", + "description": ( + "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性," + "包括对社交活动的兴趣、对人群的态度、社交互动中的主动程度以及在群体中的影响力。" + "高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我,并往往在群体中发挥领导作用;" + "低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。" + ), "trait_words": ["热情", "活力", "社交", "主动"], "subfactors": { "合群性": "个体愿意与他人聚在一起,即接近人群的倾向;高分表现乐群、好交际,低分表现封闭、独处", @@ -73,9 +109,12 @@ FACTOR_DESCRIPTIONS = { } }, "神经质": { - "description": "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、挫折和日常生活挑战时的情绪稳定性和适应能 \ - 力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度,以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波 \ - 动较大;低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。", + "description": ( + "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、挫折和" + "日常生活挑战时的情绪稳定性和适应能力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度," + "以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波动较大;" + "低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。" + ), "trait_words": ["稳定", "沉着", "从容", "坚韧"], "subfactors": { "焦虑": "个体体验焦虑感的个体差异;高分表现坐立不安,低分表现平静", @@ -86,9 +125,12 @@ FACTOR_DESCRIPTIONS = { } }, "严谨性": { - "description": "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、学习等目标性活动中的自我约束和行为管理能 \ - 力。它涉及到个体的责任感、自律性、计划性、条理性以及完成任务的态度。高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的 \ - 努力精神;低分者则可能表现出随意性强、缺乏规划、做事马虎或易放弃的特点。", + "description": ( + "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、学习等" + "目标性活动中的自我约束和行为管理能力。它涉及到个体的责任感、自律性、计划性、条理性以及" + "完成任务的态度。高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的" + "努力精神;低分者则可能表现出随意性强、缺乏规划、做事马虎或易放弃的特点。" + ), "trait_words": ["负责", "自律", "条理", "勤奋"], "subfactors": { "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;高分表现有责任心、负责任,低分表现推卸责任、逃避处罚", @@ -99,9 +141,12 @@ FACTOR_DESCRIPTIONS = { } }, "开放性": { - "description": "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。这个维度体现了个体在认知和体验方面的 \ - 广度、深度和灵活性。它包括对艺术的欣赏能力、对知识的求知欲、想象力的丰富程度,以及对冒险和创新的态度。高分者往往具有丰富的想象力、广泛的 \ - 兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、传统,喜欢熟悉和常规的事物。", + "description": ( + "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。" + "这个维度体现了个体在认知和体验方面的广度、深度和灵活性。它包括对艺术的欣赏能力、" + "对知识的求知欲、想象力的丰富程度,以及对冒险和创新的态度。高分者往往具有丰富的想象力、" + "广泛的兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、传统,喜欢熟悉和常规的事物。" + ), "trait_words": ["创新", "好奇", "艺术", "冒险"], "subfactors": { "幻想": "个体富于幻想和想象的水平;高分表现想象力丰富,低分表现想象力匮乏", @@ -112,9 +157,13 @@ FACTOR_DESCRIPTIONS = { } }, "宜人性": { - "description": "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。这个维度主要关注个体与他人互动时的态度和行为特 \ - 征,包括对他人的信任程度、同理心水平、助人意愿以及在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人 \ - 建立和谐关系;低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑他人感受。", + "description": ( + "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。这个维度主要" + "关注个体与他人互动时的态度和行为特征,包括对他人的信任程度、同理心水平、助人意愿以及" + "在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人" + "建立和谐关系;低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑" + "他人感受。" + ), "trait_words": ["友善", "同理", "信任", "合作"], "subfactors": { "信任": "个体对他人和/或他人言论的相信程度;高分表现信任他人,低分表现怀疑", From 82e7cf7a3235c629da134742dbd0f27b6c3ff5e2 Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Fri, 21 Mar 2025 13:38:00 +0800 Subject: [PATCH 11/21] =?UTF-8?q?=E8=BF=87Ruff=E6=A3=80=E6=B5=8B3?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/personality/questionnaire.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/src/plugins/personality/questionnaire.py b/src/plugins/personality/questionnaire.py index 0366b1c27..3e1a7897e 100644 --- a/src/plugins/personality/questionnaire.py +++ b/src/plugins/personality/questionnaire.py @@ -119,7 +119,8 @@ FACTOR_DESCRIPTIONS = { "subfactors": { "焦虑": "个体体验焦虑感的个体差异;高分表现坐立不安,低分表现平静", "抑郁": "个体体验抑郁情感的个体差异;高分表现郁郁寡欢,低分表现平静", - "敏感多疑": "个体常常关注自己的内心活动,行为和过于意识人对自己的看法、评价;高分表现敏感多疑,低分表现淡定、自信", + "敏感多疑": "个体常常关注自己的内心活动,行为和过于意识人对自己的看法、评价;" + "高分表现敏感多疑,低分表现淡定、自信", "脆弱性": "个体在危机或困难面前无力、脆弱的特点;高分表现无能、易受伤、逃避,低分表现坚强", "愤怒-敌意": "个体准备体验愤怒,及相关情绪的状态;高分表现暴躁易怒,低分表现平静" } @@ -133,7 +134,8 @@ FACTOR_DESCRIPTIONS = { ), "trait_words": ["负责", "自律", "条理", "勤奋"], "subfactors": { - "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;高分表现有责任心、负责任,低分表现推卸责任、逃避处罚", + "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;" + "高分表现有责任心、负责任,低分表现推卸责任、逃避处罚", "自我控制": "个体约束自己的能力,及自始至终的坚持性;高分表现自制、有毅力,低分表现冲动、无毅力", "审慎性": "个体在采取具体行动前的心理状态;高分表现谨慎、小心,低分表现鲁莽、草率", "条理性": "个体处理事务和工作的秩序,条理和逻辑性;高分表现整洁、有秩序,低分表现混乱、遗漏", From 7cad7786cc2c956464dbae331bf9f16f78ab286d Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=98=A5=E6=B2=B3=E6=99=B4?= Date: Fri, 21 Mar 2025 13:41:43 +0800 Subject: [PATCH 12/21] =?UTF-8?q?style:=20=E4=BB=A3=E7=A0=81=E6=A0=BC?= =?UTF-8?q?=E5=BC=8F=E5=8C=96=EF=BC=8C=E4=BF=AE=E5=A4=8D=E7=BC=A9=E8=BF=9B?= =?UTF-8?q?=E9=97=AE=E9=A2=98?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude --- bot.py | 14 +- src/plugins/chat/__init__.py | 7 +- src/plugins/chat/bot.py | 32 ++-- src/plugins/personality/big5_test.py | 59 +++---- src/plugins/personality/combined_test.py | 188 ++++++++++----------- src/plugins/personality/questionnaire.py | 90 ++++++---- src/plugins/personality/renqingziji.py | 57 +++---- src/plugins/personality/scene.py | 91 +++++----- webui.py | 202 +++++++++++------------ 9 files changed, 374 insertions(+), 366 deletions(-) diff --git a/bot.py b/bot.py index 88c07939b..30714e846 100644 --- a/bot.py +++ b/bot.py @@ -204,8 +204,8 @@ def check_eula(): eula_confirmed = True eula_updated = False if eula_new_hash == os.getenv("EULA_AGREE"): - eula_confirmed = True - eula_updated = False + eula_confirmed = True + eula_updated = False # 检查隐私条款确认文件是否存在 if privacy_confirm_file.exists(): @@ -214,14 +214,16 @@ def check_eula(): if privacy_new_hash == confirmed_content: privacy_confirmed = True privacy_updated = False - if privacy_new_hash == os.getenv("PRIVACY_AGREE"): - privacy_confirmed = True - privacy_updated = False + if privacy_new_hash == os.getenv("PRIVACY_AGREE"): + privacy_confirmed = True + privacy_updated = False # 如果EULA或隐私条款有更新,提示用户重新确认 if eula_updated or privacy_updated: print("EULA或隐私条款内容已更新,请在阅读后重新确认,继续运行视为同意更新后的以上两款协议") - print(f'输入"同意"或"confirmed"或设置环境变量"EULA_AGREE={eula_new_hash}"和"PRIVACY_AGREE={privacy_new_hash}"继续运行') + print( + f'输入"同意"或"confirmed"或设置环境变量"EULA_AGREE={eula_new_hash}"和"PRIVACY_AGREE={privacy_new_hash}"继续运行' + ) while True: user_input = input().strip().lower() if user_input in ["同意", "confirmed"]: diff --git a/src/plugins/chat/__init__.py b/src/plugins/chat/__init__.py index a54f781a0..c6072cb55 100644 --- a/src/plugins/chat/__init__.py +++ b/src/plugins/chat/__init__.py @@ -92,12 +92,13 @@ async def _(bot: Bot): @msg_in.handle() async def _(bot: Bot, event: MessageEvent, state: T_State): - #处理合并转发消息 + # 处理合并转发消息 if "forward" in event.message: - await chat_bot.handle_forward_message(event , bot) - else : + await chat_bot.handle_forward_message(event, bot) + else: await chat_bot.handle_message(event, bot) + @notice_matcher.handle() async def _(bot: Bot, event: NoticeEvent, state: T_State): logger.debug(f"收到通知:{event}") diff --git a/src/plugins/chat/bot.py b/src/plugins/chat/bot.py index d30940f97..24b7bdbff 100644 --- a/src/plugins/chat/bot.py +++ b/src/plugins/chat/bot.py @@ -418,13 +418,12 @@ class ChatBot: # 用户屏蔽,不区分私聊/群聊 if event.user_id in global_config.ban_user_id: return - + if isinstance(event, GroupMessageEvent): if event.group_id: if event.group_id not in global_config.talk_allowed_groups: return - # 获取合并转发消息的详细信息 forward_info = await bot.get_forward_msg(message_id=event.message_id) messages = forward_info["messages"] @@ -434,17 +433,17 @@ class ChatBot: for node in messages: # 提取发送者昵称 nickname = node["sender"].get("nickname", "未知用户") - + # 递归处理消息内容 - message_content = await self.process_message_segments(node["message"],layer=0) - + message_content = await self.process_message_segments(node["message"], layer=0) + # 拼接为【昵称】+ 内容 processed_messages.append(f"【{nickname}】{message_content}") # 组合所有消息 combined_message = "\n".join(processed_messages) combined_message = f"合并转发消息内容:\n{combined_message}" - + # 构建用户信息(使用转发消息的发送者) user_info = UserInfo( user_id=event.user_id, @@ -456,11 +455,7 @@ class ChatBot: # 构建群聊信息(如果是群聊) group_info = None if isinstance(event, GroupMessageEvent): - group_info = GroupInfo( - group_id=event.group_id, - group_name=None, - platform="qq" - ) + group_info = GroupInfo(group_id=event.group_id, group_name=None, platform="qq") # 创建消息对象 message_cq = MessageRecvCQ( @@ -475,19 +470,19 @@ class ChatBot: # 进入标准消息处理流程 await self.message_process(message_cq) - async def process_message_segments(self, segments: list,layer:int) -> str: + async def process_message_segments(self, segments: list, layer: int) -> str: """递归处理消息段""" parts = [] for seg in segments: - part = await self.process_segment(seg,layer+1) + part = await self.process_segment(seg, layer + 1) parts.append(part) return "".join(parts) - async def process_segment(self, seg: dict , layer:int) -> str: + async def process_segment(self, seg: dict, layer: int) -> str: """处理单个消息段""" seg_type = seg["type"] - if layer > 3 : - #防止有那种100层转发消息炸飞麦麦 + if layer > 3: + # 防止有那种100层转发消息炸飞麦麦 return "【转发消息】" if seg_type == "text": return seg["data"]["text"] @@ -504,13 +499,14 @@ class ChatBot: nested_messages.append("合并转发消息内容:") for node in nested_nodes: nickname = node["sender"].get("nickname", "未知用户") - content = await self.process_message_segments(node["message"],layer=layer) + content = await self.process_message_segments(node["message"], layer=layer) # nested_messages.append('-' * layer) nested_messages.append(f"{'--' * layer}【{nickname}】{content}") # nested_messages.append(f"{'--' * layer}合并转发第【{layer}】层结束") return "\n".join(nested_messages) else: return f"[{seg_type}]" - + + # 创建全局ChatBot实例 chat_bot = ChatBot() diff --git a/src/plugins/personality/big5_test.py b/src/plugins/personality/big5_test.py index 80114ec36..c66e6ec4e 100644 --- a/src/plugins/personality/big5_test.py +++ b/src/plugins/personality/big5_test.py @@ -15,17 +15,14 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.scene import get_scene_by_factor,get_all_scenes,PERSONALITY_SCENES -from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS,FACTOR_DESCRIPTIONS -from src.plugins.personality.offline_llm import LLMModel - +from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS, FACTOR_DESCRIPTIONS # noqa: E402 class BigFiveTest: def __init__(self): self.questions = PERSONALITY_QUESTIONS self.factors = FACTOR_DESCRIPTIONS - + def run_test(self): """运行测试并收集答案""" print("\n欢迎参加中国大五人格测试!") @@ -37,17 +34,17 @@ class BigFiveTest: print("5 = 比较符合") print("6 = 完全符合") print("\n请认真阅读每个描述,选择最符合您实际情况的选项。\n") - + # 创建题目序号到题目的映射 - questions_map = {q['id']: q for q in self.questions} - + questions_map = {q["id"]: q for q in self.questions} + # 获取所有题目ID并随机打乱顺序 question_ids = list(questions_map.keys()) random.shuffle(question_ids) - + answers = {} total_questions = len(question_ids) - + for i, question_id in enumerate(question_ids, 1): question = questions_map[question_id] while True: @@ -61,52 +58,43 @@ class BigFiveTest: print("请输入1-6之间的数字!") except ValueError: print("请输入有效的数字!") - + return self.calculate_scores(answers) - + def calculate_scores(self, answers): """计算各维度得分""" results = {} - factor_questions = { - "外向性": [], - "神经质": [], - "严谨性": [], - "开放性": [], - "宜人性": [] - } - + factor_questions = {"外向性": [], "神经质": [], "严谨性": [], "开放性": [], "宜人性": []} + # 将题目按因子分类 for q in self.questions: - factor_questions[q['factor']].append(q) - + factor_questions[q["factor"]].append(q) + # 计算每个维度的得分 for factor, questions in factor_questions.items(): total_score = 0 for q in questions: - score = answers[q['id']] + score = answers[q["id"]] # 处理反向计分题目 - if q['reverse_scoring']: + if q["reverse_scoring"]: score = 7 - score # 6分量表反向计分为7减原始分 total_score += score - + # 计算平均分 avg_score = round(total_score / len(questions), 2) - results[factor] = { - "得分": avg_score, - "题目数": len(questions), - "总分": total_score - } - + results[factor] = {"得分": avg_score, "题目数": len(questions), "总分": total_score} + return results def get_factor_description(self, factor): """获取因子的详细描述""" return self.factors[factor] + def main(): test = BigFiveTest() results = test.run_test() - + print("\n测试结果:") print("=" * 50) for factor, data in results.items(): @@ -114,9 +102,10 @@ def main(): print(f"平均分: {data['得分']} (总分: {data['总分']}, 题目数: {data['题目数']})") print("-" * 30) description = test.get_factor_description(factor) - print("维度说明:", description['description'][:100] + "...") - print("\n特征词:", ", ".join(description['trait_words'])) + print("维度说明:", description["description"][:100] + "...") + print("\n特征词:", ", ".join(description["trait_words"])) print("=" * 50) - + + if __name__ == "__main__": main() diff --git a/src/plugins/personality/combined_test.py b/src/plugins/personality/combined_test.py index a842847fb..b08fb458a 100644 --- a/src/plugins/personality/combined_test.py +++ b/src/plugins/personality/combined_test.py @@ -1,4 +1,4 @@ -from typing import Dict, List +from typing import Dict import json import os from pathlib import Path @@ -14,16 +14,17 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.big5_test import BigFiveTest -from src.plugins.personality.renqingziji import PersonalityEvaluator_direct -from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS, PERSONALITY_QUESTIONS +from src.plugins.personality.big5_test import BigFiveTest # noqa: E402 +from src.plugins.personality.renqingziji import PersonalityEvaluator_direct # noqa: E402 +from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS, PERSONALITY_QUESTIONS # noqa: E402 + class CombinedPersonalityTest: def __init__(self): self.big5_test = BigFiveTest() self.scenario_test = PersonalityEvaluator_direct() self.dimensions = ["开放性", "严谨性", "外向性", "宜人性", "神经质"] - + def run_combined_test(self): """运行组合测试""" print("\n=== 人格特征综合评估系统 ===") @@ -32,12 +33,12 @@ class CombinedPersonalityTest: print("2. 情景反应测评(15个场景)") print("\n两种测评完成后,将对比分析结果的异同。") input("\n准备好开始第一部分(问卷测评)了吗?按回车继续...") - + # 运行问卷测试 print("\n=== 第一部分:问卷测评 ===") print("本部分采用六级评分,请根据每个描述与您的符合程度进行打分:") print("1 = 完全不符合") - print("2 = 比较不符合") + print("2 = 比较不符合") print("3 = 有点不符合") print("4 = 有点符合") print("5 = 比较符合") @@ -47,42 +48,39 @@ class CombinedPersonalityTest: print("2. 根据您想要扮演的角色特征来回答") print("\n无论选择哪种方式,请保持一致并认真回答每个问题。") input("\n按回车开始答题...") - + questionnaire_results = self.run_questionnaire() - + # 转换问卷结果格式以便比较 - questionnaire_scores = { - factor: data["得分"] - for factor, data in questionnaire_results.items() - } - + questionnaire_scores = {factor: data["得分"] for factor, data in questionnaire_results.items()} + # 运行情景测试 print("\n=== 第二部分:情景反应测评 ===") print("接下来,您将面对一系列具体场景,请描述您在每个场景中可能的反应。") print("每个场景都会评估不同的人格维度,共15个场景。") print("您可以选择提供自己的真实反应,也可以选择扮演一个您创作的角色来回答。") input("\n准备好开始了吗?按回车继续...") - + scenario_results = self.run_scenario_test() - + # 比较和展示结果 self.compare_and_display_results(questionnaire_scores, scenario_results) - + # 保存结果 self.save_results(questionnaire_scores, scenario_results) def run_questionnaire(self): """运行问卷测试部分""" # 创建题目序号到题目的映射 - questions_map = {q['id']: q for q in PERSONALITY_QUESTIONS} - + questions_map = {q["id"]: q for q in PERSONALITY_QUESTIONS} + # 获取所有题目ID并随机打乱顺序 question_ids = list(questions_map.keys()) random.shuffle(question_ids) - + answers = {} total_questions = len(question_ids) - + for i, question_id in enumerate(question_ids, 1): question = questions_map[question_id] while True: @@ -97,48 +95,38 @@ class CombinedPersonalityTest: print("请输入1-6之间的数字!") except ValueError: print("请输入有效的数字!") - + # 每10题显示一次进度 if i % 10 == 0: - print(f"\n已完成 {i}/{total_questions} 题 ({int(i/total_questions*100)}%)") - + print(f"\n已完成 {i}/{total_questions} 题 ({int(i / total_questions * 100)}%)") + return self.calculate_questionnaire_scores(answers) - + def calculate_questionnaire_scores(self, answers): """计算问卷测试的维度得分""" results = {} - factor_questions = { - "外向性": [], - "神经质": [], - "严谨性": [], - "开放性": [], - "宜人性": [] - } - + factor_questions = {"外向性": [], "神经质": [], "严谨性": [], "开放性": [], "宜人性": []} + # 将题目按因子分类 for q in PERSONALITY_QUESTIONS: - factor_questions[q['factor']].append(q) - + factor_questions[q["factor"]].append(q) + # 计算每个维度的得分 for factor, questions in factor_questions.items(): total_score = 0 for q in questions: - score = answers[q['id']] + score = answers[q["id"]] # 处理反向计分题目 - if q['reverse_scoring']: + if q["reverse_scoring"]: score = 7 - score # 6分量表反向计分为7减原始分 total_score += score - + # 计算平均分 avg_score = round(total_score / len(questions), 2) - results[factor] = { - "得分": avg_score, - "题目数": len(questions), - "总分": total_score - } - + results[factor] = {"得分": avg_score, "题目数": len(questions), "总分": total_score} + return results - + def run_scenario_test(self): """运行情景测试部分""" final_scores = {"开放性": 0, "严谨性": 0, "外向性": 0, "宜人性": 0, "神经质": 0} @@ -160,11 +148,7 @@ class CombinedPersonalityTest: continue print("\n正在评估您的描述...") - scores = self.scenario_test.evaluate_response( - scenario_data["场景"], - response, - scenario_data["评估维度"] - ) + scores = self.scenario_test.evaluate_response(scenario_data["场景"], response, scenario_data["评估维度"]) # 更新分数 for dimension, score in scores.items(): @@ -178,7 +162,7 @@ class CombinedPersonalityTest: # 每5个场景显示一次总进度 if i % 5 == 0: - print(f"\n已完成 {i}/{len(scenarios)} 个场景 ({int(i/len(scenarios)*100)}%)") + print(f"\n已完成 {i}/{len(scenarios)} 个场景 ({int(i / len(scenarios) * 100)}%)") if i < len(scenarios): input("\n按回车继续下一个场景...") @@ -186,11 +170,8 @@ class CombinedPersonalityTest: # 计算平均分 for dimension in final_scores: if dimension_counts[dimension] > 0: - final_scores[dimension] = round( - final_scores[dimension] / dimension_counts[dimension], - 2 - ) - + final_scores[dimension] = round(final_scores[dimension] / dimension_counts[dimension], 2) + return final_scores def compare_and_display_results(self, questionnaire_scores: Dict, scenario_scores: Dict): @@ -199,39 +180,43 @@ class CombinedPersonalityTest: print("\n" + "=" * 60) print(f"{'维度':<8} {'问卷得分':>10} {'情景得分':>10} {'差异':>10} {'差异程度':>10}") print("-" * 60) - + # 收集每个维度的得分用于统计分析 questionnaire_values = [] scenario_values = [] diffs = [] - + for dimension in self.dimensions: q_score = questionnaire_scores[dimension] s_score = scenario_scores[dimension] diff = round(abs(q_score - s_score), 2) - + questionnaire_values.append(q_score) scenario_values.append(s_score) diffs.append(diff) - + # 计算差异程度 diff_level = "低" if diff < 0.5 else "中" if diff < 1.0 else "高" print(f"{dimension:<8} {q_score:>10.2f} {s_score:>10.2f} {diff:>10.2f} {diff_level:>10}") - + print("=" * 60) - + # 计算整体统计指标 mean_diff = sum(diffs) / len(diffs) std_diff = (sum((x - mean_diff) ** 2 for x in diffs) / (len(diffs) - 1)) ** 0.5 - + # 计算效应量 (Cohen's d) - pooled_std = ((sum((x - sum(questionnaire_values)/len(questionnaire_values))**2 for x in questionnaire_values) + - sum((x - sum(scenario_values)/len(scenario_values))**2 for x in scenario_values)) / - (2 * len(self.dimensions) - 2)) ** 0.5 - + pooled_std = ( + ( + sum((x - sum(questionnaire_values) / len(questionnaire_values)) ** 2 for x in questionnaire_values) + + sum((x - sum(scenario_values) / len(scenario_values)) ** 2 for x in scenario_values) + ) + / (2 * len(self.dimensions) - 2) + ) ** 0.5 + if pooled_std != 0: cohens_d = abs(mean_diff / pooled_std) - + # 解释效应量 if cohens_d < 0.2: effect_size = "微小" @@ -241,41 +226,43 @@ class CombinedPersonalityTest: effect_size = "中等" else: effect_size = "大" - + # 对所有维度进行整体t检验 t_stat, p_value = stats.ttest_rel(questionnaire_values, scenario_values) - print(f"\n整体统计分析:") + print("\n整体统计分析:") print(f"平均差异: {mean_diff:.3f}") print(f"差异标准差: {std_diff:.3f}") print(f"效应量(Cohen's d): {cohens_d:.3f}") print(f"效应量大小: {effect_size}") print(f"t统计量: {t_stat:.3f}") print(f"p值: {p_value:.3f}") - + if p_value < 0.05: print("结论: 两种测评方法的结果存在显著差异 (p < 0.05)") else: print("结论: 两种测评方法的结果无显著差异 (p >= 0.05)") - + print("\n维度说明:") for dimension in self.dimensions: print(f"\n{dimension}:") desc = FACTOR_DESCRIPTIONS[dimension] print(f"定义:{desc['description']}") print(f"特征词:{', '.join(desc['trait_words'])}") - + # 分析显著差异 significant_diffs = [] for dimension in self.dimensions: diff = abs(questionnaire_scores[dimension] - scenario_scores[dimension]) if diff >= 1.0: # 差异大于等于1分视为显著 - significant_diffs.append({ - "dimension": dimension, - "diff": diff, - "questionnaire": questionnaire_scores[dimension], - "scenario": scenario_scores[dimension] - }) - + significant_diffs.append( + { + "dimension": dimension, + "diff": diff, + "questionnaire": questionnaire_scores[dimension], + "scenario": scenario_scores[dimension], + } + ) + if significant_diffs: print("\n\n显著差异分析:") print("-" * 40) @@ -284,9 +271,9 @@ class CombinedPersonalityTest: print(f"问卷得分:{diff['questionnaire']:.2f}") print(f"情景得分:{diff['scenario']:.2f}") print(f"差异值:{diff['diff']:.2f}") - + # 分析可能的原因 - if diff['questionnaire'] > diff['scenario']: + if diff["questionnaire"] > diff["scenario"]: print("可能原因:在问卷中的自我评价较高,但在具体情景中的表现较为保守。") else: print("可能原因:在具体情景中表现出更多该维度特征,而在问卷自评时较为保守。") @@ -297,38 +284,37 @@ class CombinedPersonalityTest: "测试时间": datetime.now().strftime("%Y-%m-%d %H:%M:%S"), "问卷测评结果": questionnaire_scores, "情景测评结果": scenario_scores, - "维度说明": FACTOR_DESCRIPTIONS + "维度说明": FACTOR_DESCRIPTIONS, } - + # 确保目录存在 os.makedirs("results", exist_ok=True) - + # 生成带时间戳的文件名 filename = f"results/personality_combined_{datetime.now().strftime('%Y%m%d_%H%M%S')}.json" - + # 保存到文件 with open(filename, "w", encoding="utf-8") as f: json.dump(results, f, ensure_ascii=False, indent=2) - + print(f"\n完整的测评结果已保存到:{filename}") + def load_existing_results(): """检查并加载已有的测试结果""" results_dir = "results" if not os.path.exists(results_dir): return None - + # 获取所有personality_combined开头的文件 - result_files = [f for f in os.listdir(results_dir) - if f.startswith("personality_combined_") and f.endswith(".json")] - + result_files = [f for f in os.listdir(results_dir) if f.startswith("personality_combined_") and f.endswith(".json")] + if not result_files: return None - + # 按文件修改时间排序,获取最新的结果文件 - latest_file = max(result_files, - key=lambda f: os.path.getmtime(os.path.join(results_dir, f))) - + latest_file = max(result_files, key=lambda f: os.path.getmtime(os.path.join(results_dir, f))) + print(f"\n发现已有的测试结果:{latest_file}") try: with open(os.path.join(results_dir, latest_file), "r", encoding="utf-8") as f: @@ -338,24 +324,26 @@ def load_existing_results(): print(f"读取结果文件时出错:{str(e)}") return None + def main(): test = CombinedPersonalityTest() - + # 检查是否存在已有结果 existing_results = load_existing_results() - + if existing_results: print("\n=== 使用已有测试结果进行分析 ===") print(f"测试时间:{existing_results['测试时间']}") - + questionnaire_scores = existing_results["问卷测评结果"] scenario_scores = existing_results["情景测评结果"] - + # 直接进行结果对比分析 test.compare_and_display_results(questionnaire_scores, scenario_scores) else: print("\n未找到已有的测试结果,开始新的测试...") test.run_combined_test() + if __name__ == "__main__": - main() \ No newline at end of file + main() diff --git a/src/plugins/personality/questionnaire.py b/src/plugins/personality/questionnaire.py index 4afff1185..8e965061d 100644 --- a/src/plugins/personality/questionnaire.py +++ b/src/plugins/personality/questionnaire.py @@ -1,5 +1,9 @@ -# 人格测试问卷题目 王孟成, 戴晓阳, & 姚树桥. (2011). 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. 中国临床心理学杂志, 19(04), Article 04. -# 王孟成, 戴晓阳, & 姚树桥. (2010). 中国大五人格问卷的初步编制Ⅰ:理论框架与信度分析. 中国临床心理学杂志, 18(05), Article 05. +# 人格测试问卷题目 +# 王孟成, 戴晓阳, & 姚树桥. (2011). +# 中国大五人格问卷的初步编制Ⅲ:简式版的制定及信效度检验. 中国临床心理学杂志, 19(04), Article 04. + +# 王孟成, 戴晓阳, & 姚树桥. (2010). +# 中国大五人格问卷的初步编制Ⅰ:理论框架与信度分析. 中国临床心理学杂志, 18(05), Article 05. PERSONALITY_QUESTIONS = [ # 神经质维度 (F1) @@ -11,7 +15,6 @@ PERSONALITY_QUESTIONS = [ {"id": 6, "content": "在面对压力时,我有种快要崩溃的感觉", "factor": "神经质", "reverse_scoring": False}, {"id": 7, "content": "我常担忧一些无关紧要的事情", "factor": "神经质", "reverse_scoring": False}, {"id": 8, "content": "我常常感到内心不踏实", "factor": "神经质", "reverse_scoring": False}, - # 严谨性维度 (F2) {"id": 9, "content": "在工作上,我常只求能应付过去便可", "factor": "严谨性", "reverse_scoring": True}, {"id": 10, "content": "一旦确定了目标,我会坚持努力地实现它", "factor": "严谨性", "reverse_scoring": False}, @@ -21,9 +24,13 @@ PERSONALITY_QUESTIONS = [ {"id": 14, "content": "我喜欢一开头就把事情计划好", "factor": "严谨性", "reverse_scoring": False}, {"id": 15, "content": "我工作或学习很勤奋", "factor": "严谨性", "reverse_scoring": False}, {"id": 16, "content": "我是个倾尽全力做事的人", "factor": "严谨性", "reverse_scoring": False}, - # 宜人性维度 (F3) - {"id": 17, "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈),我仍然相信人性总的来说是善良的", "factor": "宜人性", "reverse_scoring": False}, + { + "id": 17, + "content": "尽管人类社会存在着一些阴暗的东西(如战争、罪恶、欺诈),我仍然相信人性总的来说是善良的", + "factor": "宜人性", + "reverse_scoring": False, + }, {"id": 18, "content": "我觉得大部分人基本上是心怀善意的", "factor": "宜人性", "reverse_scoring": False}, {"id": 19, "content": "虽然社会上有骗子,但我觉得大部分人还是可信的", "factor": "宜人性", "reverse_scoring": False}, {"id": 20, "content": "我不太关心别人是否受到不公正的待遇", "factor": "宜人性", "reverse_scoring": True}, @@ -31,7 +38,6 @@ PERSONALITY_QUESTIONS = [ {"id": 22, "content": "我常为那些遭遇不幸的人感到难过", "factor": "宜人性", "reverse_scoring": False}, {"id": 23, "content": "我是那种只照顾好自己,不替别人担忧的人", "factor": "宜人性", "reverse_scoring": True}, {"id": 24, "content": "当别人向我诉说不幸时,我常感到难过", "factor": "宜人性", "reverse_scoring": False}, - # 开放性维度 (F4) {"id": 25, "content": "我的想象力相当丰富", "factor": "开放性", "reverse_scoring": False}, {"id": 26, "content": "我头脑中经常充满生动的画面", "factor": "开放性", "reverse_scoring": False}, @@ -39,9 +45,18 @@ PERSONALITY_QUESTIONS = [ {"id": 28, "content": "我喜欢冒险", "factor": "开放性", "reverse_scoring": False}, {"id": 29, "content": "我是个勇于冒险,突破常规的人", "factor": "开放性", "reverse_scoring": False}, {"id": 30, "content": "我身上具有别人没有的冒险精神", "factor": "开放性", "reverse_scoring": False}, - {"id": 31, "content": "我渴望学习一些新东西,即使它们与我的日常生活无关", "factor": "开放性", "reverse_scoring": False}, - {"id": 32, "content": "我很愿意也很容易接受那些新事物、新观点、新想法", "factor": "开放性", "reverse_scoring": False}, - + { + "id": 31, + "content": "我渴望学习一些新东西,即使它们与我的日常生活无关", + "factor": "开放性", + "reverse_scoring": False, + }, + { + "id": 32, + "content": "我很愿意也很容易接受那些新事物、新观点、新想法", + "factor": "开放性", + "reverse_scoring": False, + }, # 外向性维度 (F5) {"id": 33, "content": "我喜欢参加社交与娱乐聚会", "factor": "外向性", "reverse_scoring": False}, {"id": 34, "content": "我对人多的聚会感到乏味", "factor": "外向性", "reverse_scoring": True}, @@ -50,61 +65,78 @@ PERSONALITY_QUESTIONS = [ {"id": 37, "content": "有我在的场合一般不会冷场", "factor": "外向性", "reverse_scoring": False}, {"id": 38, "content": "我希望成为领导者而不是被领导者", "factor": "外向性", "reverse_scoring": False}, {"id": 39, "content": "在一个团体中,我希望处于领导地位", "factor": "外向性", "reverse_scoring": False}, - {"id": 40, "content": "别人多认为我是一个热情和友好的人", "factor": "外向性", "reverse_scoring": False} + {"id": 40, "content": "别人多认为我是一个热情和友好的人", "factor": "外向性", "reverse_scoring": False}, ] # 因子维度说明 FACTOR_DESCRIPTIONS = { "外向性": { - "description": "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性,包括对社交活动的兴趣、对人群的态度、社交互动中的主动程度以及在群体中的影响力。高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我,并往往在群体中发挥领导作用;低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。", + "description": "反映个体神经系统的强弱和动力特征。外向性主要表现为个体在人际交往和社交活动中的倾向性," + "包括对社交活动的兴趣、" + "对人群的态度、社交互动中的主动程度以及在群体中的影响力。高分者倾向于积极参与社交活动,乐于与人交往,善于表达自我," + "并往往在群体中发挥领导作用;低分者则倾向于独处,不喜欢热闹的社交场合,表现出内向、安静的特征。", "trait_words": ["热情", "活力", "社交", "主动"], "subfactors": { "合群性": "个体愿意与他人聚在一起,即接近人群的倾向;高分表现乐群、好交际,低分表现封闭、独处", "热情": "个体对待别人时所表现出的态度;高分表现热情好客,低分表现冷淡", "支配性": "个体喜欢指使、操纵他人,倾向于领导别人的特点;高分表现好强、发号施令,低分表现顺从、低调", - "活跃": "个体精力充沛,活跃、主动性等特点;高分表现活跃,低分表现安静" - } + "活跃": "个体精力充沛,活跃、主动性等特点;高分表现活跃,低分表现安静", + }, }, "神经质": { - "description": "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、挫折和日常生活挑战时的情绪稳定性和适应能力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度,以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波动较大;低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。", + "description": "反映个体情绪的状态和体验内心苦恼的倾向性。这个维度主要关注个体在面对压力、" + "挫折和日常生活挑战时的情绪稳定性和适应能力。它包含了对焦虑、抑郁、愤怒等负面情绪的敏感程度," + "以及个体对这些情绪的调节和控制能力。高分者容易体验负面情绪,对压力较为敏感,情绪波动较大;" + "低分者则表现出较强的情绪稳定性,能够较好地应对压力和挫折。", "trait_words": ["稳定", "沉着", "从容", "坚韧"], "subfactors": { "焦虑": "个体体验焦虑感的个体差异;高分表现坐立不安,低分表现平静", "抑郁": "个体体验抑郁情感的个体差异;高分表现郁郁寡欢,低分表现平静", - "敏感多疑": "个体常常关注自己的内心活动,行为和过于意识人对自己的看法、评价;高分表现敏感多疑,低分表现淡定、自信", + "敏感多疑": "个体常常关注自己的内心活动,行为和过于意识人对自己的看法、评价;高分表现敏感多疑," + "低分表现淡定、自信", "脆弱性": "个体在危机或困难面前无力、脆弱的特点;高分表现无能、易受伤、逃避,低分表现坚强", - "愤怒-敌意": "个体准备体验愤怒,及相关情绪的状态;高分表现暴躁易怒,低分表现平静" - } + "愤怒-敌意": "个体准备体验愤怒,及相关情绪的状态;高分表现暴躁易怒,低分表现平静", + }, }, "严谨性": { - "description": "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、学习等目标性活动中的自我约束和行为管理能力。它涉及到个体的责任感、自律性、计划性、条理性以及完成任务的态度。高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的努力精神;低分者则可能表现出随意性强、缺乏规划、做事马虎或易放弃的特点。", + "description": "反映个体在目标导向行为上的组织、坚持和动机特征。这个维度体现了个体在工作、" + "学习等目标性活动中的自我约束和行为管理能力。它涉及到个体的责任感、自律性、计划性、条理性以及完成任务的态度。" + "高分者往往表现出强烈的责任心、良好的组织能力、谨慎的决策风格和持续的努力精神;低分者则可能表现出随意性强、" + "缺乏规划、做事马虎或易放弃的特点。", "trait_words": ["负责", "自律", "条理", "勤奋"], "subfactors": { - "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;高分表现有责任心、负责任,低分表现推卸责任、逃避处罚", + "责任心": "个体对待任务和他人认真负责,以及对自己承诺的信守;高分表现有责任心、负责任," + "低分表现推卸责任、逃避处罚", "自我控制": "个体约束自己的能力,及自始至终的坚持性;高分表现自制、有毅力,低分表现冲动、无毅力", "审慎性": "个体在采取具体行动前的心理状态;高分表现谨慎、小心,低分表现鲁莽、草率", "条理性": "个体处理事务和工作的秩序,条理和逻辑性;高分表现整洁、有秩序,低分表现混乱、遗漏", - "勤奋": "个体工作和学习的努力程度及为达到目标而表现出的进取精神;高分表现勤奋、刻苦,低分表现懒散" - } + "勤奋": "个体工作和学习的努力程度及为达到目标而表现出的进取精神;高分表现勤奋、刻苦,低分表现懒散", + }, }, "开放性": { - "description": "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。这个维度体现了个体在认知和体验方面的广度、深度和灵活性。它包括对艺术的欣赏能力、对知识的求知欲、想象力的丰富程度,以及对冒险和创新的态度。高分者往往具有丰富的想象力、广泛的兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、传统,喜欢熟悉和常规的事物。", + "description": "反映个体对新异事物、新观念和新经验的接受程度,以及在思维和行为方面的创新倾向。" + "这个维度体现了个体在认知和体验方面的广度、深度和灵活性。它包括对艺术的欣赏能力、对知识的求知欲、想象力的丰富程度," + "以及对冒险和创新的态度。高分者往往具有丰富的想象力、广泛的兴趣、开放的思维方式和创新的倾向;低分者则倾向于保守、" + "传统,喜欢熟悉和常规的事物。", "trait_words": ["创新", "好奇", "艺术", "冒险"], "subfactors": { "幻想": "个体富于幻想和想象的水平;高分表现想象力丰富,低分表现想象力匮乏", "审美": "个体对于艺术和美的敏感与热爱程度;高分表现富有艺术气息,低分表现一般对艺术不敏感", "好奇心": "个体对未知事物的态度;高分表现兴趣广泛、好奇心浓,低分表现兴趣少、无好奇心", "冒险精神": "个体愿意尝试有风险活动的个体差异;高分表现好冒险,低分表现保守", - "价值观念": "个体对新事物、新观念、怪异想法的态度;高分表现开放、坦然接受新事物,低分则相反" - } + "价值观念": "个体对新事物、新观念、怪异想法的态度;高分表现开放、坦然接受新事物,低分则相反", + }, }, "宜人性": { - "description": "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。这个维度主要关注个体与他人互动时的态度和行为特征,包括对他人的信任程度、同理心水平、助人意愿以及在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人建立和谐关系;低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑他人感受。", + "description": "反映个体在人际关系中的亲和倾向,体现了对他人的关心、同情和合作意愿。" + "这个维度主要关注个体与他人互动时的态度和行为特征,包括对他人的信任程度、同理心水平、" + "助人意愿以及在人际冲突中的处理方式。高分者通常表现出友善、富有同情心、乐于助人的特质,善于与他人建立和谐关系;" + "低分者则可能表现出较少的人际关注,在社交互动中更注重自身利益,较少考虑他人感受。", "trait_words": ["友善", "同理", "信任", "合作"], "subfactors": { "信任": "个体对他人和/或他人言论的相信程度;高分表现信任他人,低分表现怀疑", "体贴": "个体对别人的兴趣和需要的关注程度;高分表现体贴、温存,低分表现冷漠、不在乎", - "同情": "个体对处于不利地位的人或物的态度;高分表现富有同情心,低分表现冷漠" - } - } -} \ No newline at end of file + "同情": "个体对处于不利地位的人或物的态度;高分表现富有同情心,低分表现冷漠", + }, + }, +} diff --git a/src/plugins/personality/renqingziji.py b/src/plugins/personality/renqingziji.py index b3a3e267e..4b1fb3b69 100644 --- a/src/plugins/personality/renqingziji.py +++ b/src/plugins/personality/renqingziji.py @@ -1,10 +1,12 @@ -''' -The definition of artificial personality in this paper follows the dispositional para-digm and adapts a definition of personality developed for humans [17]: -Personality for a human is the "whole and organisation of relatively stable tendencies and patterns of experience and -behaviour within one person (distinguishing it from other persons)". This definition is modified for artificial personality: -Artificial personality describes the relatively stable tendencies and patterns of behav-iour of an AI-based machine that -can be designed by developers and designers via different modalities, such as language, creating the impression -of individuality of a humanized social agent when users interact with the machine.''' +""" +The definition of artificial personality in this paper follows the dispositional para-digm and adapts a definition of +personality developed for humans [17]: +Personality for a human is the "whole and organisation of relatively stable tendencies and patterns of experience and +behaviour within one person (distinguishing it from other persons)". This definition is modified for artificial +personality: +Artificial personality describes the relatively stable tendencies and patterns of behav-iour of an AI-based machine that +can be designed by developers and designers via different modalities, such as language, creating the impression +of individuality of a humanized social agent when users interact with the machine.""" from typing import Dict, List import json @@ -13,9 +15,9 @@ from pathlib import Path from dotenv import load_dotenv import sys -''' +""" 第一种方案:基于情景评估的人格测定 -''' +""" current_dir = Path(__file__).resolve().parent project_root = current_dir.parent.parent.parent env_path = project_root / ".env.prod" @@ -23,9 +25,9 @@ env_path = project_root / ".env.prod" root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.plugins.personality.scene import get_scene_by_factor,get_all_scenes,PERSONALITY_SCENES -from src.plugins.personality.questionnaire import PERSONALITY_QUESTIONS,FACTOR_DESCRIPTIONS -from src.plugins.personality.offline_llm import LLMModel +from src.plugins.personality.scene import get_scene_by_factor, PERSONALITY_SCENES # noqa: E402 +from src.plugins.personality.questionnaire import FACTOR_DESCRIPTIONS # noqa: E402 +from src.plugins.personality.offline_llm import LLMModel # noqa: E402 # 加载环境变量 if env_path.exists(): @@ -40,32 +42,31 @@ class PersonalityEvaluator_direct: def __init__(self): self.personality_traits = {"开放性": 0, "严谨性": 0, "外向性": 0, "宜人性": 0, "神经质": 0} self.scenarios = [] - + # 为每个人格特质获取对应的场景 for trait in PERSONALITY_SCENES: scenes = get_scene_by_factor(trait) if not scenes: continue - + # 从每个维度选择3个场景 import random + scene_keys = list(scenes.keys()) selected_scenes = random.sample(scene_keys, min(3, len(scene_keys))) - + for scene_key in selected_scenes: scene = scenes[scene_key] - + # 为每个场景添加评估维度 # 主维度是当前特质,次维度随机选择一个其他特质 other_traits = [t for t in PERSONALITY_SCENES if t != trait] secondary_trait = random.choice(other_traits) - - self.scenarios.append({ - "场景": scene["scenario"], - "评估维度": [trait, secondary_trait], - "场景编号": scene_key - }) - + + self.scenarios.append( + {"场景": scene["scenario"], "评估维度": [trait, secondary_trait], "场景编号": scene_key} + ) + self.llm = LLMModel() def evaluate_response(self, scenario: str, response: str, dimensions: List[str]) -> Dict[str, float]: @@ -78,9 +79,9 @@ class PersonalityEvaluator_direct: desc = FACTOR_DESCRIPTIONS.get(dim, "") if desc: dimension_descriptions.append(f"- {dim}:{desc}") - + dimensions_text = "\n".join(dimension_descriptions) - + prompt = f"""请根据以下场景和用户描述,评估用户在大五人格模型中的相关维度得分(1-6分)。 场景描述: @@ -178,11 +179,7 @@ def main(): print(f"测试场景数:{dimension_counts[trait]}") # 保存结果 - result = { - "final_scores": final_scores, - "dimension_counts": dimension_counts, - "scenarios": evaluator.scenarios - } + result = {"final_scores": final_scores, "dimension_counts": dimension_counts, "scenarios": evaluator.scenarios} # 确保目录存在 os.makedirs("results", exist_ok=True) diff --git a/src/plugins/personality/scene.py b/src/plugins/personality/scene.py index 936b07a3e..0ce094a36 100644 --- a/src/plugins/personality/scene.py +++ b/src/plugins/personality/scene.py @@ -1,4 +1,4 @@ -from typing import Dict, List +from typing import Dict PERSONALITY_SCENES = { "外向性": { @@ -8,7 +8,7 @@ PERSONALITY_SCENES = { 同事:「嗨!你是新来的同事吧?我是市场部的小林。」 同事看起来很友善,还主动介绍说:「待会午饭时间,我们部门有几个人准备一起去楼下新开的餐厅,你要一起来吗?可以认识一下其他同事。」""", - "explanation": "这个场景通过职场社交情境,观察个体对于新环境、新社交圈的态度和反应倾向。" + "explanation": "这个场景通过职场社交情境,观察个体对于新环境、新社交圈的态度和反应倾向。", }, "场景2": { "scenario": """在大学班级群里,班长发起了一个组织班级联谊活动的投票: @@ -16,7 +16,7 @@ PERSONALITY_SCENES = { 班长:「大家好!下周末我们准备举办一次班级联谊活动,地点在学校附近的KTV。想请大家报名参加,也欢迎大家邀请其他班级的同学!」 已经有几个同学在群里积极响应,有人@你问你要不要一起参加。""", - "explanation": "通过班级活动场景,观察个体对群体社交活动的参与意愿。" + "explanation": "通过班级活动场景,观察个体对群体社交活动的参与意愿。", }, "场景3": { "scenario": """你在社交平台上发布了一条动态,收到了很多陌生网友的评论和私信: @@ -24,13 +24,14 @@ PERSONALITY_SCENES = { 网友A:「你说的这个观点很有意思!想和你多交流一下。」 网友B:「我也对这个话题很感兴趣,要不要建个群一起讨论?」""", - "explanation": "通过网络社交场景,观察个体对线上社交的态度。" + "explanation": "通过网络社交场景,观察个体对线上社交的态度。", }, "场景4": { "scenario": """你暗恋的对象今天主动来找你: -对方:「那个...我最近在准备一个演讲比赛,听说你口才很好。能不能请你帮我看看演讲稿,顺便给我一些建议?如果你有时间的话,可以一起吃个饭聊聊。」""", - "explanation": "通过恋爱情境,观察个体在面对心仪对象时的社交表现。" +对方:「那个...我最近在准备一个演讲比赛,听说你口才很好。能不能请你帮我看看演讲稿,顺便给我一些建议?""" + """如果你有时间的话,可以一起吃个饭聊聊。」""", + "explanation": "通过恋爱情境,观察个体在面对心仪对象时的社交表现。", }, "场景5": { "scenario": """在一次线下读书会上,主持人突然点名让你分享读后感: @@ -38,18 +39,18 @@ PERSONALITY_SCENES = { 主持人:「听说你对这本书很有见解,能不能和大家分享一下你的想法?」 现场有二十多个陌生的读书爱好者,都期待地看着你。""", - "explanation": "通过即兴发言场景,观察个体的社交表现欲和公众表达能力。" - } + "explanation": "通过即兴发言场景,观察个体的社交表现欲和公众表达能力。", + }, }, - "神经质": { "场景1": { - "scenario": """你正在准备一个重要的项目演示,这关系到你的晋升机会。就在演示前30分钟,你收到了主管发来的消息: + "scenario": """你正在准备一个重要的项目演示,这关系到你的晋升机会。""" + """就在演示前30分钟,你收到了主管发来的消息: 主管:「临时有个变动,CEO也会来听你的演示。他对这个项目特别感兴趣。」 正当你准备回复时,主管又发来一条:「对了,能不能把演示时间压缩到15分钟?CEO下午还有其他安排。你之前准备的是30分钟的版本对吧?」""", - "explanation": "这个场景通过突发的压力情境,观察个体在面对计划外变化时的情绪反应和调节能力。" + "explanation": "这个场景通过突发的压力情境,观察个体在面对计划外变化时的情绪反应和调节能力。", }, "场景2": { "scenario": """期末考试前一天晚上,你收到了好朋友发来的消息: @@ -57,7 +58,7 @@ PERSONALITY_SCENES = { 好朋友:「不好意思这么晚打扰你...我看你平时成绩很好,能不能帮我解答几个问题?我真的很担心明天的考试。」 你看了看时间,已经是晚上11点,而你原本计划的复习还没完成。""", - "explanation": "通过考试压力场景,观察个体在时间紧张时的情绪管理。" + "explanation": "通过考试压力场景,观察个体在时间紧张时的情绪管理。", }, "场景3": { "scenario": """你在社交媒体上发表的一个观点引发了争议,有不少人开始批评你: @@ -67,7 +68,7 @@ PERSONALITY_SCENES = { 网友B:「建议楼主先去补补课再来发言。」 评论区里的负面评论越来越多,还有人开始人身攻击。""", - "explanation": "通过网络争议场景,观察个体面对批评时的心理承受能力。" + "explanation": "通过网络争议场景,观察个体面对批评时的心理承受能力。", }, "场景4": { "scenario": """你和恋人约好今天一起看电影,但在约定时间前半小时,对方发来消息: @@ -77,7 +78,7 @@ PERSONALITY_SCENES = { 二十分钟后,对方又发来消息:「可能要再等等,抱歉!」 电影快要开始了,但对方还是没有出现。""", - "explanation": "通过恋爱情境,观察个体对不确定性的忍耐程度。" + "explanation": "通过恋爱情境,观察个体对不确定性的忍耐程度。", }, "场景5": { "scenario": """在一次重要的小组展示中,你的组员在演示途中突然卡壳了: @@ -85,10 +86,9 @@ PERSONALITY_SCENES = { 组员小声对你说:「我忘词了,接下来的部分是什么来着...」 台下的老师和同学都在等待,气氛有些尴尬。""", - "explanation": "通过公开场合的突发状况,观察个体的应急反应和压力处理能力。" - } + "explanation": "通过公开场合的突发状况,观察个体的应急反应和压力处理能力。", + }, }, - "严谨性": { "场景1": { "scenario": """你是团队的项目负责人,刚刚接手了一个为期两个月的重要项目。在第一次团队会议上: @@ -98,7 +98,7 @@ PERSONALITY_SCENES = { 小张:「要不要先列个时间表?不过感觉太详细的计划也没必要,点到为止就行。」 小李:「客户那边说如果能提前完成有奖励,我觉得我们可以先做快一点的部分。」""", - "explanation": "这个场景通过项目管理情境,体现个体在工作方法、计划性和责任心方面的特征。" + "explanation": "这个场景通过项目管理情境,体现个体在工作方法、计划性和责任心方面的特征。", }, "场景2": { "scenario": """期末小组作业,组长让大家分工完成一份研究报告。在截止日期前三天: @@ -108,7 +108,7 @@ PERSONALITY_SCENES = { 组员B:「我这边可能还要一天才能完成,最近太忙了。」 组员C发来一份没有任何引用出处、可能存在抄袭的内容:「我写完了,你们看看怎么样?」""", - "explanation": "通过学习场景,观察个体对学术规范和质量要求的重视程度。" + "explanation": "通过学习场景,观察个体对学术规范和质量要求的重视程度。", }, "场景3": { "scenario": """你在一个兴趣小组的群聊中,大家正在讨论举办一次线下活动: @@ -118,7 +118,7 @@ PERSONALITY_SCENES = { 成员B:「对啊,随意一点挺好的。」 成员C:「人来了自然就热闹了。」""", - "explanation": "通过活动组织场景,观察个体对活动计划的态度。" + "explanation": "通过活动组织场景,观察个体对活动计划的态度。", }, "场景4": { "scenario": """你和恋人计划一起去旅游,对方说: @@ -126,7 +126,7 @@ PERSONALITY_SCENES = { 恋人:「我们就随心而行吧!订个目的地,其他的到了再说,这样更有意思。」 距离出发还有一周时间,但机票、住宿和具体行程都还没有确定。""", - "explanation": "通过旅行规划场景,观察个体的计划性和对不确定性的接受程度。" + "explanation": "通过旅行规划场景,观察个体的计划性和对不确定性的接受程度。", }, "场景5": { "scenario": """在一个重要的团队项目中,你发现一个同事的工作存在明显错误: @@ -134,18 +134,19 @@ PERSONALITY_SCENES = { 同事:「差不多就行了,反正领导也看不出来。」 这个错误可能不会立即造成问题,但长期来看可能会影响项目质量。""", - "explanation": "通过工作质量场景,观察个体对细节和标准的坚持程度。" - } + "explanation": "通过工作质量场景,观察个体对细节和标准的坚持程度。", + }, }, - "开放性": { "场景1": { "scenario": """周末下午,你的好友小美兴致勃勃地给你打电话: -小美:「我刚发现一个特别有意思的沉浸式艺术展!不是传统那种挂画的展览,而是把整个空间都变成了艺术品。观众要穿特制的服装,还要带上VR眼镜,好像还有AI实时互动!」 +小美:「我刚发现一个特别有意思的沉浸式艺术展!不是传统那种挂画的展览,而是把整个空间都变成了艺术品。""" + """观众要穿特制的服装,还要带上VR眼镜,好像还有AI实时互动!」 -小美继续说:「虽然票价不便宜,但听说体验很独特。网上评价两极分化,有人说是前所未有的艺术革新,也有人说是哗众取宠。要不要周末一起去体验一下?」""", - "explanation": "这个场景通过新型艺术体验,反映个体对创新事物的接受程度和尝试意愿。" +小美继续说:「虽然票价不便宜,但听说体验很独特。网上评价两极分化,有人说是前所未有的艺术革新,也有人说是哗众取宠。""" + """要不要周末一起去体验一下?」""", + "explanation": "这个场景通过新型艺术体验,反映个体对创新事物的接受程度和尝试意愿。", }, "场景2": { "scenario": """在一节创意写作课上,老师提出了一个特别的作业: @@ -153,15 +154,16 @@ PERSONALITY_SCENES = { 老师:「下周的作业是用AI写作工具协助创作一篇小说。你们可以自由探索如何与AI合作,打破传统写作方式。」 班上随即展开了激烈讨论,有人认为这是对创作的亵渎,也有人对这种新形式感到兴奋。""", - "explanation": "通过新技术应用场景,观察个体对创新学习方式的态度。" + "explanation": "通过新技术应用场景,观察个体对创新学习方式的态度。", }, "场景3": { "scenario": """在社交媒体上,你看到一个朋友分享了一种新的生活方式: -「最近我在尝试'数字游牧'生活,就是一边远程工作一边环游世界。没有固定住所,住青旅或短租,认识来自世界各地的朋友。虽然有时会很不稳定,但这种自由的生活方式真的很棒!」 +「最近我在尝试'数字游牧'生活,就是一边远程工作一边环游世界。""" + """没有固定住所,住青旅或短租,认识来自世界各地的朋友。虽然有时会很不稳定,但这种自由的生活方式真的很棒!」 评论区里争论不断,有人向往这种生活,也有人觉得太冒险。""", - "explanation": "通过另类生活方式,观察个体对非传统选择的态度。" + "explanation": "通过另类生活方式,观察个体对非传统选择的态度。", }, "场景4": { "scenario": """你的恋人突然提出了一个想法: @@ -169,7 +171,7 @@ PERSONALITY_SCENES = { 恋人:「我们要不要尝试一下开放式关系?就是在保持彼此关系的同时,也允许和其他人发展感情。现在国外很多年轻人都这样。」 这个提议让你感到意外,你之前从未考虑过这种可能性。""", - "explanation": "通过感情观念场景,观察个体对非传统关系模式的接受度。" + "explanation": "通过感情观念场景,观察个体对非传统关系模式的接受度。", }, "场景5": { "scenario": """在一次朋友聚会上,大家正在讨论未来职业规划: @@ -179,10 +181,9 @@ PERSONALITY_SCENES = { 朋友B:「我想去学习生物科技,准备转行做人造肉研发。」 朋友C:「我在考虑加入一个区块链创业项目,虽然风险很大。」""", - "explanation": "通过职业选择场景,观察个体对新兴领域的探索意愿。" - } + "explanation": "通过职业选择场景,观察个体对新兴领域的探索意愿。", + }, }, - "宜人性": { "场景1": { "scenario": """在回家的公交车上,你遇到这样一幕: @@ -194,7 +195,7 @@ PERSONALITY_SCENES = { 年轻人B:「现在的老年人真是...我看她包里还有菜,肯定是去菜市场买完菜回来的,这么多人都不知道叫子女开车接送。」 就在这时,老奶奶一个趔趄,差点摔倒。她扶住了扶手,但包里的东西洒了一些出来。""", - "explanation": "这个场景通过公共场合的助人情境,体现个体的同理心和对他人需求的关注程度。" + "explanation": "这个场景通过公共场合的助人情境,体现个体的同理心和对他人需求的关注程度。", }, "场景2": { "scenario": """在班级群里,有同学发起为生病住院的同学捐款: @@ -204,7 +205,7 @@ PERSONALITY_SCENES = { 同学B:「我觉得这是他家里的事,我们不方便参与吧。」 同学C:「但是都是同学一场,帮帮忙也是应该的。」""", - "explanation": "通过同学互助场景,观察个体的助人意愿和同理心。" + "explanation": "通过同学互助场景,观察个体的助人意愿和同理心。", }, "场景3": { "scenario": """在一个网络讨论组里,有人发布了求助信息: @@ -215,7 +216,7 @@ PERSONALITY_SCENES = { 「生活本来就是这样,想开点!」 「你这样子太消极了,要积极面对。」 「谁还没点烦心事啊,过段时间就好了。」""", - "explanation": "通过网络互助场景,观察个体的共情能力和安慰方式。" + "explanation": "通过网络互助场景,观察个体的共情能力和安慰方式。", }, "场景4": { "scenario": """你的恋人向你倾诉工作压力: @@ -223,7 +224,7 @@ PERSONALITY_SCENES = { 恋人:「最近工作真的好累,感觉快坚持不下去了...」 但今天你也遇到了很多烦心事,心情也不太好。""", - "explanation": "通过感情关系场景,观察个体在自身状态不佳时的关怀能力。" + "explanation": "通过感情关系场景,观察个体在自身状态不佳时的关怀能力。", }, "场景5": { "scenario": """在一次团队项目中,新来的同事小王因为经验不足,造成了一个严重的错误。在部门会议上: @@ -231,27 +232,29 @@ PERSONALITY_SCENES = { 主管:「这个错误造成了很大的损失,是谁负责的这部分?」 小王看起来很紧张,欲言又止。你知道是他造成的错误,同时你也是这个项目的共同负责人。""", - "explanation": "通过职场情境,观察个体在面对他人过错时的态度和处理方式。" - } - } + "explanation": "通过职场情境,观察个体在面对他人过错时的态度和处理方式。", + }, + }, } + def get_scene_by_factor(factor: str) -> Dict: """ 根据人格因子获取对应的情景测试 - + Args: factor (str): 人格因子名称 - + Returns: Dict: 包含情景描述的字典 """ return PERSONALITY_SCENES.get(factor, None) + def get_all_scenes() -> Dict: """ 获取所有情景测试 - + Returns: Dict: 所有情景测试的字典 """ diff --git a/webui.py b/webui.py index b598df7c0..60ffa4805 100644 --- a/webui.py +++ b/webui.py @@ -4,11 +4,14 @@ import toml import signal import sys import requests + try: from src.common.logger import get_module_logger + logger = get_module_logger("webui") except ImportError: from loguru import logger + # 检查并创建日志目录 log_dir = "logs/webui" if not os.path.exists(log_dir): @@ -24,11 +27,13 @@ import ast from packaging import version from decimal import Decimal + def signal_handler(signum, frame): """处理 Ctrl+C 信号""" logger.info("收到终止信号,正在关闭 Gradio 服务器...") sys.exit(0) + # 注册信号处理器 signal.signal(signal.SIGINT, signal_handler) @@ -44,10 +49,10 @@ if not os.path.exists(".env.prod"): raise FileNotFoundError("环境配置文件 .env.prod 不存在,请检查配置文件路径") config_data = toml.load("config/bot_config.toml") -#增加对老版本配置文件支持 +# 增加对老版本配置文件支持 LEGACY_CONFIG_VERSION = version.parse("0.0.1") -#增加最低支持版本 +# 增加最低支持版本 MIN_SUPPORT_VERSION = version.parse("0.0.8") MIN_SUPPORT_MAIMAI_VERSION = version.parse("0.5.13") @@ -66,7 +71,7 @@ else: HAVE_ONLINE_STATUS_VERSION = version.parse("0.0.9") -#定义意愿模式可选项 +# 定义意愿模式可选项 WILLING_MODE_CHOICES = [ "classical", "dynamic", @@ -74,11 +79,10 @@ WILLING_MODE_CHOICES = [ ] - - -#添加WebUI配置文件版本 +# 添加WebUI配置文件版本 WEBUI_VERSION = version.parse("0.0.9") + # ============================================== # env环境配置文件读取部分 def parse_env_config(config_file): @@ -204,7 +208,7 @@ MODEL_PROVIDER_LIST = parse_model_providers(env_config_data) # env读取保存结束 # ============================================== -#获取在线麦麦数量 +# 获取在线麦麦数量 def get_online_maimbot(url="http://hyybuth.xyz:10058/api/clients/details", timeout=10): @@ -331,19 +335,19 @@ def format_list_to_str(lst): # env保存函数 def save_trigger( - server_address, - server_port, - final_result_list, - t_mongodb_host, - t_mongodb_port, - t_mongodb_database_name, - t_console_log_level, - t_file_log_level, - t_default_console_log_level, - t_default_file_log_level, - t_api_provider, - t_api_base_url, - t_api_key, + server_address, + server_port, + final_result_list, + t_mongodb_host, + t_mongodb_port, + t_mongodb_database_name, + t_console_log_level, + t_file_log_level, + t_default_console_log_level, + t_default_file_log_level, + t_api_provider, + t_api_base_url, + t_api_key, ): final_result_lists = format_list_to_str(final_result_list) env_config_data["env_HOST"] = server_address @@ -412,12 +416,12 @@ def save_bot_config(t_qqbot_qq, t_nickname, t_nickname_final_result): # 监听滑块的值变化,确保总和不超过 1,并显示警告 def adjust_personality_greater_probabilities( - t_personality_1_probability, t_personality_2_probability, t_personality_3_probability + t_personality_1_probability, t_personality_2_probability, t_personality_3_probability ): total = ( - Decimal(str(t_personality_1_probability)) - + Decimal(str(t_personality_2_probability)) - + Decimal(str(t_personality_3_probability)) + Decimal(str(t_personality_1_probability)) + + Decimal(str(t_personality_2_probability)) + + Decimal(str(t_personality_3_probability)) ) if total > Decimal("1.0"): warning_message = ( @@ -428,12 +432,12 @@ def adjust_personality_greater_probabilities( def adjust_personality_less_probabilities( - t_personality_1_probability, t_personality_2_probability, t_personality_3_probability + t_personality_1_probability, t_personality_2_probability, t_personality_3_probability ): total = ( - Decimal(str(t_personality_1_probability)) - + Decimal(str(t_personality_2_probability)) - + Decimal(str(t_personality_3_probability)) + Decimal(str(t_personality_1_probability)) + + Decimal(str(t_personality_2_probability)) + + Decimal(str(t_personality_3_probability)) ) if total < Decimal("1.0"): warning_message = ( @@ -445,9 +449,7 @@ def adjust_personality_less_probabilities( def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) + - Decimal(str(t_model_2_probability)) + - Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) ) if total > Decimal("1.0"): warning_message = ( @@ -459,9 +461,7 @@ def adjust_model_greater_probabilities(t_model_1_probability, t_model_2_probabil def adjust_model_less_probabilities(t_model_1_probability, t_model_2_probability, t_model_3_probability): total = ( - Decimal(str(t_model_1_probability)) - + Decimal(str(t_model_2_probability)) - + Decimal(str(t_model_3_probability)) + Decimal(str(t_model_1_probability)) + Decimal(str(t_model_2_probability)) + Decimal(str(t_model_3_probability)) ) if total < Decimal("1.0"): warning_message = ( @@ -474,13 +474,13 @@ def adjust_model_less_probabilities(t_model_1_probability, t_model_2_probability # ============================================== # 人格保存函数 def save_personality_config( - t_prompt_personality_1, - t_prompt_personality_2, - t_prompt_personality_3, - t_prompt_schedule, - t_personality_1_probability, - t_personality_2_probability, - t_personality_3_probability, + t_prompt_personality_1, + t_prompt_personality_2, + t_prompt_personality_3, + t_prompt_schedule, + t_personality_1_probability, + t_personality_2_probability, + t_personality_3_probability, ): # 保存人格提示词 config_data["personality"]["prompt_personality"][0] = t_prompt_personality_1 @@ -501,20 +501,20 @@ def save_personality_config( def save_message_and_emoji_config( - t_min_text_length, - t_max_context_size, - t_emoji_chance, - t_thinking_timeout, - t_response_willing_amplifier, - t_response_interested_rate_amplifier, - t_down_frequency_rate, - t_ban_words_final_result, - t_ban_msgs_regex_final_result, - t_check_interval, - t_register_interval, - t_auto_save, - t_enable_check, - t_check_prompt, + t_min_text_length, + t_max_context_size, + t_emoji_chance, + t_thinking_timeout, + t_response_willing_amplifier, + t_response_interested_rate_amplifier, + t_down_frequency_rate, + t_ban_words_final_result, + t_ban_msgs_regex_final_result, + t_check_interval, + t_register_interval, + t_auto_save, + t_enable_check, + t_check_prompt, ): config_data["message"]["min_text_length"] = t_min_text_length config_data["message"]["max_context_size"] = t_max_context_size @@ -536,27 +536,27 @@ def save_message_and_emoji_config( def save_response_model_config( - t_willing_mode, - t_model_r1_probability, - t_model_r2_probability, - t_model_r3_probability, - t_max_response_length, - t_model1_name, - t_model1_provider, - t_model1_pri_in, - t_model1_pri_out, - t_model2_name, - t_model2_provider, - t_model3_name, - t_model3_provider, - t_emotion_model_name, - t_emotion_model_provider, - t_topic_judge_model_name, - t_topic_judge_model_provider, - t_summary_by_topic_model_name, - t_summary_by_topic_model_provider, - t_vlm_model_name, - t_vlm_model_provider, + t_willing_mode, + t_model_r1_probability, + t_model_r2_probability, + t_model_r3_probability, + t_max_response_length, + t_model1_name, + t_model1_provider, + t_model1_pri_in, + t_model1_pri_out, + t_model2_name, + t_model2_provider, + t_model3_name, + t_model3_provider, + t_emotion_model_name, + t_emotion_model_provider, + t_topic_judge_model_name, + t_topic_judge_model_provider, + t_summary_by_topic_model_name, + t_summary_by_topic_model_provider, + t_vlm_model_name, + t_vlm_model_provider, ): if PARSED_CONFIG_VERSION >= version.parse("0.0.10"): config_data["willing"]["willing_mode"] = t_willing_mode @@ -586,15 +586,15 @@ def save_response_model_config( def save_memory_mood_config( - t_build_memory_interval, - t_memory_compress_rate, - t_forget_memory_interval, - t_memory_forget_time, - t_memory_forget_percentage, - t_memory_ban_words_final_result, - t_mood_update_interval, - t_mood_decay_rate, - t_mood_intensity_factor, + t_build_memory_interval, + t_memory_compress_rate, + t_forget_memory_interval, + t_memory_forget_time, + t_memory_forget_percentage, + t_memory_ban_words_final_result, + t_mood_update_interval, + t_mood_decay_rate, + t_mood_intensity_factor, ): config_data["memory"]["build_memory_interval"] = t_build_memory_interval config_data["memory"]["memory_compress_rate"] = t_memory_compress_rate @@ -611,17 +611,17 @@ def save_memory_mood_config( def save_other_config( - t_keywords_reaction_enabled, - t_enable_advance_output, - t_enable_kuuki_read, - t_enable_debug_output, - t_enable_friend_chat, - t_chinese_typo_enabled, - t_error_rate, - t_min_freq, - t_tone_error_rate, - t_word_replace_rate, - t_remote_status, + t_keywords_reaction_enabled, + t_enable_advance_output, + t_enable_kuuki_read, + t_enable_debug_output, + t_enable_friend_chat, + t_chinese_typo_enabled, + t_error_rate, + t_min_freq, + t_tone_error_rate, + t_word_replace_rate, + t_remote_status, ): config_data["keywords_reaction"]["enable"] = t_keywords_reaction_enabled config_data["others"]["enable_advance_output"] = t_enable_advance_output @@ -641,9 +641,9 @@ def save_other_config( def save_group_config( - t_talk_allowed_final_result, - t_talk_frequency_down_final_result, - t_ban_user_id_final_result, + t_talk_allowed_final_result, + t_talk_frequency_down_final_result, + t_ban_user_id_final_result, ): config_data["groups"]["talk_allowed"] = t_talk_allowed_final_result config_data["groups"]["talk_frequency_down"] = t_talk_frequency_down_final_result @@ -1212,10 +1212,10 @@ with gr.Blocks(title="MaimBot配置文件编辑") as app: willing_mode = gr.Dropdown( choices=WILLING_MODE_CHOICES, value=config_data["willing"]["willing_mode"], - label="回复意愿模式" + label="回复意愿模式", ) else: - willing_mode = gr.Textbox(visible=False,value="disabled") + willing_mode = gr.Textbox(visible=False, value="disabled") with gr.Row(): model_r1_probability = gr.Slider( minimum=0, From a2c6e418436e465b4a8ed1b587beaa4a54b796c9 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=98=A5=E6=B2=B3=E6=99=B4?= Date: Fri, 21 Mar 2025 14:05:47 +0800 Subject: [PATCH 13/21] fix markdown --- docs/linux_deploy_guide_for_beginners.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/linux_deploy_guide_for_beginners.md b/docs/linux_deploy_guide_for_beginners.md index ece0a3334..1f1b0899f 100644 --- a/docs/linux_deploy_guide_for_beginners.md +++ b/docs/linux_deploy_guide_for_beginners.md @@ -320,7 +320,7 @@ sudo systemctl enable bot.service # 启动bot服务 sudo systemctl status bot.service # 检查bot服务状态 ``` -```python +```bash python bot.py # 运行麦麦 ``` From 6c3afa84c4d74b6cbcbc3e0e7b8ba56f5cf030de Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Fri, 21 Mar 2025 14:37:19 +0800 Subject: [PATCH 14/21] =?UTF-8?q?better=20=E6=9B=B4=E5=A5=BD=E7=9A=84?= =?UTF-8?q?=E8=AE=B0=E5=BF=86=E6=8A=BD=E5=8F=96=E7=AD=96=E7=95=A5=EF=BC=8C?= =?UTF-8?q?=E5=B9=B6=E4=B8=94=E7=A7=BB=E9=99=A4=E4=BA=86=E6=97=A0=E7=94=A8?= =?UTF-8?q?=E9=80=89=E9=A1=B9?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docs/installation_cute.md | 2 - docs/installation_standard.md | 2 - src/common/logger.py | 5 +- src/plugins/chat/__init__.py | 2 +- src/plugins/chat/config.py | 18 +- src/plugins/memory_system/memory.py | 79 +- .../memory_system/memory_manual_build.py | 3 +- src/plugins/memory_system/memory_test1.py | 1185 ----------------- .../memory_system/sample_distribution.py | 172 +++ src/plugins/schedule/offline_llm.py | 123 ++ .../schedule/schedule_generator copy.py | 192 +++ src/plugins/schedule/schedule_generator.py | 26 +- template.env | 3 +- template/bot_config_template.toml | 17 +- 14 files changed, 547 insertions(+), 1282 deletions(-) delete mode 100644 src/plugins/memory_system/memory_test1.py create mode 100644 src/plugins/memory_system/sample_distribution.py create mode 100644 src/plugins/schedule/offline_llm.py create mode 100644 src/plugins/schedule/schedule_generator copy.py diff --git a/docs/installation_cute.md b/docs/installation_cute.md index ca97f18e9..5eb5dfdcd 100644 --- a/docs/installation_cute.md +++ b/docs/installation_cute.md @@ -147,9 +147,7 @@ enable_check = false # 是否要检查表情包是不是合适的喵 check_prompt = "符合公序良俗" # 检查表情包的标准呢 [others] -enable_advance_output = true # 是否要显示更多的运行信息呢 enable_kuuki_read = true # 让机器人能够"察言观色"喵 -enable_debug_output = false # 是否启用调试输出喵 enable_friend_chat = false # 是否启用好友聊天喵 [groups] diff --git a/docs/installation_standard.md b/docs/installation_standard.md index dcbbf0c99..a2e60f22a 100644 --- a/docs/installation_standard.md +++ b/docs/installation_standard.md @@ -115,9 +115,7 @@ talk_frequency_down = [] # 降低回复频率的群号 ban_user_id = [] # 禁止回复的用户QQ号 [others] -enable_advance_output = true # 是否启用高级输出 enable_kuuki_read = true # 是否启用读空气功能 -enable_debug_output = false # 是否启用调试输出 enable_friend_chat = false # 是否启用好友聊天 # 模型配置 diff --git a/src/common/logger.py b/src/common/logger.py index f0b2dfe5c..2673275a5 100644 --- a/src/common/logger.py +++ b/src/common/logger.py @@ -31,9 +31,10 @@ _handler_registry: Dict[str, List[int]] = {} current_file_path = Path(__file__).resolve() LOG_ROOT = "logs" -ENABLE_ADVANCE_OUTPUT = False +ENABLE_ADVANCE_OUTPUT = os.getenv("SIMPLE_OUTPUT", "false") +print(f"ENABLE_ADVANCE_OUTPUT: {ENABLE_ADVANCE_OUTPUT}") -if ENABLE_ADVANCE_OUTPUT: +if not ENABLE_ADVANCE_OUTPUT: # 默认全局配置 DEFAULT_CONFIG = { # 日志级别配置 diff --git a/src/plugins/chat/__init__.py b/src/plugins/chat/__init__.py index a54f781a0..e73d0a230 100644 --- a/src/plugins/chat/__init__.py +++ b/src/plugins/chat/__init__.py @@ -110,7 +110,7 @@ async def build_memory_task(): """每build_memory_interval秒执行一次记忆构建""" logger.debug("[记忆构建]------------------------------------开始构建记忆--------------------------------------") start_time = time.time() - await hippocampus.operation_build_memory(chat_size=20) + await hippocampus.operation_build_memory() end_time = time.time() logger.success( f"[记忆构建]--------------------------记忆构建完成:耗时: {end_time - start_time:.2f} " diff --git a/src/plugins/chat/config.py b/src/plugins/chat/config.py index ce30b280b..d0cb18822 100644 --- a/src/plugins/chat/config.py +++ b/src/plugins/chat/config.py @@ -68,9 +68,9 @@ class BotConfig: MODEL_V3_PROBABILITY: float = 0.1 # V3模型概率 MODEL_R1_DISTILL_PROBABILITY: float = 0.1 # R1蒸馏模型概率 - enable_advance_output: bool = False # 是否启用高级输出 + # enable_advance_output: bool = False # 是否启用高级输出 enable_kuuki_read: bool = True # 是否启用读空气功能 - enable_debug_output: bool = False # 是否启用调试输出 + # enable_debug_output: bool = False # 是否启用调试输出 enable_friend_chat: bool = False # 是否启用好友聊天 mood_update_interval: float = 1.0 # 情绪更新间隔 单位秒 @@ -106,6 +106,11 @@ class BotConfig: memory_forget_time: int = 24 # 记忆遗忘时间(小时) memory_forget_percentage: float = 0.01 # 记忆遗忘比例 memory_compress_rate: float = 0.1 # 记忆压缩率 + build_memory_sample_num: int = 10 # 记忆构建采样数量 + build_memory_sample_length: int = 20 # 记忆构建采样长度 + memory_build_distribution: list = field( + default_factory=lambda: [4,2,0.6,24,8,0.4] + ) # 记忆构建分布,参数:分布1均值,标准差,权重,分布2均值,标准差,权重 memory_ban_words: list = field( default_factory=lambda: ["表情包", "图片", "回复", "聊天记录"] ) # 添加新的配置项默认值 @@ -315,6 +320,11 @@ class BotConfig: "memory_forget_percentage", config.memory_forget_percentage ) config.memory_compress_rate = memory_config.get("memory_compress_rate", config.memory_compress_rate) + if config.INNER_VERSION in SpecifierSet(">=0.0.11"): + config.memory_build_distribution = memory_config.get("memory_build_distribution", config.memory_build_distribution) + config.build_memory_sample_num = memory_config.get("build_memory_sample_num", config.build_memory_sample_num) + config.build_memory_sample_length = memory_config.get("build_memory_sample_length", config.build_memory_sample_length) + def remote(parent: dict): remote_config = parent["remote"] @@ -351,10 +361,10 @@ class BotConfig: def others(parent: dict): others_config = parent["others"] - config.enable_advance_output = others_config.get("enable_advance_output", config.enable_advance_output) + # config.enable_advance_output = others_config.get("enable_advance_output", config.enable_advance_output) config.enable_kuuki_read = others_config.get("enable_kuuki_read", config.enable_kuuki_read) if config.INNER_VERSION in SpecifierSet(">=0.0.7"): - config.enable_debug_output = others_config.get("enable_debug_output", config.enable_debug_output) + # config.enable_debug_output = others_config.get("enable_debug_output", config.enable_debug_output) config.enable_friend_chat = others_config.get("enable_friend_chat", config.enable_friend_chat) # 版本表达式:>=1.0.0,<2.0.0 diff --git a/src/plugins/memory_system/memory.py b/src/plugins/memory_system/memory.py index 07a7fb2ee..b55dcf7b3 100644 --- a/src/plugins/memory_system/memory.py +++ b/src/plugins/memory_system/memory.py @@ -18,6 +18,7 @@ from ..chat.utils import ( ) from ..models.utils_model import LLM_request from src.common.logger import get_module_logger, LogConfig, MEMORY_STYLE_CONFIG +from src.plugins.memory_system.sample_distribution import MemoryBuildScheduler # 定义日志配置 memory_config = LogConfig( @@ -195,19 +196,9 @@ class Hippocampus: return hash(f"{nodes[0]}:{nodes[1]}") def random_get_msg_snippet(self, target_timestamp: float, chat_size: int, max_memorized_time_per_msg: int) -> list: - """随机抽取一段时间内的消息片段 - Args: - - target_timestamp: 目标时间戳 - - chat_size: 抽取的消息数量 - - max_memorized_time_per_msg: 每条消息的最大记忆次数 - - Returns: - - list: 抽取出的消息记录列表 - - """ try_count = 0 - # 最多尝试三次抽取 - while try_count < 3: + # 最多尝试2次抽取 + while try_count < 2: messages = get_closest_chat_from_db(length=chat_size, timestamp=target_timestamp) if messages: # 检查messages是否均没有达到记忆次数限制 @@ -224,54 +215,37 @@ class Hippocampus: ) return messages try_count += 1 - # 三次尝试均失败 return None - def get_memory_sample(self, chat_size=20, time_frequency=None): - """获取记忆样本 - - Returns: - list: 消息记录列表,每个元素是一个消息记录字典列表 - """ + def get_memory_sample(self): # 硬编码:每条消息最大记忆次数 # 如有需求可写入global_config - if time_frequency is None: - time_frequency = {"near": 2, "mid": 4, "far": 3} max_memorized_time_per_msg = 3 - current_timestamp = datetime.datetime.now().timestamp() + # 创建双峰分布的记忆调度器 + scheduler = MemoryBuildScheduler( + n_hours1=global_config.memory_build_distribution[0], # 第一个分布均值(4小时前) + std_hours1=global_config.memory_build_distribution[1], # 第一个分布标准差 + weight1=global_config.memory_build_distribution[2], # 第一个分布权重 60% + n_hours2=global_config.memory_build_distribution[3], # 第二个分布均值(24小时前) + std_hours2=global_config.memory_build_distribution[4], # 第二个分布标准差 + weight2=global_config.memory_build_distribution[5], # 第二个分布权重 40% + total_samples=global_config.build_memory_sample_num # 总共生成10个时间点 + ) + + # 生成时间戳数组 + timestamps = scheduler.get_timestamp_array() + logger.debug(f"生成的时间戳数组: {timestamps}") + chat_samples = [] - - # 短期:1h 中期:4h 长期:24h - logger.debug("正在抽取短期消息样本") - for i in range(time_frequency.get("near")): - random_time = current_timestamp - random.randint(1, 3600) - messages = self.random_get_msg_snippet(random_time, chat_size, max_memorized_time_per_msg) + for timestamp in timestamps: + messages = self.random_get_msg_snippet(timestamp, global_config.build_memory_sample_length, max_memorized_time_per_msg) if messages: - logger.debug(f"成功抽取短期消息样本{len(messages)}条") + time_diff = (datetime.datetime.now().timestamp() - timestamp) / 3600 + logger.debug(f"成功抽取 {time_diff:.1f} 小时前的消息样本,共{len(messages)}条") chat_samples.append(messages) else: - logger.warning(f"第{i}次短期消息样本抽取失败") - - logger.debug("正在抽取中期消息样本") - for i in range(time_frequency.get("mid")): - random_time = current_timestamp - random.randint(3600, 3600 * 4) - messages = self.random_get_msg_snippet(random_time, chat_size, max_memorized_time_per_msg) - if messages: - logger.debug(f"成功抽取中期消息样本{len(messages)}条") - chat_samples.append(messages) - else: - logger.warning(f"第{i}次中期消息样本抽取失败") - - logger.debug("正在抽取长期消息样本") - for i in range(time_frequency.get("far")): - random_time = current_timestamp - random.randint(3600 * 4, 3600 * 24) - messages = self.random_get_msg_snippet(random_time, chat_size, max_memorized_time_per_msg) - if messages: - logger.debug(f"成功抽取长期消息样本{len(messages)}条") - chat_samples.append(messages) - else: - logger.warning(f"第{i}次长期消息样本抽取失败") + logger.warning(f"时间戳 {timestamp} 的消息样本抽取失败") return chat_samples @@ -372,9 +346,8 @@ class Hippocampus: ) return topic_num - async def operation_build_memory(self, chat_size=20): - time_frequency = {"near": 1, "mid": 4, "far": 4} - memory_samples = self.get_memory_sample(chat_size, time_frequency) + async def operation_build_memory(self): + memory_samples = self.get_memory_sample() for i, messages in enumerate(memory_samples, 1): all_topics = [] diff --git a/src/plugins/memory_system/memory_manual_build.py b/src/plugins/memory_system/memory_manual_build.py index 0bf276ddd..4d6596e9f 100644 --- a/src/plugins/memory_system/memory_manual_build.py +++ b/src/plugins/memory_system/memory_manual_build.py @@ -7,11 +7,9 @@ import sys import time from collections import Counter from pathlib import Path - import matplotlib.pyplot as plt import networkx as nx from dotenv import load_dotenv -from src.common.logger import get_module_logger import jieba # from chat.config import global_config @@ -19,6 +17,7 @@ import jieba root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) +from src.common.logger import get_module_logger from src.common.database import db # noqa E402 from src.plugins.memory_system.offline_llm import LLMModel # noqa E402 diff --git a/src/plugins/memory_system/memory_test1.py b/src/plugins/memory_system/memory_test1.py deleted file mode 100644 index df4f892d0..000000000 --- a/src/plugins/memory_system/memory_test1.py +++ /dev/null @@ -1,1185 +0,0 @@ -# -*- coding: utf-8 -*- -import datetime -import math -import random -import sys -import time -from collections import Counter -from pathlib import Path - -import matplotlib.pyplot as plt -import networkx as nx -from dotenv import load_dotenv -from src.common.logger import get_module_logger -import jieba - -logger = get_module_logger("mem_test") - -""" -该理论认为,当两个或多个事物在形态上具有相似性时, -它们在记忆中会形成关联。 -例如,梨和苹果在形状和都是水果这一属性上有相似性, -所以当我们看到梨时,很容易通过形态学联想记忆联想到苹果。 -这种相似性联想有助于我们对新事物进行分类和理解, -当遇到一个新的类似水果时, -我们可以通过与已有的水果记忆进行相似性匹配, -来推测它的一些特征。 - - - -时空关联性联想: -除了相似性联想,MAM 还强调时空关联性联想。 -如果两个事物在时间或空间上经常同时出现,它们也会在记忆中形成关联。 -比如,每次在公园里看到花的时候,都能听到鸟儿的叫声, -那么花和鸟儿叫声的形态特征(花的视觉形态和鸟叫的听觉形态)就会在记忆中形成关联, -以后听到鸟叫可能就会联想到公园里的花。 - -""" - -# from chat.config import global_config -sys.path.append("C:/GitHub/MaiMBot") # 添加项目根目录到 Python 路径 -from src.common.database import db # noqa E402 -from src.plugins.memory_system.offline_llm import LLMModel # noqa E402 - -# 获取当前文件的目录 -current_dir = Path(__file__).resolve().parent -# 获取项目根目录(上三层目录) -project_root = current_dir.parent.parent.parent -# env.dev文件路径 -env_path = project_root / ".env.dev" - -# 加载环境变量 -if env_path.exists(): - logger.info(f"从 {env_path} 加载环境变量") - load_dotenv(env_path) -else: - logger.warning(f"未找到环境变量文件: {env_path}") - logger.info("将使用默认配置") - - -def calculate_information_content(text): - """计算文本的信息量(熵)""" - char_count = Counter(text) - total_chars = len(text) - - entropy = 0 - for count in char_count.values(): - probability = count / total_chars - entropy -= probability * math.log2(probability) - - return entropy - - -def get_closest_chat_from_db(length: int, timestamp: str): - """从数据库中获取最接近指定时间戳的聊天记录,并记录读取次数 - - Returns: - list: 消息记录字典列表,每个字典包含消息内容和时间信息 - """ - chat_records = [] - closest_record = db.messages.find_one({"time": {"$lte": timestamp}}, sort=[("time", -1)]) - - if closest_record and closest_record.get("memorized", 0) < 4: - closest_time = closest_record["time"] - group_id = closest_record["group_id"] - # 获取该时间戳之后的length条消息,且groupid相同 - records = list( - db.messages.find({"time": {"$gt": closest_time}, "group_id": group_id}).sort("time", 1).limit(length) - ) - - # 更新每条消息的memorized属性 - for record in records: - current_memorized = record.get("memorized", 0) - if current_memorized > 3: - print("消息已读取3次,跳过") - return "" - - # 更新memorized值 - db.messages.update_one({"_id": record["_id"]}, {"$set": {"memorized": current_memorized + 1}}) - - # 添加到记录列表中 - chat_records.append( - {"text": record["detailed_plain_text"], "time": record["time"], "group_id": record["group_id"]} - ) - - return chat_records - - -class Memory_cortex: - def __init__(self, memory_graph: "Memory_graph"): - self.memory_graph = memory_graph - - def sync_memory_from_db(self): - """ - 从数据库同步数据到内存中的图结构 - 将清空当前内存中的图,并从数据库重新加载所有节点和边 - """ - # 清空当前图 - self.memory_graph.G.clear() - - # 获取当前时间作为默认时间 - default_time = datetime.datetime.now().timestamp() - - # 从数据库加载所有节点 - nodes = db.graph_data.nodes.find() - for node in nodes: - concept = node["concept"] - memory_items = node.get("memory_items", []) - # 确保memory_items是列表 - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - - # 获取时间属性,如果不存在则使用默认时间 - created_time = node.get("created_time") - last_modified = node.get("last_modified") - - # 如果时间属性不存在,则更新数据库 - if created_time is None or last_modified is None: - created_time = default_time - last_modified = default_time - # 更新数据库中的节点 - db.graph_data.nodes.update_one( - {"concept": concept}, {"$set": {"created_time": created_time, "last_modified": last_modified}} - ) - logger.info(f"为节点 {concept} 添加默认时间属性") - - # 添加节点到图中,包含时间属性 - self.memory_graph.G.add_node( - concept, memory_items=memory_items, created_time=created_time, last_modified=last_modified - ) - - # 从数据库加载所有边 - edges = db.graph_data.edges.find() - for edge in edges: - source = edge["source"] - target = edge["target"] - - # 只有当源节点和目标节点都存在时才添加边 - if source in self.memory_graph.G and target in self.memory_graph.G: - # 获取时间属性,如果不存在则使用默认时间 - created_time = edge.get("created_time") - last_modified = edge.get("last_modified") - - # 如果时间属性不存在,则更新数据库 - if created_time is None or last_modified is None: - created_time = default_time - last_modified = default_time - # 更新数据库中的边 - db.graph_data.edges.update_one( - {"source": source, "target": target}, - {"$set": {"created_time": created_time, "last_modified": last_modified}}, - ) - logger.info(f"为边 {source} - {target} 添加默认时间属性") - - self.memory_graph.G.add_edge( - source, - target, - strength=edge.get("strength", 1), - created_time=created_time, - last_modified=last_modified, - ) - - logger.success("从数据库同步记忆图谱完成") - - def calculate_node_hash(self, concept, memory_items): - """ - 计算节点的特征值 - """ - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - # 将记忆项排序以确保相同内容生成相同的哈希值 - sorted_items = sorted(memory_items) - # 组合概念和记忆项生成特征值 - content = f"{concept}:{'|'.join(sorted_items)}" - return hash(content) - - def calculate_edge_hash(self, source, target): - """ - 计算边的特征值 - """ - # 对源节点和目标节点排序以确保相同的边生成相同的哈希值 - nodes = sorted([source, target]) - return hash(f"{nodes[0]}:{nodes[1]}") - - def sync_memory_to_db(self): - """ - 检查并同步内存中的图结构与数据库 - 使用特征值(哈希值)快速判断是否需要更新 - """ - current_time = datetime.datetime.now().timestamp() - - # 获取数据库中所有节点和内存中所有节点 - db_nodes = list(db.graph_data.nodes.find()) - memory_nodes = list(self.memory_graph.G.nodes(data=True)) - - # 转换数据库节点为字典格式,方便查找 - db_nodes_dict = {node["concept"]: node for node in db_nodes} - - # 检查并更新节点 - for concept, data in memory_nodes: - memory_items = data.get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - - # 计算内存中节点的特征值 - memory_hash = self.calculate_node_hash(concept, memory_items) - - if concept not in db_nodes_dict: - # 数据库中缺少的节点,添加 - node_data = { - "concept": concept, - "memory_items": memory_items, - "hash": memory_hash, - "created_time": data.get("created_time", current_time), - "last_modified": data.get("last_modified", current_time), - } - db.graph_data.nodes.insert_one(node_data) - else: - # 获取数据库中节点的特征值 - db_node = db_nodes_dict[concept] - db_hash = db_node.get("hash", None) - - # 如果特征值不同,则更新节点 - if db_hash != memory_hash: - db.graph_data.nodes.update_one( - {"concept": concept}, - {"$set": {"memory_items": memory_items, "hash": memory_hash, "last_modified": current_time}}, - ) - - # 检查并删除数据库中多余的节点 - memory_concepts = set(node[0] for node in memory_nodes) - for db_node in db_nodes: - if db_node["concept"] not in memory_concepts: - db.graph_data.nodes.delete_one({"concept": db_node["concept"]}) - - # 处理边的信息 - db_edges = list(db.graph_data.edges.find()) - memory_edges = list(self.memory_graph.G.edges(data=True)) - - # 创建边的哈希值字典 - db_edge_dict = {} - for edge in db_edges: - edge_hash = self.calculate_edge_hash(edge["source"], edge["target"]) - db_edge_dict[(edge["source"], edge["target"])] = {"hash": edge_hash, "strength": edge.get("strength", 1)} - - # 检查并更新边 - for source, target, data in memory_edges: - edge_hash = self.calculate_edge_hash(source, target) - edge_key = (source, target) - strength = data.get("strength", 1) - - if edge_key not in db_edge_dict: - # 添加新边 - edge_data = { - "source": source, - "target": target, - "strength": strength, - "hash": edge_hash, - "created_time": data.get("created_time", current_time), - "last_modified": data.get("last_modified", current_time), - } - db.graph_data.edges.insert_one(edge_data) - else: - # 检查边的特征值是否变化 - if db_edge_dict[edge_key]["hash"] != edge_hash: - db.graph_data.edges.update_one( - {"source": source, "target": target}, - {"$set": {"hash": edge_hash, "strength": strength, "last_modified": current_time}}, - ) - - # 删除多余的边 - memory_edge_set = set((source, target) for source, target, _ in memory_edges) - for edge_key in db_edge_dict: - if edge_key not in memory_edge_set: - source, target = edge_key - db.graph_data.edges.delete_one({"source": source, "target": target}) - - logger.success("完成记忆图谱与数据库的差异同步") - - def remove_node_from_db(self, topic): - """ - 从数据库中删除指定节点及其相关的边 - - Args: - topic: 要删除的节点概念 - """ - # 删除节点 - db.graph_data.nodes.delete_one({"concept": topic}) - # 删除所有涉及该节点的边 - db.graph_data.edges.delete_many({"$or": [{"source": topic}, {"target": topic}]}) - - -class Memory_graph: - def __init__(self): - self.G = nx.Graph() # 使用 networkx 的图结构 - - def connect_dot(self, concept1, concept2): - # 避免自连接 - if concept1 == concept2: - return - - current_time = datetime.datetime.now().timestamp() - - # 如果边已存在,增加 strength - if self.G.has_edge(concept1, concept2): - self.G[concept1][concept2]["strength"] = self.G[concept1][concept2].get("strength", 1) + 1 - # 更新最后修改时间 - self.G[concept1][concept2]["last_modified"] = current_time - else: - # 如果是新边,初始化 strength 为 1 - self.G.add_edge(concept1, concept2, strength=1, created_time=current_time, last_modified=current_time) - - def add_dot(self, concept, memory): - current_time = datetime.datetime.now().timestamp() - - if concept in self.G: - # 如果节点已存在,将新记忆添加到现有列表中 - if "memory_items" in self.G.nodes[concept]: - if not isinstance(self.G.nodes[concept]["memory_items"], list): - # 如果当前不是列表,将其转换为列表 - self.G.nodes[concept]["memory_items"] = [self.G.nodes[concept]["memory_items"]] - self.G.nodes[concept]["memory_items"].append(memory) - # 更新最后修改时间 - self.G.nodes[concept]["last_modified"] = current_time - else: - self.G.nodes[concept]["memory_items"] = [memory] - self.G.nodes[concept]["last_modified"] = current_time - else: - # 如果是新节点,创建新的记忆列表 - self.G.add_node(concept, memory_items=[memory], created_time=current_time, last_modified=current_time) - - def get_dot(self, concept): - # 检查节点是否存在于图中 - if concept in self.G: - # 从图中获取节点数据 - node_data = self.G.nodes[concept] - return concept, node_data - return None - - def get_related_item(self, topic, depth=1): - if topic not in self.G: - return [], [] - - first_layer_items = [] - second_layer_items = [] - - # 获取相邻节点 - neighbors = list(self.G.neighbors(topic)) - - # 获取当前节点的记忆项 - node_data = self.get_dot(topic) - if node_data: - concept, data = node_data - if "memory_items" in data: - memory_items = data["memory_items"] - if isinstance(memory_items, list): - first_layer_items.extend(memory_items) - else: - first_layer_items.append(memory_items) - - # 只在depth=2时获取第二层记忆 - if depth >= 2: - # 获取相邻节点的记忆项 - for neighbor in neighbors: - node_data = self.get_dot(neighbor) - if node_data: - concept, data = node_data - if "memory_items" in data: - memory_items = data["memory_items"] - if isinstance(memory_items, list): - second_layer_items.extend(memory_items) - else: - second_layer_items.append(memory_items) - - return first_layer_items, second_layer_items - - @property - def dots(self): - # 返回所有节点对应的 Memory_dot 对象 - return [self.get_dot(node) for node in self.G.nodes()] - - -# 海马体 -class Hippocampus: - def __init__(self, memory_graph: Memory_graph): - self.memory_graph = memory_graph - self.memory_cortex = Memory_cortex(memory_graph) - self.llm_model = LLMModel() - self.llm_model_small = LLMModel(model_name="deepseek-ai/DeepSeek-V2.5") - self.llm_model_get_topic = LLMModel(model_name="Pro/Qwen/Qwen2.5-7B-Instruct") - self.llm_model_summary = LLMModel(model_name="Qwen/Qwen2.5-32B-Instruct") - - def get_memory_sample(self, chat_size=20, time_frequency=None): - """获取记忆样本 - - Returns: - list: 消息记录列表,每个元素是一个消息记录字典列表 - """ - if time_frequency is None: - time_frequency = {"near": 2, "mid": 4, "far": 3} - current_timestamp = datetime.datetime.now().timestamp() - chat_samples = [] - - # 短期:1h 中期:4h 长期:24h - for _ in range(time_frequency.get("near")): - random_time = current_timestamp - random.randint(1, 3600 * 4) - messages = get_closest_chat_from_db(length=chat_size, timestamp=random_time) - if messages: - chat_samples.append(messages) - - for _ in range(time_frequency.get("mid")): - random_time = current_timestamp - random.randint(3600 * 4, 3600 * 24) - messages = get_closest_chat_from_db(length=chat_size, timestamp=random_time) - if messages: - chat_samples.append(messages) - - for _ in range(time_frequency.get("far")): - random_time = current_timestamp - random.randint(3600 * 24, 3600 * 24 * 7) - messages = get_closest_chat_from_db(length=chat_size, timestamp=random_time) - if messages: - chat_samples.append(messages) - - return chat_samples - - def calculate_topic_num(self, text, compress_rate): - """计算文本的话题数量""" - information_content = calculate_information_content(text) - topic_by_length = text.count("\n") * compress_rate - topic_by_information_content = max(1, min(5, int((information_content - 3) * 2))) - topic_num = int((topic_by_length + topic_by_information_content) / 2) - print( - f"topic_by_length: {topic_by_length}, topic_by_information_content: {topic_by_information_content}, " - f"topic_num: {topic_num}" - ) - return topic_num - - async def memory_compress(self, messages: list, compress_rate=0.1): - """压缩消息记录为记忆 - - Args: - messages: 消息记录字典列表,每个字典包含text和time字段 - compress_rate: 压缩率 - - Returns: - tuple: (压缩记忆集合, 相似主题字典) - - 压缩记忆集合: set of (话题, 记忆) 元组 - - 相似主题字典: dict of {话题: [(相似主题, 相似度), ...]} - """ - if not messages: - return set(), {} - - # 合并消息文本,同时保留时间信息 - input_text = "" - time_info = "" - # 计算最早和最晚时间 - earliest_time = min(msg["time"] for msg in messages) - latest_time = max(msg["time"] for msg in messages) - - earliest_dt = datetime.datetime.fromtimestamp(earliest_time) - latest_dt = datetime.datetime.fromtimestamp(latest_time) - - # 如果是同一年 - if earliest_dt.year == latest_dt.year: - earliest_str = earliest_dt.strftime("%m-%d %H:%M:%S") - latest_str = latest_dt.strftime("%m-%d %H:%M:%S") - time_info += f"是在{earliest_dt.year}年,{earliest_str} 到 {latest_str} 的对话:\n" - else: - earliest_str = earliest_dt.strftime("%Y-%m-%d %H:%M:%S") - latest_str = latest_dt.strftime("%Y-%m-%d %H:%M:%S") - time_info += f"是从 {earliest_str} 到 {latest_str} 的对话:\n" - - for msg in messages: - input_text += f"{msg['text']}\n" - - print(input_text) - - topic_num = self.calculate_topic_num(input_text, compress_rate) - topics_response = self.llm_model_get_topic.generate_response(self.find_topic_llm(input_text, topic_num)) - - # 过滤topics - filter_keywords = ["表情包", "图片", "回复", "聊天记录"] - topics = [ - topic.strip() - for topic in topics_response[0].replace(",", ",").replace("、", ",").replace(" ", ",").split(",") - if topic.strip() - ] - filtered_topics = [topic for topic in topics if not any(keyword in topic for keyword in filter_keywords)] - - print(f"过滤后话题: {filtered_topics}") - - # 为每个话题查找相似的已存在主题 - print("\n检查相似主题:") - similar_topics_dict = {} # 存储每个话题的相似主题列表 - - for topic in filtered_topics: - # 获取所有现有节点 - existing_topics = list(self.memory_graph.G.nodes()) - similar_topics = [] - - # 对每个现有节点计算相似度 - for existing_topic in existing_topics: - # 使用jieba分词并计算余弦相似度 - topic_words = set(jieba.cut(topic)) - existing_words = set(jieba.cut(existing_topic)) - - # 计算词向量 - all_words = topic_words | existing_words - v1 = [1 if word in topic_words else 0 for word in all_words] - v2 = [1 if word in existing_words else 0 for word in all_words] - - # 计算余弦相似度 - similarity = cosine_similarity(v1, v2) - - # 如果相似度超过阈值,添加到结果中 - if similarity >= 0.6: # 设置相似度阈值 - similar_topics.append((existing_topic, similarity)) - - # 按相似度降序排序 - similar_topics.sort(key=lambda x: x[1], reverse=True) - # 只保留前5个最相似的主题 - similar_topics = similar_topics[:5] - - # 存储到字典中 - similar_topics_dict[topic] = similar_topics - - # 输出结果 - if similar_topics: - print(f"\n主题「{topic}」的相似主题:") - for similar_topic, score in similar_topics: - print(f"- {similar_topic} (相似度: {score:.3f})") - else: - print(f"\n主题「{topic}」没有找到相似主题") - - # 创建所有话题的请求任务 - tasks = [] - for topic in filtered_topics: - topic_what_prompt = self.topic_what(input_text, topic, time_info) - # 创建异步任务 - task = self.llm_model_small.generate_response_async(topic_what_prompt) - tasks.append((topic.strip(), task)) - - # 等待所有任务完成 - compressed_memory = set() - for topic, task in tasks: - response = await task - if response: - compressed_memory.add((topic, response[0])) - - return compressed_memory, similar_topics_dict - - async def operation_build_memory(self, chat_size=12): - # 最近消息获取频率 - time_frequency = {"near": 3, "mid": 8, "far": 5} - memory_samples = self.get_memory_sample(chat_size, time_frequency) - - all_topics = [] # 用于存储所有话题 - - for i, messages in enumerate(memory_samples, 1): - # 加载进度可视化 - all_topics = [] - progress = (i / len(memory_samples)) * 100 - bar_length = 30 - filled_length = int(bar_length * i // len(memory_samples)) - bar = "█" * filled_length + "-" * (bar_length - filled_length) - print(f"\n进度: [{bar}] {progress:.1f}% ({i}/{len(memory_samples)})") - - # 生成压缩后记忆 - compress_rate = 0.1 - compressed_memory, similar_topics_dict = await self.memory_compress(messages, compress_rate) - print( - f"\033[1;33m压缩后记忆数量\033[0m: {len(compressed_memory)},似曾相识的话题: {len(similar_topics_dict)}" - ) - - # 将记忆加入到图谱中 - for topic, memory in compressed_memory: - print(f"\033[1;32m添加节点\033[0m: {topic}") - self.memory_graph.add_dot(topic, memory) - all_topics.append(topic) - - # 连接相似的已存在主题 - if topic in similar_topics_dict: - similar_topics = similar_topics_dict[topic] - for similar_topic, similarity in similar_topics: - # 避免自连接 - if topic != similar_topic: - # 根据相似度设置连接强度 - strength = int(similarity * 10) # 将0.3-1.0的相似度映射到3-10的强度 - print(f"\033[1;36m连接相似节点\033[0m: {topic} 和 {similar_topic} (强度: {strength})") - # 使用相似度作为初始连接强度 - self.memory_graph.G.add_edge(topic, similar_topic, strength=strength) - - # 连接同批次的相关话题 - for i in range(len(all_topics)): - for j in range(i + 1, len(all_topics)): - print(f"\033[1;32m连接同批次节点\033[0m: {all_topics[i]} 和 {all_topics[j]}") - self.memory_graph.connect_dot(all_topics[i], all_topics[j]) - - self.memory_cortex.sync_memory_to_db() - - def forget_connection(self, source, target): - """ - 检查并可能遗忘一个连接 - - Args: - source: 连接的源节点 - target: 连接的目标节点 - - Returns: - tuple: (是否有变化, 变化类型, 变化详情) - 变化类型: 0-无变化, 1-强度减少, 2-连接移除 - """ - current_time = datetime.datetime.now().timestamp() - # 获取边的属性 - edge_data = self.memory_graph.G[source][target] - last_modified = edge_data.get("last_modified", current_time) - - # 如果连接超过7天未更新 - if current_time - last_modified > 6000: # test - # 获取当前强度 - current_strength = edge_data.get("strength", 1) - # 减少连接强度 - new_strength = current_strength - 1 - edge_data["strength"] = new_strength - edge_data["last_modified"] = current_time - - # 如果强度降为0,移除连接 - if new_strength <= 0: - self.memory_graph.G.remove_edge(source, target) - return True, 2, f"移除连接: {source} - {target} (强度降至0)" - else: - return True, 1, f"减弱连接: {source} - {target} (强度: {current_strength} -> {new_strength})" - - return False, 0, "" - - def forget_topic(self, topic): - """ - 检查并可能遗忘一个话题的记忆 - - Args: - topic: 要检查的话题 - - Returns: - tuple: (是否有变化, 变化类型, 变化详情) - 变化类型: 0-无变化, 1-记忆减少, 2-节点移除 - """ - current_time = datetime.datetime.now().timestamp() - # 获取节点的最后修改时间 - node_data = self.memory_graph.G.nodes[topic] - last_modified = node_data.get("last_modified", current_time) - - # 如果话题超过7天未更新 - if current_time - last_modified > 3000: # test - memory_items = node_data.get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - - if memory_items: - # 获取当前记忆数量 - current_count = len(memory_items) - # 随机选择一条记忆删除 - removed_item = random.choice(memory_items) - memory_items.remove(removed_item) - - if memory_items: - # 更新节点的记忆项和最后修改时间 - self.memory_graph.G.nodes[topic]["memory_items"] = memory_items - self.memory_graph.G.nodes[topic]["last_modified"] = current_time - return ( - True, - 1, - f"减少记忆: {topic} (记忆数量: {current_count} -> " - f"{len(memory_items)})\n被移除的记忆: {removed_item}", - ) - else: - # 如果没有记忆了,删除节点及其所有连接 - self.memory_graph.G.remove_node(topic) - return True, 2, f"移除节点: {topic} (无剩余记忆)\n最后一条记忆: {removed_item}" - - return False, 0, "" - - async def operation_forget_topic(self, percentage=0.1): - """ - 随机选择图中一定比例的节点和边进行检查,根据时间条件决定是否遗忘 - - Args: - percentage: 要检查的节点和边的比例,默认为0.1(10%) - """ - # 获取所有节点和边 - all_nodes = list(self.memory_graph.G.nodes()) - all_edges = list(self.memory_graph.G.edges()) - - # 计算要检查的数量 - check_nodes_count = max(1, int(len(all_nodes) * percentage)) - check_edges_count = max(1, int(len(all_edges) * percentage)) - - # 随机选择要检查的节点和边 - nodes_to_check = random.sample(all_nodes, check_nodes_count) - edges_to_check = random.sample(all_edges, check_edges_count) - - # 用于统计不同类型的变化 - edge_changes = {"weakened": 0, "removed": 0} - node_changes = {"reduced": 0, "removed": 0} - - # 检查并遗忘连接 - print("\n开始检查连接...") - for source, target in edges_to_check: - changed, change_type, details = self.forget_connection(source, target) - if changed: - if change_type == 1: - edge_changes["weakened"] += 1 - logger.info(f"\033[1;34m[连接减弱]\033[0m {details}") - elif change_type == 2: - edge_changes["removed"] += 1 - logger.info(f"\033[1;31m[连接移除]\033[0m {details}") - - # 检查并遗忘话题 - print("\n开始检查节点...") - for node in nodes_to_check: - changed, change_type, details = self.forget_topic(node) - if changed: - if change_type == 1: - node_changes["reduced"] += 1 - logger.info(f"\033[1;33m[记忆减少]\033[0m {details}") - elif change_type == 2: - node_changes["removed"] += 1 - logger.info(f"\033[1;31m[节点移除]\033[0m {details}") - - # 同步到数据库 - if any(count > 0 for count in edge_changes.values()) or any(count > 0 for count in node_changes.values()): - self.memory_cortex.sync_memory_to_db() - print("\n遗忘操作统计:") - print(f"连接变化: {edge_changes['weakened']} 个减弱, {edge_changes['removed']} 个移除") - print(f"节点变化: {node_changes['reduced']} 个减少记忆, {node_changes['removed']} 个移除") - else: - print("\n本次检查没有节点或连接满足遗忘条件") - - async def merge_memory(self, topic): - """ - 对指定话题的记忆进行合并压缩 - - Args: - topic: 要合并的话题节点 - """ - # 获取节点的记忆项 - memory_items = self.memory_graph.G.nodes[topic].get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - - # 如果记忆项不足,直接返回 - if len(memory_items) < 10: - return - - # 随机选择10条记忆 - selected_memories = random.sample(memory_items, 10) - - # 拼接成文本 - merged_text = "\n".join(selected_memories) - print(f"\n[合并记忆] 话题: {topic}") - print(f"选择的记忆:\n{merged_text}") - - # 使用memory_compress生成新的压缩记忆 - compressed_memories, _ = await self.memory_compress(selected_memories, 0.1) - - # 从原记忆列表中移除被选中的记忆 - for memory in selected_memories: - memory_items.remove(memory) - - # 添加新的压缩记忆 - for _, compressed_memory in compressed_memories: - memory_items.append(compressed_memory) - print(f"添加压缩记忆: {compressed_memory}") - - # 更新节点的记忆项 - self.memory_graph.G.nodes[topic]["memory_items"] = memory_items - print(f"完成记忆合并,当前记忆数量: {len(memory_items)}") - - async def operation_merge_memory(self, percentage=0.1): - """ - 随机检查一定比例的节点,对内容数量超过100的节点进行记忆合并 - - Args: - percentage: 要检查的节点比例,默认为0.1(10%) - """ - # 获取所有节点 - all_nodes = list(self.memory_graph.G.nodes()) - # 计算要检查的节点数量 - check_count = max(1, int(len(all_nodes) * percentage)) - # 随机选择节点 - nodes_to_check = random.sample(all_nodes, check_count) - - merged_nodes = [] - for node in nodes_to_check: - # 获取节点的内容条数 - memory_items = self.memory_graph.G.nodes[node].get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - content_count = len(memory_items) - - # 如果内容数量超过100,进行合并 - if content_count > 100: - print(f"\n检查节点: {node}, 当前记忆数量: {content_count}") - await self.merge_memory(node) - merged_nodes.append(node) - - # 同步到数据库 - if merged_nodes: - self.memory_cortex.sync_memory_to_db() - print(f"\n完成记忆合并操作,共处理 {len(merged_nodes)} 个节点") - else: - print("\n本次检查没有需要合并的节点") - - async def _identify_topics(self, text: str) -> list: - """从文本中识别可能的主题""" - topics_response = self.llm_model_get_topic.generate_response(self.find_topic_llm(text, 5)) - topics = [ - topic.strip() - for topic in topics_response[0].replace(",", ",").replace("、", ",").replace(" ", ",").split(",") - if topic.strip() - ] - return topics - - def _find_similar_topics(self, topics: list, similarity_threshold: float = 0.4, debug_info: str = "") -> list: - """查找与给定主题相似的记忆主题""" - all_memory_topics = list(self.memory_graph.G.nodes()) - all_similar_topics = [] - - for topic in topics: - if debug_info: - pass - - topic_vector = text_to_vector(topic) - - for memory_topic in all_memory_topics: - memory_vector = text_to_vector(memory_topic) - all_words = set(topic_vector.keys()) | set(memory_vector.keys()) - v1 = [topic_vector.get(word, 0) for word in all_words] - v2 = [memory_vector.get(word, 0) for word in all_words] - similarity = cosine_similarity(v1, v2) - - if similarity >= similarity_threshold: - all_similar_topics.append((memory_topic, similarity)) - - return all_similar_topics - - def _get_top_topics(self, similar_topics: list, max_topics: int = 5) -> list: - """获取相似度最高的主题""" - seen_topics = set() - top_topics = [] - - for topic, score in sorted(similar_topics, key=lambda x: x[1], reverse=True): - if topic not in seen_topics and len(top_topics) < max_topics: - seen_topics.add(topic) - top_topics.append((topic, score)) - - return top_topics - - async def memory_activate_value(self, text: str, max_topics: int = 5, similarity_threshold: float = 0.3) -> int: - """计算输入文本对记忆的激活程度""" - logger.info(f"[记忆激活]识别主题: {await self._identify_topics(text)}") - - identified_topics = await self._identify_topics(text) - if not identified_topics: - return 0 - - all_similar_topics = self._find_similar_topics( - identified_topics, similarity_threshold=similarity_threshold, debug_info="记忆激活" - ) - - if not all_similar_topics: - return 0 - - top_topics = self._get_top_topics(all_similar_topics, max_topics) - - if len(top_topics) == 1: - topic, score = top_topics[0] - memory_items = self.memory_graph.G.nodes[topic].get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - content_count = len(memory_items) - penalty = 1.0 / (1 + math.log(content_count + 1)) - - activation = int(score * 50 * penalty) - print( - f"\033[1;32m[记忆激活]\033[0m 单主题「{topic}」- 相似度: {score:.3f}, 内容数: {content_count}, " - f"激活值: {activation}" - ) - return activation - - matched_topics = set() - topic_similarities = {} - - for memory_topic, _similarity in top_topics: - memory_items = self.memory_graph.G.nodes[memory_topic].get("memory_items", []) - if not isinstance(memory_items, list): - memory_items = [memory_items] if memory_items else [] - content_count = len(memory_items) - penalty = 1.0 / (1 + math.log(content_count + 1)) - - for input_topic in identified_topics: - topic_vector = text_to_vector(input_topic) - memory_vector = text_to_vector(memory_topic) - all_words = set(topic_vector.keys()) | set(memory_vector.keys()) - v1 = [topic_vector.get(word, 0) for word in all_words] - v2 = [memory_vector.get(word, 0) for word in all_words] - sim = cosine_similarity(v1, v2) - if sim >= similarity_threshold: - matched_topics.add(input_topic) - adjusted_sim = sim * penalty - topic_similarities[input_topic] = max(topic_similarities.get(input_topic, 0), adjusted_sim) - print( - f"\033[1;32m[记忆激活]\033[0m 主题「{input_topic}」-> " - f"「{memory_topic}」(内容数: {content_count}, " - f"相似度: {adjusted_sim:.3f})" - ) - - topic_match = len(matched_topics) / len(identified_topics) - average_similarities = sum(topic_similarities.values()) / len(topic_similarities) if topic_similarities else 0 - - activation = int((topic_match + average_similarities) / 2 * 100) - print( - f"\033[1;32m[记忆激活]\033[0m 匹配率: {topic_match:.3f}, 平均相似度: {average_similarities:.3f}, " - f"激活值: {activation}" - ) - - return activation - - async def get_relevant_memories( - self, text: str, max_topics: int = 5, similarity_threshold: float = 0.4, max_memory_num: int = 5 - ) -> list: - """根据输入文本获取相关的记忆内容""" - identified_topics = await self._identify_topics(text) - - all_similar_topics = self._find_similar_topics( - identified_topics, similarity_threshold=similarity_threshold, debug_info="记忆检索" - ) - - relevant_topics = self._get_top_topics(all_similar_topics, max_topics) - - relevant_memories = [] - for topic, score in relevant_topics: - first_layer, _ = self.memory_graph.get_related_item(topic, depth=1) - if first_layer: - if len(first_layer) > max_memory_num / 2: - first_layer = random.sample(first_layer, max_memory_num // 2) - for memory in first_layer: - relevant_memories.append({"topic": topic, "similarity": score, "content": memory}) - - relevant_memories.sort(key=lambda x: x["similarity"], reverse=True) - - if len(relevant_memories) > max_memory_num: - relevant_memories = random.sample(relevant_memories, max_memory_num) - - return relevant_memories - - def find_topic_llm(self, text, topic_num): - prompt = ( - f"这是一段文字:{text}。请你从这段话中总结出{topic_num}个关键的概念,可以是名词,动词,或者特定人物,帮我列出来," - f"用逗号,隔开,尽可能精简。只需要列举{topic_num}个话题就好,不要有序号,不要告诉我其他内容。" - ) - return prompt - - def topic_what(self, text, topic, time_info): - prompt = ( - f'这是一段文字,{time_info}:{text}。我想让你基于这段文字来概括"{topic}"这个概念,帮我总结成一句自然的话,' - f"可以包含时间和人物,以及具体的观点。只输出这句话就好" - ) - return prompt - - -def segment_text(text): - """使用jieba进行文本分词""" - seg_text = list(jieba.cut(text)) - return seg_text - - -def text_to_vector(text): - """将文本转换为词频向量""" - words = segment_text(text) - vector = {} - for word in words: - vector[word] = vector.get(word, 0) + 1 - return vector - - -def cosine_similarity(v1, v2): - """计算两个向量的余弦相似度""" - dot_product = sum(a * b for a, b in zip(v1, v2)) - norm1 = math.sqrt(sum(a * a for a in v1)) - norm2 = math.sqrt(sum(b * b for b in v2)) - if norm1 == 0 or norm2 == 0: - return 0 - return dot_product / (norm1 * norm2) - - -def visualize_graph_lite(memory_graph: Memory_graph, color_by_memory: bool = False): - # 设置中文字体 - plt.rcParams["font.sans-serif"] = ["SimHei"] # 用来正常显示中文标签 - plt.rcParams["axes.unicode_minus"] = False # 用来正常显示负号 - - G = memory_graph.G - - # 创建一个新图用于可视化 - H = G.copy() - - # 过滤掉内容数量小于2的节点 - nodes_to_remove = [] - for node in H.nodes(): - memory_items = H.nodes[node].get("memory_items", []) - memory_count = len(memory_items) if isinstance(memory_items, list) else (1 if memory_items else 0) - if memory_count < 2: - nodes_to_remove.append(node) - - H.remove_nodes_from(nodes_to_remove) - - # 如果没有符合条件的节点,直接返回 - if len(H.nodes()) == 0: - print("没有找到内容数量大于等于2的节点") - return - - # 计算节点大小和颜色 - node_colors = [] - node_sizes = [] - nodes = list(H.nodes()) - - # 获取最大记忆数用于归一化节点大小 - max_memories = 1 - for node in nodes: - memory_items = H.nodes[node].get("memory_items", []) - memory_count = len(memory_items) if isinstance(memory_items, list) else (1 if memory_items else 0) - max_memories = max(max_memories, memory_count) - - # 计算每个节点的大小和颜色 - for node in nodes: - # 计算节点大小(基于记忆数量) - memory_items = H.nodes[node].get("memory_items", []) - memory_count = len(memory_items) if isinstance(memory_items, list) else (1 if memory_items else 0) - # 使用指数函数使变化更明显 - ratio = memory_count / max_memories - size = 400 + 2000 * (ratio**2) # 增大节点大小 - node_sizes.append(size) - - # 计算节点颜色(基于连接数) - degree = H.degree(node) - if degree >= 30: - node_colors.append((1.0, 0, 0)) # 亮红色 (#FF0000) - else: - # 将1-10映射到0-1的范围 - color_ratio = (degree - 1) / 29.0 if degree > 1 else 0 - # 使用蓝到红的渐变 - red = min(0.9, color_ratio) - blue = max(0.0, 1.0 - color_ratio) - node_colors.append((red, 0, blue)) - - # 绘制图形 - plt.figure(figsize=(16, 12)) # 减小图形尺寸 - pos = nx.spring_layout( - H, - k=1, # 调整节点间斥力 - iterations=100, # 增加迭代次数 - scale=1.5, # 减小布局尺寸 - weight="strength", - ) # 使用边的strength属性作为权重 - - nx.draw( - H, - pos, - with_labels=True, - node_color=node_colors, - node_size=node_sizes, - font_size=12, # 保持增大的字体大小 - font_family="SimHei", - font_weight="bold", - edge_color="gray", - width=1.5, - ) # 统一的边宽度 - - title = """记忆图谱可视化(仅显示内容≥2的节点) -节点大小表示记忆数量 -节点颜色:蓝(弱连接)到红(强连接)渐变,边的透明度表示连接强度 -连接强度越大的节点距离越近""" - plt.title(title, fontsize=16, fontfamily="SimHei") - plt.show() - - -async def main(): - # 初始化数据库 - logger.info("正在初始化数据库连接...") - start_time = time.time() - - test_pare = { - "do_build_memory": True, - "do_forget_topic": False, - "do_visualize_graph": True, - "do_query": False, - "do_merge_memory": False, - } - - # 创建记忆图 - memory_graph = Memory_graph() - - # 创建海马体 - hippocampus = Hippocampus(memory_graph) - - # 从数据库同步数据 - hippocampus.memory_cortex.sync_memory_from_db() - - end_time = time.time() - logger.info(f"\033[32m[加载海马体耗时: {end_time - start_time:.2f} 秒]\033[0m") - - # 构建记忆 - if test_pare["do_build_memory"]: - logger.info("开始构建记忆...") - chat_size = 20 - await hippocampus.operation_build_memory(chat_size=chat_size) - - end_time = time.time() - logger.info( - f"\033[32m[构建记忆耗时: {end_time - start_time:.2f} 秒,chat_size={chat_size},chat_count = 16]\033[0m" - ) - - if test_pare["do_forget_topic"]: - logger.info("开始遗忘记忆...") - await hippocampus.operation_forget_topic(percentage=0.01) - - end_time = time.time() - logger.info(f"\033[32m[遗忘记忆耗时: {end_time - start_time:.2f} 秒]\033[0m") - - if test_pare["do_merge_memory"]: - logger.info("开始合并记忆...") - await hippocampus.operation_merge_memory(percentage=0.1) - - end_time = time.time() - logger.info(f"\033[32m[合并记忆耗时: {end_time - start_time:.2f} 秒]\033[0m") - - if test_pare["do_visualize_graph"]: - # 展示优化后的图形 - logger.info("生成记忆图谱可视化...") - print("\n生成优化后的记忆图谱:") - visualize_graph_lite(memory_graph) - - if test_pare["do_query"]: - # 交互式查询 - while True: - query = input("\n请输入新的查询概念(输入'退出'以结束):") - if query.lower() == "退出": - break - - items_list = memory_graph.get_related_item(query) - if items_list: - first_layer, second_layer = items_list - if first_layer: - print("\n直接相关的记忆:") - for item in first_layer: - print(f"- {item}") - if second_layer: - print("\n间接相关的记忆:") - for item in second_layer: - print(f"- {item}") - else: - print("未找到相关记忆。") - - -if __name__ == "__main__": - import asyncio - - asyncio.run(main()) diff --git a/src/plugins/memory_system/sample_distribution.py b/src/plugins/memory_system/sample_distribution.py new file mode 100644 index 000000000..1d285f7b4 --- /dev/null +++ b/src/plugins/memory_system/sample_distribution.py @@ -0,0 +1,172 @@ +import numpy as np +import matplotlib.pyplot as plt +from scipy import stats +import time +from datetime import datetime, timedelta + +class DistributionVisualizer: + def __init__(self, mean=0, std=1, skewness=0, sample_size=10): + """ + 初始化分布可视化器 + + 参数: + mean (float): 期望均值 + std (float): 标准差 + skewness (float): 偏度 + sample_size (int): 样本大小 + """ + self.mean = mean + self.std = std + self.skewness = skewness + self.sample_size = sample_size + self.samples = None + + def generate_samples(self): + """生成具有指定参数的样本""" + if self.skewness == 0: + # 对于无偏度的情况,直接使用正态分布 + self.samples = np.random.normal(loc=self.mean, scale=self.std, size=self.sample_size) + else: + # 使用 scipy.stats 生成具有偏度的分布 + self.samples = stats.skewnorm.rvs(a=self.skewness, + loc=self.mean, + scale=self.std, + size=self.sample_size) + + def get_weighted_samples(self): + """获取加权后的样本数列""" + if self.samples is None: + self.generate_samples() + # 将样本值乘以样本大小 + return self.samples * self.sample_size + + def get_statistics(self): + """获取分布的统计信息""" + if self.samples is None: + self.generate_samples() + + return { + "均值": np.mean(self.samples), + "标准差": np.std(self.samples), + "实际偏度": stats.skew(self.samples) + } + +class MemoryBuildScheduler: + def __init__(self, + n_hours1, std_hours1, weight1, + n_hours2, std_hours2, weight2, + total_samples=50): + """ + 初始化记忆构建调度器 + + 参数: + n_hours1 (float): 第一个分布的均值(距离现在的小时数) + std_hours1 (float): 第一个分布的标准差(小时) + weight1 (float): 第一个分布的权重 + n_hours2 (float): 第二个分布的均值(距离现在的小时数) + std_hours2 (float): 第二个分布的标准差(小时) + weight2 (float): 第二个分布的权重 + total_samples (int): 要生成的总时间点数量 + """ + # 归一化权重 + total_weight = weight1 + weight2 + self.weight1 = weight1 / total_weight + self.weight2 = weight2 / total_weight + + self.n_hours1 = n_hours1 + self.std_hours1 = std_hours1 + self.n_hours2 = n_hours2 + self.std_hours2 = std_hours2 + self.total_samples = total_samples + self.base_time = datetime.now() + + def generate_time_samples(self): + """生成混合分布的时间采样点""" + # 根据权重计算每个分布的样本数 + samples1 = int(self.total_samples * self.weight1) + samples2 = self.total_samples - samples1 + + # 生成两个正态分布的小时偏移 + hours_offset1 = np.random.normal( + loc=self.n_hours1, + scale=self.std_hours1, + size=samples1 + ) + + hours_offset2 = np.random.normal( + loc=self.n_hours2, + scale=self.std_hours2, + size=samples2 + ) + + # 合并两个分布的偏移 + hours_offset = np.concatenate([hours_offset1, hours_offset2]) + + # 将偏移转换为实际时间戳(使用绝对值确保时间点在过去) + timestamps = [self.base_time - timedelta(hours=abs(offset)) for offset in hours_offset] + + # 按时间排序(从最早到最近) + return sorted(timestamps) + + def get_timestamp_array(self): + """返回时间戳数组""" + timestamps = self.generate_time_samples() + return [int(t.timestamp()) for t in timestamps] + +def print_time_samples(timestamps, show_distribution=True): + """打印时间样本和分布信息""" + print(f"\n生成的{len(timestamps)}个时间点分布:") + print("序号".ljust(5), "时间戳".ljust(25), "距现在(小时)") + print("-" * 50) + + now = datetime.now() + time_diffs = [] + + for i, timestamp in enumerate(timestamps, 1): + hours_diff = (now - timestamp).total_seconds() / 3600 + time_diffs.append(hours_diff) + print(f"{str(i).ljust(5)} {timestamp.strftime('%Y-%m-%d %H:%M:%S').ljust(25)} {hours_diff:.2f}") + + # 打印统计信息 + print("\n统计信息:") + print(f"平均时间偏移:{np.mean(time_diffs):.2f}小时") + print(f"标准差:{np.std(time_diffs):.2f}小时") + print(f"最早时间:{min(timestamps).strftime('%Y-%m-%d %H:%M:%S')} ({max(time_diffs):.2f}小时前)") + print(f"最近时间:{max(timestamps).strftime('%Y-%m-%d %H:%M:%S')} ({min(time_diffs):.2f}小时前)") + + if show_distribution: + # 计算时间分布的直方图 + hist, bins = np.histogram(time_diffs, bins=40) + print("\n时间分布(每个*代表一个时间点):") + for i in range(len(hist)): + if hist[i] > 0: + print(f"{bins[i]:6.1f}-{bins[i+1]:6.1f}小时: {'*' * int(hist[i])}") + +# 使用示例 +if __name__ == "__main__": + # 创建一个双峰分布的记忆调度器 + scheduler = MemoryBuildScheduler( + n_hours1=12, # 第一个分布均值(12小时前) + std_hours1=8, # 第一个分布标准差 + weight1=0.7, # 第一个分布权重 70% + n_hours2=36, # 第二个分布均值(36小时前) + std_hours2=24, # 第二个分布标准差 + weight2=0.3, # 第二个分布权重 30% + total_samples=50 # 总共生成50个时间点 + ) + + # 生成时间分布 + timestamps = scheduler.generate_time_samples() + + # 打印结果,包含分布可视化 + print_time_samples(timestamps, show_distribution=True) + + # 打印时间戳数组 + timestamp_array = scheduler.get_timestamp_array() + print("\n时间戳数组(Unix时间戳):") + print("[", end="") + for i, ts in enumerate(timestamp_array): + if i > 0: + print(", ", end="") + print(ts, end="") + print("]") \ No newline at end of file diff --git a/src/plugins/schedule/offline_llm.py b/src/plugins/schedule/offline_llm.py new file mode 100644 index 000000000..e4dc23f93 --- /dev/null +++ b/src/plugins/schedule/offline_llm.py @@ -0,0 +1,123 @@ +import asyncio +import os +import time +from typing import Tuple, Union + +import aiohttp +import requests +from src.common.logger import get_module_logger + +logger = get_module_logger("offline_llm") + + +class LLMModel: + def __init__(self, model_name="deepseek-ai/DeepSeek-V3", **kwargs): + self.model_name = model_name + self.params = kwargs + self.api_key = os.getenv("SILICONFLOW_KEY") + self.base_url = os.getenv("SILICONFLOW_BASE_URL") + + if not self.api_key or not self.base_url: + raise ValueError("环境变量未正确加载:SILICONFLOW_KEY 或 SILICONFLOW_BASE_URL 未设置") + + logger.info(f"API URL: {self.base_url}") # 使用 logger 记录 base_url + + def generate_response(self, prompt: str) -> Union[str, Tuple[str, str]]: + """根据输入的提示生成模型的响应""" + headers = {"Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json"} + + # 构建请求体 + data = { + "model": self.model_name, + "messages": [{"role": "user", "content": prompt}], + "temperature": 0.5, + **self.params, + } + + # 发送请求到完整的 chat/completions 端点 + api_url = f"{self.base_url.rstrip('/')}/chat/completions" + logger.info(f"Request URL: {api_url}") # 记录请求的 URL + + max_retries = 3 + base_wait_time = 15 # 基础等待时间(秒) + + for retry in range(max_retries): + try: + response = requests.post(api_url, headers=headers, json=data) + + if response.status_code == 429: + wait_time = base_wait_time * (2**retry) # 指数退避 + logger.warning(f"遇到请求限制(429),等待{wait_time}秒后重试...") + time.sleep(wait_time) + continue + + response.raise_for_status() # 检查其他响应状态 + + result = response.json() + if "choices" in result and len(result["choices"]) > 0: + content = result["choices"][0]["message"]["content"] + reasoning_content = result["choices"][0]["message"].get("reasoning_content", "") + return content, reasoning_content + return "没有返回结果", "" + + except Exception as e: + if retry < max_retries - 1: # 如果还有重试机会 + wait_time = base_wait_time * (2**retry) + logger.error(f"[回复]请求失败,等待{wait_time}秒后重试... 错误: {str(e)}") + time.sleep(wait_time) + else: + logger.error(f"请求失败: {str(e)}") + return f"请求失败: {str(e)}", "" + + logger.error("达到最大重试次数,请求仍然失败") + return "达到最大重试次数,请求仍然失败", "" + + async def generate_response_async(self, prompt: str) -> Union[str, Tuple[str, str]]: + """异步方式根据输入的提示生成模型的响应""" + headers = {"Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json"} + + # 构建请求体 + data = { + "model": self.model_name, + "messages": [{"role": "user", "content": prompt}], + "temperature": 0.5, + **self.params, + } + + # 发送请求到完整的 chat/completions 端点 + api_url = f"{self.base_url.rstrip('/')}/chat/completions" + logger.info(f"Request URL: {api_url}") # 记录请求的 URL + + max_retries = 3 + base_wait_time = 15 + + async with aiohttp.ClientSession() as session: + for retry in range(max_retries): + try: + async with session.post(api_url, headers=headers, json=data) as response: + if response.status == 429: + wait_time = base_wait_time * (2**retry) # 指数退避 + logger.warning(f"遇到请求限制(429),等待{wait_time}秒后重试...") + await asyncio.sleep(wait_time) + continue + + response.raise_for_status() # 检查其他响应状态 + + result = await response.json() + if "choices" in result and len(result["choices"]) > 0: + content = result["choices"][0]["message"]["content"] + reasoning_content = result["choices"][0]["message"].get("reasoning_content", "") + return content, reasoning_content + return "没有返回结果", "" + + except Exception as e: + if retry < max_retries - 1: # 如果还有重试机会 + wait_time = base_wait_time * (2**retry) + logger.error(f"[回复]请求失败,等待{wait_time}秒后重试... 错误: {str(e)}") + await asyncio.sleep(wait_time) + else: + logger.error(f"请求失败: {str(e)}") + return f"请求失败: {str(e)}", "" + + logger.error("达到最大重试次数,请求仍然失败") + return "达到最大重试次数,请求仍然失败", "" diff --git a/src/plugins/schedule/schedule_generator copy.py b/src/plugins/schedule/schedule_generator copy.py new file mode 100644 index 000000000..7ebc00a54 --- /dev/null +++ b/src/plugins/schedule/schedule_generator copy.py @@ -0,0 +1,192 @@ +import datetime +import json +import re +import os +import sys +from typing import Dict, Union + +from nonebot import get_driver + +# 添加项目根目录到 Python 路径 +root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) +sys.path.append(root_path) + +# from src.plugins.chat.config import global_config +from src.common.database import db # 使用正确的导入语法 +from src.plugins.schedule.offline_llm import LLMModel +from src.common.logger import get_module_logger + +logger = get_module_logger("scheduler") + + +class ScheduleGenerator: + enable_output: bool = True + + def __init__(self): + # 使用离线LLM模型 + self.llm_scheduler = LLMModel(model_name="Pro/deepseek-ai/DeepSeek-V3", temperature=0.9) + self.today_schedule_text = "" + self.today_schedule = {} + self.tomorrow_schedule_text = "" + self.tomorrow_schedule = {} + self.yesterday_schedule_text = "" + self.yesterday_schedule = {} + + async def initialize(self): + today = datetime.datetime.now() + tomorrow = datetime.datetime.now() + datetime.timedelta(days=1) + yesterday = datetime.datetime.now() - datetime.timedelta(days=1) + + self.today_schedule_text, self.today_schedule = await self.generate_daily_schedule(target_date=today) + self.tomorrow_schedule_text, self.tomorrow_schedule = await self.generate_daily_schedule( + target_date=tomorrow, read_only=True + ) + self.yesterday_schedule_text, self.yesterday_schedule = await self.generate_daily_schedule( + target_date=yesterday, read_only=True + ) + + async def generate_daily_schedule( + self, target_date: datetime.datetime = None, read_only: bool = False + ) -> Dict[str, str]: + date_str = target_date.strftime("%Y-%m-%d") + weekday = target_date.strftime("%A") + + schedule_text = str + + existing_schedule = db.schedule.find_one({"date": date_str}) + if existing_schedule: + if self.enable_output: + logger.debug(f"{date_str}的日程已存在:") + schedule_text = existing_schedule["schedule"] + # print(self.schedule_text) + + elif not read_only: + logger.debug(f"{date_str}的日程不存在,准备生成新的日程。") + prompt = ( + f"""我是{global_config.BOT_NICKNAME},{global_config.PROMPT_SCHEDULE_GEN},请为我生成{date_str}({weekday})的日程安排,包括:""" + + """ + 1. 早上的学习和工作安排 + 2. 下午的活动和任务 + 3. 晚上的计划和休息时间 + 请按照时间顺序列出具体时间点和对应的活动,用一个时间点而不是时间段来表示时间,用JSON格式返回日程表, + 仅返回内容,不要返回注释,不要添加任何markdown或代码块样式,时间采用24小时制, + 格式为{"时间": "活动","时间": "活动",...}。""" + ) + + try: + schedule_text, _ = self.llm_scheduler.generate_response(prompt) + db.schedule.insert_one({"date": date_str, "schedule": schedule_text}) + self.enable_output = True + except Exception as e: + logger.error(f"生成日程失败: {str(e)}") + schedule_text = "生成日程时出错了" + # print(self.schedule_text) + else: + if self.enable_output: + logger.debug(f"{date_str}的日程不存在。") + schedule_text = "忘了" + + return schedule_text, None + + schedule_form = self._parse_schedule(schedule_text) + return schedule_text, schedule_form + + def _parse_schedule(self, schedule_text: str) -> Union[bool, Dict[str, str]]: + """解析日程文本,转换为时间和活动的字典""" + try: + reg = r"\{(.|\r|\n)+\}" + matched = re.search(reg, schedule_text)[0] + schedule_dict = json.loads(matched) + return schedule_dict + except json.JSONDecodeError: + logger.exception("解析日程失败: {}".format(schedule_text)) + return False + + def _parse_time(self, time_str: str) -> str: + """解析时间字符串,转换为时间""" + return datetime.datetime.strptime(time_str, "%H:%M") + + def get_current_task(self) -> str: + """获取当前时间应该进行的任务""" + current_time = datetime.datetime.now().strftime("%H:%M") + + # 找到最接近当前时间的任务 + closest_time = None + min_diff = float("inf") + + # 检查今天的日程 + if not self.today_schedule: + return "摸鱼" + for time_str in self.today_schedule.keys(): + diff = abs(self._time_diff(current_time, time_str)) + if closest_time is None or diff < min_diff: + closest_time = time_str + min_diff = diff + + # 检查昨天的日程中的晚间任务 + if self.yesterday_schedule: + for time_str in self.yesterday_schedule.keys(): + if time_str >= "20:00": # 只考虑晚上8点之后的任务 + # 计算与昨天这个时间点的差异(需要加24小时) + diff = abs(self._time_diff(current_time, time_str)) + if diff < min_diff: + closest_time = time_str + min_diff = diff + return closest_time, self.yesterday_schedule[closest_time] + + if closest_time: + return closest_time, self.today_schedule[closest_time] + return "摸鱼" + + def _time_diff(self, time1: str, time2: str) -> int: + """计算两个时间字符串之间的分钟差""" + if time1 == "24:00": + time1 = "23:59" + if time2 == "24:00": + time2 = "23:59" + t1 = datetime.datetime.strptime(time1, "%H:%M") + t2 = datetime.datetime.strptime(time2, "%H:%M") + diff = int((t2 - t1).total_seconds() / 60) + # 考虑时间的循环性 + if diff < -720: + diff += 1440 # 加一天的分钟 + elif diff > 720: + diff -= 1440 # 减一天的分钟 + # print(f"时间1[{time1}]: 时间2[{time2}],差值[{diff}]分钟") + return diff + + def print_schedule(self): + """打印完整的日程安排""" + if not self._parse_schedule(self.today_schedule_text): + logger.warning("今日日程有误,将在下次运行时重新生成") + db.schedule.delete_one({"date": datetime.datetime.now().strftime("%Y-%m-%d")}) + else: + logger.info("=== 今日日程安排 ===") + for time_str, activity in self.today_schedule.items(): + logger.info(f"时间[{time_str}]: 活动[{activity}]") + logger.info("==================") + self.enable_output = False + + +async def main(): + # 使用示例 + scheduler = ScheduleGenerator() + await scheduler.initialize() + scheduler.print_schedule() + print("\n当前任务:") + print(await scheduler.get_current_task()) + + print("昨天日程:") + print(scheduler.yesterday_schedule) + print("今天日程:") + print(scheduler.today_schedule) + print("明天日程:") + print(scheduler.tomorrow_schedule) + +# 当作为组件导入时使用的实例 +bot_schedule = ScheduleGenerator() + +if __name__ == "__main__": + import asyncio + # 当直接运行此文件时执行 + asyncio.run(main()) diff --git a/src/plugins/schedule/schedule_generator.py b/src/plugins/schedule/schedule_generator.py index 11db6664d..3fabfa389 100644 --- a/src/plugins/schedule/schedule_generator.py +++ b/src/plugins/schedule/schedule_generator.py @@ -1,12 +1,15 @@ import datetime import json import re +import os +import sys from typing import Dict, Union from nonebot import get_driver -from src.plugins.chat.config import global_config +# 添加项目根目录到 Python 路径 +from src.plugins.chat.config import global_config from ...common.database import db # 使用正确的导入语法 from ..models.utils_model import LLM_request from src.common.logger import get_module_logger @@ -165,24 +168,5 @@ class ScheduleGenerator: logger.info(f"时间[{time_str}]: 活动[{activity}]") logger.info("==================") self.enable_output = False - - -# def main(): -# # 使用示例 -# scheduler = ScheduleGenerator() -# # new_schedule = scheduler.generate_daily_schedule() -# scheduler.print_schedule() -# print("\n当前任务:") -# print(scheduler.get_current_task()) - -# print("昨天日程:") -# print(scheduler.yesterday_schedule) -# print("今天日程:") -# print(scheduler.today_schedule) -# print("明天日程:") -# print(scheduler.tomorrow_schedule) - -# if __name__ == "__main__": -# main() - +# 当作为组件导入时使用的实例 bot_schedule = ScheduleGenerator() diff --git a/template.env b/template.env index 6791c5842..934a331d0 100644 --- a/template.env +++ b/template.env @@ -1,8 +1,6 @@ HOST=127.0.0.1 PORT=8080 -ENABLE_ADVANCE_OUTPUT=false - # 插件配置 PLUGINS=["src2.plugins.chat"] @@ -31,6 +29,7 @@ CHAT_ANY_WHERE_KEY= SILICONFLOW_KEY= # 定义日志相关配置 +SIMPLE_OUTPUT=true # 精简控制台输出格式 CONSOLE_LOG_LEVEL=INFO # 自定义日志的默认控制台输出日志级别 FILE_LOG_LEVEL=DEBUG # 自定义日志的默认文件输出日志级别 DEFAULT_CONSOLE_LOG_LEVEL=SUCCESS # 原生日志的控制台输出日志级别(nonebot就是这一类) diff --git a/template/bot_config_template.toml b/template/bot_config_template.toml index ec2b5fbd4..e5cf1df86 100644 --- a/template/bot_config_template.toml +++ b/template/bot_config_template.toml @@ -1,5 +1,5 @@ [inner] -version = "0.0.10" +version = "0.0.11" #以下是给开发人员阅读的,一般用户不需要阅读 #如果你想要修改配置文件,请在修改后将version的值进行变更 @@ -66,12 +66,15 @@ model_r1_distill_probability = 0.1 # 麦麦回答时选择次要回复模型3 max_response_length = 1024 # 麦麦回答的最大token数 [willing] -willing_mode = "classical" -# willing_mode = "dynamic" -# willing_mode = "custom" +willing_mode = "classical" # 回复意愿模式 经典模式 +# willing_mode = "dynamic" # 动态模式(可能不兼容) +# willing_mode = "custom" # 自定义模式(可自行调整 [memory] build_memory_interval = 2000 # 记忆构建间隔 单位秒 间隔越低,麦麦学习越多,但是冗余信息也会增多 +build_memory_distribution = [4,2,0.6,24,8,0.4] # 记忆构建分布,参数:分布1均值,标准差,权重,分布2均值,标准差,权重 +build_memory_sample_num = 10 # 采样数量,数值越高记忆采样次数越多 +build_memory_sample_length = 20 # 采样长度,数值越高一段记忆内容越丰富 memory_compress_rate = 0.1 # 记忆压缩率 控制记忆精简程度 建议保持默认,调高可以获得更多信息,但是冗余信息也会增多 forget_memory_interval = 1000 # 记忆遗忘间隔 单位秒 间隔越低,麦麦遗忘越频繁,记忆更精简,但更难学习 @@ -109,9 +112,7 @@ tone_error_rate=0.2 # 声调错误概率 word_replace_rate=0.006 # 整词替换概率 [others] -enable_advance_output = false # 是否启用高级输出 enable_kuuki_read = true # 是否启用读空气功能 -enable_debug_output = false # 是否启用调试输出 enable_friend_chat = false # 是否启用好友聊天 [groups] @@ -120,9 +121,9 @@ talk_allowed = [ 123, ] #可以回复消息的群 talk_frequency_down = [] #降低回复频率的群 -ban_user_id = [] #禁止回复消息的QQ号 +ban_user_id = [] #禁止回复和读取消息的QQ号 -[remote] #测试功能,发送统计信息,主要是看全球有多少只麦麦 +[remote] #发送统计信息,主要是看全球有多少只麦麦 enable = true From 432104f582b99d8e61aa56bf1da993f3a4a725d1 Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Fri, 21 Mar 2025 14:49:59 +0800 Subject: [PATCH 15/21] =?UTF-8?q?fix=20=E4=BF=AE=E6=AD=A3=E6=A0=BC?= =?UTF-8?q?=E5=BC=8F=E9=98=B2=E6=AD=A2=E6=9B=B9=E9=A3=9E?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/config.py | 15 ++++++++++++--- src/plugins/memory_system/memory.py | 6 +++++- src/plugins/memory_system/memory_manual_build.py | 2 +- src/plugins/memory_system/sample_distribution.py | 2 -- src/plugins/schedule/schedule_generator copy.py | 9 ++++----- src/plugins/schedule/schedule_generator.py | 2 -- 6 files changed, 22 insertions(+), 14 deletions(-) diff --git a/src/plugins/chat/config.py b/src/plugins/chat/config.py index d0cb18822..17b3cfece 100644 --- a/src/plugins/chat/config.py +++ b/src/plugins/chat/config.py @@ -321,9 +321,18 @@ class BotConfig: ) config.memory_compress_rate = memory_config.get("memory_compress_rate", config.memory_compress_rate) if config.INNER_VERSION in SpecifierSet(">=0.0.11"): - config.memory_build_distribution = memory_config.get("memory_build_distribution", config.memory_build_distribution) - config.build_memory_sample_num = memory_config.get("build_memory_sample_num", config.build_memory_sample_num) - config.build_memory_sample_length = memory_config.get("build_memory_sample_length", config.build_memory_sample_length) + config.memory_build_distribution = memory_config.get( + "memory_build_distribution", + config.memory_build_distribution + ) + config.build_memory_sample_num = memory_config.get( + "build_memory_sample_num", + config.build_memory_sample_num + ) + config.build_memory_sample_length = memory_config.get( + "build_memory_sample_length", + config.build_memory_sample_length + ) def remote(parent: dict): diff --git a/src/plugins/memory_system/memory.py b/src/plugins/memory_system/memory.py index b55dcf7b3..1f69dd3cf 100644 --- a/src/plugins/memory_system/memory.py +++ b/src/plugins/memory_system/memory.py @@ -239,7 +239,11 @@ class Hippocampus: chat_samples = [] for timestamp in timestamps: - messages = self.random_get_msg_snippet(timestamp, global_config.build_memory_sample_length, max_memorized_time_per_msg) + messages = self.random_get_msg_snippet( + timestamp, + global_config.build_memory_sample_length, + max_memorized_time_per_msg + ) if messages: time_diff = (datetime.datetime.now().timestamp() - timestamp) / 3600 logger.debug(f"成功抽取 {time_diff:.1f} 小时前的消息样本,共{len(messages)}条") diff --git a/src/plugins/memory_system/memory_manual_build.py b/src/plugins/memory_system/memory_manual_build.py index 4d6596e9f..b575f455e 100644 --- a/src/plugins/memory_system/memory_manual_build.py +++ b/src/plugins/memory_system/memory_manual_build.py @@ -17,7 +17,7 @@ import jieba root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -from src.common.logger import get_module_logger +from src.common.logger import get_module_logger # noqa: E402 from src.common.database import db # noqa E402 from src.plugins.memory_system.offline_llm import LLMModel # noqa E402 diff --git a/src/plugins/memory_system/sample_distribution.py b/src/plugins/memory_system/sample_distribution.py index 1d285f7b4..dbe4b88a4 100644 --- a/src/plugins/memory_system/sample_distribution.py +++ b/src/plugins/memory_system/sample_distribution.py @@ -1,7 +1,5 @@ import numpy as np -import matplotlib.pyplot as plt from scipy import stats -import time from datetime import datetime, timedelta class DistributionVisualizer: diff --git a/src/plugins/schedule/schedule_generator copy.py b/src/plugins/schedule/schedule_generator copy.py index 7ebc00a54..eff0a08d6 100644 --- a/src/plugins/schedule/schedule_generator copy.py +++ b/src/plugins/schedule/schedule_generator copy.py @@ -5,16 +5,15 @@ import os import sys from typing import Dict, Union -from nonebot import get_driver # 添加项目根目录到 Python 路径 root_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "../../..")) sys.path.append(root_path) -# from src.plugins.chat.config import global_config -from src.common.database import db # 使用正确的导入语法 -from src.plugins.schedule.offline_llm import LLMModel -from src.common.logger import get_module_logger +from src.common.database import db # noqa: E402 +from src.common.logger import get_module_logger # noqa: E402 +from src.plugins.schedule.offline_llm import LLMModel # noqa: E402 +from src.plugins.chat.config import global_config # noqa: E402 logger = get_module_logger("scheduler") diff --git a/src/plugins/schedule/schedule_generator.py b/src/plugins/schedule/schedule_generator.py index 3fabfa389..d58211215 100644 --- a/src/plugins/schedule/schedule_generator.py +++ b/src/plugins/schedule/schedule_generator.py @@ -1,8 +1,6 @@ import datetime import json import re -import os -import sys from typing import Dict, Union from nonebot import get_driver From 7c50e333692c878807d66c58db2986e7ac90b0cd Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Fri, 21 Mar 2025 16:24:28 +0800 Subject: [PATCH 16/21] =?UTF-8?q?better=20=E6=9B=B4=E5=A5=BD=E7=9A=84logge?= =?UTF-8?q?r=E6=A0=BC=E5=BC=8F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/common/logger.py | 21 ++++---- src/plugins/chat/bot.py | 2 +- src/plugins/chat/message_sender.py | 2 +- src/plugins/chat/utils.py | 17 +++---- src/plugins/chat/utils_image.py | 2 +- src/plugins/memory_system/memory.py | 77 ++++++++++++++++++++--------- 6 files changed, 73 insertions(+), 48 deletions(-) diff --git a/src/common/logger.py b/src/common/logger.py index 2673275a5..91f1a1da0 100644 --- a/src/common/logger.py +++ b/src/common/logger.py @@ -31,10 +31,10 @@ _handler_registry: Dict[str, List[int]] = {} current_file_path = Path(__file__).resolve() LOG_ROOT = "logs" -ENABLE_ADVANCE_OUTPUT = os.getenv("SIMPLE_OUTPUT", "false") -print(f"ENABLE_ADVANCE_OUTPUT: {ENABLE_ADVANCE_OUTPUT}") +SIMPLE_OUTPUT = os.getenv("SIMPLE_OUTPUT", "false") +print(f"SIMPLE_OUTPUT: {SIMPLE_OUTPUT}") -if not ENABLE_ADVANCE_OUTPUT: +if not SIMPLE_OUTPUT: # 默认全局配置 DEFAULT_CONFIG = { # 日志级别配置 @@ -86,7 +86,6 @@ MEMORY_STYLE_CONFIG = { }, } -# 海马体日志样式配置 SENDER_STYLE_CONFIG = { "advanced": { "console_format": ( @@ -153,17 +152,17 @@ CHAT_STYLE_CONFIG = { "file_format": ("{time:YYYY-MM-DD HH:mm:ss} | {level: <8} | {extra[module]: <15} | 见闻 | {message}"), }, "simple": { - "console_format": ("{time:MM-DD HH:mm} | 见闻 | {message}"), + "console_format": ("{time:MM-DD HH:mm} | 见闻 | {message}"), # noqa: E501 "file_format": ("{time:YYYY-MM-DD HH:mm:ss} | {level: <8} | {extra[module]: <15} | 见闻 | {message}"), }, } -# 根据ENABLE_ADVANCE_OUTPUT选择配置 -MEMORY_STYLE_CONFIG = MEMORY_STYLE_CONFIG["advanced"] if ENABLE_ADVANCE_OUTPUT else MEMORY_STYLE_CONFIG["simple"] -TOPIC_STYLE_CONFIG = TOPIC_STYLE_CONFIG["advanced"] if ENABLE_ADVANCE_OUTPUT else TOPIC_STYLE_CONFIG["simple"] -SENDER_STYLE_CONFIG = SENDER_STYLE_CONFIG["advanced"] if ENABLE_ADVANCE_OUTPUT else SENDER_STYLE_CONFIG["simple"] -LLM_STYLE_CONFIG = LLM_STYLE_CONFIG["advanced"] if ENABLE_ADVANCE_OUTPUT else LLM_STYLE_CONFIG["simple"] -CHAT_STYLE_CONFIG = CHAT_STYLE_CONFIG["advanced"] if ENABLE_ADVANCE_OUTPUT else CHAT_STYLE_CONFIG["simple"] +# 根据SIMPLE_OUTPUT选择配置 +MEMORY_STYLE_CONFIG = MEMORY_STYLE_CONFIG["simple"] if SIMPLE_OUTPUT else MEMORY_STYLE_CONFIG["advanced"] +TOPIC_STYLE_CONFIG = TOPIC_STYLE_CONFIG["simple"] if SIMPLE_OUTPUT else TOPIC_STYLE_CONFIG["advanced"] +SENDER_STYLE_CONFIG = SENDER_STYLE_CONFIG["simple"] if SIMPLE_OUTPUT else SENDER_STYLE_CONFIG["advanced"] +LLM_STYLE_CONFIG = LLM_STYLE_CONFIG["simple"] if SIMPLE_OUTPUT else LLM_STYLE_CONFIG["advanced"] +CHAT_STYLE_CONFIG = CHAT_STYLE_CONFIG["simple"] if SIMPLE_OUTPUT else CHAT_STYLE_CONFIG["advanced"] def is_registered_module(record: dict) -> bool: diff --git a/src/plugins/chat/bot.py b/src/plugins/chat/bot.py index 24b7bdbff..38450f903 100644 --- a/src/plugins/chat/bot.py +++ b/src/plugins/chat/bot.py @@ -154,7 +154,7 @@ class ChatBot: ) # 开始思考的时间点 thinking_time_point = round(time.time(), 2) - logger.info(f"开始思考的时间点: {thinking_time_point}") + # logger.debug(f"开始思考的时间点: {thinking_time_point}") think_id = "mt" + str(thinking_time_point) thinking_message = MessageThinking( message_id=think_id, diff --git a/src/plugins/chat/message_sender.py b/src/plugins/chat/message_sender.py index 741cc2889..d79e9e7ab 100644 --- a/src/plugins/chat/message_sender.py +++ b/src/plugins/chat/message_sender.py @@ -220,7 +220,7 @@ class MessageManager: message_timeout = container.get_timeout_messages() if message_timeout: - logger.warning(f"发现{len(message_timeout)}条超时消息") + logger.debug(f"发现{len(message_timeout)}条超时消息") for msg in message_timeout: if msg == message_earliest: continue diff --git a/src/plugins/chat/utils.py b/src/plugins/chat/utils.py index 8b728ee4d..1563ea526 100644 --- a/src/plugins/chat/utils.py +++ b/src/plugins/chat/utils.py @@ -76,18 +76,11 @@ def calculate_information_content(text): def get_closest_chat_from_db(length: int, timestamp: str): - """从数据库中获取最接近指定时间戳的聊天记录 - - Args: - length: 要获取的消息数量 - timestamp: 时间戳 - - Returns: - list: 消息记录列表,每个记录包含时间和文本信息 - """ + # print(f"获取最接近指定时间戳的聊天记录,长度: {length}, 时间戳: {timestamp}") + # print(f"当前时间: {timestamp},转换后时间: {time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(timestamp))}") chat_records = [] closest_record = db.messages.find_one({"time": {"$lte": timestamp}}, sort=[("time", -1)]) - + # print(f"最接近的记录: {closest_record}") if closest_record: closest_time = closest_record["time"] chat_id = closest_record["chat_id"] # 获取chat_id @@ -102,7 +95,9 @@ def get_closest_chat_from_db(length: int, timestamp: str): .sort("time", 1) .limit(length) ) - + # print(f"获取到的记录: {chat_records}") + length = len(chat_records) + # print(f"获取到的记录长度: {length}") # 转换记录格式 formatted_records = [] for record in chat_records: diff --git a/src/plugins/chat/utils_image.py b/src/plugins/chat/utils_image.py index ea0c160eb..521795024 100644 --- a/src/plugins/chat/utils_image.py +++ b/src/plugins/chat/utils_image.py @@ -112,7 +112,7 @@ class ImageManager: # 查询缓存的描述 cached_description = self._get_description_from_db(image_hash, "emoji") if cached_description: - logger.info(f"缓存表情包描述: {cached_description}") + logger.debug(f"缓存表情包描述: {cached_description}") return f"[表情包:{cached_description}]" # 调用AI获取描述 diff --git a/src/plugins/memory_system/memory.py b/src/plugins/memory_system/memory.py index 1f69dd3cf..f5012c828 100644 --- a/src/plugins/memory_system/memory.py +++ b/src/plugins/memory_system/memory.py @@ -26,6 +26,11 @@ memory_config = LogConfig( console_format=MEMORY_STYLE_CONFIG["console_format"], file_format=MEMORY_STYLE_CONFIG["file_format"], ) +# print(f"memory_config: {memory_config}") +# print(f"MEMORY_STYLE_CONFIG: {MEMORY_STYLE_CONFIG}") +# print(f"MEMORY_STYLE_CONFIG['console_format']: {MEMORY_STYLE_CONFIG['console_format']}") +# print(f"MEMORY_STYLE_CONFIG['file_format']: {MEMORY_STYLE_CONFIG['file_format']}") + logger = get_module_logger("memory_system", config=memory_config) @@ -198,13 +203,15 @@ class Hippocampus: def random_get_msg_snippet(self, target_timestamp: float, chat_size: int, max_memorized_time_per_msg: int) -> list: try_count = 0 # 最多尝试2次抽取 - while try_count < 2: + while try_count < 3: messages = get_closest_chat_from_db(length=chat_size, timestamp=target_timestamp) if messages: + # print(f"抽取到的消息: {messages}") # 检查messages是否均没有达到记忆次数限制 for message in messages: if message["memorized_times"] >= max_memorized_time_per_msg: messages = None + # print(f"抽取到的消息提取次数达到限制,跳过") break if messages: # 成功抽取短期消息样本 @@ -235,8 +242,10 @@ class Hippocampus: # 生成时间戳数组 timestamps = scheduler.get_timestamp_array() - logger.debug(f"生成的时间戳数组: {timestamps}") - + # logger.debug(f"生成的时间戳数组: {timestamps}") + # print(f"生成的时间戳数组: {timestamps}") + # print(f"时间戳的实际时间: {[time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(ts)) for ts in timestamps]}") + logger.info(f"回忆往事: {[time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(ts)) for ts in timestamps]}") chat_samples = [] for timestamp in timestamps: messages = self.random_get_msg_snippet( @@ -247,18 +256,14 @@ class Hippocampus: if messages: time_diff = (datetime.datetime.now().timestamp() - timestamp) / 3600 logger.debug(f"成功抽取 {time_diff:.1f} 小时前的消息样本,共{len(messages)}条") + # print(f"成功抽取 {time_diff:.1f} 小时前的消息样本,共{len(messages)}条") chat_samples.append(messages) else: - logger.warning(f"时间戳 {timestamp} 的消息样本抽取失败") + logger.debug(f"时间戳 {timestamp} 的消息样本抽取失败") return chat_samples async def memory_compress(self, messages: list, compress_rate=0.1): - """压缩消息记录为记忆 - - Returns: - tuple: (压缩记忆集合, 相似主题字典) - """ if not messages: return set(), {} @@ -291,15 +296,23 @@ class Hippocampus: topics_response = await self.llm_topic_judge.generate_response(self.find_topic_llm(input_text, topic_num)) # 过滤topics + # 从配置文件获取需要过滤的关键词列表 filter_keywords = global_config.memory_ban_words + + # 将topics_response[0]中的中文逗号、顿号、空格都替换成英文逗号 + # 然后按逗号分割成列表,并去除每个topic前后的空白字符 topics = [ topic.strip() for topic in topics_response[0].replace(",", ",").replace("、", ",").replace(" ", ",").split(",") if topic.strip() ] + + # 过滤掉包含禁用关键词的topic + # any()检查topic中是否包含任何一个filter_keywords中的关键词 + # 只保留不包含禁用关键词的topic filtered_topics = [topic for topic in topics if not any(keyword in topic for keyword in filter_keywords)] - logger.info(f"过滤后话题: {filtered_topics}") + logger.debug(f"过滤后话题: {filtered_topics}") # 创建所有话题的请求任务 tasks = [] @@ -309,31 +322,42 @@ class Hippocampus: tasks.append((topic.strip(), task)) # 等待所有任务完成 - compressed_memory = set() + # 初始化压缩后的记忆集合和相似主题字典 + compressed_memory = set() # 存储压缩后的(主题,内容)元组 similar_topics_dict = {} # 存储每个话题的相似主题列表 + + # 遍历每个主题及其对应的LLM任务 for topic, task in tasks: response = await task if response: + # 将主题和LLM生成的内容添加到压缩记忆中 compressed_memory.add((topic, response[0])) - # 为每个话题查找相似的已存在主题 + + # 为当前主题寻找相似的已存在主题 existing_topics = list(self.memory_graph.G.nodes()) similar_topics = [] + # 计算当前主题与每个已存在主题的相似度 for existing_topic in existing_topics: + # 使用jieba分词,将主题转换为词集合 topic_words = set(jieba.cut(topic)) existing_words = set(jieba.cut(existing_topic)) - all_words = topic_words | existing_words - v1 = [1 if word in topic_words else 0 for word in all_words] - v2 = [1 if word in existing_words else 0 for word in all_words] + # 构建词向量用于计算余弦相似度 + all_words = topic_words | existing_words # 所有不重复的词 + v1 = [1 if word in topic_words else 0 for word in all_words] # 当前主题的词向量 + v2 = [1 if word in existing_words else 0 for word in all_words] # 已存在主题的词向量 + # 计算余弦相似度 similarity = cosine_similarity(v1, v2) - if similarity >= 0.6: + # 如果相似度超过阈值,添加到相似主题列表 + if similarity >= 0.7: similar_topics.append((existing_topic, similarity)) + # 按相似度降序排序,只保留前3个最相似的主题 similar_topics.sort(key=lambda x: x[1], reverse=True) - similar_topics = similar_topics[:5] + similar_topics = similar_topics[:3] similar_topics_dict[topic] = similar_topics return compressed_memory, similar_topics_dict @@ -352,7 +376,8 @@ class Hippocampus: async def operation_build_memory(self): memory_samples = self.get_memory_sample() - + all_added_nodes = [] + all_added_edges = [] for i, messages in enumerate(memory_samples, 1): all_topics = [] # 加载进度可视化 @@ -364,12 +389,13 @@ class Hippocampus: compress_rate = global_config.memory_compress_rate compressed_memory, similar_topics_dict = await self.memory_compress(messages, compress_rate) - logger.info(f"压缩后记忆数量: {len(compressed_memory)},似曾相识的话题: {len(similar_topics_dict)}") + logger.debug(f"压缩后记忆数量: {compressed_memory},似曾相识的话题: {similar_topics_dict}") current_time = datetime.datetime.now().timestamp() - + logger.debug(f"添加节点: {', '.join(topic for topic, _ in compressed_memory)}") + all_added_nodes.extend(topic for topic, _ in compressed_memory) + for topic, memory in compressed_memory: - logger.info(f"添加节点: {topic}") self.memory_graph.add_dot(topic, memory) all_topics.append(topic) @@ -379,7 +405,8 @@ class Hippocampus: for similar_topic, similarity in similar_topics: if topic != similar_topic: strength = int(similarity * 10) - logger.info(f"连接相似节点: {topic} 和 {similar_topic} (强度: {strength})") + logger.debug(f"连接相似节点: {topic} 和 {similar_topic} (强度: {strength})") + all_added_edges.append(f"{topic}-{similar_topic}") self.memory_graph.G.add_edge( topic, similar_topic, @@ -391,9 +418,13 @@ class Hippocampus: # 连接同批次的相关话题 for i in range(len(all_topics)): for j in range(i + 1, len(all_topics)): - logger.info(f"连接同批次节点: {all_topics[i]} 和 {all_topics[j]}") + logger.debug(f"连接同批次节点: {all_topics[i]} 和 {all_topics[j]}") + all_added_edges.append(f"{all_topics[i]}-{all_topics[j]}") self.memory_graph.connect_dot(all_topics[i], all_topics[j]) + logger.success(f"更新记忆: {', '.join(all_added_nodes)}") + logger.success(f"强化连接: {', '.join(all_added_edges)}") + # logger.success(f"强化连接: {', '.join(all_added_edges)}") self.sync_memory_to_db() def sync_memory_to_db(self): From 74f5bc2328b8300314fdc38d03cea0a9658ca8aa Mon Sep 17 00:00:00 2001 From: UnCLAS-Prommer Date: Fri, 21 Mar 2025 16:44:59 +0800 Subject: [PATCH 17/21] =?UTF-8?q?=E6=9B=B4=E6=96=B0requirements.txt?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- requirements.txt | Bin 658 -> 672 bytes 1 file changed, 0 insertions(+), 0 deletions(-) diff --git a/requirements.txt b/requirements.txt index 1e9e5ff25b8c4ccae9904607247966efcd269ab7..0dfd751484930ec11fed6da3b69ff72e6f5be121 100644 GIT binary patch delta 22 dcmbQlx`1`VBqlyy1}=tThGd3Jh60941^_ Date: Fri, 21 Mar 2025 16:59:46 +0800 Subject: [PATCH 18/21] =?UTF-8?q?fix=20=E7=A7=BB=E9=99=A4/n?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/prompt_builder.py | 26 +++++++++++++------------- 1 file changed, 13 insertions(+), 13 deletions(-) diff --git a/src/plugins/chat/prompt_builder.py b/src/plugins/chat/prompt_builder.py index 379aa4624..4ef8b6283 100644 --- a/src/plugins/chat/prompt_builder.py +++ b/src/plugins/chat/prompt_builder.py @@ -141,21 +141,21 @@ class PromptBuilder: logger.debug(f"知识检索耗时: {(end_time - start_time):.3f}秒") prompt = f""" -今天是{current_date},现在是{current_time},你今天的日程是:\ -``\n -{bot_schedule.today_schedule}\n -``\n -{prompt_info}\n -{memory_prompt}\n -{chat_target}\n -{chat_talking_prompt}\n -现在"{sender_name}"说的:\n -``\n -{message_txt}\n -``\n +今天是{current_date},现在是{current_time},你今天的日程是: +`` +{bot_schedule.today_schedule} +`` +{prompt_info} +{memory_prompt} +{chat_target} +{chat_talking_prompt} +现在"{sender_name}"说的: +`` +{message_txt} +`` 引起了你的注意,{relation_prompt_all}{mood_prompt}\n `` -你的网名叫{global_config.BOT_NICKNAME},{prompt_personality}。 +你的网名叫{global_config.BOT_NICKNAME},有人也叫你{"/".join(global_config.BOT_ALIAS_NAMES)},{prompt_personality},{prompt_personality}。 正在{bot_schedule_now_activity}的你同时也在一边{chat_target_2},现在请你读读之前的聊天记录,然后给出日常且口语化的回复,平淡一些, 尽量简短一些。{keywords_reaction_prompt}请注意把握聊天内容,不要刻意突出自身学科背景,不要回复的太有条理,可以有个性。 {prompt_ger} From e5d19d4bd91d16e36faf1513e804128cd677c540 Mon Sep 17 00:00:00 2001 From: DrSmoothl <1787882683@qq.com> Date: Fri, 21 Mar 2025 17:32:50 +0800 Subject: [PATCH 19/21] =?UTF-8?q?fix:=20=E4=BF=AE=E5=A4=8DwebUI=E6=9C=AA?= =?UTF-8?q?=E6=AD=A3=E7=A1=AE=E5=A4=84=E7=90=86=E8=A1=8C=E6=9C=AB=E6=B3=A8?= =?UTF-8?q?=E9=87=8A=E7=9A=84=E9=97=AE=E9=A2=98?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- webui.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/webui.py b/webui.py index 60ffa4805..a3b7eab64 100644 --- a/webui.py +++ b/webui.py @@ -98,10 +98,14 @@ def parse_env_config(config_file): # 逐行处理配置 for line in lines: line = line.strip() - # 忽略空行和注释 + # 忽略空行和注释行 if not line or line.startswith("#"): continue + # 处理行尾注释 + if "#" in line: + line = line.split("#")[0].strip() + # 拆分键值对 key, value = line.split("=", 1) From a47266abd29e21f3457b5be4bc66346474a3dfff Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Fri, 21 Mar 2025 17:44:18 +0800 Subject: [PATCH 20/21] =?UTF-8?q?better=20=E6=9B=B4=E5=A5=BD=E7=9A=84llm?= =?UTF-8?q?=E4=BD=BF=E7=94=A8=E7=BB=9F=E8=AE=A1?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/config.py | 2 -- src/plugins/chat/emoji_manager.py | 8 +++--- src/plugins/chat/llm_generator.py | 15 ++++++++--- src/plugins/chat/utils.py | 4 +-- src/plugins/memory_system/memory.py | 12 +++++++-- src/plugins/models/utils_model.py | 3 ++- template/bot_config_template.toml | 42 +++++++++++++++++------------ 7 files changed, 54 insertions(+), 32 deletions(-) diff --git a/src/plugins/chat/config.py b/src/plugins/chat/config.py index 17b3cfece..151aa5724 100644 --- a/src/plugins/chat/config.py +++ b/src/plugins/chat/config.py @@ -56,7 +56,6 @@ class BotConfig: llm_reasoning: Dict[str, str] = field(default_factory=lambda: {}) llm_reasoning_minor: Dict[str, str] = field(default_factory=lambda: {}) llm_normal: Dict[str, str] = field(default_factory=lambda: {}) - llm_normal_minor: Dict[str, str] = field(default_factory=lambda: {}) llm_topic_judge: Dict[str, str] = field(default_factory=lambda: {}) llm_summary_by_topic: Dict[str, str] = field(default_factory=lambda: {}) llm_emotion_judge: Dict[str, str] = field(default_factory=lambda: {}) @@ -235,7 +234,6 @@ class BotConfig: "llm_reasoning", "llm_reasoning_minor", "llm_normal", - "llm_normal_minor", "llm_topic_judge", "llm_summary_by_topic", "llm_emotion_judge", diff --git a/src/plugins/chat/emoji_manager.py b/src/plugins/chat/emoji_manager.py index e3a6b77af..57c2b0b85 100644 --- a/src/plugins/chat/emoji_manager.py +++ b/src/plugins/chat/emoji_manager.py @@ -38,9 +38,9 @@ class EmojiManager: def __init__(self): self._scan_task = None - self.vlm = LLM_request(model=global_config.vlm, temperature=0.3, max_tokens=1000, request_type="image") + self.vlm = LLM_request(model=global_config.vlm, temperature=0.3, max_tokens=1000, request_type="emoji") self.llm_emotion_judge = LLM_request( - model=global_config.llm_emotion_judge, max_tokens=600, temperature=0.8, request_type="image" + model=global_config.llm_emotion_judge, max_tokens=600, temperature=0.8, request_type="emoji" ) # 更高的温度,更少的token(后续可以根据情绪来调整温度) def _ensure_emoji_dir(self): @@ -111,7 +111,7 @@ class EmojiManager: if not text_for_search: logger.error("无法获取文本的情绪") return None - text_embedding = await get_embedding(text_for_search) + text_embedding = await get_embedding(text_for_search, request_type="emoji") if not text_embedding: logger.error("无法获取文本的embedding") return None @@ -310,7 +310,7 @@ class EmojiManager: logger.info(f"[检查] 表情包检查通过: {check}") if description is not None: - embedding = await get_embedding(description) + embedding = await get_embedding(description, request_type="emoji") # 准备数据库记录 emoji_record = { "filename": filename, diff --git a/src/plugins/chat/llm_generator.py b/src/plugins/chat/llm_generator.py index 80daa250b..556f36e2e 100644 --- a/src/plugins/chat/llm_generator.py +++ b/src/plugins/chat/llm_generator.py @@ -32,10 +32,17 @@ class ResponseGenerator: temperature=0.7, max_tokens=1000, stream=True, + request_type="response", + ) + self.model_v3 = LLM_request( + model=global_config.llm_normal, temperature=0.7, max_tokens=3000, request_type="response" + ) + self.model_r1_distill = LLM_request( + model=global_config.llm_reasoning_minor, temperature=0.7, max_tokens=3000, request_type="response" + ) + self.model_sum = LLM_request( + model=global_config.llm_summary_by_topic, temperature=0.7, max_tokens=3000, request_type="relation" ) - self.model_v3 = LLM_request(model=global_config.llm_normal, temperature=0.7, max_tokens=3000) - self.model_r1_distill = LLM_request(model=global_config.llm_reasoning_minor, temperature=0.7, max_tokens=3000) - self.model_v25 = LLM_request(model=global_config.llm_normal_minor, temperature=0.7, max_tokens=3000) self.current_model_type = "r1" # 默认使用 R1 self.current_model_name = "unknown model" @@ -175,7 +182,7 @@ class ResponseGenerator: """ # 调用模型生成结果 - result, _, _ = await self.model_v25.generate_response(prompt) + result, _, _ = await self.model_sum.generate_response(prompt) result = result.strip() # 解析模型输出的结果 diff --git a/src/plugins/chat/utils.py b/src/plugins/chat/utils.py index 1563ea526..fd940a645 100644 --- a/src/plugins/chat/utils.py +++ b/src/plugins/chat/utils.py @@ -55,9 +55,9 @@ def is_mentioned_bot_in_message(message: MessageRecv) -> bool: return False -async def get_embedding(text): +async def get_embedding(text, request_type="embedding"): """获取文本的embedding向量""" - llm = LLM_request(model=global_config.embedding, request_type="embedding") + llm = LLM_request(model=global_config.embedding, request_type=request_type) # return llm.get_embedding_sync(text) return await llm.get_embedding(text) diff --git a/src/plugins/memory_system/memory.py b/src/plugins/memory_system/memory.py index f5012c828..ece8de748 100644 --- a/src/plugins/memory_system/memory.py +++ b/src/plugins/memory_system/memory.py @@ -174,9 +174,9 @@ class Memory_graph: class Hippocampus: def __init__(self, memory_graph: Memory_graph): self.memory_graph = memory_graph - self.llm_topic_judge = LLM_request(model=global_config.llm_topic_judge, temperature=0.5, request_type="topic") + self.llm_topic_judge = LLM_request(model=global_config.llm_topic_judge, temperature=0.5, request_type="memory") self.llm_summary_by_topic = LLM_request( - model=global_config.llm_summary_by_topic, temperature=0.5, request_type="topic" + model=global_config.llm_summary_by_topic, temperature=0.5, request_type="memory" ) def get_all_node_names(self) -> list: @@ -375,6 +375,8 @@ class Hippocampus: return topic_num async def operation_build_memory(self): + logger.debug("------------------------------------开始构建记忆--------------------------------------") + start_time = time.time() memory_samples = self.get_memory_sample() all_added_nodes = [] all_added_edges = [] @@ -426,6 +428,12 @@ class Hippocampus: logger.success(f"强化连接: {', '.join(all_added_edges)}") # logger.success(f"强化连接: {', '.join(all_added_edges)}") self.sync_memory_to_db() + + end_time = time.time() + logger.success( + f"--------------------------记忆构建完成:耗时: {end_time - start_time:.2f} " + "秒--------------------------" + ) def sync_memory_to_db(self): """检查并同步内存中的图结构与数据库""" diff --git a/src/plugins/models/utils_model.py b/src/plugins/models/utils_model.py index 91e43fd4f..975bcaf7b 100644 --- a/src/plugins/models/utils_model.py +++ b/src/plugins/models/utils_model.py @@ -581,7 +581,8 @@ class LLM_request: completion_tokens=completion_tokens, total_tokens=total_tokens, user_id="system", # 可以根据需要修改 user_id - request_type="embedding", # 请求类型为 embedding + # request_type="embedding", # 请求类型为 embedding + request_type=self.request_type, # 请求类型为 text endpoint="/embeddings", # API 端点 ) return result["data"][0].get("embedding", None) diff --git a/template/bot_config_template.toml b/template/bot_config_template.toml index e5cf1df86..bf7118d12 100644 --- a/template/bot_config_template.toml +++ b/template/bot_config_template.toml @@ -128,52 +128,60 @@ enable = true #下面的模型若使用硅基流动则不需要更改,使用ds官方则改成.env.prod自定义的宏,使用自定义模型则选择定位相似的模型自己填写 -#推理模型: +#推理模型 + [model.llm_reasoning] #回复模型1 主要回复模型 name = "Pro/deepseek-ai/DeepSeek-R1" +# name = "Qwen/QwQ-32B" provider = "SILICONFLOW" -pri_in = 0 #模型的输入价格(非必填,可以记录消耗) -pri_out = 0 #模型的输出价格(非必填,可以记录消耗) +pri_in = 4 #模型的输入价格(非必填,可以记录消耗) +pri_out = 16 #模型的输出价格(非必填,可以记录消耗) [model.llm_reasoning_minor] #回复模型3 次要回复模型 name = "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" provider = "SILICONFLOW" +pri_in = 1.26 #模型的输入价格(非必填,可以记录消耗) +pri_out = 1.26 #模型的输出价格(非必填,可以记录消耗) #非推理模型 [model.llm_normal] #V3 回复模型2 次要回复模型 name = "Pro/deepseek-ai/DeepSeek-V3" provider = "SILICONFLOW" +pri_in = 2 #模型的输入价格(非必填,可以记录消耗) +pri_out = 8 #模型的输出价格(非必填,可以记录消耗) -[model.llm_normal_minor] #V2.5 -name = "deepseek-ai/DeepSeek-V2.5" -provider = "SILICONFLOW" - -[model.llm_emotion_judge] #主题判断 0.7/m +[model.llm_emotion_judge] #表情包判断 name = "Qwen/Qwen2.5-14B-Instruct" provider = "SILICONFLOW" +pri_in = 0.7 +pri_out = 0.7 -[model.llm_topic_judge] #主题判断:建议使用qwen2.5 7b +[model.llm_topic_judge] #记忆主题判断:建议使用qwen2.5 7b name = "Pro/Qwen/Qwen2.5-7B-Instruct" provider = "SILICONFLOW" +pri_in = 0 +pri_out = 0 -[model.llm_summary_by_topic] #建议使用qwen2.5 32b 及以上 +[model.llm_summary_by_topic] #概括模型,建议使用qwen2.5 32b 及以上 name = "Qwen/Qwen2.5-32B-Instruct" provider = "SILICONFLOW" -pri_in = 0 -pri_out = 0 +pri_in = 1.26 +pri_out = 1.26 -[model.moderation] #内容审核 未启用 +[model.moderation] #内容审核,开发中 name = "" provider = "SILICONFLOW" -pri_in = 0 -pri_out = 0 +pri_in = 1.0 +pri_out = 2.0 # 识图模型 -[model.vlm] #图像识别 0.35/m -name = "Pro/Qwen/Qwen2-VL-7B-Instruct" +[model.vlm] #图像识别 +name = "Pro/Qwen/Qwen2.5-VL-7B-Instruct" provider = "SILICONFLOW" +pri_in = 0.35 +pri_out = 0.35 #嵌入模型 From a7278a37c77787ab24a466f201a8b6b2d3277325 Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Fri, 21 Mar 2025 17:59:13 +0800 Subject: [PATCH 21/21] =?UTF-8?q?better=20cmd=E6=B8=85=E7=90=86=E5=A4=A7?= =?UTF-8?q?=E5=B8=88?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/plugins/chat/utils.py | 2 +- src/plugins/chat/utils_image.py | 2 +- src/plugins/memory_system/memory.py | 16 ++++++++++++---- 3 files changed, 14 insertions(+), 6 deletions(-) diff --git a/src/plugins/chat/utils.py b/src/plugins/chat/utils.py index fd940a645..cc53db623 100644 --- a/src/plugins/chat/utils.py +++ b/src/plugins/chat/utils.py @@ -314,7 +314,7 @@ def split_into_sentences_w_remove_punctuation(text: str) -> List[str]: sentence = sentence.replace(",", " ").replace(",", " ") sentences_done.append(sentence) - logger.info(f"处理后的句子: {sentences_done}") + logger.debug(f"处理后的句子: {sentences_done}") return sentences_done diff --git a/src/plugins/chat/utils_image.py b/src/plugins/chat/utils_image.py index 521795024..7e20b35db 100644 --- a/src/plugins/chat/utils_image.py +++ b/src/plugins/chat/utils_image.py @@ -184,7 +184,7 @@ class ImageManager: logger.warning(f"虽然生成了描述,但是找到缓存图片描述 {cached_description}") return f"[图片:{cached_description}]" - logger.info(f"描述是{description}") + logger.debug(f"描述是{description}") if description is None: logger.warning("AI未能生成图片描述") diff --git a/src/plugins/memory_system/memory.py b/src/plugins/memory_system/memory.py index ece8de748..6efbddd56 100644 --- a/src/plugins/memory_system/memory.py +++ b/src/plugins/memory_system/memory.py @@ -379,6 +379,7 @@ class Hippocampus: start_time = time.time() memory_samples = self.get_memory_sample() all_added_nodes = [] + all_connected_nodes = [] all_added_edges = [] for i, messages in enumerate(memory_samples, 1): all_topics = [] @@ -396,6 +397,7 @@ class Hippocampus: current_time = datetime.datetime.now().timestamp() logger.debug(f"添加节点: {', '.join(topic for topic, _ in compressed_memory)}") all_added_nodes.extend(topic for topic, _ in compressed_memory) + # all_connected_nodes.extend(topic for topic, _ in similar_topics_dict) for topic, memory in compressed_memory: self.memory_graph.add_dot(topic, memory) @@ -407,8 +409,13 @@ class Hippocampus: for similar_topic, similarity in similar_topics: if topic != similar_topic: strength = int(similarity * 10) + logger.debug(f"连接相似节点: {topic} 和 {similar_topic} (强度: {strength})") all_added_edges.append(f"{topic}-{similar_topic}") + + all_connected_nodes.append(topic) + all_connected_nodes.append(similar_topic) + self.memory_graph.G.add_edge( topic, similar_topic, @@ -425,7 +432,8 @@ class Hippocampus: self.memory_graph.connect_dot(all_topics[i], all_topics[j]) logger.success(f"更新记忆: {', '.join(all_added_nodes)}") - logger.success(f"强化连接: {', '.join(all_added_edges)}") + logger.debug(f"强化连接: {', '.join(all_added_edges)}") + logger.info(f"强化连接节点: {', '.join(all_connected_nodes)}") # logger.success(f"强化连接: {', '.join(all_added_edges)}") self.sync_memory_to_db() @@ -860,10 +868,9 @@ class Hippocampus: async def memory_activate_value(self, text: str, max_topics: int = 5, similarity_threshold: float = 0.3) -> int: """计算输入文本对记忆的激活程度""" - logger.info(f"识别主题: {await self._identify_topics(text)}") - # 识别主题 identified_topics = await self._identify_topics(text) + if not identified_topics: return 0 @@ -924,7 +931,8 @@ class Hippocampus: # 计算最终激活值 activation = int((topic_match + average_similarities) / 2 * 100) - logger.info(f"匹配率: {topic_match:.3f}, 平均相似度: {average_similarities:.3f}, 激活值: {activation}") + + logger.info(f"识别主题: {identified_topics}, 匹配率: {topic_match:.3f}, 激活值: {activation}") return activation