Merge branch 'debug' of https://github.com/SengokuCola/MaiMBot into debug

This commit is contained in:
HYY1116
2025-03-02 19:48:31 +08:00
6 changed files with 43 additions and 19 deletions

26
.env Normal file
View File

@@ -0,0 +1,26 @@
# 您不应该修改默认值,这个文件被仓库索引,请修改.env.prod
ENVIRONMENT=prod
HOST=127.0.0.1
PORT=8080
COMMAND_START=["/"]
# 插件配置
PLUGINS=["src2.plugins.chat"]
# 默认配置
MONGODB_HOST=127.0.0.1
MONGODB_PORT=27017
DATABASE_NAME=MegBot
MONGODB_USERNAME = "" # 默认空值
MONGODB_PASSWORD = "" # 默认空值
MONGODB_AUTH_SOURCE = "" # 默认空值
#key and url
CHAT_ANY_WHERE_KEY=
SILICONFLOW_KEY=
CHAT_ANY_WHERE_BASE_URL=https://api.chatanywhere.tech/v1
SILICONFLOW_BASE_URL=https://api.siliconflow.cn/v1/
DEEP_SEEK_KEY=
DEEP_SEEK_BASE_URL=https://api.deepseek.com/v1

3
.gitignore vendored
View File

@@ -14,7 +14,7 @@ reasoning_content.bat
reasoning_window.bat
queue_update.txt
memory_graph.gml
.env.dev
.env.*
# Byte-compiled / optimized / DLL files
@@ -147,7 +147,6 @@ celerybeat.pid
*.sage.py
# Environments
.env
.venv
env/
venv/

View File

@@ -2,8 +2,7 @@ FROM nonebot/nb-cli:latest
WORKDIR /
COPY . /MaiMBot/
WORKDIR /MaiMBot
RUN mv config/env.example config/.env \
&& mv config/bot_config_toml config/bot_config.toml
RUN mv config/bot_config_toml config/bot_config.toml
RUN pip install --upgrade -r requirements.txt
VOLUME [ "/MaiMBot/config" ]
EXPOSE 8080

View File

@@ -27,7 +27,7 @@ services:
- mongodb:/data/db
- mongodbCONFIG:/data/configdb
image: mongo:latest
maimbot:
container_name: maimbot
environment:
@@ -41,8 +41,8 @@ services:
volumes:
- maimbotCONFIG:/MaiMBot/config
- maimbotDATA:/MaiMBot/data
- ./.env.prod:/MaiMBot/.env.prod
image: sengokucola/maimbot:latest
volumes:
maimbotCONFIG:
@@ -51,4 +51,5 @@ volumes:
napcatCONFIG:
mongodb:
mongodbCONFIG:

Binary file not shown.

View File

@@ -20,7 +20,7 @@ class LLMModel:
self.model_name = model_name
self.params = kwargs
async def generate_response(self, prompt: str) -> Tuple[str, str]:
def generate_response(self, prompt: str) -> Tuple[str, str]:
"""根据输入的提示生成模型的响应"""
headers = {
"Authorization": f"Bearer {self.api_key}",
@@ -39,17 +39,16 @@ class LLMModel:
api_url = f"{self.base_url.rstrip('/')}/chat/completions"
try:
async with aiohttp.ClientSession() as session:
async with session.post(api_url, headers=headers, json=data) as response:
response.raise_for_status() # 检查响应状态
result = await response.json()
if "choices" in result and len(result["choices"]) > 0:
content = result["choices"][0]["message"]["content"]
reasoning_content = result["choices"][0]["message"].get("reasoning_content", "")
return content, reasoning_content # 返回内容和推理内容
return "没有返回结果", "" # 返回两个值
response = requests.post(api_url, headers=headers, json=data)
response.raise_for_status() # 检查响应状态
result = response.json()
if "choices" in result and len(result["choices"]) > 0:
content = result["choices"][0]["message"]["content"]
reasoning_content = result["choices"][0]["message"].get("reasoning_content", "")
return content, reasoning_content # 返回内容和推理内容
return "没有返回结果", "" # 返回两个值
except Exception as e:
return f"请求失败: {str(e)}", "" # 返回错误信息和空字符串