fix: 更新模型配置和环境变量,调整版本号和模型参数

This commit is contained in:
墨梓柒
2025-07-27 13:37:21 +08:00
parent 2335ec6577
commit e240fb92ca
7 changed files with 194 additions and 323 deletions

View File

@@ -229,120 +229,84 @@ show_prompt = false # 是否显示prompt
[model]
model_max_output_length = 1024 # 模型单次返回的最大token数
model_max_output_length = 800 # 模型单次返回的最大token数
#------------必填:组件模型------------
#------------模型任务配置------------
# 所有模型名称需要对应 model_config.toml 中配置的模型名称
[model.utils] # 在麦麦的一些组件中使用的模型,例如表情包模块,取名模块,关系模块,是麦麦必须的模型
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2 #模型的输入价格(非必填,可以记录消耗)
pri_out = 8 #模型的输出价格(非必填,可以记录消耗)
#默认temp 0.2 如果你使用的是老V3或者其他模型请自己修改temp参数
temp = 0.2 #模型的温度新V3建议0.1-0.3
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.2 # 模型温度新V3建议0.1-0.3
max_tokens = 800 # 最大输出token数
[model.utils_small] # 在麦麦的一些组件中使用的小模型,消耗量较大,建议使用速度较快的小模型
# 强烈建议使用免费的小模型
name = "Qwen/Qwen3-8B"
provider = "SILICONFLOW"
pri_in = 0
pri_out = 0
temp = 0.7
model_name = "qwen3-8b" # 对应 model_config.toml 中的模型名称
temperature = 0.7
max_tokens = 800
enable_thinking = false # 是否启用思考
[model.replyer_1] # 首要回复模型,还用于表达器和表达方式学习
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2 #模型的输入价格(非必填,可以记录消耗)
pri_out = 8 #模型的输出价格(非必填,可以记录消耗)
#默认temp 0.2 如果你使用的是老V3或者其他模型请自己修改temp参数
temp = 0.2 #模型的温度新V3建议0.1-0.3
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.2 # 模型温度新V3建议0.1-0.3
max_tokens = 800
[model.replyer_2] # 次要回复模型
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2 #模型的输入价格(非必填,可以记录消耗)
pri_out = 8 #模型的输出价格(非必填,可以记录消耗)
#默认temp 0.2 如果你使用的是老V3或者其他模型请自己修改temp参数
temp = 0.2 #模型的温度新V3建议0.1-0.3
model_name = "siliconflow-deepseek-r1" # 对应 model_config.toml 中的模型名称
temperature = 0.7 # 模型温度
max_tokens = 800
[model.planner] #决策:负责决定麦麦该做什么的模型
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2
pri_out = 8
temp = 0.3
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.3
max_tokens = 800
[model.emotion] #负责麦麦的情绪变化
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2
pri_out = 8
temp = 0.3
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.3
max_tokens = 800
[model.memory] # 记忆模型
name = "Qwen/Qwen3-30B-A3B"
provider = "SILICONFLOW"
pri_in = 0.7
pri_out = 2.8
temp = 0.7
model_name = "qwen3-30b" # 对应 model_config.toml 中的模型名称
temperature = 0.7
max_tokens = 800
enable_thinking = false # 是否启用思考
[model.vlm] # 图像识别模型
name = "Pro/Qwen/Qwen2.5-VL-7B-Instruct"
provider = "SILICONFLOW"
pri_in = 0.35
pri_out = 0.35
model_name = "qwen2.5-vl-72b" # 对应 model_config.toml 中的模型名称
max_tokens = 800
[model.voice] # 语音识别模型
name = "FunAudioLLM/SenseVoiceSmall"
provider = "SILICONFLOW"
pri_in = 0
pri_out = 0
model_name = "sensevoice-small" # 对应 model_config.toml 中的模型名称
[model.tool_use] #工具调用模型,需要使用支持工具调用的模型
name = "Qwen/Qwen3-14B"
provider = "SILICONFLOW"
pri_in = 0.5
pri_out = 2
temp = 0.7
model_name = "qwen3-14b" # 对应 model_config.toml 中的模型名称
temperature = 0.7
max_tokens = 800
enable_thinking = false # 是否启用思考qwen3 only
#嵌入模型
[model.embedding]
name = "BAAI/bge-m3"
provider = "SILICONFLOW"
pri_in = 0
pri_out = 0
model_name = "bge-m3" # 对应 model_config.toml 中的模型名称
#------------LPMM知识库模型------------
[model.lpmm_entity_extract] # 实体提取模型
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2
pri_out = 8
temp = 0.2
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.2
max_tokens = 800
[model.lpmm_rdf_build] # RDF构建模型
name = "Pro/deepseek-ai/DeepSeek-V3"
provider = "SILICONFLOW"
pri_in = 2
pri_out = 8
temp = 0.2
model_name = "siliconflow-deepseek-v3" # 对应 model_config.toml 中的模型名称
temperature = 0.2
max_tokens = 800
[model.lpmm_qa] # 问答模型
name = "Qwen/Qwen3-30B-A3B"
provider = "SILICONFLOW"
pri_in = 0.7
pri_out = 2.8
temp = 0.7
model_name = "deepseek-r1-distill-qwen-32b" # 对应 model_config.toml 中的模型名称
temperature = 0.7
max_tokens = 800
enable_thinking = false # 是否启用思考
[maim_message]
auth_token = [] # 认证令牌用于API验证为空则不启用验证
# 以下项目若要使用需要打开use_custom并单独配置maim_message的服务器

View File

@@ -1,5 +1,5 @@
[inner]
version = "0.1.0"
version = "0.1.1"
# 配置文件版本号迭代规则同bot_config.toml
@@ -50,21 +50,79 @@ price_out = 8.0
#可选若无该字段默认值为false
#force_stream_mode = true
#[[models]]
#model_identifier = "deepseek-reasoner"
#name = "deepseek-r1"
#api_provider = "DeepSeek"
#model_flags = ["text", "tool_calling", "reasoning"]
#price_in = 4.0
#price_out = 16.0
#
#[[models]]
#model_identifier = "BAAI/bge-m3"
#name = "siliconflow-bge-m3"
#api_provider = "SiliconFlow"
#model_flags = ["text", "embedding"]
#price_in = 0
#price_out = 0
[[models]]
model_identifier = "deepseek-reasoner"
name = "deepseek-r1"
api_provider = "DeepSeek"
model_flags = [ "text", "tool_calling", "reasoning",]
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-V3"
name = "siliconflow-deepseek-v3"
api_provider = "SiliconFlow"
price_in = 2.0
price_out = 8.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-R1"
name = "siliconflow-deepseek-r1"
api_provider = "SiliconFlow"
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B"
name = "deepseek-r1-distill-qwen-32b"
api_provider = "SiliconFlow"
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Qwen/Qwen3-8B"
name = "qwen3-8b"
api_provider = "SiliconFlow"
price_in = 0
price_out = 0
[[models]]
model_identifier = "Qwen/Qwen3-14B"
name = "qwen3-14b"
api_provider = "SiliconFlow"
price_in = 0.5
price_out = 2.0
[[models]]
model_identifier = "Qwen/Qwen3-30B-A3B"
name = "qwen3-30b"
api_provider = "SiliconFlow"
price_in = 0.7
price_out = 2.8
[[models]]
model_identifier = "Qwen/Qwen2.5-VL-72B-Instruct"
name = "qwen2.5-vl-72b"
api_provider = "SiliconFlow"
model_flags = [ "vision", "text",]
price_in = 4.13
price_out = 4.13
[[models]]
model_identifier = "FunAudioLLM/SenseVoiceSmall"
name = "sensevoice-small"
api_provider = "SiliconFlow"
model_flags = [ "audio",]
price_in = 0
price_out = 0
[[models]]
model_identifier = "BAAI/bge-m3"
name = "bge-m3"
api_provider = "SiliconFlow"
model_flags = [ "text", "embedding",]
price_in = 0
price_out = 0
[task_model_usage]

View File

@@ -1,5 +1,5 @@
[inner]
version = "0.1.0"
version = "0.1.1"
# 配置文件版本号迭代规则同bot_config.toml
@@ -50,27 +50,85 @@ price_out = 8.0
#可选若无该字段默认值为false
#force_stream_mode = true
#[[models]]
#model_identifier = "deepseek-reasoner"
#name = "deepseek-r1"
#api_provider = "DeepSeek"
#model_flags = ["text", "tool_calling", "reasoning"]
#price_in = 4.0
#price_out = 16.0
#
#[[models]]
#model_identifier = "BAAI/bge-m3"
#name = "siliconflow-bge-m3"
#api_provider = "SiliconFlow"
#model_flags = ["text", "embedding"]
#price_in = 0
#price_out = 0
[[models]]
model_identifier = "deepseek-reasoner"
name = "deepseek-r1"
api_provider = "DeepSeek"
model_flags = [ "text", "tool_calling", "reasoning",]
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-V3"
name = "siliconflow-deepseek-v3"
api_provider = "SiliconFlow"
price_in = 2.0
price_out = 8.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-R1"
name = "siliconflow-deepseek-r1"
api_provider = "SiliconFlow"
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B"
name = "deepseek-r1-distill-qwen-32b"
api_provider = "SiliconFlow"
price_in = 4.0
price_out = 16.0
[[models]]
model_identifier = "Qwen/Qwen3-8B"
name = "qwen3-8b"
api_provider = "SiliconFlow"
price_in = 0
price_out = 0
[[models]]
model_identifier = "Qwen/Qwen3-14B"
name = "qwen3-14b"
api_provider = "SiliconFlow"
price_in = 0.5
price_out = 2.0
[[models]]
model_identifier = "Qwen/Qwen3-30B-A3B"
name = "qwen3-30b"
api_provider = "SiliconFlow"
price_in = 0.7
price_out = 2.8
[[models]]
model_identifier = "Qwen/Qwen2.5-VL-72B-Instruct"
name = "qwen2.5-vl-72b"
api_provider = "SiliconFlow"
model_flags = [ "vision", "text",]
price_in = 4.13
price_out = 4.13
[[models]]
model_identifier = "FunAudioLLM/SenseVoiceSmall"
name = "sensevoice-small"
api_provider = "SiliconFlow"
model_flags = [ "audio",]
price_in = 0
price_out = 0
[[models]]
model_identifier = "BAAI/bge-m3"
name = "bge-m3"
api_provider = "SiliconFlow"
model_flags = [ "text", "embedding",]
price_in = 0
price_out = 0
[task_model_usage]
#llm_reasoning = {model="deepseek-r1", temperature=0.8, max_tokens=1024, max_retry=0}
#llm_normal = {model="deepseek-r1", max_tokens=1024, max_retry=0}
#embedding = "siliconflow-bge-m3"
llm_reasoning = {model="deepseek-r1", temperature=0.8, max_tokens=1024, max_retry=0}
llm_normal = {model="deepseek-r1", max_tokens=1024, max_retry=0}
embedding = "siliconflow-bge-m3"
#schedule = [
# "deepseek-v3",
# "deepseek-r1",

View File

@@ -1,16 +1,2 @@
HOST=127.0.0.1
PORT=8000
#key and url
SILICONFLOW_BASE_URL=https://api.siliconflow.cn/v1/
DEEP_SEEK_BASE_URL=https://api.deepseek.com/v1
CHAT_ANY_WHERE_BASE_URL=https://api.chatanywhere.tech/v1
BAILIAN_BASE_URL = https://dashscope.aliyuncs.com/compatible-mode/v1
xxxxxxx_BASE_URL=https://xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
# 定义你要用的api的key(需要去对应网站申请哦)
DEEP_SEEK_KEY=
CHAT_ANY_WHERE_KEY=
SILICONFLOW_KEY=
BAILIAN_KEY =
xxxxxxx_KEY=
PORT=8000