Merge pull request #441 from Maple127667/main-fix

token统计部分
This commit is contained in:
SengokuCola
2025-03-17 17:23:18 +08:00
committed by GitHub
7 changed files with 43 additions and 13 deletions

View File

@@ -38,9 +38,9 @@ class EmojiManager:
def __init__(self):
self._scan_task = None
self.vlm = LLM_request(model=global_config.vlm, temperature=0.3, max_tokens=1000)
self.vlm = LLM_request(model=global_config.vlm, temperature=0.3, max_tokens=1000,request_type = 'image')
self.llm_emotion_judge = LLM_request(
model=global_config.llm_emotion_judge, max_tokens=600, temperature=0.8
model=global_config.llm_emotion_judge, max_tokens=600, temperature=0.8,request_type = 'image'
) # 更高的温度更少的token后续可以根据情绪来调整温度
def _ensure_emoji_dir(self):

View File

@@ -14,7 +14,7 @@ config = driver.config
class TopicIdentifier:
def __init__(self):
self.llm_topic_judge = LLM_request(model=global_config.llm_topic_judge)
self.llm_topic_judge = LLM_request(model=global_config.llm_topic_judge,request_type = 'topic')
async def identify_topic_llm(self, text: str) -> Optional[List[str]]:
"""识别消息主题,返回主题列表"""

View File

@@ -54,7 +54,7 @@ def is_mentioned_bot_in_message(message: MessageRecv) -> bool:
async def get_embedding(text):
"""获取文本的embedding向量"""
llm = LLM_request(model=global_config.embedding)
llm = LLM_request(model=global_config.embedding,request_type = 'embedding')
# return llm.get_embedding_sync(text)
return await llm.get_embedding(text)

View File

@@ -37,7 +37,7 @@ class ImageManager:
self._ensure_description_collection()
self._ensure_image_dir()
self._initialized = True
self._llm = LLM_request(model=global_config.vlm, temperature=0.4, max_tokens=1000)
self._llm = LLM_request(model=global_config.vlm, temperature=0.4, max_tokens=1000,request_type = 'image')
def _ensure_image_dir(self):
"""确保图像存储目录存在"""