fix:更新统计信息罢了
This commit is contained in:
@@ -26,10 +26,10 @@ class ResponseGenerator:
|
||||
model=global_config.llm_reasoning,
|
||||
temperature=0.7,
|
||||
max_tokens=3000,
|
||||
request_type="response",
|
||||
request_type="response_reasoning",
|
||||
)
|
||||
self.model_normal = LLM_request(
|
||||
model=global_config.llm_normal, temperature=0.8, max_tokens=256, request_type="response"
|
||||
model=global_config.llm_normal, temperature=0.8, max_tokens=256, request_type="response_reasoning"
|
||||
)
|
||||
|
||||
self.model_sum = LLM_request(
|
||||
|
||||
@@ -148,7 +148,7 @@ class ThinkFlowChat:
|
||||
|
||||
if groupinfo.group_id not in global_config.talk_allowed_groups:
|
||||
return
|
||||
logger.info("使用思维流聊天模式")
|
||||
# logger.info("使用思维流聊天模式")
|
||||
|
||||
# 创建聊天流
|
||||
chat = await chat_manager.get_or_create_stream(
|
||||
|
||||
@@ -23,7 +23,7 @@ logger = get_module_logger("llm_generator", config=llm_config)
|
||||
class ResponseGenerator:
|
||||
def __init__(self):
|
||||
self.model_normal = LLM_request(
|
||||
model=global_config.llm_normal, temperature=0.8, max_tokens=256, request_type="response"
|
||||
model=global_config.llm_normal, temperature=0.8, max_tokens=256, request_type="response_heartflow"
|
||||
)
|
||||
|
||||
self.model_sum = LLM_request(
|
||||
|
||||
Reference in New Issue
Block a user