better:新增config项目

This commit is contained in:
SengokuCola
2025-04-14 22:09:00 +08:00
parent 53332791e4
commit d9f191705f
11 changed files with 51 additions and 31 deletions

View File

@@ -29,7 +29,7 @@ class ResponseGenerator:
request_type="response_reasoning",
)
self.model_normal = LLM_request(
model=global_config.llm_normal, temperature=0.8, max_tokens=256, request_type="response_reasoning"
model=global_config.llm_normal, temperature=global_config.llm_normal["temp"], max_tokens=256, request_type="response_reasoning"
)
self.model_sum = LLM_request(

View File

@@ -26,7 +26,7 @@ logger = get_module_logger("llm_generator", config=llm_config)
class ResponseGenerator:
def __init__(self):
self.model_normal = LLM_request(
model=global_config.llm_normal, temperature=0.15, max_tokens=256, request_type="response_heartflow"
model=global_config.llm_normal, temperature=global_config.llm_normal["temp"], max_tokens=256, request_type="response_heartflow"
)
self.model_sum = LLM_request(
@@ -49,7 +49,7 @@ class ResponseGenerator:
if random.random() > 0:
checked = False
current_model = self.model_normal
current_model.temperature = 0.3 * arousal_multiplier # 激活度越高,温度越高
current_model.temperature = global_config.llm_normal["temp"] * arousal_multiplier # 激活度越高,温度越高
model_response = await self._generate_response_with_model(
message, current_model, thinking_id, mode="normal"
)
@@ -58,13 +58,13 @@ class ResponseGenerator:
else:
checked = True
current_model = self.model_normal
current_model.temperature = 0.3 * arousal_multiplier # 激活度越高,温度越高
current_model.temperature = global_config.llm_normal["temp"] * arousal_multiplier # 激活度越高,温度越高
print(f"生成{message.processed_plain_text}回复温度是:{current_model.temperature}")
model_response = await self._generate_response_with_model(
message, current_model, thinking_id, mode="simple"
)
current_model.temperature = 0.3
current_model.temperature = global_config.llm_normal["temp"]
model_checked_response = await self._check_response_with_model(
message, model_response, current_model, thinking_id
)