refactor(llm): 精简故障转移执行器的调用逻辑
`_execute_with_failover` 方法在 `raise_on_failure=True` 时,保证了要么返回有效结果,要么抛出异常。 因此,先前在调用该方法后的 `if result:` 判断和 `raise RuntimeError` 语句是冗余且不可达的。本次重构移除了这些不必要的代码,直接返回该方法的调用结果,使代码更加简洁和清晰。
This commit is contained in:
@@ -209,12 +209,7 @@ class LLMRequest:
|
||||
)
|
||||
return content, (reasoning_content, model_info.name, tool_calls)
|
||||
|
||||
result = await self._execute_with_failover(request_callable=request_logic, raise_on_failure=True)
|
||||
if result:
|
||||
return result
|
||||
|
||||
# 这段代码理论上不可达,因为 raise_on_failure=True 会抛出异常
|
||||
raise RuntimeError("图片响应生成失败,所有模型均尝试失败。")
|
||||
return await self._execute_with_failover(request_callable=request_logic, raise_on_failure=True)
|
||||
|
||||
async def generate_response_for_voice(self, voice_base64: str) -> Optional[str]:
|
||||
"""
|
||||
@@ -237,8 +232,7 @@ class LLMRequest:
|
||||
return response.content or None
|
||||
|
||||
# 对于语音识别,如果所有模型都失败,我们可能不希望程序崩溃,而是返回None
|
||||
result = await self._execute_with_failover(request_callable=request_logic, raise_on_failure=False)
|
||||
return result
|
||||
return await self._execute_with_failover(request_callable=request_logic, raise_on_failure=False)
|
||||
|
||||
async def generate_response_async(
|
||||
self,
|
||||
@@ -507,12 +501,7 @@ class LLMRequest:
|
||||
|
||||
return embedding, model_info.name
|
||||
|
||||
result = await self._execute_with_failover(request_callable=request_logic, raise_on_failure=True)
|
||||
if result:
|
||||
return result
|
||||
|
||||
# 这段代码理论上不可达,因为 raise_on_failure=True 会抛出异常
|
||||
raise RuntimeError("获取 embedding 失败,所有模型均尝试失败。")
|
||||
return await self._execute_with_failover(request_callable=request_logic, raise_on_failure=True)
|
||||
|
||||
def _model_scheduler(
|
||||
self, failed_models: set | None = None
|
||||
@@ -674,7 +663,7 @@ class LLMRequest:
|
||||
cannot_retry_msg=f"任务-'{task_name}' 模型-'{model_name}': 连接异常,超过最大重试次数,请检查网络连接状态或URL是否正确",
|
||||
)
|
||||
elif isinstance(e, ReqAbortException):
|
||||
logger.warning(f"任务-'{task_name}' 模型-'{model_name}': 请求被中断,详细信息-{str(e.message)}")
|
||||
logger.warning(f"任务-'{task_name}' 模型-'{model_name}': 请求被中断,详细信息-{e}")
|
||||
return -1, None # 不再重试请求该模型
|
||||
elif isinstance(e, RespNotOkException):
|
||||
return self._handle_resp_not_ok(
|
||||
@@ -688,7 +677,7 @@ class LLMRequest:
|
||||
)
|
||||
elif isinstance(e, RespParseException):
|
||||
# 响应解析错误
|
||||
logger.error(f"任务-'{task_name}' 模型-'{model_name}': 响应解析错误,错误信息-{e.message}")
|
||||
logger.error(f"任务-'{task_name}' 模型-'{model_name}': 响应解析错误,错误信息-{e}")
|
||||
logger.debug(f"附加内容: {str(e.ext_info)}")
|
||||
return -1, None # 不再重试请求该模型
|
||||
else:
|
||||
|
||||
Reference in New Issue
Block a user