增加extra_params功能
This commit is contained in:
@@ -83,6 +83,7 @@ class BaseClient:
|
|||||||
| None = None,
|
| None = None,
|
||||||
async_response_parser: Callable[[ChatCompletion], tuple[APIResponse, tuple[int, int, int]]] | None = None,
|
async_response_parser: Callable[[ChatCompletion], tuple[APIResponse, tuple[int, int, int]]] | None = None,
|
||||||
interrupt_flag: asyncio.Event | None = None,
|
interrupt_flag: asyncio.Event | None = None,
|
||||||
|
extra_params: dict[str, Any] | None = None,
|
||||||
) -> APIResponse:
|
) -> APIResponse:
|
||||||
"""
|
"""
|
||||||
获取对话响应
|
获取对话响应
|
||||||
@@ -103,6 +104,7 @@ class BaseClient:
|
|||||||
self,
|
self,
|
||||||
model_info: ModelInfo,
|
model_info: ModelInfo,
|
||||||
embedding_input: str,
|
embedding_input: str,
|
||||||
|
extra_params: dict[str, Any] | None = None,
|
||||||
) -> APIResponse:
|
) -> APIResponse:
|
||||||
"""
|
"""
|
||||||
获取文本嵌入
|
获取文本嵌入
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
raise DeprecationWarning("Genimi Client is not fully available yet.")
|
||||||
import asyncio
|
import asyncio
|
||||||
import io
|
import io
|
||||||
from collections.abc import Iterable
|
from collections.abc import Iterable
|
||||||
|
|||||||
@@ -396,6 +396,7 @@ class OpenaiClient(BaseClient):
|
|||||||
Callable[[ChatCompletion], tuple[APIResponse, Optional[tuple[int, int, int]]]]
|
Callable[[ChatCompletion], tuple[APIResponse, Optional[tuple[int, int, int]]]]
|
||||||
] = None,
|
] = None,
|
||||||
interrupt_flag: asyncio.Event | None = None,
|
interrupt_flag: asyncio.Event | None = None,
|
||||||
|
extra_params: dict[str, Any] | None = None,
|
||||||
) -> APIResponse:
|
) -> APIResponse:
|
||||||
"""
|
"""
|
||||||
获取对话响应
|
获取对话响应
|
||||||
@@ -434,6 +435,7 @@ class OpenaiClient(BaseClient):
|
|||||||
max_tokens=max_tokens,
|
max_tokens=max_tokens,
|
||||||
stream=True,
|
stream=True,
|
||||||
response_format=NOT_GIVEN,
|
response_format=NOT_GIVEN,
|
||||||
|
extra_body=extra_params,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
while not req_task.done():
|
while not req_task.done():
|
||||||
@@ -455,6 +457,7 @@ class OpenaiClient(BaseClient):
|
|||||||
max_tokens=max_tokens,
|
max_tokens=max_tokens,
|
||||||
stream=False,
|
stream=False,
|
||||||
response_format=NOT_GIVEN,
|
response_format=NOT_GIVEN,
|
||||||
|
extra_body=extra_params,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
while not req_task.done():
|
while not req_task.done():
|
||||||
@@ -487,6 +490,7 @@ class OpenaiClient(BaseClient):
|
|||||||
self,
|
self,
|
||||||
model_info: ModelInfo,
|
model_info: ModelInfo,
|
||||||
embedding_input: str,
|
embedding_input: str,
|
||||||
|
extra_params: dict[str, Any] | None = None,
|
||||||
) -> APIResponse:
|
) -> APIResponse:
|
||||||
"""
|
"""
|
||||||
获取文本嵌入
|
获取文本嵌入
|
||||||
@@ -498,6 +502,7 @@ class OpenaiClient(BaseClient):
|
|||||||
raw_response = await self.client.embeddings.create(
|
raw_response = await self.client.embeddings.create(
|
||||||
model=model_info.model_identifier,
|
model=model_info.model_identifier,
|
||||||
input=embedding_input,
|
input=embedding_input,
|
||||||
|
extra_body=extra_params,
|
||||||
)
|
)
|
||||||
except APIConnectionError as e:
|
except APIConnectionError as e:
|
||||||
raise NetworkConnectionError() from e
|
raise NetworkConnectionError() from e
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
[inner]
|
[inner]
|
||||||
version = "1.1.0"
|
version = "1.1.1"
|
||||||
|
|
||||||
# 配置文件版本号迭代规则同bot_config.toml
|
# 配置文件版本号迭代规则同bot_config.toml
|
||||||
|
|
||||||
@@ -59,6 +59,8 @@ name = "qwen3-8b"
|
|||||||
api_provider = "SiliconFlow"
|
api_provider = "SiliconFlow"
|
||||||
price_in = 0
|
price_in = 0
|
||||||
price_out = 0
|
price_out = 0
|
||||||
|
[models.extra_params] # 可选的额外参数配置
|
||||||
|
enable_thinking = false # 不启用思考
|
||||||
|
|
||||||
[[models]]
|
[[models]]
|
||||||
model_identifier = "Qwen/Qwen3-14B"
|
model_identifier = "Qwen/Qwen3-14B"
|
||||||
@@ -66,6 +68,8 @@ name = "qwen3-14b"
|
|||||||
api_provider = "SiliconFlow"
|
api_provider = "SiliconFlow"
|
||||||
price_in = 0.5
|
price_in = 0.5
|
||||||
price_out = 2.0
|
price_out = 2.0
|
||||||
|
[models.extra_params] # 可选的额外参数配置
|
||||||
|
enable_thinking = false # 不启用思考
|
||||||
|
|
||||||
[[models]]
|
[[models]]
|
||||||
model_identifier = "Qwen/Qwen3-30B-A3B"
|
model_identifier = "Qwen/Qwen3-30B-A3B"
|
||||||
@@ -73,6 +77,8 @@ name = "qwen3-30b"
|
|||||||
api_provider = "SiliconFlow"
|
api_provider = "SiliconFlow"
|
||||||
price_in = 0.7
|
price_in = 0.7
|
||||||
price_out = 2.8
|
price_out = 2.8
|
||||||
|
[models.extra_params] # 可选的额外参数配置
|
||||||
|
enable_thinking = false # 不启用思考
|
||||||
|
|
||||||
[[models]]
|
[[models]]
|
||||||
model_identifier = "Qwen/Qwen2.5-VL-72B-Instruct"
|
model_identifier = "Qwen/Qwen2.5-VL-72B-Instruct"
|
||||||
|
|||||||
Reference in New Issue
Block a user