From 16a704e01f7967d7a9c0ee16998c2dd6da0cbcd9 Mon Sep 17 00:00:00 2001 From: SengokuCola <1026294844@qq.com> Date: Mon, 9 Jun 2025 00:33:28 +0800 Subject: [PATCH] fix: remove token --- src/chat/focus_chat/info_processors/relationship_processor.py | 1 - src/chat/focus_chat/info_processors/self_processor.py | 1 - src/chat/focus_chat/info_processors/tool_processor.py | 1 - src/chat/focus_chat/info_processors/working_memory_processor.py | 1 - 4 files changed, 4 deletions(-) diff --git a/src/chat/focus_chat/info_processors/relationship_processor.py b/src/chat/focus_chat/info_processors/relationship_processor.py index d3654502f..656f01a0f 100644 --- a/src/chat/focus_chat/info_processors/relationship_processor.py +++ b/src/chat/focus_chat/info_processors/relationship_processor.py @@ -94,7 +94,6 @@ class RelationshipProcessor(BaseProcessor): self.llm_model = LLMRequest( model=global_config.model.relation, - max_tokens=800, request_type="focus.relationship", ) diff --git a/src/chat/focus_chat/info_processors/self_processor.py b/src/chat/focus_chat/info_processors/self_processor.py index 450afdbab..36dc3c950 100644 --- a/src/chat/focus_chat/info_processors/self_processor.py +++ b/src/chat/focus_chat/info_processors/self_processor.py @@ -56,7 +56,6 @@ class SelfProcessor(BaseProcessor): self.llm_model = LLMRequest( model=global_config.model.relation, - max_tokens=800, request_type="focus.processor.self_identify", ) diff --git a/src/chat/focus_chat/info_processors/tool_processor.py b/src/chat/focus_chat/info_processors/tool_processor.py index 5edad5fff..cf31f4418 100644 --- a/src/chat/focus_chat/info_processors/tool_processor.py +++ b/src/chat/focus_chat/info_processors/tool_processor.py @@ -43,7 +43,6 @@ class ToolProcessor(BaseProcessor): self.log_prefix = f"[{subheartflow_id}:ToolExecutor] " self.llm_model = LLMRequest( model=global_config.model.focus_tool_use, - max_tokens=500, request_type="focus.processor.tool", ) self.structured_info = [] diff --git a/src/chat/focus_chat/info_processors/working_memory_processor.py b/src/chat/focus_chat/info_processors/working_memory_processor.py index d40b3c93b..9eb848089 100644 --- a/src/chat/focus_chat/info_processors/working_memory_processor.py +++ b/src/chat/focus_chat/info_processors/working_memory_processor.py @@ -61,7 +61,6 @@ class WorkingMemoryProcessor(BaseProcessor): self.llm_model = LLMRequest( model=global_config.model.planner, - max_tokens=800, request_type="focus.processor.working_memory", )