From 0583495548f8fcd8f3ad46b3eb68fdec08e5dc7d Mon Sep 17 00:00:00 2001 From: jxxghp Date: Thu, 30 Apr 2026 17:10:14 +0800 Subject: [PATCH] refactor: remove legacy disable_thinking and reasoning_effort parameters from LLM helper and related tests --- app/agent/llm/helper.py | 32 ++++--------------------------- app/api/endpoints/llm.py | 8 +------- tests/test_llm_helper_testcall.py | 3 --- tests/test_system_llm_test.py | 8 -------- 4 files changed, 5 insertions(+), 46 deletions(-) diff --git a/app/agent/llm/helper.py b/app/agent/llm/helper.py index 5d70ffda..c9f8b9f0 100644 --- a/app/agent/llm/helper.py +++ b/app/agent/llm/helper.py @@ -471,14 +471,9 @@ class LLMHelper: def _resolve_thinking_level( cls, thinking_level: str | None = None, - disable_thinking: bool | None = None, - reasoning_effort: str | None = None, ) -> str | None: """ 统一兼容新旧 thinking 参数。 - - 新前端只会传 `thinking_level`,但测试和部分旧调用仍可能带 - `disable_thinking` / `reasoning_effort`,这里集中做一次归一化。 """ def _normalize(value: str | None) -> str | None: @@ -504,12 +499,7 @@ class LLMHelper: if normalized_thinking_level: return normalized_thinking_level - legacy_effort = _normalize(reasoning_effort) - if disable_thinking: - return "off" - if disable_thinking is False: - return legacy_effort or "auto" - return legacy_effort + return "off" @classmethod async def get_llm( @@ -518,10 +508,8 @@ class LLMHelper: provider: str | None = None, model: str | None = None, thinking_level: str | None = None, - disable_thinking: bool | None = None, - reasoning_effort: str | None = None, - api_key: str | None = None, - base_url: str | None = None, + api_key: str | None = settings.LLM_API_KEY, + base_url: str | None = settings.LLM_BASE_URL, ): """ 获取LLM实例 @@ -532,13 +520,7 @@ class LLMHelper: 是否启用思考模式)。支持的级别包括 "off"(关闭)、"auto"(自动)、"minimal"、"low"、"medium"、"high"、"max"/"xhigh"(最大)。 不同模型对思考模式的支持和表现不同,具体映射关系请 参考代码实现。对于不支持思考模式的模型,该参数将被忽略。 - :param disable_thinking: 兼容旧参数,若传入则会被转换为新的 - `thinking_level` 语义。 - :param reasoning_effort: 兼容旧参数,若传入则会被转换为新的 - `thinking_level` 语义。 - :param api_key: API Key,默认为 - 配置项LLM_API_KEY。对于某些提供商( - 如 DeepSeek),可能需要同时提供 base_url。 + :param api_key: API Key,默认为配置项LLM_API_KEY。对于某些提供商(如 DeepSeek),可能需要同时提供 base_url。 :param base_url: API Base URL,默认为配置项LLM_BASE_URL。 :return: LLM实例 """ @@ -546,8 +528,6 @@ class LLMHelper: model_name = model if model is not None else settings.LLM_MODEL normalized_thinking_level = cls._resolve_thinking_level( thinking_level=thinking_level, - disable_thinking=disable_thinking, - reasoning_effort=reasoning_effort, ) try: # 延迟导入,避免单测在最小 stub 环境下 import `llm.py` 时被 provider @@ -718,8 +698,6 @@ class LLMHelper: provider: str | None = None, model: str | None = None, thinking_level: str | None = None, - disable_thinking: bool | None = None, - reasoning_effort: str | None = None, api_key: str | None = None, base_url: str | None = None, ) -> dict: @@ -734,8 +712,6 @@ class LLMHelper: provider=provider_name, model=model_name, thinking_level=thinking_level, - disable_thinking=disable_thinking, - reasoning_effort=reasoning_effort, api_key=api_key, base_url=base_url, ) diff --git a/app/api/endpoints/llm.py b/app/api/endpoints/llm.py index 08819e0e..369b59c5 100644 --- a/app/api/endpoints/llm.py +++ b/app/api/endpoints/llm.py @@ -28,8 +28,6 @@ class LlmTestRequest(BaseModel): provider: Optional[str] = None model: Optional[str] = None thinking_level: Optional[str] = None - disable_thinking: Optional[bool] = None - reasoning_effort: Optional[str] = None api_key: Optional[str] = None base_url: Optional[str] = None @@ -231,9 +229,7 @@ async def llm_test( enabled=settings.AI_AGENT_ENABLE, provider=settings.LLM_PROVIDER, model=settings.LLM_MODEL, - thinking_level=getattr(settings, "LLM_THINKING_LEVEL", None), - disable_thinking=getattr(settings, "LLM_DISABLE_THINKING", None), - reasoning_effort=getattr(settings, "LLM_REASONING_EFFORT", None), + thinking_level=settings.LLM_THINKING_LEVEL, api_key=settings.LLM_API_KEY, base_url=settings.LLM_BASE_URL, ) @@ -265,8 +261,6 @@ async def llm_test( provider=payload.provider, model=payload.model, thinking_level=payload.thinking_level, - disable_thinking=payload.disable_thinking, - reasoning_effort=payload.reasoning_effort, api_key=payload.api_key, base_url=payload.base_url, ) diff --git a/tests/test_llm_helper_testcall.py b/tests/test_llm_helper_testcall.py index 81a7362d..4747d77d 100644 --- a/tests/test_llm_helper_testcall.py +++ b/tests/test_llm_helper_testcall.py @@ -84,8 +84,6 @@ class LlmHelperTestCallTest(unittest.TestCase): provider="deepseek", model="deepseek-chat", thinking_level=None, - disable_thinking=None, - reasoning_effort=None, api_key="sk-test", base_url="https://api.deepseek.com", ) @@ -132,7 +130,6 @@ class LlmHelperTestCallTest(unittest.TestCase): llm_module.LLMHelper.get_llm( provider="openai", model="kimi-k2.6", - disable_thinking=True, api_key="sk-test", base_url="https://kimi.example.com/v1", ) diff --git a/tests/test_system_llm_test.py b/tests/test_system_llm_test.py index 6714a706..b9866d8b 100644 --- a/tests/test_system_llm_test.py +++ b/tests/test_system_llm_test.py @@ -136,8 +136,6 @@ class LlmTestEndpointTest(unittest.TestCase): provider="deepseek", model="deepseek-chat", thinking_level="max", - disable_thinking=None, - reasoning_effort=None, api_key="sk-test", base_url="https://api.deepseek.com", ) @@ -183,8 +181,6 @@ class LlmTestEndpointTest(unittest.TestCase): provider="openai", model="gpt-4.1-mini", thinking_level="high", - disable_thinking=None, - reasoning_effort=None, api_key="sk-live", base_url="https://example.com/v1", ) @@ -205,8 +201,6 @@ class LlmTestEndpointTest(unittest.TestCase): enabled=True, provider="deepseek", model="deepseek-v4-pro", - disable_thinking=False, - reasoning_effort="xhigh", api_key="sk-live", base_url="https://api.deepseek.com", ) @@ -223,8 +217,6 @@ class LlmTestEndpointTest(unittest.TestCase): provider="deepseek", model="deepseek-v4-pro", thinking_level=None, - disable_thinking=False, - reasoning_effort="xhigh", api_key="sk-live", base_url="https://api.deepseek.com", )