feat: add reasoning_effort config to enable LLM thinking mode

This commit is contained in:
Re-bin
2026-02-28 17:18:05 +00:00
parent 0a5f3b6194
commit f9d72e2e74
9 changed files with 23 additions and 2 deletions

View File

@@ -31,6 +31,7 @@ class OpenAICodexProvider(LLMProvider):
model: str | None = None,
max_tokens: int = 4096,
temperature: float = 0.7,
reasoning_effort: str | None = None,
) -> LLMResponse:
model = model or self.default_model
system_prompt, input_items = _convert_messages(messages)