diff --git a/openevolve/llm/openai.py b/openevolve/llm/openai.py index 7477e5b349..f7d0648de2 100644 --- a/openevolve/llm/openai.py +++ b/openevolve/llm/openai.py @@ -155,9 +155,11 @@ async def generate_with_context( "model": self.model, "messages": formatted_messages, "temperature": kwargs.get("temperature", self.temperature), - "top_p": kwargs.get("top_p", self.top_p), "max_tokens": kwargs.get("max_tokens", self.max_tokens), } + top_p = kwargs.get("top_p", self.top_p) + if top_p is not None: + params["top_p"] = top_p # Handle reasoning_effort for open source reasoning models. reasoning_effort = kwargs.get("reasoning_effort", self.reasoning_effort)