From 67531d9663b62fda5070452ddd69263d2cd8ad29 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Peter=20St=C3=B6ckli?= Date: Thu, 19 Mar 2026 06:34:50 +0000 Subject: [PATCH] don't fail on unknown default model --- src/seclab_taskflow_agent/agent.py | 11 +++++------ 1 file changed, 5 insertions(+), 6 deletions(-) diff --git a/src/seclab_taskflow_agent/agent.py b/src/seclab_taskflow_agent/agent.py index 437efd1..9c1ae1e 100644 --- a/src/seclab_taskflow_agent/agent.py +++ b/src/seclab_taskflow_agent/agent.py @@ -41,10 +41,7 @@ case AI_API_ENDPOINT_ENUM.AI_API_OPENAI: default_model = "gpt-4o" case _: - raise ValueError( - f"Unsupported Model Endpoint: {api_endpoint}\n" - f"Supported endpoints: {[e.to_url() for e in AI_API_ENDPOINT_ENUM]}" - ) + default_model = "please-set-default-model-via-env" DEFAULT_MODEL = os.getenv("COPILOT_DEFAULT_MODEL", default=default_model) @@ -105,7 +102,8 @@ def __init__( async def on_handoff( self, context: RunContextWrapper[TContext], agent: Agent[TContext], source: Agent[TContext] ) -> None: - logging.debug(f"TaskAgentHooks on_handoff: {source.name} -> {agent.name}") + logging.debug( + f"TaskAgentHooks on_handoff: {source.name} -> {agent.name}") if self._on_handoff: await self._on_handoff(context, agent, source) @@ -169,7 +167,8 @@ def _ToolsToFinalOutputFunction( name=name, instructions=instructions, tool_use_behavior=_ToolsToFinalOutputFunction if exclude_from_context else "run_llm_again", - model=OpenAIChatCompletionsModel(model=model, openai_client=client), + model=OpenAIChatCompletionsModel( + model=model, openai_client=client), handoffs=handoffs, mcp_servers=mcp_servers, model_settings=model_settings or ModelSettings(),