Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Feature Request: Allow Separate Models for Tool Execution and Final Response in OpenAI Agent SDK #684

Copy link
Copy link
Open
@ciliamadani

Description

@ciliamadani
Issue body actions

✨ Feature Request: Allow Separate Models for Tool Execution and Final Response in OpenAI Agent SDK

Summary

Please add support in the OpenAI Agent SDK to split model usage between:

  • One model for tool reasoning/execution (e.g., gpt-3.5-turbo)
  • Another model for final response generation (e.g., o1)

Motivation

In many real-world agent workflows, the agent needs to:

  1. Execute one or more tools (e.g., retrieve context, classify intent, generate a prompt)
  2. Use the output to generate a high-quality final response

Using a single high-end model (like o1) for all steps increases cost unnecessarily.
Using a cheaper model (like gpt-3.5-turbo) for all steps reduces the final response quality.


Proposed API

Agent(
    name="Vector Assistant",
    reasoning_model="gpt-3.5-turbo",   # Handles all tool execution and logic
    response_model="o1",           # Used only once, after tools run
    ...
)

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or requestNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions

      Morty Proxy This is a proxified and sanitized view of the page, visit original site.