[GH-ISSUE #9659] Compatibility with new OpenAI responses API #6303

Closed
opened 2026-04-12 17:45:23 -05:00 by GiteaMirror · 16 comments
Owner

Originally created by @pamelafox on GitHub (Mar 11, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/9659

Originally assigned to: @drifkin, @ParthSareen on GitHub.

OpenAI just announced the responses API:
https://www.youtube.com/live/hciNKcLwSes
https://platform.openai.com/docs/guides/responses-vs-chat-completions

They are not currently deprecating chat completions, however.
I'm curious what you think about a compatibility layer with the Responses API, or if you're going to stick with chat completions compatibility only.

Originally created by @pamelafox on GitHub (Mar 11, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/9659 Originally assigned to: @drifkin, @ParthSareen on GitHub. OpenAI just announced the responses API: https://www.youtube.com/live/hciNKcLwSes https://platform.openai.com/docs/guides/responses-vs-chat-completions They are not currently deprecating chat completions, however. I'm curious what you think about a compatibility layer with the Responses API, or if you're going to stick with chat completions compatibility only.
GiteaMirror added the feature request label 2026-04-12 17:45:23 -05:00
Author
Owner

@ParthSareen commented on GitHub (Mar 13, 2025):

Hi @pamelafox,

The team is aware of the changes and we keep a close eye on what is "standardized". Our approach always has been to provide the simplest experience for which we have our own API and SDKs.

We'll see what is worthwhile and what users need as the landscape evolves.

<!-- gh-comment-id:2722269343 --> @ParthSareen commented on GitHub (Mar 13, 2025): Hi @pamelafox, The team is aware of the changes and we keep a close eye on what is "standardized". Our approach always has been to provide the simplest experience for which we have our own API and SDKs. We'll see what is worthwhile and what users need as the landscape evolves.
Author
Owner

@PierreMesure commented on GitHub (Mar 17, 2025):

Great that the issue has already been posted and that you are following if the Responses API becomes an industry standard!

<!-- gh-comment-id:2728892463 --> @PierreMesure commented on GitHub (Mar 17, 2025): Great that the issue has already been posted and that you are following if the Responses API becomes an industry standard!
Author
Owner

@OriNachum commented on GitHub (Apr 17, 2025):

This is now blocking Codex, as it expects Responses API.

<!-- gh-comment-id:2813680394 --> @OriNachum commented on GitHub (Apr 17, 2025): This is now blocking Codex, as it expects Responses API.
Author
Owner

@ParthSareen commented on GitHub (Apr 17, 2025):

We're looking into it!

<!-- gh-comment-id:2813691630 --> @ParthSareen commented on GitHub (Apr 17, 2025): We're looking into it!
Author
Owner

@cgaspard commented on GitHub (Apr 23, 2025):

+1 we need this

<!-- gh-comment-id:2825409078 --> @cgaspard commented on GitHub (Apr 23, 2025): +1 we need this
Author
Owner

@OriNachum commented on GitHub (Apr 24, 2025):

@cgaspard

I built this (and verified it works) for Responses API integration.

https://github.com/teabranch/openai-responses-server

This Responses API is beyond just notation. It is a state shift.
Codex works with it (Can see here in a docker: https://github.com/OriNachum/autonomous-intelligence/tree/main/baby-tau ), so ToolUse works (Running commands)

But also in plan to add: Web Search, Code Interpreter, Computer use, Upload files and files search, MCPs, and more.

It is under MIT license.

Additionally, I plan on finding a good open source models of different sizes and train them with Tool use, so they are compatible and fit different local devices.

Does this help you?
Which features you need most?
Feel free to open issues

Edit: name changed to openai-responses-server and moved to my group's ownership for better support.
It was also deployed as a pypi package

<!-- gh-comment-id:2826340187 --> @OriNachum commented on GitHub (Apr 24, 2025): @cgaspard I built this (and verified it works) for Responses API integration. https://github.com/teabranch/openai-responses-server This Responses API is beyond just notation. It is a state shift. Codex works with it (Can see here in a docker: https://github.com/OriNachum/autonomous-intelligence/tree/main/baby-tau ), so ToolUse works (Running commands) But also in plan to add: Web Search, Code Interpreter, Computer use, Upload files and files search, MCPs, and more. It is under MIT license. Additionally, I plan on finding a good open source models of different sizes and train them with Tool use, so they are compatible and fit different local devices. Does this help you? Which features you need most? Feel free to open issues Edit: name changed to openai-responses-server and moved to my group's ownership for better support. It was also deployed as [a pypi package](https://pypi.org/project/openai-responses-server)
Author
Owner

@gorbypark commented on GitHub (Jun 26, 2025):

The new OpenAI Agents SDK for typescript (and I assume for Python) require this to work with ollama as well.

<!-- gh-comment-id:3009122802 --> @gorbypark commented on GitHub (Jun 26, 2025): The new OpenAI Agents SDK for typescript (and I assume for Python) require this to work with ollama as well.
Author
Owner

@Powerkrieger commented on GitHub (Aug 26, 2025):

Is this being worked on?

<!-- gh-comment-id:3224758916 --> @Powerkrieger commented on GitHub (Aug 26, 2025): Is this being worked on?
Author
Owner

@ZJCODE commented on GitHub (Aug 27, 2025):

will this be supported?

<!-- gh-comment-id:3227957864 --> @ZJCODE commented on GitHub (Aug 27, 2025): will this be supported?
Author
Owner

@pawelkaliniakit commented on GitHub (Sep 3, 2025):

OpenAI Agents SDK - python lib needs it.

<!-- gh-comment-id:3249051618 --> @pawelkaliniakit commented on GitHub (Sep 3, 2025): OpenAI Agents SDK - python lib needs it.
Author
Owner

@Powerkrieger commented on GitHub (Sep 18, 2025):

There is an issue in vllm targeting the same problem. There is some more discussion there as to whether it should actually be implemented by vllm, with reasoning against it that applies to ollama aswell.

Multiple approaches appending wrappers to provide the responses endpoint are mentioned there, including LlamaStack and Open-Responses-Server.

Since Ollama staff might decide to not implement this endpoint at all or for some time, anyone needing this might be interested in those repos or at least the discussion. Disclaimer: I have not tried any of those approaches yet.

<!-- gh-comment-id:3306473649 --> @Powerkrieger commented on GitHub (Sep 18, 2025): There is an [issue in vllm](https://github.com/vllm-project/vllm/issues/14721) targeting the same problem. There is some more discussion there as to whether it should actually be implemented by vllm, with reasoning against it that applies to ollama aswell. Multiple approaches appending wrappers to provide the responses endpoint are mentioned there, including [LlamaStack](https://github.com/llamastack/llama-stack) and [Open-Responses-Server](https://github.com/teabranch/open-responses-server). Since Ollama staff might decide to not implement this endpoint at all or for some time, anyone needing this might be interested in those repos or at least the discussion. Disclaimer: I have not tried any of those approaches yet.
Author
Owner

@Powerkrieger commented on GitHub (Sep 19, 2025):

Having tried (or tried to try lol) both, I can say for just using it as a wrapper, OpenResponsesServer is easier to tackle this problem, as it is straightforward in being setup. LlamaStacks documentation is lacking at the initial steps so it took me waaay too long to get anything done.

<!-- gh-comment-id:3311840316 --> @Powerkrieger commented on GitHub (Sep 19, 2025): Having tried (or tried to try lol) both, I can say for just using it as a wrapper, OpenResponsesServer is easier to tackle this problem, as it is straightforward in being setup. LlamaStacks documentation is lacking at the initial steps so it took me waaay too long to get anything done.
Author
Owner

@realharry commented on GitHub (Oct 19, 2025):

Considering the importance (or, dominance) of OpenAI in LLM development, Ollama API is becoming less and less useful without the full support of OpenAI APIs. I'm also one of the people who are waiting for this to be implemented in Ollama. 🙏

<!-- gh-comment-id:3419161282 --> @realharry commented on GitHub (Oct 19, 2025): Considering the importance (or, dominance) of OpenAI in LLM development, Ollama API is becoming less and less useful without the full support of OpenAI APIs. I'm also one of the people who are waiting for this to be implemented in Ollama. 🙏
Author
Owner

@reza8i commented on GitHub (Nov 5, 2025):

If you do want to use the Responses API you can use Hugging Face’s Responses.js proxy to convert Chat Completions to Responses API.

Alternatively, for basic examples , use their example python server

https://github.com/openai/gpt-oss?tab=readme-ov-file#responses-api

<!-- gh-comment-id:3490514612 --> @reza8i commented on GitHub (Nov 5, 2025): If you do want to use the Responses API you can use Hugging Face’s Responses.js proxy to convert Chat Completions to Responses API. Alternatively, for basic examples , use their example python server https://github.com/openai/gpt-oss?tab=readme-ov-file#responses-api
Author
Owner

@wigging commented on GitHub (Nov 14, 2025):

Any updates on Ollama supporting the Responses API?

<!-- gh-comment-id:3534353884 --> @wigging commented on GitHub (Nov 14, 2025): Any updates on Ollama supporting the Responses API?
Author
Owner

@drifkin commented on GitHub (Dec 12, 2025):

The initial implementation in #13351 is merged, will be in the next release

<!-- gh-comment-id:3644316078 --> @drifkin commented on GitHub (Dec 12, 2025): The initial implementation in #13351 is merged, will be in the next release
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#6303