Responses API
Some OpenAI models only support the Responses API and cannot be called via the Chat Completions endpoint. Chuizi.AI handles this transparently: you send a standard /v1/chat/completions request, and the gateway automatically converts it to a Responses API call when the target model requires it.
You do not need to learn the Responses API format. Your existing OpenAI SDK code works as-is.
Which Models Use the Responses API
The following models are Responses API-only. They do not support the Chat Completions endpoint on upstream providers:
| Model | Description |
|---|---|
openai/gpt-5-codex | GPT-5 Codex -- optimized for code generation |
openai/gpt-5.1-codex | GPT-5.1 Codex -- improved code generation |
openai/gpt-5.1-codex-mini | GPT-5.1 Codex Mini -- faster, lighter code model |
openai/gpt-5-pro | GPT-5 Pro -- advanced reasoning |
Bare model names (e.g., gpt-5-codex, gpt-5.1-codex-mini) also work as aliases.
All other models (GPT-4o, GPT-5, o3, o4-mini, etc.) continue to use Chat Completions normally.
How the Translation Works
When you send a request to /v1/chat/completions targeting a Responses API-only model, Chuizi.AI:
- Converts your request:
messagesarray becomes Responses APIinputitems. System messages become theinstructionsfield. - Forwards to upstream: The converted request goes to the provider's Responses API endpoint.
- Converts the response back: The Responses API output is translated back into a standard Chat Completions response (with
choices,usage, etc.).
This all happens server-side. From your perspective, the request and response look identical to any other Chat Completions call.
Next Steps
- Choosing a Protocol — Compare all supported protocols and when to use each
- Chat Completions API — Full reference for the Chat Completions endpoint
- Codex CLI Integration — Set up Codex CLI with Chuizi.AI