feat(openai): add gpt-5.4-pro to model map#22954
Conversation
- Add gpt-5.4-pro and gpt-5.4-pro-2026-03-05 snapshot - Input: $30/1M tokens, Output: $180/1M tokens - 1.05M context window, 128K max output tokens - Priority pricing for >272K input tokens (2x input, 1.5x output) - Supports reasoning.effort: medium, high, xhigh - Responses API, Chat Completions, Batch endpoints Made-with: Cursor
|
The latest updates on your projects. Learn more about Vercel for GitHub.
|
Greptile SummaryThis PR adds two new OpenAI model entries — Key observations:
Confidence Score: 3/5
|
| Filename | Overview |
|---|---|
| model_prices_and_context_window.json | Adds gpt-5.4-pro and gpt-5.4-pro-2026-03-05 model entries. Pricing and context window values look correct, but both entries are missing input_cost_per_token_batches and output_cost_per_token_batches despite listing /v1/batch as a supported endpoint. |
| litellm/model_prices_and_context_window_backup.json | Backup copy of the main pricing file — identical changes with the same missing batch pricing fields issue. |
Flowchart
%%{init: {'theme': 'neutral'}}%%
flowchart TD
A[Caller invokes gpt-5.4-pro] --> B{Endpoint?}
B -->|/v1/chat/completions| C[responses mode handler]
B -->|/v1/responses| C
B -->|/v1/batch| D[Batch handler]
C --> E{Priority tier?}
E -->|input ≤ 272K tokens| F[Standard pricing\n$30 input / $180 output per 1M]
E -->|input > 272K tokens| G[Priority pricing\n$60 input / $270 output per 1M]
D --> H[Batch pricing\n⚠ Missing: fields not set\ninput_cost_per_token_batches\noutput_cost_per_token_batches]
Last reviewed commit: 0505034
| "gpt-5.4-pro": { | ||
| "cache_read_input_token_cost": 3e-06, | ||
| "cache_read_input_token_cost_priority": 6e-06, | ||
| "input_cost_per_token": 3e-05, | ||
| "input_cost_per_token_priority": 6e-05, | ||
| "litellm_provider": "openai", | ||
| "max_input_tokens": 1050000, | ||
| "max_output_tokens": 128000, | ||
| "max_tokens": 128000, | ||
| "mode": "responses", | ||
| "output_cost_per_token": 1.8e-04, | ||
| "output_cost_per_token_priority": 2.7e-04, | ||
| "supported_endpoints": [ | ||
| "/v1/chat/completions", | ||
| "/v1/batch", | ||
| "/v1/responses" | ||
| ], | ||
| "supported_modalities": [ | ||
| "text", | ||
| "image" | ||
| ], | ||
| "supported_output_modalities": [ | ||
| "text" | ||
| ], | ||
| "supports_function_calling": true, | ||
| "supports_native_streaming": true, | ||
| "supports_parallel_function_calling": true, | ||
| "supports_pdf_input": true, | ||
| "supports_prompt_caching": true, | ||
| "supports_reasoning": true, | ||
| "supports_response_schema": false, | ||
| "supports_system_messages": true, | ||
| "supports_tool_choice": true, | ||
| "supports_service_tier": true, | ||
| "supports_vision": true, | ||
| "supports_web_search": true, | ||
| "supports_none_reasoning_effort": false, | ||
| "supports_xhigh_reasoning_effort": true | ||
| }, |
There was a problem hiding this comment.
Missing batch pricing fields
The model lists /v1/batch in supported_endpoints but omits input_cost_per_token_batches and output_cost_per_token_batches. All other models in this file that expose the batch endpoint include these fields — for example, gpt-5-pro at line 21123–21130 has input_cost_per_token_batches: 7.5e-06 and output_cost_per_token_batches: 6e-05, and azure/o3-pro (lines 4759–4766) has both batch cost fields alongside mode: "responses" and /v1/batch support.
Without these fields, LiteLLM cannot accurately track or report costs for batch requests submitted to this model. Based on OpenAI's typical 50% batch discount, the missing values would be:
input_cost_per_token_batches:1.5e-05(i.e. $15/1M)output_cost_per_token_batches:9e-05(i.e. $90/1M)
The same omission applies to the gpt-5.4-pro-2026-03-05 snapshot entry at line 21082.
|
Hey @Sameerlite, I raised this issue: #23014 about |
Summary
Adds GPT-5.4 pro to the LiteLLM model map.
Changes
Model specs
Made with Cursor