Skip to main content
OpenAIFastUltra

GPT-5.4

OpenAI's newest flagship model with 1M context window and 128K output tokens. Delivers top-tier reasoning across all domains with adjustable reasoning effort levels from none to xhigh.

5 credits
per request
1M token context window
128K max output tokens
Adjustable reasoning (none/low/medium/high/xhigh)
Function calling, web search, file search, computer use
Native vision support

Use with AI Assistant

Copy usage instructions for Claude, ChatGPT, or other AI

llms.txt

Model Specifications

Context Window
1M
tokens
Max Output
128K
tokens
Training Cutoff
2025-08
Compatible SDK
OpenAI

Capabilities

Vision
Function Calling
Streaming
JSON Mode
System Prompt

Token Pricing (per 1M tokens)

Token TypeCreditsUSD Equivalent
Input Tokens2,500$2.50
Output Tokens15,000$15.00

* 1 credit โ‰ˆ $0.001 (actual charges may vary based on usage)

Quick Start

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-5.4",
  "messages": [
    {
      "role": "system",
      "content": "You are an expert analyst with deep domain knowledge."
    },
    {
      "role": "user",
      "content": "Compare the architectural trade-offs between microservices and monolithic architectures for a fintech startup processing 10M transactions per day."
    }
  ],
  "reasoning_effort": "high",
  "max_completion_tokens": 4000
}'

Parameters

ParameterTypeRequiredDefaultDescription
messagesarrayYes-Array of message objects with role and content
modelstringYesgpt-5.4Model identifier
max_completion_tokensintegerNo4096Maximum tokens in response (up to 128000). Note: use max_completion_tokens, not max_tokens
reasoning_effortstringNomediumReasoning effort level: none, low, medium, high, or xhigh
nonelowmediumhighxhigh
temperaturefloatNo1.0Sampling temperature (0-2)
streambooleanNofalseEnable Server-Sent Events streaming
top_pfloatNo1.0Nucleus sampling threshold (0-1)

Examples

Advanced Reasoning

Complex multi-step analysis with GPT-5.4

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-5.4",
  "messages": [
    {
      "role": "system",
      "content": "You are an expert analyst with deep domain knowledge."
    },
    {
      "role": "user",
      "content": "Compare the architectural trade-offs between microservices and monolithic architectures for a fintech startup processing 10M transactions per day."
    }
  ],
  "reasoning_effort": "high",
  "max_completion_tokens": 4000
}'

Code Generation

Generate production-quality code with GPT-5.4

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-5.4",
  "messages": [
    {
      "role": "system",
      "content": "You are a senior software engineer. Write clean, well-tested, production-ready code."
    },
    {
      "role": "user",
      "content": "Implement a rate limiter using the token bucket algorithm in Go with Redis backend."
    }
  ],
  "temperature": 0.3,
  "max_completion_tokens": 4000
}'

Tips & Best Practices

1Use reasoning_effort 'xhigh' for the most complex reasoning tasks
2128K output tokens enables very long-form content generation
3Supports computer use and web search tools for agentic workflows
4Lower temperature (0.2-0.5) recommended for coding and analytical tasks

Use Cases

Complex multi-step reasoning and analysis
Advanced code generation and architecture design
Long document processing and summarization
Agentic workflows with tool use
Enterprise-grade content processing

Model Info

ProviderOpenAI
Version2026-03
CategoryLLM
Price5 credits

API Endpoint

POST /llm/openai/v1/chat/completions
Try in PlaygroundBack to Docs