Skip to main content
Core.Today
|
OpenAIFastStandard

GPT-4.1 Nano

OpenAI's fastest and cheapest model. Optimized for classification, autocompletion, and low-latency tasks. Ultra-affordable at $0.10/1M input tokens.

1 credits
per 1K tokens (avg)
1M token context window
32K max output tokens
Ultra-low latency
Lowest cost in GPT-4.1 family
Native vision (text + image)
Function calling & JSON mode

Use with AI Assistant

Copy usage instructions for Claude, ChatGPT, or other AI

llms.txt

Model Specifications

Context Window
1.0M
tokens
Max Output
33K
tokens
Training Cutoff
2024-05
Compatible SDK
OpenAI

Capabilities

Vision
Function Calling
Streaming
JSON Mode
System Prompt

Token Pricing (per 1M tokens)

Token TypeCreditsUSD Equivalent
Input Tokens100$0.10
Output Tokens400$0.40
Cached Tokens25$0.03

* 1 credit โ‰ˆ $0.001 (actual charges may vary based on usage)

Quick Start

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-4.1-nano",
  "messages": [
    {
      "role": "system",
      "content": "Classify the sentiment as positive, negative, or neutral. Respond with JSON: {\"sentiment\": \"...\", \"confidence\": 0.0}"
    },
    {
      "role": "user",
      "content": "The new update is amazing! Everything runs so much smoother now."
    }
  ],
  "response_format": {
    "type": "json_object"
  },
  "max_tokens": 50
}'

Parameters

ParameterTypeRequiredDefaultDescription
messagesarrayYes-Array of message objects with role and content
modelstringYesgpt-4.1-nanoModel identifier
temperaturefloatNo1.0Sampling temperature (0-2). Lower = more focused, higher = more creative
max_tokensintegerNo4096Maximum tokens in response (up to 32768)
streambooleanNofalseEnable Server-Sent Events streaming
response_formatobjectNo-Format of response: { type: 'json_object' } for JSON mode
toolsarrayNo-List of tools (functions) the model can call
top_pfloatNo1.0Nucleus sampling threshold (0-1)

Examples

Text Classification

Ultra-fast sentiment classification

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-4.1-nano",
  "messages": [
    {
      "role": "system",
      "content": "Classify the sentiment as positive, negative, or neutral. Respond with JSON: {\"sentiment\": \"...\", \"confidence\": 0.0}"
    },
    {
      "role": "user",
      "content": "The new update is amazing! Everything runs so much smoother now."
    }
  ],
  "response_format": {
    "type": "json_object"
  },
  "max_tokens": 50
}'

Quick Autocomplete

Low-latency text completion

curl -X POST "https://api.core.today/llm/openai/v1/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gpt-4.1-nano",
  "messages": [
    {
      "role": "system",
      "content": "Complete the user's sentence naturally. Keep it brief."
    },
    {
      "role": "user",
      "content": "The main advantage of microservices architecture is"
    }
  ],
  "temperature": 0.3,
  "max_tokens": 100
}'

Tips & Best Practices

1Best for high-volume, low-latency tasks like classification
2At $0.10/1M input tokens, ideal for cost-sensitive applications
3Use for content moderation and routing pipelines
4Great for autocomplete and suggestion features
5Supports same capabilities as GPT-4.1 (vision, tools, JSON)
6Consider GPT-4.1 Mini for tasks requiring more reasoning

Use Cases

Real-time classification and routing
Autocomplete and suggestions
High-frequency API calls
Content moderation
Simple chatbot interactions
Quick data validation

Model Info

ProviderOpenAI
Version2025-04-14
CategoryLLM
Price1 credits

API Endpoint

POST /llm/openai/v1/chat/completions
Try in PlaygroundBack to Docs