Skip to main content
Core.Today
|
GoogleFastStandard

Gemini 3.1 Flash Lite Preview

Ultra-lightweight variant of Gemini 3.1 Flash. The most cost-effective Gemini model with support for cached input and audio input. Ideal for high-throughput, budget-conscious applications.

100 credits
per request
Most cost-effective Gemini model
Cached input token support
Audio input support
Fast inference speed
Function calling & JSON mode

Use with AI Assistant

Copy usage instructions for Claude, ChatGPT, or other AI

llms.txt

Model Specifications

Context Window
1M
tokens
Max Output
66K
tokens
Training Cutoff
January 2025
Compatible SDK
OpenAI, Google AI

Capabilities

Vision
Function Calling
Streaming
JSON Mode
System Prompt

Token Pricing (per 1M tokens)

Token TypeCreditsUSD Equivalent
Input Tokens250$0.25
Output Tokens1,500$1.50

* 1 credit โ‰ˆ $0.001 (actual charges may vary based on usage)

Quick Start

curl -X POST "https://api.core.today/llm/gemini/v1beta/openai/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gemini-3.1-flash-lite-preview",
  "messages": [
    {
      "role": "system",
      "content": "Classify the following text as: spam, not_spam. Respond with only the label."
    },
    {
      "role": "user",
      "content": "Congratulations! You have been selected for a special prize. Click here to claim now!"
    }
  ],
  "max_tokens": 50,
  "temperature": 0
}'

Parameters

ParameterTypeRequiredDefaultDescription
messagesarrayYes-Array of message objects (OpenAI format)
temperaturefloatNo1Sampling temperature (0-2)
top_pfloatNo0.95Nucleus sampling parameter
max_tokensintegerNo-Maximum output tokens
streambooleanNofalseEnable Server-Sent Events streaming

Examples

Quick Classification

Lightweight text classification with Flash Lite

curl -X POST "https://api.core.today/llm/gemini/v1beta/openai/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gemini-3.1-flash-lite-preview",
  "messages": [
    {
      "role": "system",
      "content": "Classify the following text as: spam, not_spam. Respond with only the label."
    },
    {
      "role": "user",
      "content": "Congratulations! You have been selected for a special prize. Click here to claim now!"
    }
  ],
  "max_tokens": 50,
  "temperature": 0
}'

Tips & Best Practices

1Most affordable Gemini model at $0.25/$1.50 per M tokens
2Use cached input tokens for repeated context to further reduce costs
3Ideal for high-volume classification and routing tasks
4Supports audio input for voice-based applications

Use Cases

High-volume text processing
Real-time chat applications
Quick classification and routing
Lightweight data extraction
Audio transcription and understanding

Model Info

ProviderGoogle
Version3.1-preview
CategoryLLM
Price100 credits

API Endpoint

POST /llm/gemini/v1beta/openai/chat/completions
Try in PlaygroundBack to Docs