Skip to main content
Core.Today
|
GoogleFastStandard

Gemini 2.0 Flash Lite

Ultra-lightweight version of Gemini 2.0 Flash optimized for maximum speed and minimal cost. Perfect for high-volume, latency-sensitive applications.

0.5 credits
per 1K tokens (avg)
Ultra-fast inference
Minimal cost per request
128K context window
Text generation
High throughput

Use with AI Assistant

Copy usage instructions for Claude, ChatGPT, or other AI

llms.txt

Model Specifications

Context Window
128K
tokens
Max Output
8K
tokens
Training Cutoff
2024-08
Compatible SDK
OpenAI, Google AI

Capabilities

Vision
Function Calling
Streaming
JSON Mode
System Prompt

Token Pricing (per 1M tokens)

Token TypeCreditsUSD Equivalent
Input Tokens150$0.15
Output Tokens600$0.60

* 1 credit โ‰ˆ $0.001 (actual charges may vary based on usage)

Quick Start

curl -X POST "https://api.core.today/llm/gemini/v1beta/openai/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gemini-2.0-flash-lite",
  "messages": [
    {
      "role": "system",
      "content": "Classify the sentiment of the text as positive, negative, or neutral. Respond with just the label."
    },
    {
      "role": "user",
      "content": "The product works great but the delivery was slow."
    }
  ],
  "max_tokens": 10,
  "temperature": 0
}'

Parameters

ParameterTypeRequiredDefaultDescription
messagesarrayYes-Array of message objects (OpenAI format)
temperaturefloatNo1Sampling temperature (0-2)
max_tokensintegerNo-Maximum output tokens

Examples

Quick Classification

Fast text classification

curl -X POST "https://api.core.today/llm/gemini/v1beta/openai/chat/completions" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer cdt_your_api_key" \
  -d '{
  "model": "gemini-2.0-flash-lite",
  "messages": [
    {
      "role": "system",
      "content": "Classify the sentiment of the text as positive, negative, or neutral. Respond with just the label."
    },
    {
      "role": "user",
      "content": "The product works great but the delivery was slow."
    }
  ],
  "max_tokens": 10,
  "temperature": 0
}'

Tips & Best Practices

1Most cost-effective Gemini model
2Best for simple, high-volume tasks
3Use temperature 0 for deterministic classification
4Ideal for real-time applications requiring low latency

Use Cases

High-volume chatbots
Real-time classification
Content filtering
Simple data extraction
Batch processing

Model Info

ProviderGoogle
Version2.0
CategoryLLM
Price0.5 credits

API Endpoint

POST /llm/gemini/v1beta/openai/chat/completions
Try in PlaygroundBack to Docs