google/gemini-2.0-flash-001
1,048,576 context · $0.10/M input tokens · $0.40/M output tokens
Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It...
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="google/gemini-2.0-flash-001",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)google gemini-2.0-flash-001
| Specification | Value |
|---|---|
| Provider | |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 1048576 tokens |
| Max Output | 8192 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.1 |
| Output | $0.4 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: google/gemini-2.0-flash-001
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="google/gemini-2.0-flash-001",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "google/gemini-2.0-flash-001",
"messages": [{"role": "user", "content": "Hello!"}]
}'
google/gemini-2.0-flash-001
Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It...
Input
$0.1 /M
Output
$0.4 /M
Konteks
1049K
Output Maks.
8K
Vision
Didukung
Penggunaan Tool
Didukung
Akses Gemini 2.0 Flash 001 melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Buka PlaygroundHarga di WaveSpeedAI: $0.10 per juta token input dan $0.40 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
Gemini 2.0 Flash 001 mendukung hingga 1049K token konteks dengan hingga 8K token output per permintaan.
Ya. WaveSpeedAI menyediakan Gemini 2.0 Flash 001 melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji Gemini 2.0 Flash 001.