mistralai/codestral-2508
256,000 context · $0.30/M input tokens · $0.90/M output tokens
Mistral's cutting-edge language model for coding released end of July 2025. Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation. Blog Post
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="mistralai/codestral-2508",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)mistralai codestral-2508
| Specification | Value |
|---|---|
| Provider | Mistralai |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 256000 tokens |
| Max Output | 4096 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.3 |
| Output | $0.10 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: mistralai/codestral-2508
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="mistralai/codestral-2508",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "mistralai/codestral-2508",
"messages": [{"role": "user", "content": "Hello!"}]
}'
mistralai/codestral-2508
Mistral's cutting-edge language model for coding released end of July 2025. Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation. [Blog Post](https://mistral.ai/news/codestral-25-08)
Input
$0.3 /M
Output
$0.9 /M
Konteks
256K
Output Maks.
4K
Penggunaan Tool
Didukung
Akses Codestral 2508 melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Harga di WaveSpeedAI: $0.30 per juta token input dan $0.90 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
Codestral 2508 mendukung hingga 256K token konteks dengan hingga 4K token output per permintaan.
Ya. WaveSpeedAI menyediakan Codestral 2508 melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji Codestral 2508.