mistralai/devstral-small
131,072 context · $0.10/M input tokens · $0.30/M output tokens
Devstral Small 1.1 is a 24B parameter open-weight language model for software engineering agents, developed by Mistral AI in collaboration with All Hands AI. Finetuned from Mistral Small 3.1 and...
Pago por uso
Sin costos iniciales, paga solo por lo que uses
Usa los siguientes ejemplos de código para integrar con nuestra API:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="mistralai/devstral-small",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)Devstral Small 1
Devstral Small 1.1 is a 24B parameter open-weight language model for software engineering agents, developed by Mistral AI in collaboration with All Hands AI. Finetuned from Mistral Small 3.1 and released under the Apache 2.0 license, it features a 128k token context window and supports both Mistral-style function calling and XML output formats.
Designed for agentic coding workflows, Devstral Small 1.1 is optimized for tasks such as codebase exploration, multi-file edits, and integration into autonomous development agents like OpenHands and Cline. It achieves 53.6% on SWE-Bench Verified, surpassing all other open models on this benchmark, while remaining lightweight enough to run on a single 4090 GPU or Apple silicon machine. The model uses a Tekken tokenizer with a 131k vocabulary and is deployable via vLLM, Transformers, Ollama, LM Studio, and other OpenAI-compatible runtimes.
| Specification | Value |
|---|---|
| Provider | Mistralai |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 131072 tokens |
| Max Output | tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.1 |
| Output | $0.3 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: mistralai/devstral-small
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="mistralai/devstral-small",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "mistralai/devstral-small",
"messages": [{"role": "user", "content": "Hello!"}]
}'
mistralai/devstral-small
Devstral Small 1.1 is a 24B parameter open-weight language model for software engineering agents, developed by Mistral AI in collaboration with All Hands AI. Finetuned from Mistral Small 3.1 and...
Entrada
$0.1 /M
Salida
$0.3 /M
Contexto
131K
Uso de herramientas
Compatible
Accede a Devstral Small mediante nuestra API unificada — compatible con OpenAI, sin arranques en frío, precios transparentes.
Precios en WaveSpeedAI: $0.10 por millón de tokens de entrada y $0.30 por millón de tokens de salida. El prompt caching y el procesamiento por lotes se facturan por separado y reducen el coste efectivo en cargas largas y repetitivas.
Devstral Small admite hasta 131K tokens de contexto y hasta — tokens de salida por solicitud.
Sí. WaveSpeedAI expone Devstral Small a través de un endpoint compatible con OpenAI en https://llm.wavespeed.ai/v1. Apunta el SDK oficial de OpenAI a esta base URL con tu clave API de WaveSpeedAI — sin más cambios de código.
Inicia sesión en WaveSpeedAI, crea una clave API en Access Keys y envía una solicitud a https://llm.wavespeed.ai/v1/chat/completions con el id de modelo mostrado arriba. Las cuentas nuevas reciben créditos gratuitos para evaluar Devstral Small antes de pagar por token.