minimax/minimax-m2
196,608 context · $0.30/M input tokens · $1.20/M output tokens
MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning,...
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="minimax/minimax-m2",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows
MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows.
| Specification | Value |
|---|---|
| Provider | Minimax |
| Model Type | Large Language Model (LLM) |
| Architecture | MoE (Mixture of Experts) |
| Context Window | 196608 tokens |
| Max Output | 65536 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.3 |
| Output | $1.1 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: minimax/minimax-m2
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="minimax/minimax-m2",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "minimax/minimax-m2",
"messages": [{"role": "user", "content": "Hello!"}]
}'
minimax/minimax-m2
MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning,...
Input
$0.3 /M
Output
$1.2 /M
Konteks
197K
Output Maks.
197K
Penggunaan Tool
Didukung
Akses Minimax M2 melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Harga di WaveSpeedAI: $0.30 per juta token input dan $1.20 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
Minimax M2 mendukung hingga 197K token konteks dengan hingga 197K token output per permintaan.
Ya. WaveSpeedAI menyediakan Minimax M2 melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji Minimax M2.