qwen/qwen3.5-122b-a10b
262,144 context · $0.40/M input tokens · $3.20/M output tokens
The Qwen3.5 122B-A10B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. In terms of overall performance, this model is second only to Qwen3.5-397B-A17B. Its text capabilities significantly outperform those of Qwen3-235B-2507, and its visual capabilities surpass those of Qwen3-VL-235B.
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="qwen/qwen3.5-122b-a10b",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)The Qwen3.5 122B-A10B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixtur
The Qwen3.5 122B-A10B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. In terms of overall performance, this model is second only to Qwen3.5-397B-A17B. Its text capabilities significantly outperform those of Qwen3-235B-2507, and its visual capabilities surpass those of Qwen3-VL-235B.
| Specification | Value |
|---|---|
| Provider | Qwen |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 262144 tokens |
| Max Output | 65536 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.3 |
| Output | $2.1 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: qwen/qwen3.5-122b-a10b
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="qwen/qwen3.5-122b-a10b",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "qwen/qwen3.5-122b-a10b",
"messages": [{"role": "user", "content": "Hello!"}]
}'
qwen/qwen3.5-122b-a10b
The Qwen3.5 122B-A10B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. In terms of overall performance, this model is second only to Qwen3.5-397B-A17B. Its text capabilities significantly outperform those of Qwen3-235B-2507, and its visual capabilities surpass those of Qwen3-VL-235B.
Input
$0.4 /M
Output
$3.2 /M
Konteks
262K
Output Maks.
66K
Vision
Didukung
Penggunaan Tool
Didukung
Akses Qwen3.5 122b A10b melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Buka PlaygroundHarga di WaveSpeedAI: $0.40 per juta token input dan $3.20 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
Qwen3.5 122b A10b mendukung hingga 262K token konteks dengan hingga 66K token output per permintaan.
Ya. WaveSpeedAI menyediakan Qwen3.5 122b A10b melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji Qwen3.5 122b A10b.