qwen/qwen3-max-thinking
262,144 context · $1.20/M input tokens · $6.00/M output tokens
Qwen3-Max-Thinking is the flagship reasoning model in the Qwen3 series, designed for high-stakes cognitive tasks that require deep, multi-step reasoning. By significantly scaling model capacity and reinforcement learning compute, it delivers major gains in factual accuracy, complex reasoning, instruction following, alignment with human preferences, and agentic behavior.
按量付费
无需预付费用,仅按实际使用量付费
使用以下代码示例接入我们的 API:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="qwen/qwen3-max-thinking",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)Qwen3-Max-Thinking is the flagship reasoning model in the Qwen3 series, designed for high-stakes cognitive tasks that require deep, multi-step reasoni
Qwen3-Max-Thinking is the flagship reasoning model in the Qwen3 series, designed for high-stakes cognitive tasks that require deep, multi-step reasoning. By significantly scaling model capacity and reinforcement learning compute, it delivers major gains in factual accuracy, complex reasoning, instruction following, alignment with human preferences, and agentic behavior.
| Specification | Value |
|---|---|
| Provider | Qwen |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 262144 tokens |
| Max Output | 32768 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.8 |
| Output | $3.9 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: qwen/qwen3-max-thinking
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="qwen/qwen3-max-thinking",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "qwen/qwen3-max-thinking",
"messages": [{"role": "user", "content": "Hello!"}]
}'
qwen/qwen3-max-thinking
Qwen3-Max-Thinking is the flagship reasoning model in the Qwen3 series, designed for high-stakes cognitive tasks that require deep, multi-step reasoning. By significantly scaling model capacity and reinforcement learning compute, it delivers major gains in factual accuracy, complex reasoning, instruction following, alignment with human preferences, and agentic behavior.
输入
$1.2 /M
输出
$6 /M
上下文
262K
最大输出
33K
工具调用
支持
通过我们的统一 API 接入 Qwen3 Max Thinking — 兼容 OpenAI、无冷启动、透明计费。
打开 PlaygroundWaveSpeedAI 定价:输入每百万 token $1.20,输出每百万 token $6.00。Prompt 缓存和批处理单独计费,可显著降低长上下文、高重复任务的实际成本。
Qwen3 Max Thinking 单次请求最多支持 262K 上下文 token,输出最多 33K token。
是的。WaveSpeedAI 通过 https://llm.wavespeed.ai/v1 的 OpenAI 兼容端点提供 Qwen3 Max Thinking。把官方 OpenAI SDK 的 base URL 指向该地址,使用 WaveSpeedAI 的 API Key 即可,无需任何其他代码改动。
登录 WaveSpeedAI,在 Access Keys 中生成 API Key,使用上方显示的 model id 向 https://llm.wavespeed.ai/v1/chat/completions 发送请求。新账户可获得免费额度,用于试用 Qwen3 Max Thinking。