anthropic/claude-opus-4.6
1,000,000 context · $5.00/M input$4.50/M input · $25.00/M output$22.50/M output10% off
Opus 4.6 is Anthropic’s strongest model for coding and long-running professional tasks. It is built for agents that operate across entire workflows rather than single prompts, making it especially effective for large codebases, complex refactors, and multi-step debugging that unfolds over time. The model shows deeper contextual understanding, stronger problem decomposition, and greater reliability on hard engineering tasks than prior generations.Beyond coding, Opus 4.6 excels at sustained knowledge work. It produces near-production-ready documents, plans, and analyses in a single pass, and maintains coherence across very long outputs and extended sessions. This makes it a strong default for tasks that require persistence, judgment, and follow-through, such as technical design, migration planning, and end-to-end project execution.
按用量付費
無需預付費用,僅按實際使用量付費
使用以下程式碼範例整合我們的 API:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="anthropic/claude-opus-4.6",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)Opus 4.6 is Anthropic’s strongest model for coding and long-running professional tasks. It is built for agents that operate across entire workflows ra
Opus 4.6 is Anthropic’s strongest model for coding and long-running professional tasks. It is built for agents that operate across entire workflows rather than single prompts, making it especially effective for large codebases, complex refactors, and multi-step debugging that unfolds over time. The model shows deeper contextual understanding, stronger problem decomposition, and greater reliability on hard engineering tasks than prior generations.Beyond coding, Opus 4.6 excels at sustained knowledge work. It produces near-production-ready documents, plans, and analyses in a single pass, and maintains coherence across very long outputs and extended sessions. This makes it a strong default for tasks that require persistence, judgment, and follow-through, such as technical design, migration planning, and end-to-end project execution.
| Specification | Value |
|---|---|
| Provider | Anthropic |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 1000000 tokens |
| Max Output | 128000 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $5.0 |
| Output | $25.0 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: anthropic/claude-opus-4.6
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="anthropic/claude-opus-4.6",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "anthropic/claude-opus-4.6",
"messages": [{"role": "user", "content": "Hello!"}]
}'
anthropic/claude-opus-4.6
Opus 4.6 is Anthropic’s strongest model for coding and long-running professional tasks. It is built for agents that operate across entire workflows rather than single prompts, making it especially effective for large codebases, complex refactors, and multi-step debugging that unfolds over time. The model shows deeper contextual understanding, stronger problem decomposition, and greater reliability on hard engineering tasks than prior generations.Beyond coding, Opus 4.6 excels at sustained knowledge work. It produces near-production-ready documents, plans, and analyses in a single pass, and maintains coherence across very long outputs and extended sessions. This makes it a strong default for tasks that require persistence, judgment, and follow-through, such as technical design, migration planning, and end-to-end project execution.
輸入
$5$4.50 /M
輸出
$25$22.50 /M
折扣
10% 折扣
上下文
1000K
最大輸出
128K
Vision
支援
工具調用
支援
WaveSpeedAI 定價:輸入每百萬 token $4.50,輸出每百萬 token $22.50。Prompt 快取與批次處理分別計費,可顯著降低長上下文、高重複任務的實際成本。
Claude Opus 4.6 每次請求最多支援 1000K 上下文 token,輸出最多 128K token。
是的。WaveSpeedAI 透過 https://llm.wavespeed.ai/v1 的 OpenAI 相容端點提供 Claude Opus 4.6。將官方 OpenAI SDK 的 base URL 指向該位址,使用 WaveSpeedAI 的 API Key 即可,無需其他程式碼變更。
登入 WaveSpeedAI,在 Access Keys 建立 API Key,使用上方顯示的 model id 向 https://llm.wavespeed.ai/v1/chat/completions 發送請求。新帳號將獲得免費額度,用於試用 Claude Opus 4.6。