deepseek/deepseek-r1
64,000 context · $0.70/M input tokens · $2.50/M output tokens
DeepSeek R1 is here: Performance on par with OpenAI o1, but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass....
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="deepseek/deepseek-r1",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)DeepSeek R1 is here: Performance on par with OpenAI o1, but open-sourced and with fully open reasoning tokens
DeepSeek R1 is here: Performance on par with OpenAI o1, but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass.
Fully open-source model & technical report.
MIT licensed: Distill & commercialize freely!
| Specification | Value |
|---|---|
| Provider | Deepseek |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 64000 tokens |
| Max Output | 16000 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.8 |
| Output | $2.7 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: deepseek/deepseek-r1
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="deepseek/deepseek-r1",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "deepseek/deepseek-r1",
"messages": [{"role": "user", "content": "Hello!"}]
}'
deepseek/deepseek-r1
DeepSeek R1 is here: Performance on par with OpenAI o1, but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass....
Input
$0.7 /M
Output
$2.5 /M
Konteks
64K
Output Maks.
16K
Penggunaan Tool
Didukung
Akses DeepSeek R1 melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Buka PlaygroundHarga di WaveSpeedAI: $0.70 per juta token input dan $2.50 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
DeepSeek R1 mendukung hingga 64K token konteks dengan hingga 16K token output per permintaan.
Ya. WaveSpeedAI menyediakan DeepSeek R1 melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji DeepSeek R1.