openai/gpt-oss-safeguard-20b
131,072 context · $0.07/M input tokens · $0.30/M output tokens
gpt-oss-safeguard-20b is a safety reasoning model from OpenAI built upon gpt-oss-20b. This open-weight, 21B-parameter Mixture-of-Experts (MoE) model offers lower latency for safety tasks like content classification, LLM filtering, and trust...
Bayar sesuai pemakaian
Tanpa biaya di muka, bayar hanya sesuai penggunaan
Gunakan contoh kode berikut untuk integrasi dengan API kami:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-oss-safeguard-20b",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)openai gpt-oss-safeguard-20b
| Specification | Value |
|---|---|
| Provider | Openai |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 131072 tokens |
| Max Output | 65536 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.1 |
| Output | $0.3 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: openai/gpt-oss-safeguard-20b
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-oss-safeguard-20b",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "openai/gpt-oss-safeguard-20b",
"messages": [{"role": "user", "content": "Hello!"}]
}'
openai/gpt-oss-safeguard-20b
gpt-oss-safeguard-20b is a safety reasoning model from OpenAI built upon gpt-oss-20b. This open-weight, 21B-parameter Mixture-of-Experts (MoE) model offers lower latency for safety tasks like content classification, LLM filtering, and trust...
Input
$0.075 /M
Output
$0.3 /M
Konteks
131K
Output Maks.
66K
Penggunaan Tool
Didukung
Akses GPT Oss Safeguard 20b melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.
Harga di WaveSpeedAI: $0.07 per juta token input dan $0.30 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.
GPT Oss Safeguard 20b mendukung hingga 131K token konteks dengan hingga 66K token output per permintaan.
Ya. WaveSpeedAI menyediakan GPT Oss Safeguard 20b melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.
Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji GPT Oss Safeguard 20b.