Seedance 2.0 | Special Offer ✦ 10% OFF NOW | Ends May 13 (UTC+0)
alibaba
qwen/qwen3-coder

qwen/qwen3-coder

262,144 context · $0.22/M input tokens · $1.00/M output tokens

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over...

Harga

Bayar sesuai pemakaian

Tanpa biaya di muka, bayar hanya sesuai penggunaan

Input$0.22 / M Tokens
Output$1.00 / M Tokens

Penggunaan API

Gunakan contoh kode berikut untuk integrasi dengan API kami:

from openai import OpenAI

client = OpenAI(
    api_key="YOUR_API_KEY",
    base_url="https://llm.wavespeed.ai/v1"
)

response = client.chat.completions.create(
    model="qwen/qwen3-coder",
    messages=[
        {"role": "user", "content": "Hello!"}
    ]
)

print(response.choices[0].message.content)

Pengenalan Model

Qwen qwen3-coder

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts).

Pricing for the Alibaba endpoints varies by context length. Once a request is greater than 128k input tokens, the higher pricing is used.


Why It Looks Great

  • Large Language Model architecture for efficient processing
  • 262144 context window for long document handling
  • Competitive pricing at $0.2/$1.1 per million tokens

Key Features

  • Context Window: 262144 tokens
  • Max Output: N/A tokens
  • Vision: Supported
  • Function Calling: Supported

Specifications

SpecificationValue
ProviderQwen
Model TypeLarge Language Model (LLM)
ArchitectureN/A
Context Window262144 tokens
Max Outputtokens
InputText
OutputText
VisionSupported
Function CallingSupported

Pricing

Token TypeCost per Million Tokens
Input$0.2
Output$1.1

How to Use

  1. Write your prompt — describe the task, provide context, and specify desired output format.
  2. Submit — the model processes your request and returns the response.

API Integration

Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: qwen/qwen3-coder


API Usage

Python SDK

from openai import OpenAI

client = OpenAI(
    api_key="YOUR_API_KEY",
    base_url="https://llm.wavespeed.ai/v1"
)

response = client.chat.completions.create(
    model="qwen/qwen3-coder",
    messages=[
        {"role": "user", "content": "Hello!"}
    ]
)

print(response.choices[0].message.content)

cURL

curl https://llm.wavespeed.ai/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -d '{
    "model": "qwen/qwen3-coder",
    "messages": [{"role": "user", "content": "Hello!"}]
  }'

Notes

  • Model: qwen/qwen3-coder
  • Provider: Qwen

Info

Penyediaalibaba
Tipellm

Fitur yang Didukung

Input
Teks
Output
Teks
Konteks262,144
Output Maks-
Vision-
Function Calling✓ Didukung

Panduan Akses API

Base URLhttps://llm.wavespeed.ai/v1
API Endpointchat/completions
Model IDqwen/qwen3-coder

Qwen3 Coder API

qwen/qwen3-coder

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over...

Input

$0.22 /M

Output

$1 /M

Konteks

262K

Penggunaan Tool

Didukung

Coba Qwen3 Coder di WaveSpeedAI

Akses Qwen3 Coder melalui API terpadu kami — kompatibel dengan OpenAI, tanpa cold start, harga transparan.

Pertanyaan Umum tentang Qwen3 Coder

Berapa biaya Qwen3 Coder melalui API?+

Harga di WaveSpeedAI: $0.22 per juta token input dan $1.00 per juta token output. Prompt caching dan batch processing ditagih terpisah dan mengurangi biaya efektif pada beban kerja yang panjang dan berulang.

Berapa context window Qwen3 Coder?+

Qwen3 Coder mendukung hingga 262K token konteks dengan hingga — token output per permintaan.

Apakah Qwen3 Coder kompatibel dengan OpenAI?+

Ya. WaveSpeedAI menyediakan Qwen3 Coder melalui endpoint yang kompatibel dengan OpenAI di https://llm.wavespeed.ai/v1. Arahkan OpenAI SDK resmi ke base URL ini dengan API key WaveSpeedAI Anda — tanpa perubahan kode lainnya.

Bagaimana memulai dengan Qwen3 Coder?+

Masuk ke WaveSpeedAI, buat API key di Access Keys, lalu kirim permintaan ke https://llm.wavespeed.ai/v1/chat/completions dengan model id seperti ditampilkan di atas. Akun baru menerima kredit gratis untuk menguji Qwen3 Coder.

API LLM terkait