google/gemma-3-27b-it
131,072 context · $0.08/M input tokens · $0.16/M output tokens
Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities,...
Pay-per-use
Nessun costo iniziale, paga solo per ciò che usi
Usa i seguenti esempi di codice per integrare la nostra API:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="google/gemma-3-27b-it",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)Gemma 3 introduces multimodality, supporting vision-language input and text outputs
Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 27B is Google's latest open source model, successor to Gemma 2
| Specification | Value |
|---|---|
| Provider | |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 128000 tokens |
| Max Output | 65536 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $0.0 |
| Output | $0.2 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: google/gemma-3-27b-it
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="google/gemma-3-27b-it",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "google/gemma-3-27b-it",
"messages": [{"role": "user", "content": "Hello!"}]
}'
google/gemma-3-27b-it
Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities,...
Input
$0.08 /M
Output
$0.16 /M
Contesto
131K
Output max
16K
Vision
Supportato
Uso strumenti
Supportato
Accedi a Gemma 3 27b It tramite la nostra API unificata — compatibile con OpenAI, senza cold start, prezzi trasparenti.
Prezzi su WaveSpeedAI: $0.08 per milione di token in input e $0.16 per milione di token in output. Prompt caching e batch processing sono fatturati separatamente e riducono il costo effettivo su carichi lunghi e ripetitivi.
Gemma 3 27b It supporta fino a 131K token di contesto e fino a 16K token di output per richiesta.
Sì. WaveSpeedAI espone Gemma 3 27b It tramite un endpoint compatibile con OpenAI all'indirizzo https://llm.wavespeed.ai/v1. Punta l'SDK ufficiale di OpenAI a questa base URL con la tua API key WaveSpeedAI — senza altre modifiche al codice.
Accedi a WaveSpeedAI, crea una API key in Access Keys, poi invia una richiesta a https://llm.wavespeed.ai/v1/chat/completions con il model id mostrato sopra. I nuovi account ricevono crediti gratuiti per testare Gemma 3 27b It.