就緒
{
"output": "Machine learning is a branch of artificial intelligence (AI) and computer science that focuses on using data and algorithms to imitate how humans learn, gradually improving its accuracy."
}$0.006每次運行·~166 / $1
NVIDIA Nemotron-3 Nano Omni Text is a lightweight text-generation model for prompt-based language understanding and response generation. Provide an English prompt, and the model can generate answers, summaries, structured outputs, explanations, and other text-based responses with controllable length and sampling behavior.
Fast text generation Generate responses quickly for chat, automation, summarization, and general language tasks.
Flexible response control
Adjust max_tokens, temperature, and top_p to balance response length, determinism, and creativity.
Optional system steering
Use system_prompt to guide tone, structure, formatting, or task behavior for more controlled outputs.
Reasoning mode options
Choose between no_think and think depending on your preferred response mode and workflow.
Production-ready API Suitable for assistants, content tools, automation pipelines, internal workflows, and structured text generation tasks.
| Parameter | Required | Description |
|---|---|---|
| prompt | Yes | English text prompt sent to the model. |
| system_prompt | No | Optional system prompt used to steer behavior, tone, or response style. |
| reasoning_mode | No | Reasoning mode: no_think (default) or think. |
| max_tokens | No | Maximum number of tokens to generate. Default: 1024. |
| temperature | No | Sampling temperature. Lower values are more deterministic. Default: 0.7. |
| top_p | No | Nucleus sampling probability mass. Default: 0.95. |
no_think or think depending on your workflow.max_tokens, temperature, and top_p.Summarize the following product requirements into a concise executive brief with key goals, risks, and next steps.
Billed by configured max_tokens.
| Max Tokens | Cost |
|---|---|
| 1000 | $0.006 |
| 1024 | $0.0061 |
| 2000 | $0.012 |
| 4000 | $0.024 |
| 8000 | $0.048 |
max_tokens value.max_tokens increases cost linearly.prompt, system_prompt, reasoning_mode, temperature, and top_p do not change pricing directly.system_prompt when you need consistent tone, role behavior, or formatting rules.temperature lower when you want more stable and deterministic results.max_tokens only when you need longer outputs, since pricing is tied to that value.top_p and temperature carefully together to balance creativity and control.prompt is the only required field.prompt must be written in English.reasoning_mode = no_think, max_tokens = 1024, temperature = 0.7, and top_p = 0.95.max_tokens, not on other generation settings.