
Novita AI
AI-powered creative tools and model hosting. Requesty routes to 34 Novita AI models starting at $0.02 per 1M input tokens with context windows up to 1.0M tokens. One API key, OpenAI-compatible SDK, no markup.
Flagship model
moonshotai/kimi-k2-instructMMLU Pro
82.3%
GPQA
70.0%
HumanEval
89.9%
SWE-Bench
65.8%
All Novita AI models
| Model | Context | Max Output | Input/1M | Output/1M | Capabilities | SWE-Bench |
|---|---|---|---|---|---|---|
GLM-5 | 203K | 131K | $1.00 | $3.20 | π§ π§β‘ | β |
minimax/minimax-m2.7 | 200K | 128K | $0.30 | $1.20 | ππ§ π§β‘ | β |
deepseek-v3.2 | 164K | 66K | $0.27 | $0.40 | π§ π§β‘ | β |
qwen/qwen3.5-397b-a17b | 262K | 66K | $0.60 | $3.60 | ππ§ π§ | β |
zai-org/glm-4.5 | 131K | β | $0.60 | $2.20 | π§ | β |
zai-org/glm-4.6 | 205K | 131K | $0.60 | $2.20 | π§ | β |
moonshotai/kimi-k2-instruct | 131K | β | $0.57 | $2.30 | π§ | 66% |
deepseek-v3-turbo | 128K | β | $0.40 | $1.30 | π§ | 42% |
deepseek_v3 | 64K | β | $0.89 | $0.89 | π§ | β |
deepseek-v3-0324 | 128K | β | $0.40 | $1.30 | π§ | β |
qwen/qwen-2.5-72b-instruct | 32K | β | $0.38 | $0.40 | π§ | β |
qwen/qwen3-235b-a22b-fp8 | 128K | β | $0.20 | $0.80 | β | |
qwen/qwen2.5-vl-72b-instruct | 96K | β | $0.80 | $0.80 | β | |
meta-llama/llama-3-8b-instruct | 8K | β | $0.04 | $0.04 | β | |
nousresearch/hermes-2-pro-llama-3-8b | 8K | β | $0.14 | $0.14 | β | |
meta-llama/llama-3.2-3b-instruct | 33K | β | $0.03 | $0.05 | π§ | β |
wizardlm-2-8x22b | 66K | β | $0.62 | $0.62 | β | |
meta-llama/llama-3.1-8b-instruct | 16K | β | $0.05 | $0.05 | β | |
meta-llama/llama-3.3-70b-instruct | 131K | β | $0.39 | $0.39 | π§ | 23% |
meta-llama/llama-3-70b-instruct | 8K | β | $0.51 | $0.74 | β | |
meta-llama/llama-3.2-1b-instruct | 131K | β | $0.02 | $0.02 | β | |
deepseek-prover-v2-671b | 160K | β | $0.70 | $2.50 | β | |
deepseek-r1-distill-llama-70b | 32K | β | $0.80 | $0.80 | β | |
deepseek-r1-distill-qwen-32b | 13K | β | $0.30 | $0.30 | π§ | β |
deepseek-r1 | 64K | β | $4.00 | $4.00 | π§ | 49% |
deepseek-r1-turbo | 64K | β | $0.70 | $2.50 | π§ | 49% |
meta-llama/llama-4-maverick-17b-128e-instruct-fp8 | 1.0M | 1.0M | $0.20 | $0.85 | β | |
deepseek-r1-distill-qwen-14b | 128K | β | $0.15 | $0.15 | π§ | β |
sao10k/l31-70b-euryale-v2.2 | 16K | β | $1.48 | $1.48 | β | |
Sao10K/L3-8B-Stheno-v3.2 | 8K | β | $0.05 | $0.05 | β | |
sao10k/l3-70b-euryale-v2.1 | 16K | β | $1.48 | $1.48 | β | |
sao10k/l3-8b-lunaris | 8K | β | $0.05 | $0.05 | β | |
mistralai/mistral-nemo | 131K | β | $0.17 | $0.17 | β | |
gryphe/mythomax-l2-13b | 4K | β | $0.09 | $0.09 | π§ | β |
About Novita AI on Requesty
How many Novita AI models are available through Requesty?
Requesty routes to 34 Novita AI models including regional variants, with pricing synced in real time to the upstream provider.
What is the cheapest Novita AI model?
The cheapest Novita AI model starts at $0.02 per million input tokens. See the pricing column in the table below for full per-model rates.
Does Requesty add markup on Novita AI pricing?
No. Requesty passes through exactly what Novita AI charges. You pay the same per-token rates as going direct β plus you get smart routing, caching, analytics, and one unified API for 400+ models.
Is my data used to train Novita AI models?
Novita AI's training policy varies by product and tier. See their privacy policy for specifics, and contact Requesty for enterprise-grade data controls.
Where are Novita AI models hosted?
Novita AI models are hosted in πΊπΈ US. Some models are available in additional regions through AWS Bedrock, Azure, or Google Vertex AI β filter by region on the Novita AI rows in the models explorer.
