Requesty
DeepInfra Inc. logo

DeepInfra Inc.

Serverless inference for machine learning models. Requesty routes to 19 DeepInfra Inc. models starting at $0.02 per 1M input tokens with context windows up to 262K tokens. One API key, OpenAI-compatible SDK, no markup.

MMLU Pro
68.9%
GPQA
50.5%
HumanEval
88.4%
SWE-Bench
23.3%

All DeepInfra Inc. models

ModelContextMax OutputInput/1MOutput/1MCapabilitiesSWE-Bench
Kimi K2.5
262K131K$0.45$2.25
πŸ‘πŸ”§βš‘
β€”
deepseek-ai/DeepSeek-V3.1
164Kβ€”$0.30$1.00
πŸ”§
β€”
zai-org/GLM-4.5-Air
131K4K$0.20$1.10
πŸ”§
β€”
zai-org/GLM-4.5
131K4K$0.60$2.20
πŸ”§
β€”
Qwen/Qwen3-Coder-480B-A35B-Instruct
262Kβ€”$0.40$1.60
πŸ”§
β€”
phi-4
16Kβ€”$0.07$0.14
β€”
Qwen/Qwen2.5-72B-Instruct
131Kβ€”$0.23$0.40
πŸ”§
β€”
Qwen/Qwen3-32B
40Kβ€”$0.10$0.30
πŸ”§
β€”
Qwen/Qwen2.5-Coder-32B-Instruct
16Kβ€”$0.07$0.16
πŸ”§
β€”
Qwen/Qwen3-235B-A22B
40K4K$0.20$0.60
πŸ”§
β€”
meta-llama/Llama-3.3-70B-Instruct
131Kβ€”$0.23$0.40
πŸ”§
23%
meta-llama/Meta-Llama-3.1-405B-Instruct
131Kβ€”$0.80$0.80
β€”
meta-llama/Llama-3.3-70B-Instruct-Turbo
131Kβ€”$0.12$0.30
πŸ”§
β€”
meta-llama/Llama-3.2-90B-Vision-Instruct
131K4K$0.35$0.40
β€”
meta-llama/Meta-Llama-3.1-70B-Instruct
131Kβ€”$0.23$0.40
πŸ”§
β€”
deepseek-ai/DeepSeek-V3
128K8K$0.85$0.90
πŸ”§
42%
deepseek-ai/DeepSeek-R1-Distill-Llama-70B
64K8K$0.23$0.69
β€”
deepseek-ai/DeepSeek-R1
64K8K$0.85$2.50
πŸ”§
49%
meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
131Kβ€”$0.02$0.05
πŸ”§
β€”

About DeepInfra Inc. on Requesty

How many DeepInfra Inc. models are available through Requesty?
Requesty routes to 19 DeepInfra Inc. models including regional variants, with pricing synced in real time to the upstream provider.
What is the cheapest DeepInfra Inc. model?
The cheapest DeepInfra Inc. model starts at $0.02 per million input tokens. See the pricing column in the table below for full per-model rates.
Does Requesty add markup on DeepInfra Inc. pricing?
No. Requesty passes through exactly what DeepInfra Inc. charges. You pay the same per-token rates as going direct β€” plus you get smart routing, caching, analytics, and one unified API for 400+ models.
Is my data used to train DeepInfra Inc. models?
DeepInfra Inc.'s terms state that API data is not used for training. See their privacy policy for the authoritative statement.
Where are DeepInfra Inc. models hosted?
DeepInfra Inc. models are hosted in πŸ‡ΊπŸ‡Έ US. Some models are available in additional regions through AWS Bedrock, Azure, or Google Vertex AI β€” filter by region on the DeepInfra Inc. rows in the models explorer.