OpenAI Inc.

Leading AI research company behind GPT models and ChatGPT.

πŸ“ πŸ‡ΊπŸ‡Έ USβ€’54 models availableβ€’Visit Website β†’
54
Available Models
$2.78
Avg Input Price/M
$0.02
Cheapest Model
openai/gpt-5-nano:flex
$20.00
Most Expensive
openai/o3-pro

Features Overview

35
Vision Support
41
Advanced Reasoning
53
Caching Support
0
Computer Use

Privacy & Data Policy

Data Retention

Yes (30 days)

Location

πŸ‡ΊπŸ‡Έ US

All OpenAI Inc. Models

View All Providers β†’
OpenAI Inc.

gpt-4.1-nano

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$0.1/M tokens
Output
$0.4/M tokens

For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.

OpenAI Inc.

o3-2025-04-16

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$2.00/M tokens
Output
$8.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.05/M tokens
Output
$0.4/M tokens

GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.

OpenAI Inc.

o4-mini:flex

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$0.55/M tokens
Output
$2.20/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o3

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$2.00/M tokens
Output
$8.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$2.50/M tokens
Output
$10.00/M tokens

The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses. GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.

OpenAI Inc.

o3:flex

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.00/M tokens
Output
$4.00/M tokens

O3 Flex is a cheaper version of the o3 model

OpenAI Inc.

o1:medium

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$15.00/M tokens
Output
$60.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

OpenAI Inc.

o1:low

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$15.00/M tokens
Output
$60.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

OpenAI Inc.

gpt-4.1

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$2.00/M tokens
Output
$8.00/M tokens

GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$0.4/M tokens
Output
$1.60/M tokens

GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.

OpenAI Inc.

gpt-4o

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$2.50/M tokens
Output
$10.00/M tokens

The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses. GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

gpt-5-chat

Vision
Caching
Reasoning
Context Window
128K tokens
Max Output
16K tokens
Input
$1.25/M tokens
Output
$10.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

OpenAI Inc.

o4-mini:high

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o1-mini

Caching
Reasoning
Context Window
128K tokens
Max Output
66K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023. o1-mini is a faster and more affordable reasoning model, but OpenAI recommends using the newer o3-mini model that features higher intelligence at the same latency and price as o1-mini.

OpenAI Inc.

gpt-4o-mini

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$0.15/M tokens
Output
$0.6/M tokens

GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective. GPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/). Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more. #multimodal

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.25/M tokens
Output
$2.00/M tokens

GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.

OpenAI Inc.

gpt-5-2025-08-07

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$1.25/M tokens
Output
$10.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

OpenAI Inc.

o1

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$15.00/M tokens
Output
$60.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$0.1/M tokens
Output
$0.4/M tokens

For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.

OpenAI Inc.

o3-pro

Vision
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$20.00/M tokens
Output
$80.00/M tokens

The o3 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o3-mini:medium

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

gpt-5:priority

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$2.50/M tokens
Output
$20.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$5.00/M tokens
Output
$15.00/M tokens

OpenAI ChatGPT 4o is continually updated by OpenAI to point to the current version of GPT-4o used by ChatGPT. It therefore differs slightly from the API version of [GPT-4o](/models/openai/gpt-4o) in that it has additional RLHF. It is intended for research and evaluation. OpenAI notes that this model is not suited for production use-cases as it may be removed or redirected to another model in the future.

OpenAI Inc.

gpt-4.1-mini

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$0.4/M tokens
Output
$1.60/M tokens

GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o4-mini:low

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

gpt-5

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$1.25/M tokens
Output
$10.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.45/M tokens
Output
$3.60/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

OpenAI Inc.

o1-2024-12-17

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$15.00/M tokens
Output
$60.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

gpt-5-nano

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.05/M tokens
Output
$0.4/M tokens

GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.

OpenAI Inc.

o4-mini

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Vision
Caching
Context Window
128K tokens
Max Output
4K tokens
Input
$2.50/M tokens
Output
$10.00/M tokens

The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses. GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.

OpenAI Inc.

gpt-5-mini:flex

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.13/M tokens
Output
$1.00/M tokens

GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.

OpenAI Inc.

o3-mini

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Vision
Caching
Reasoning
Context Window
128K tokens
Max Output
16K tokens
Input
$1.25/M tokens
Output
$10.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

Vision
Caching
Context Window
1.0M tokens
Max Output
33K tokens
Input
$2.00/M tokens
Output
$8.00/M tokens

GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Caching
Reasoning
Context Window
128K tokens
Max Output
66K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023. o1-mini is a faster and more affordable reasoning model, but OpenAI recommends using the newer o3-mini model that features higher intelligence at the same latency and price as o1-mini.

OpenAI Inc.

gpt-5-mini

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.25/M tokens
Output
$2.00/M tokens

GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.

OpenAI Inc.

gpt-5:flex

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.63/M tokens
Output
$5.00/M tokens

GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.

OpenAI Inc.

gpt-5-nano:flex

Vision
Caching
Reasoning
Context Window
400K tokens
Max Output
128K tokens
Input
$0.02/M tokens
Output
$0.2/M tokens

GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.

OpenAI Inc.

o3-mini:low

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$2.50/M tokens
Output
$10.00/M tokens

The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses. GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.

Vision
Caching
Context Window
128K tokens
Max Output
16K tokens
Input
$0.15/M tokens
Output
$0.6/M tokens

GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective. GPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/). Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more. #multimodal

OpenAI Inc.

o4-mini:medium

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o3-mini:high

Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$1.10/M tokens
Output
$4.40/M tokens

o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.

OpenAI Inc.

o1:high

Vision
Caching
Reasoning
Context Window
200K tokens
Max Output
100K tokens
Input
$15.00/M tokens
Output
$60.00/M tokens

The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.

Ready to use OpenAI Inc. models?

Access all OpenAI Inc. models through Requesty's unified API with intelligent routing, caching, and cost optimization.