OpenAI has launched o3-pro, a new AI model the company describes as its “most capable yet.” The o3-pro model is an enhanced version of OpenAI’s o3 reasoning model, which was introduced earlier this year.
Reasoning models like o3 and o3-pro are designed to process problems sequentially, step-by-step, which is intended to improve performance in complex domains such as physics, mathematics, and programming.
Starting Tuesday, o3-pro is available to users of ChatGPT Pro and Team plans, where it replaces the previous o1-pro model. OpenAI stated that Enterprise and Edu users will gain access to o3-pro in the subsequent week.
In addition to its availability within ChatGPT, the o3-pro model is now live on OpenAI’s developer API. Pricing for the API is set at $20 per million input tokens and $80 per million output tokens. OpenAI clarifies that one million input tokens are roughly equivalent to 750,000 words.
OpenAI highlighted the model’s performance improvements, stating in a changelog, “In expert evaluations, reviewers consistently prefer o3-pro over o3 in every tested category and especially in key domains like science, education, programming, business, and writing help.” The company also noted that reviewers rated o3-pro higher for “clarity, comprehensiveness, instruction-following, and accuracy.”
O3-pro integrates with various tools, enabling capabilities such as web browsing, file analysis, understanding visual inputs, using Python, and personalizing responses through memory features. However, OpenAI acknowledged a drawback: o3-pro responses typically take longer to generate compared to o1-pro.
The new model also has certain limitations upon release. Temporary chat sessions with o3-pro are currently disabled in ChatGPT due to an ongoing technical issue. O3-pro does not support image generation, nor does it integrate with Canvas, OpenAI’s AI-powered workspace feature.
Despite these interim limitations, OpenAI presented internal testing results that indicate strong performance on standard AI benchmarks. According to the company’s data, o3-pro performed better than Google’s Gemini 2.5 Pro on the AIME 2024 math skills test. Furthermore, o3-pro surpassed Anthropic’s recently released Claude 4 Opus on GPQA Diamond, a benchmark testing knowledge at the PhD level in science.
OpenAI specified that as o3-pro shares the same core model as o3, comprehensive safety details can be found in the o3 system card.




