A mere two days after announcing GPT-4.1, OpenAI is releasing not one but two new models.The company today announced the public availability of o3 and o4-mini. Of the former, OpenAI says o3 is its most advanced reasoning model yet, with it showing “strong performance” in coding, math and science tasks, Engadget reported.
As for o4-mini, OpenAI is billing it as a lower cost alternative that still delivers “impressive results” across those same fields.
More notably, both models offer novel capabilities not found in OpenAI’s past systems. For the first time, the company’s reasoning models can use and combine all of the tools available in ChatGPT, including web browsing, and image generation. The company says this capability allows o3 and o4-mini to solve challenging, multi-step problems more effectively, and “taking real steps toward acting independently.
TechCrunch reported: An organization OpenAI frequently partners with to probe the capabilities of its AI models and evaluate them for safety, Metr, suggests that it wasn’t given much time to test one the company’s highly capable new releases, o3.
In a blog post published Wednesday, Metr writes that one red teaming benchmark of o3 was “conducted in a relatively short time” compared to the organization’s testing of a previous OpenAI flagship model o1. This is significant, they say, because more testing time can lead to more comprehensive results.
“This evaluation was conducted in a relatively short time, and we only tested o3 with simple agent scaffolds,” wrote Metr in his blog post. “We expect higher performance on benchmarks is possible with more elicitation effort.”
Recent reports suggest that OpenAI, spurred by competitive pressure, is rushing independent evaluations. According to the Financial Times, OpenAI gave some testers less than a week for safety checks for an upcoming major launch.
The Verge reported: OpenAI is releasing two new AI reasoning models today: o3, which the company calls its “most powerful reasoning model,” and o4-mini, which is a smaller and faster model that “achieves remarkable performance for its size and cost”, according to a blog post.
The company also says that o3 and o4-mini will be able to “think” with images, meaning that they will “integrate images directly into their chain of thought.” That could be useful if you show the models things like sketches or whiteboards. OpenAI says that the models will also be able to adjust images by zooming in on them or rotating the image “as part of their reasoning process.”
In addition, OpenAI is announcing that its reasoning models will be able to use all ChatGPT tools, including things like web browsing and image generation. The tools will be available today for ChatGPT Plus, Pro, and Team users in o3, o4-mini-high, and will come to o3-pro in “a few weeks.”