img {
display: block;
margin: 0 auto;
}
.aligncenter {
text-align: center;
}
Shantanu Kumar / Pexels
After months of speculation and anticipation, OpenAI has finally unveiled the production version of its advanced reasoning model, Project Strawberry, which has now been rechristened “o1.” Alongside it comes a “mini” version (just like GPT-4o did), though this one offers quicker and more responsive interactions at the expense of relying on a larger knowledge base. It seems that o1 showcases a blend of technological advancements. It is the first in OpenAI’s line of reasoning models designed to utilize human-like deduction to answer complex questions on various subjects such as science, coding, and math at a speed that surpasses humans. For instance, during testing, o1 was presented with a qualifying exam for the International Mathematics Olympiad. While its predecessor, GPT-4o, could only correctly solve 13% of the problems presented, o1 managed to get an impressive 83% right. In an online Codeforces competition, o1 scored in the 89th percentile. Additionally, o1 can respond to queries that have stumped previous models, like “which is bigger, 9.11 or 9.9?” However, the company makes it clear that this release is only a preview of the neophyte model’s full potential. The new o1, as per OpenAI’s research lead, Jerry Tworek, who told The Verge, has been trained using a completely new optimization algorithm and a specially tailored new training dataset. By employing a combination of reinforcement learning and “chain of thought” reasoning, o1 reportedly yields more accurate inferences than its predecessor. “We have noticed that this model hallucinates less,” Tworek said, yet he cautioned, “we can’t claim that we have completely eliminated hallucinations.” Both ChatGPT-Plus and Teams subscribers will be able to test out o1 and o1-mini starting today. Enterprise and Edu subscribers are expected to have access by next week. The company states that o1-mini will eventually be made available to free-tier users, though it didn’t specify a timeline. Developers will notice a significant increase in the API pricing for o1 compared to GPT-4o. Access to o1 will cost $15 per million input tokens (compared to $5 per million for GPT-4o) and $60 per million output tokens, which is four times more than the $5 per million fee for 4o. The real question now is whether the new model thinks the word “strawberry” contains two R’s or three.