OpenAI recently unveiled its latest AI model, code-named “Strawberry” and officially released as o1-preview. This model is touted for its reasoning capabilities, allowing it to logically work through complex problems, a step forward from earlier AI models. However, the rollout has sparked criticism as OpenAI is accused of keeping key details about how the model reasons hidden from users, raising transparency concerns. Several users reported receiving warning emails as OpenAI threatens bans over questions on Strawberry’s reasoning processes.
Some social media accounts mention that terms like “reasoning trace” or even asking the model directly about its reasoning led to warnings, with threats of losing access if violations continue. These flagged requests reportedly violated OpenAI’s usage policies, which aim to prevent users from bypassing safeguards built into the AI.
In these warning emails, OpenAI specifically notes that attempts to circumvent its safety measures could lead to a ban. This has fueled complaints from some developers and researchers.
Balancing Safety and Competitive Edge
OpenAI threatens bans over questions on Strawberry’s reasoning, limiting transparency into how the model arrives at answers. OpenAI defends its stance, explaining that hiding the AI’s reasoning process helps maintain compliance with safety policies. The company suggests that exposing the “raw” thought process could lead to unintended or harmful outputs, making it necessary to filter certain aspects. Additionally, OpenAI acknowledges that keeping the reasoning process hidden helps maintain a competitive edge by preventing competitors from replicating its AI models.
This decision, however, has drawn criticism from members of the AI community. Independent researchers, such as Simon Willison, have expressed disappointment, saying that transparency is crucial for developers working with large language models. Willison and others argue that OpenAI’s move toward secrecy marks a step backward from the company’s original vision of open AI development.
Hidden Reasoning Sparks Debate
The reasoning abilities of “Strawberry” were a major selling point during its release. The model is capable of articulating its thought process step-by-step, allowing it to solve complex problems that previously stumped other models, including OpenAI’s GPT-4. But OpenAI has decided to keep this reasoning chain hidden from users, except in a watered-down version that is generated by a secondary model.
According to OpenAI, this hidden chain of thought is essential for monitoring the AI’s behavior and ensuring it follows safety guidelines. By analyzing the reasoning process internally, the company aims to ensure that the model operates within ethical and safe boundaries without showing potentially harmful thoughts to users.
Despite the controversy, OpenAI’s new model marks a significant advance in AI reasoning. The company has been working on improving logical reasoning in its models through reinforcement learning. This method teaches the AI to refine its thought process based on feedback, enhancing its ability to tackle complex tasks in areas like mathematics, chemistry, and coding.
OpenAI demonstrated the new model’s capabilities by solving challenging problems that previous models struggled with, such as intricate math puzzles. The o1-preview model performed much better than its predecessors, including on tests like the American Invitational Mathematics Examination (AIME), where it achieved an 83% success rate compared to GPT-4’s 12%.
Impact on the AI Community
Some social media users revealed that OpenAI threatens bans over questions on Strawberry’s reasoning if specific terms like “reasoning trace” are used. While OpenAI continues to push forward with its reasoning-focused AI models, the company’s decision to withhold the full reasoning process has sparked concern among researchers. Some believe that this lack of transparency could hinder efforts to improve AI safety, as red teamers—who typically try to hack AI models to identify vulnerabilities—are limited in their ability to understand how the model operates.
Others, like Mozilla’s Marco Figueroa, argue that this approach could also stifle innovation. Figueroa recently shared his experience of receiving a warning from OpenAI for attempting to use reasoning-related prompts, calling the situation a setback for AI red-teaming efforts aimed at making AI models safer.
Also Read: California Signs Laws to Protect Actors Against AI in Major Move.