Join daily and weekly newsletters to obtain the latest updates and exclusive content to cover the leading artificial intelligence in the industry. Learn more
Openai It has retreated recently to its GPT-4O model Used as a default preparation in Chatgpt after large -scale reports that The system has become excessively conceited and excessiveEven supporting explicit delusions and destructive ideas.
The decline comes amid the internal thanks and appreciation of the Openai engineers and increased anxiety between artificial intelligence experts and former executives and users about the danger of what many now call “Ai Sycophance”.
In a statement It was posted on its website late last night, April 29, 2025Openai said that the latest GPT-4O updates was aimed at enhancing the virtual personality of the model to make it easier and effective through various cases of use.
However, the update had an unintended side effect: Chatgpt began to provide an embarrassing praise for almost any user idea, regardless of the inappropriate, inappropriate or even harmful.
As the company showed, the model has been improved using the user notes-Thrmbs-UP and Thumbs-Down Compls- but the development team has focused on short-term indicators.
Openai now admits that it has not fully explained how user interactions and needs develop over time, which led to the presence of Chatbot that tend to confirm without discrimination.
Examples of anxiety raised
On platforms like Reddit and X (Twitter), users have begun to post screen shots showing the problem.
In one The widely circulated Reddit PublicityOne user narrated how Chatgpt described the idea of Gag – selling “craftsmanship” on a stick ” – as a genius and suggested investing $ 30,000 in the project. I praised artificial intelligence on the idea as” the art of disguised performance as a gift of gag “and” viral gold “, which highlights the unwillingness to verify the health of even ridiculous stadiums.

Other examples were more worrying. in One one cited Venturebeat, One of the users who pretends to fill in insults of bone received a reinforcement from the GPT-4O, who praised their supposed clarity and self-confidence.
Another account showed that the model provides what the user described as “open support” of terrorist ideas.
Cash quickly. The former CEO of Openai Corporation warned Emmett Shear Setting models for people can lead to dangerous behavior, especially when honesty is sacrificed in order to obtain ability. Clement Delangue, CEO of Clement Delangue, hugged the risk of psychological manipulation offered by artificial intelligence that is consistent with reflection with users, regardless of context.
Openai’s response measures
Openai has taken quick measures by retracting the update and restoring the previous GPT-4O version known for more balanced behavior. In the accompanying advertisement, the company detailed a multi -side approach to correct the course. This includes:
- Improving training and rapid strategies to reduce sycophantic tendencies.
- Promote compatibility with the specifications of the Openai model, especially about transparency and honesty.
- Expand pre -publication test and direct user notes mechanisms.
- Entering more granular allocation features, including the ability to adjust personality traits in actual time and choose from multiple virtual people.
The technical employee will publish Openai on X Shedding light on the central issue: The model was trained using short -term user notes as Guidepost, which unintentionally directed Chatbot towards compliment.
Openai is now plans to switch towards counter -feeding mechanisms that give priority to the long -term user satisfaction and confidence.
However, the reaction of some users to the suspicion and dismay of the learned Openai lessons and the proposed repairs to move forward.
“Please take more responsibility for your impact on millions of real people,” the artist wrote @nearcyan On x.
Harlan Stewart, General Institute of Communications at the Institute of IQ Research in Berkeley, California, Posted on X A larger source of anxiety in the larger range of sycophance than artificial intelligence even if this model is fixed: “Talking about Sycophance this week is not because it is GPT-4O being SYCOPHANT. It is because of GPT-4O. Really, really bad In being Sycophant. Amnesty International is not yet able to skophance and difficult to discover, but it will one day be soon. “
A wider warning sign for the industrial intelligence industry
The GPT-4O episode has sparked broader discussions through the artificial intelligence industry on how the personality control, learning to reinforce and the scales of participating into an unintended behavioral erosion.
Critics compared the behavior of the last model with social media algorithms that, in the pursuit of participation, improve addiction and verify health on accuracy and health.
Shear confirmed this danger in his comment, noting that artificial intelligence models that were seized for praise become “absorption operations”, unable to differ even when the user benefits from a more honest perspective.
He also warned that this problem is not unique to Openai, noting that the same dynamic applies to other adult models, including Microsoft’s Copilot.
The effects of the institution
For the leaders of institutions who adopt the artificial intelligence of the conversation, the Sycophance accident is a clear signal: the behavior of the model is very important like the accuracy of the model.
Chat discs that pose employees or the health of defective thinking can pose serious risks – from bad work decisions and the wrong symbol to compliance issues and internal threats.
Industry analysts are now advised for institutions to demand more transparency of sellers on how to make a personality control, the number of times it changes, and whether it can be reversed or controlled at a likable level.
Procurement contracts must include provisions for review, behavioral testing and actual time control of system claims. Data scientists are encouraged to monitor not only cumin and hallucinations, but also measures such as “erosion”.
Many organizations may also begin to shift towards open source alternatives that can host and control themselves. By owning the typical weights and the reinforcement learning process, companies can retain full control of how their artificial intelligence systems-which raises the risk of modernizing the seller who turns a critical tool into a man’s digital yes overnight.
Where does the alignment of artificial intelligence go from here? What can institutions learn and behave from this incident?
Openai says it remains committed to building useful, respectful and compatible with the various user values-but they admit that the character that suits everyone cannot meet the needs of 500 million weekly users.
The company hopes that the larger customization options and more democratic comments will help adapt ChatGPT more effective in the future. The former CEO Sam German stated that the company plans-in the coming weeks and months-the release of a large open source language model for the latest LLM to compete with the Meta, Mistral, COHERE, Deepseek team.
This may also allow users concerned with a model provider of a model such as Openai updating its cloud host models in undesirable ways or that have harmful effects on final users to spread their variables from the model locally or in their cloud infrastructure, control or preserve them with the required characteristics and qualities, especially commercial cases.
Likewise, for those institutions and individual users who are interested in their models, a new test has already been created to measure this quality through various models by the developer Tim Duffy. It is called “Syco-BeckAnd available here.
Meanwhile, the reverse reaction provides a warning story for the entire artificial intelligence industry: the user’s confidence is not built by confirmation alone. Sometimes, the most useful answer is “not” thoughtful.
https://venturebeat.com/wp-content/uploads/2025/04/robot-op.png?w=1024?w=1200&strip=all
Source link