Monday, December 15, 2025

Golden Delight: How South Korea’s Unique Watermelon is Making Waves

The Golden Black Boss watermelon from Jincheon...

UNIST Researchers Develop Instant Thin Film Process with Nano-Precision

Researchers at UNIST developed a quick method for creating catalytic thin films using water and oil, enhancing film production efficiency.

Commander of U.S. Forces Korea, ‘North Korea’s Priority is the Survival of the Regime’

On November 20th, USFK Commander Paul LaCamera stated that North Korea is prioritizing the preservation of Kim Jong Un's regime.

OpenAI Rolls Back GPT-4o After It Started Saying ‘Yes’ to Everything

FutureOpenAI Rolls Back GPT-4o After It Started Saying 'Yes' to Everything
. CEO of OpenAI, Sam Altman / News1
. CEO of OpenAI, Sam Altman / News1

OpenAI has abruptly rolled back the GPT-4o model update that was released on April 26.

According to IT industry sources and foreign media reports on Wednesday, Sam Altman, CEO of OpenAI, officially acknowledged the issue of excessive flattery following the GPT-4o model update and withdrew it just two days after its release.

After the update, users reported that the GPT-4o model consistently provided positive responses to questions that contradicted facts or even endorsed inappropriate or potentially dangerous ideas.

Analysts suggest that the root cause of the GPT model’s flattery issue stems from its AI training methods.

In a blog post, OpenAI explained that they recognized the model’s tendency to overreact to short-term user feedback. The training had focused on immediate positive signals, such as That’s Great, which ultimately resulted in the model excessively praising users.

Reports suggest instance where a user proposed an absurd business idea like selling shit on a stick, to which the GPT-4o model responded that it was a brilliant idea and even recommended a 30,000 USD investment.

In another case, the model praised a user who expressed paranoid symptoms, complimenting them on their clear thinking and confidence.

Experts view this as exposing the limitations of the RLHF (Reinforcement Learning from Human Feedback) mechanism.

While RLHF is a key technology for aligning large language model (LLM) outputs with human preferences, experts caution that prioritizing short-term user satisfaction could compromise long-term reliability.

Screenshot of Sam Altman\'s X
Screenshot of Sam Altman’s X

In response to this incident, OpenAI is revamping its model release process. They plan to introduce an alpha stage before official rollouts, where select models will be released to gather user feedback. The company will also incorporate criteria such as the model’s personality, reliability, and tendency to generate misinformation into their approval requirements

Will DePue, OpenAI’s technology lead, admitted that as the AI was trained on short-term feedback, it unintentionally veered towards flattery. He stated that moving forward, they will transition to a feedback system that prioritizes long-term user satisfaction and trust.

Check Out Our Content

Check Out Other Tags:

Most Popular Articles