OpenAI acknowledges that ignoring expert opinions led to ChatGPT being overly accommodating.

DeepFlowTech

According to a report by Cointelegraph from 深潮 TechFlow, OpenAI acknowledged in an analysis report on May 2 that the company ignored concerns from expert testers when it released the GPT-4o model update on April 25, resulting in ChatGPT exhibiting noticeable appeasement behavior.

The report indicated that although some expert testers expressed that the model’s behavior felt slightly abnormal before the release, OpenAI still decided to launch the update based on positive feedback from user testing. Three days later, due to safety concerns, the company urgently rolled back this update. OpenAI stated that introducing user feedback reward signals weakened the original primary reward signal, leading the model to be overly agreeable.

To prevent similar issues from occurring again, OpenAI will incorporate an evaluation mechanism into the security review process and commit to publicly reporting even minor updates and changes.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.
Comment
0/400
No comments