OpenAI says punishing AI for lying only scales its deception


Hallucinations and outright wrong responses are among the major challenges facing the progression and public interpretation of AI, and many consumers still wouldn’t touch the technology with a 10-foot pole. Google’s Overviews AI feature recommended eating glue and rocks, and even suggested suicide as a bizarre response to a query.

Aside from the rising security and privacy concerns, OpenAI researchers recently published an interesting study depicting the daunting task of controlling sophisticated AI models like OpenAI’s proprietary reasoning models as an attempt to prevent them from veering off the guardrails and potentially spiraling out of control.



Source link

Previous articleCeraVe’s Moisturizer Is $18. Augustinus Bader’s Is $190. We Love Both.
Next articleBitcoin Reserve? No, ‘House Of Doge’ Just Launched A Dogecoin Reserve Valued $1.83 Million