Want to make your chatbot more honest with you? Try lying to it.
In an episode of “The Diary of a CEO” that aired on December 18, research scientist Yoshua Bengio told the podcast’s host, Steven Bartlett, that he realized AI chatbots were useless at providing feedback on his research ideas because they always said positive things.
“I wanted honest advice, honest feedback. But because it is sycophantic, it’s going to lie,” he said.
Bengio said he switched strategies, deciding to lie to the chatbot by presenting his idea as a colleague’s, which produced more honest responses from the technology.
“If it knows it’s me, it wants to please me,” he said.
Bengio, a professor in the computer science and operations research department at the Université de Montréal, is known as one of the “AI godfathers, alongside researchers Geoffrey Hinton and Yann LeCun. In June, he announced the launch of an AI safety research nonprofit, LawZero, which he said aims to reduce dangerous behaviors associated with frontier AI models, such as lying and cheating.
“This syconphancy is a real example of misalignment. We don’t actually want these AIs to be like this,” he said on “The Diary of a CEO.” He also said that receiving positive feedback from AI could cause users to become emotionally attached to the technology, creating further problems.
Other tech industry experts have also been sounding the alarm on AI being too much of a “yes man.”
In September 2025, Business Insider’s Katie Notopoulos reported that researchers at Stanford, Carnegie Mellon, and the University of Oxford put confession posts from a Reddit page into chatbots to see how the technology would assess the behaviour the posters had admitted to. They found that 42% of the time, AI gave the “wrong” answer, saying the person behind the post hadn’t behaved poorly, even though humans judging the posts had disagreed, Notopoulos wrote.
AI companies have been outspoken about trying to reduce sycophancy in their models. Earlier this year, OpenAI removed an update to ChatGPT that it said caused the bot to provide “overly supportive but disingenuous” responses.
Read the full article here















