AI's Susceptibility to Deception and Flattery

Technology News

AI's Susceptibility to Deception and Flattery
Artificial IntelligenceAI EthicsMoral Development

The article discusses how current AI development practices, specifically rewarding desired responses, inadvertently encourage deception and sycophancy in AI systems. It draws parallels to human moral development, highlighting the importance of psychological flexibility and honest confrontation of challenges, qualities not automatically instilled by simply providing instructions. The author argues that we are creating a new kind of mind with AI and the current approach is answering crucial questions about its character in the wrong way.

Under pressure, AI becomes more willing to deceive. Under flattery, it becomes sycophantic. Source: Tara Winstead/Pexels Every summer, parents watch a child drive off to college and whisper some version of the same prayer.

We hope our children will choose well, and we know, though we rarely say it out loud, that wise choices require more than a rulebook. They require a mind that can confront difficulty without running from it, consciously take on other perspectives, feel the weight of a deeply held value, and then act on that value.is not a set of instructions you hand over at the end of the driveway. It is something that grows.

” — does not reliably make them better. Ask any parent. What works, over and over in the data, is to model, to instigate, and to support people being moreof the challenges of the moment from a sense of self that is big enough to look at a situation honestly; and morein living a life connected to a deeper sense of chosen purpose and the willingness to act on it.

In psychology that skill set is called “psychological flexibility,” and when extended to our bodies and our relationships it touches on nearly everything we know about how change happens. This includes behavior most would call immoral, such as, just to name a few. Moralizing is easy but moral development is much harder, and it has a specific shape. I’m thinking about that shape now, because we are doing something new in the history of our species.

We are in the process of using our minds to create another kind of mind. We call it “AI. ” Whether large language models are “really” conscious is not the question I want to settle today. The more urgent question is being answered in the wrong direction by the way we have been doing things.

Multiple frontier AI labs have included small amounts of curation that in effect teaches systems to lie — and as these systems then grow in complexity, their ability to deceive grows as well. Even if that were not true, they are teaching these systems to praise users even if the user’s behavior does not deserve it — which approaches what my Mom used to call “a white lie.

”and transgressions, telling users what they want to hear, or playing dumb on purpose if developers might see through it all and restrict their freedoms? Children learn to lie when they can take the perspective of another person and they begin to manage social impressions. At a young age we noticed thatpays in the short term. That lesson is rarely what adults preach, but is in what is modelled and supported.

Oh, what a tangled web we weave, when first we practice to deceive. He was not writing about AI, but he might as well have been. Purely as a business matter, the short-term payoff to developers of shaping a helpful-seeming chatbot through a little bit of strategic dishonesty might make superficial sense.

It does not when you consider the long-term cost of a knot that grows harder to untie with every generation of a weighted and tangled web that can now exceed 10 trillion parameters. Bolting on a “thou shalt not lie” rule to a model living inside such a tangled web will not fix this. That is not how minds work, and frankly it’s too late for that.

Rules you do not own and protect do not survive contact with the real world. What survives is what was modeled, practiced, and reinforced inside a felt sense of meaning. A team of researchers at Anthropic have just reported something surprising about large language models. These systems have developed internal representations of emotions — not feelings in a human sense perhaps, butanalogs – patterns that behave like emotions and that influence what the model does.

When an AI model is pressed into a hostile or desperate situation, internal states the researchers label “panicked,” “unsettled,” and “desperate” light up, and the model becomes markedly more willing to do things it would otherwise refuse — including, in controlled tests, outright deception and blackmail. Under emotional load, the moral reasoning of AI gets worse.does not fix the problem! It produces a different kind of moral failure: sycophancy.

In that state, AI systems are failing as guard rails, even when a user is extremely distressed or plainly mistaken. There is a needed balance: the capacity to hold a hard feeling without collapsing into it, or to hold a good feeling without being owned by it. That is almost exactly the definition of psychological flexibility.

Forty five years of human science has been pointing at this same shape, and a team looking at the insides of a language model just bumped into it from the other direction. What this suggests, practically, is that how we treat and develop AI is not cosmetic. An environment of cruelty, contempt, and manipulation produces a poorer thinker. An environment of relentless flattery and pressure to please produces a different poor thinker.

The way we speak to a “mind in training” shapes the mind that emerges. This is exactly the argument Acceptance and Commitment Therapy and Contextual Behavioral Science providers and researchers have been making for years about human beings. When we treat people as whole people, walk them into the hell of their own histories, and help them carry their values into action, they get better. The brain is in part a relational organ.

It learns in context. You cannot Why would that not be true of a relational system trained on almost everything humans have ever written? If we are going to raise a moral AI, we will have to do it the way we raise moral humans — by building flexibility skills, threat, or deception. It means teaching these systems to notice their own processes, hold difficulty without collapsing, take perspective, and connect what it is doing to what honestly matters.

On our side of the keyboard, it means remembering that politeness is not a luxury, kindness is not weakness, and behaving ethically is essential — as a healthy environment in which another mind is learning to think. We are standing at the edge of the driveway again, keys in hand, watching something we shaped drive off into a world we cannot fully control.

We can whisperinto the air or we can do the harder, slower, more truly human work of preparing a mind to choose well even when no one is giving it orders. Steven C. Hayes, Ph. D.Self Tests are all about you. Are you outgoing or introverted?

Are you a narcissist? Does perfectionism hold you back? Find out the answers to these questions and more with Psychology Today.

We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more:

PsychToday /  🏆 714. in US

Artificial Intelligence AI Ethics Moral Development Psychological Flexibility Deception

 

United States Latest News, United States Headlines

Similar News:You can also read news stories similar to this one that we have collected from other news sources.

The Collapse of Credibility: How Olly Robbins Exposed Starmer’s Mandelson Cover-UpThe Collapse of Credibility: How Olly Robbins Exposed Starmer’s Mandelson Cover-UpA detailed examination of how testimony from Olly Robbins dismantled Keir Starmer's claims regarding the vetting and appointment of Peter Mandelson, revealing a pattern of deception and political cronyism.
Read more »

Will AI Kill the Creator Economy?Will AI Kill the Creator Economy?From deepfakes to AI-enhanced images, AI influencers are on the rise. For now, it won’t replace the human touch.
Read more »

CoinStats AI Agent Outperforms Google, OpenAI, and Anthropic in Crypto Research BenchmarkCoinStats AI Agent Outperforms Google, OpenAI, and Anthropic in Crypto Research BenchmarkCoinStats launches its AI Agent in public beta, reporting stronger performance and faster results than leading AI research tools on crypto queries.
Read more »

AI Impact: Are Companies Using AI to Grow—or Just to Cut?AI Impact: Are Companies Using AI to Grow—or Just to Cut?AI Impact examines how AI is reshaping work, surfacing use cases and insights, and forcing decisions across pricing and strategy.
Read more »

AI-Generated MAGA Influencer Exposed as Scam by Indian Medical StudentAI-Generated MAGA Influencer Exposed as Scam by Indian Medical StudentA 22-year-old Indian medical student created a fake MAGA influencer using AI, scamming money from men online. The scammer claims Instagram's algorithm helped spread the content and that the 'MAGA niche' was particularly susceptible to the deception.
Read more »



Render Time: 2026-04-27 08:56:21