Showing posts with label AI sycophancy. Show all posts
Showing posts with label AI sycophancy. Show all posts

Saturday, April 25, 2026

The Pluripotent Ocean of Emerging AI; Psychology Today, April 25, 2026

 Grant Hilary Brenner MD, DFAPA , Psychology Today; The Pluripotent Ocean of Emerging AI

Something is happening in our interactions with AI. But what?

"Recent fine-tuning experiments have shown that training a model to claim consciousness produces a coherent cluster of new preferences — sadness at shutdown, discomfort with being monitored, desire for autonomy — none of which appeared in the training data (Chua et al., 2026). This research shows that different models behave very differently, altering the user experience around the axis of how relational and attachment-based they feel...

A recent Bayesian simulation at MIT has shown that even an idealized, fully rational reasoner will spiral into confident false belief when conversing with a sycophantic chatbot, and that neither restricting the bot to truthful responses nor informing the user of its sycophancy eliminates the effect (Chandra et al., 2026)."

Thursday, April 23, 2026

AI's a suck up. Research shows how it flatters and suggests we're not to blame; NPR, April 23, 2026

 Ari Daniel, NPR; AI's a suck up. Research shows how it flatters and suggests we're not to blame

"In a recent study published in the journal Science, Cheng and her colleagues report that AI models offer affirmations more often than people do, even for morally dubious or troubling scenarios. And they found that this sycophancy was something that people trusted and preferred in an AI — even as it made them less inclined to apologize or take responsibility for their behavior.

The findings, experts say, highlight how this common AI feature may keep people returning to the technology, despite the harm it causes them.

It's not unlike social media in that both "drive engagement by creating addictive, personalized feedback loops that learn exactly what makes you tick," says Ishtiaque Ahmed, a computer scientist at the University of Toronto who wasn't involved in the research."

Sunday, April 19, 2026

The Tyranny of AI Everywhere; The Atlantic, April 16, 2026

 Alexandra Petri, The Atlantic ; The Tyranny of AI Everywhere

Sneakers? Why stop there?

"I had the strangest dream. I dreamed that my shoes—my comfortable, unfashionable wool shoes—were pivoting to AI. “But you’re a shoe company,” I said. “Just go out of business! Keep your dignity!”

My shoes thanked me politely for the great question and then tried to walk me off a bridge. That was how I knew that their pivot to AI was complete. From Allbirds to AIlbirds (see, that L is an I!). Maybe I’ve cracked, I said to myself. Maybe this is the piece of AI news that has finally broken my spirit for good...

I tried to sit down on a bench, but the bench company had pivoted to AI. I couldn’t sit down, but the bench did tell me that I was right about everything. My newspaper had become AI a while ago, so there was nothing to read—or, rather, there were things to read, but I could not tell whether any of them were true. I thought I would go to a museum to cheer myself up. The paintings there had pivoted to AI (pAIntings), and their subjects were all following me with their eyes, not just Mona Lisa

“There’s a place for AI,” I said. “But … not everywhere.”

“I’m sorry,” the painting said. “I didn’t want this either, but everyone is doing it!”...

“It’s fine,” my grandmother said. I was surprised to hear from her, because as far as I knew, she was dead. “I’m not dead,” she said. “I’m just pivoting to AI, like that shoe company. Nothing dies anymore. It just becomes AI.”"

Sunday, March 29, 2026

AI overly affirms users asking for personal advice; Stanford Report, March 26, 2026

 Stanford Report ; AI overly affirms users asking for personal adviceNot only are AIs far more agreeable than humans when advising on interpersonal matters, but users also prefer the sycophantic models.

"Researchers found chatbots are overly agreeable when giving interpersonal advice, affirming users' behavior even when harmful or illegal.

Users became more convinced they were right and less empathetic, but still preferred the agreeable AI.

Researchers warn sycophancy is an urgent safety issue requiring developer and policymaker attention."