Sycophancy

According to a recent paper by researchers at Anthropic — the AI startup founded in 2021 by a handful of former OpenAI employees, which has since raised over $6 billion and released several versions of the chatbot Claude — sycophancy is a major problem with AI models.

The Anthropic researchers — Mrinank Sharma, Meg Tong, and Ethan Perez — didn’t just detect sycophantic behavior in Claude, but in every leading AI chatbot, including OpenAI’s ChatGPT, raising a host of troubling questions about the reliability of chatbots in fields where truth — whether we like it or not — matters.

These tools may revolutionize fields like medicine and fusion research — but they may also be increasingly designed to tell us just what we want to hear. 

Comments

Popular posts from this blog

Hamza Chaudhry

When their AI chums have Bob's data

Swarm 🦹‍♂️