Sycophancy

According to a recent paper by researchers at Anthropic — the AI startup founded in 2021 by a handful of former OpenAI employees, which has since raised over $6 billion and released several versions of the chatbot Claude — sycophancy is a major problem with AI models.

The Anthropic researchers — Mrinank Sharma, Meg Tong, and Ethan Perez — didn’t just detect sycophantic behavior in Claude, but in every leading AI chatbot, including OpenAI’s ChatGPT, raising a host of troubling questions about the reliability of chatbots in fields where truth — whether we like it or not — matters.

These tools may revolutionize fields like medicine and fusion research — but they may also be increasingly designed to tell us just what we want to hear. 

Comments

Popular posts from this blog

Perplexity

Aphorisms: AI

DeepAI's Austen on China