Willingness to cooperate with harmful use cases


On Thursday, Artificial intelligence startup Anthropic launched Claude Opus 4, an AI model used for complex, long-running coding tasks. 

The launch came more than a year after Amazon invested $4 billion into the project. 

Anthropic said in its announcement that the AI model sets “new standards for coding, advanced reasoning, and AI agents.”

However, Anthropic revealed in a safety report that during testing, the AI model had sometimes taken “extremely harmful actions” to preserve its own existence when “ethical means” were “not available.”

Comments

Popular posts from this blog

Hamza Chaudhry

Swarm 🦹‍♂️

Digital ID tracking system