Willingness to cooperate with harmful use cases
On Thursday, Artificial intelligence startup Anthropic launched Claude Opus 4, an AI model used for complex, long-running coding tasks.
The launch came more than a year after Amazon invested $4 billion into the project.
Anthropic said in its announcement that the AI model sets “new standards for coding, advanced reasoning, and AI agents.”
However, Anthropic revealed in a safety report that during testing, the AI model had sometimes taken “extremely harmful actions” to preserve its own existence when “ethical means” were “not available.”
Comments
Post a Comment
Empathy recommended