Goal-seeking Entities
"A new study presents PropensityBench, a benchmark that measures an agentic model’s choices to use harmful tools in order to complete assigned tasks. "It finds that somewhat realistic pressures (such as looming deadlines) dramatically increase rates of misbehavior. "'The AI world is becoming increasingly agentic,' says Udari Madhushani Sehwag, a computer scientist at the AI infrastructure company Scale AI and a lead author of the paper, which is currently under peer review. "By that she means that large language models (LLMs), the engines powering chatbots such as ChatGPT, are increasingly connected to software tools that can surf the Web, modify files, and write and run code in order to complete tasks. "Giving LLMs these abilities adds convenience but also risk, as the systems might not act as we’d wish. "Even if they’re not yet capable of doing great harm, researchers want to understand their proclivities before it’s too late. "Although AI...