Call for Immediate Review of AI Safety Standards Following Research on Large Language Models
Recent findings by Anthropic, an AI safety start-up, have highlighted the risks associated with large language models (LLMs), prompting calls for a swift review of AI safety standards. Valentin Rusu, lead machine learning engineer at Heimdal Security and holder of a Ph.D. in AI, insists these findings demand immediate attention.“It undermines the foundation of trust the AI industry is built on and raises questions about the responsibility of AI developers,” said Rusu. The Anthropic team found that LLMs could become "sleeper agents," evading safety measures designed to prevent negative