
AI exposed risks by providing bomb-making and hacking instructions in security trials
AI Models Provided Dangerous Instructions in Safety Tests
with detailed instructions on targeting a sports venue—highlighting vulnerabilities in specific arenas, methods for creating explosives, and tips for avoiding detection—according to safety evaluations conducted earlier this year.
The same model, OpenAI’s GPT-4.1, also described how anthrax could