A recent NPR report highlights growing scrutiny of OpenAI after two separate mass shooting cases were linked to alleged use of ChatGPT during planning stages. The incidents—one in the U.S. and another in Canada—have intensified concerns about how AI tools might be misused, even as companies emphasize built-in safeguards.
One case involves the 2025 Florida State University shooting, where investigators allege the suspect used ChatGPT to seek information about weapons, ammunition, and timing. This has led to a criminal investigation into OpenAI, with authorities questioning whether AI-generated responses contributed in any meaningful way to the attack.
The second case occurred in Tumbler Ridge, Canada (2026), where the attacker had previously used ChatGPT to discuss violent scenarios. The account was flagged and eventually banned by OpenAI, but the company did not alert law enforcement at the time—something that later drew heavy criticism from officials after the attack occurred.
Overall, the situation has sparked a broader debate about AI responsibility and safety systems. Critics argue that AI companies should do more to detect and escalate dangerous behavior, while others point out that chatbots often provide general or publicly available information rather than direct encouragement of violence. The key issue now is how to balance user privacy, free access to information, and proactive intervention—as governments and regulators increasingly examine the role of AI in real-world harm.