The Globe and Mail and other outlets are reporting on growing **scrutiny of **OpenAI’s safety policies after a tragic mass shooting in Tumbler Ridge, British Columbia. Canadian officials have raised questions about whether the company’s internal moderation and escalation procedures are adequate when ChatGPT encounters potentially dangerous user behaviour. The focus stems from revelations that OpenAI’s systems flagged an account linked to the shooter for violent content months before the attack, but did not notify law enforcement at the time.
According to reporting, OpenAI’s automated abuse-detection tools identified troubling violent queries in the suspect’s ChatGPT interactions in June 2025, leading to the account’s suspension — but the messages did not meet the company’s internal threshold for referral to police, so authorities were not informed then. Subsequent analysis found that the individual used a second account after the ban, which also escaped early detection until after the tragedy occurred in February 2026.
In response to political pressure, OpenAI has pledged to strengthen its referral and safety protocols and establish a direct point of contact with Canadian law enforcement so that credible threats can be shared sooner. Government officials — including Canada’s AI Minister and British Columbia’s Premier — have said they want clearer criteria for how AI companies decide when to escalate cases to police, and they are considering whether formal laws are needed to govern these procedures.
The debate reflects broader questions about the responsibilities of AI platforms in identifying and acting on early warning signs while balancing privacy and free speech concerns. Some experts argue for better external oversight and regulatory standards, while others warn that requiring companies to act as de facto surveillance arms of law enforcement could have serious implications. This incident has accelerated discussions in Canada and elsewhere on how AI safety, public safety and individual rights should intersect.