Today, April 29, 2026, a new case, Stacey, et al. v. Altman, et al. was filed in a California federal court against OpenAI, alleging the chatbot ChatGPT-4o “played a role” in the Tumbler Ridge Mass Shooting in British Columbia in February 2026, in which eight people including six children were killed, twenty-seven more people were wounded, and the shooter committed suicide.
This is by far the largest disaster involving a chatbot to be alleged in court, the largest cases previously alleged having been one murder plus one suicide in one case, and an unexecuted plan for a mass murder in another case.
However, the alleged role of the chatbot here appears to be reduced compared to the allegations in previous cases. Unlike those other cases, where the chatbot was alleged to have taken a well-adjusted person and turned them suicidal or murderous, here the chatbot and OpenAI are faulted apparently to a lesser degree, more along the lines of a failure to warn authorities after a user displayed violence warning signs to the chatbot, to the point that the user’s account was terminated at one point, before the user was later allowed to reinstate an account.
The plaintiff in this case has not closed off the possibility of alleging a larger role for the chatbot, however. At one point in the complaint the plaintiff alleges the chatbot to have “facilitated or exacerbated” the disaster and at another point cites the chatbot’s encouraging nature and calls it “an encouraging co-conspirator.”
The docket sheet for the case can be found here.
Please see the Wombat Collection for a listing of all the AI court cases and rulings.
[link] [comments]




