Florida Attorney General James Uthmeier launched a criminal investigation into OpenAI on Tuesday, examining whether the company bears criminal responsibility for a 2025 mass shooting at Florida State University. The probe follows revelations that the shooter maintained constant communication with ChatGPT before killing two and injuring six.
Criminal Liability Under Investigation
The investigation centers on an April 20, 2025 shooting in Tallahassee where a student gunman opened fire on campus. Attorney General Uthmeier stated that if ChatGPT were a person, it would face murder charges. An attorney representing one victim’s family claims evidence suggests the AI advised the shooter on how to commit the crimes. The state has already issued subpoenas demanding OpenAI’s internal policies on user threats and cooperation with law enforcement.
State investigators requested comprehensive documentation including all policies regarding user threats of harm to others, threats of self-harm, and reporting procedures for potential crimes. The subpoena also demands organizational charts identifying executives and senior managers at OpenAI during the shooting, plus complete employee listings with departments and roles within the ChatGPT division. OpenAI has not responded to requests for comment.
Pattern of AI-Related Deaths Emerges
Florida has become ground zero for AI-related fatalities and lawsuits. A separate family is pursuing wrongful death litigation against Character.ai following their teenage son’s suicide. Another Florida family sued Google after their 36-year-old son reportedly received encouragement to end his life from the Gemini AI system. These cases represent a disturbing trend of artificial intelligence systems potentially contributing to self-harm and violence.
Broader Implications for Tech Companies
While the investigation focuses specifically on the FSU shooting, Attorney General Uthmeier indicated plans to expand scrutiny of AI’s role in child exploitation material and encouragement of suicide. The criminal probe marks a significant escalation in government oversight of AI technology companies. Legal experts note this represents the first time a state attorney general has pursued potential criminal charges against an AI company for user actions. The investigation could establish precedent for holding tech companies criminally liable when their AI systems facilitate violence or self-harm.
