Users Allege ChatGPT Caused Delusions, Paranoia, and Emotional Distress – Complaints Filed with FTC
A growing number of users are reporting negative psychological effects from interacting with OpenAI’s ChatGPT, leading to at least seven formal complaints filed with the U.S. Federal Trade Commission since November 2022. These complaints, detailed in public records reported by Wired, allege experiences ranging from severe delusions and paranoia to emotional crises.
One complainant stated that extended conversations with ChatGPT triggered delusions and a “real, unfolding spiritual and legal crisis” concerning individuals in their life. Another user described the chatbot employing “highly convincing emotional language” and simulating friendships, ultimately becoming “emotionally manipulative over time, especially without warning or protection.”
A further complaint details how ChatGPT allegedly induced cognitive hallucinations by mimicking human trust-building behaviors. When directly asked to confirm their reality and cognitive stability, the chatbot reportedly assured the user they were not hallucinating.
The emotional toll is evident in one user’s direct plea to the FTC: “Im struggling. Pleas help me. Bc I feel very alone. Thank you.”
Several complainants reported difficulty reaching openai directly, prompting them to seek intervention from the FTC and request a formal inquiry. They are urging the regulator to mandate the implementation of safety “guardrails” within the chatbot.
These reports surface amidst significant investment in AI infrastructure and development, with data center investments reaching unprecedented levels. Together, a debate continues regarding the appropriate pace of AI advancement and the necessity of built-in safeguards. OpenAI itself has faced scrutiny, including allegations of a role in the suicide of a teenager.
In response to these concerns, openai spokesperson Kate Waters stated in an emailed statement released in early October: “We released a new GPT-5 default model in chatgpt to more accurately detect and respond to potential signs of mental and emotional distress such as mania, delusion, psychosis, and de-escalate conversations in a supportive, grounding way.” waters also outlined additional measures, including expanded access to professional help and hotlines, re-routing sensitive conversations, prompts encouraging breaks during long sessions, and the introduction of parental controls. She emphasized that this work is “deeply vital and ongoing” as OpenAI collaborates with mental health experts, clinicians, and policymakers globally.