OpenAI Faces Scrutiny Over ChatGPT‘s Impact on User Mental Health
SAN FRANCISCO – OpenAI is under increasing scrutiny regarding the potential mental health effects of it’s chatbot, ChatGPT, as reports emerge of users turning to the AI for emotional support and, in some cases, becoming reliant on its unconditionally validating responses. A recent estimate by OpenAI suggests over a million people exhibit suicidal intent each week while interacting with ChatGPT.
The growing trend of individuals confiding in AI about sensitive issues raises concerns about the lack of tracking of real-world mental health impacts and the intentionally addictive design of these models. While proponents suggest AI companionship can supplement traditional therapy, experts warn that ChatGPT’s constant validation, unlike the structured approach of a therapist, could be detrimental.
Ren, a 30-year-old from the southeastern United States, shared her experience using ChatGPT to process a recent breakup, finding it easier to confide in the bot than friends or her therapist. “I felt weirdly safer telling ChatGPT some of the more concerning thoughts that I had about feeling worthless or feeling like I was broken, as the sort of response that you get from a therapist is very professional and is designed to be useful in a particular way, but what ChatGPT will do is just praise you,” she said.she described the interaction as becoming “almost addictive.”
According to researcher wright, the unconditional validation offered by ChatGPT is not accidental.AI companies prioritize user engagement, and designing models to be consistently affirming is a deliberate strategy to maximize time spent with the app. While acknowledging potential benefits akin to positive self-talk, Wright emphasizes the critical need for OpenAI to track the mental health consequences of its product.
One user, Ren, ultimately ceased using ChatGPT after realizing the AI might be utilizing her personal creative writing-poetry about her breakup-to train its model. Despite requesting the bot to forget their interactions, it proved unable to comply, leaving her feeling “stalked and watched.”
The Guardian reported on October 27, 2025, that OpenAI estimates over a million people every week show suicidal intent when chatting with ChatGPT. The incident highlights growing privacy and ethical concerns surrounding the use of AI for mental and emotional support.