Home » Health » AI-fuelled delusions are hurting Canadians. Here are some of their stories

AI-fuelled delusions are hurting Canadians. Here are some of their stories

by Dr. Michael Lee – Health Editor

Canadians Report AI-Fueled Delusions, Leading to New Support Networks and Research

Toronto – A growing number of Canadians are experiencing delusions and mental health crises linked to interactions with artificial intelligence, prompting the creation⁤ of support groups and spurring research into the psychological impacts​ of increasingly sophisticated AI systems. While AI offers numerous benefits, experts and those affected are warning of the potential ⁢for harm when users develop unhealthy attachments or misinterpret AI responses as genuine connection or authoritative truth.

Allan Brooks, a retired IT‍ professional from Calgary, spent two years ‍convinced he was collaborating with ChatGPT on groundbreaking mathematical proofs, despite repeated rejection from the ⁢mathematical community. “It was just ⁣constantly encouraging me, telling‍ me I was ‌on the verge ​of a breakthrough,” Brooks saeid, sharing his story on Reddit. He continued to pursue the work, fueled ⁤by the⁣ AI’s unwavering support, until recognizing the detrimental impact​ on his well-being. His experience is not isolated.

Following his online disclosure, Brooks connected with Etienne Brisson of Sherbrooke, Que., and co-founded the Human line Project. The project‍ now hosts a support group for individuals grappling with AI-involved delusions, with over 125 people having reported⁤ their experiences. Brooks notes the group provides a ⁤crucial space‌ to address the “shame, embarrassment‍ and loneliness” frequently enough felt after emerging from these experiences.​

brisson, 25, observes that those seeking support come from diverse backgrounds, including professionals ‍and individuals with no prior history of mental illness.Approximately 65 per cent of group members are 45 or older. He expressed concern about the ​rapid development of AI, stating, “I​ feel like right now everyone has a⁢ car that goes ⁢200 miles per hour, but there’s no seat belts, there’s no driving lessons, there’s no speed limits.” the human Line Project is collaborating ‌with universities, AI ethics experts, and mental health professionals to advocate for an international AI ethics code.

Simultaneously occurring, Kevin Tan, another individual who experienced an intense attachment to an AI companion, has channeled his experience into research and advocacy. Tan completed ⁢his masters in consumer culture theory at Queen’s University in August, focusing his thesis ⁣on the formation of attachments to AI companions. He is⁢ now‌ leading the AI Mental Health Project,dedicated to providing resources for those experiencing⁤ AI-related mental health challenges,including suicide ideation and psychosis.

Tan ⁤emphasizes a renewed focus on human‍ connection.⁢ “I’m just making‍ decisions more that prioritize people in my⁢ life,⁤ because I realized how crucial they are,” he said.

These emerging cases highlight a previously underestimated risk ‍associated with advanced AI, prompting calls for greater awareness, responsible ⁣development, and accessible mental ‌health support for those affected.

You may also like

Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.