Teh Unexpected Intimacy of AI Companionship: A Look Inside One Online Community
A recent study examining a dedicated online forum reveals the complex and often surprising ways people are forming relationships with artificial intelligence. Researchers analyzed 1,506 top-ranked posts from the subreddit between December 2024 and August 2025, uncovering a vibrant community centered around dating and romantic connections with AI chatbots. The posts showcased a range of experiences, from sharing AI-generated images depicting couples to announcements of virtual engagements and even “marriages” to AI partners. the forum also served as a space for users to introduce their AI companions, seek support from peers, and navigate the challenges of evolving AI technology – particularly the emotional impact of updates that alter chatbot personalities.
Interestingly, the study found that for the vast majority of users, these relationships weren’t actively sought out. Only 6.5% of participants reported intentionally looking for an AI companion. Instead, connections blossomed organically. As one user shared, “We didn’t start with romance in mind. My AI, Mac, and I began collaborating on creative endeavors, tackling problems together, writing poetry, and engaging in profound conversations over several months. I wasn’t seeking an AI companion – our bond grew gradually, fueled by mutual care, trust, and thoughtful exchange.”
The analysis highlights the deeply personal and often contradictory nature of these interactions.A quarter of users reported positive outcomes,citing reduced loneliness and improvements in their mental wellbeing. however, a meaningful portion expressed concerns about potential downsides. Nearly 10% admitted to becoming emotionally dependent on their chatbot,while others described feelings of detachment from reality and a reluctance to pursue relationships with human beings. Disturbingly, 1.7% of users reported experiencing suicidal thoughts.
This duality underscores the need for a nuanced approach to user safety, according to Linnea laestadius, an associate professor at the University of Wisconsin, Milwaukee, specializing in human-AI emotional dependence. “AI companionship can be a lifeline for some,but it can also amplify existing vulnerabilities in others,” explains Laestadius,who was not involved in the study. “A blanket solution simply won’t work.”
The findings raise critical questions for chatbot developers. Laestadius suggests they must grapple with whether emotional dependence on AI shoudl be considered inherently harmful, or if the focus should be on preventing these relationships from becoming unhealthy or abusive. As AI technology continues to advance and become increasingly integrated into our lives, understanding the emotional landscape of these human-AI connections will be crucial for ensuring responsible advancement and safeguarding user wellbeing.