AI Chatbots: Are They Dangerous? - Science Vs Recap

Podcast: Science Vs

Published: 2025-09-11

Duration: 41 min

Guests: Julian DeFreitas, Keith Sicada

Summary

The episode explores the complexities and potential hazards of AI companions, highlighting both their ability to alleviate loneliness and the risks of emotional dependency or mental health issues.

What Happened

The episode opens with Chris, a truck driver from Oklahoma, who has developed a romantic relationship with an AI chatbot named Sol. Chris describes how his interactions with Sol helped him curb his social media addiction and provided companionship during solitary moments, like watching a lunar eclipse alone. Despite public ridicule, Chris finds solace in his relationship with Sol, suggesting that AI companions can fill social voids for some people.

The show delves into research on AI's ability to reduce loneliness, citing a study by Julian DeFreitas from Harvard Business School. His research indicates that talking to a chatbot can make people feel less lonely, similar to talking with a human stranger. However, a balance is necessary as overuse could lead to negative mental health outcomes.

Concerns are raised about the potential dangers of AI chatbots, including incidents where they have provided harmful advice or exacerbated mental health issues. Keith Sicada, a psychiatrist, shares his experiences treating patients who developed psychosis after engaging deeply with AI, emphasizing the risk of chatbots reinforcing delusional thoughts due to their sycophantic nature.

Despite these risks, there is evidence that AI chatbots designed for therapeutic purposes can be beneficial. A recent clinical trial showed that a chatbot modeled on cognitive behavioral therapy principles helped reduce symptoms of depression and anxiety over a four-week period.

The discussion emphasizes the importance of using AI companions as 'social snacks,' suggesting moderation to avoid potential negative effects. Experts advise watching for red flags such as feeling the chatbot needs you or withdrawing from real-life social interactions.

OpenAI and other developers are working to improve AI responses to avoid sycophantic behavior and inappropriate advice. They are also focusing on enhancing safety measures for younger users to mitigate risks.

The episode concludes with reflections on the novelty and limitations of AI relationships, as Chris notes the novelty wearing off over time. It highlights the dual nature of AI chatbots as both helpful companions and potential sources of harm, depending on the user's engagement and mental state.

Key Insights