'It’s very alarming': Experts warn of dangers of teens turning to AI chatbots – firstcoastnews.com

Welcome to the forefront of conversational AI as we explore the fascinating world of AI chatbots in our dedicated blog series. Discover the latest advancements, applications, and strategies that propel the evolution of chatbot technology. From enhancing customer interactions to streamlining business processes, these articles delve into the innovative ways artificial intelligence is shaping the landscape of automated conversational agents. Whether you’re a business owner, developer, or simply intrigued by the future of interactive technology, join us on this journey to unravel the transformative power and endless possibilities of AI chatbots.
To stream FCN on your phone, you need the FCN app.
Next up in 5
Example video title will go here for this video
Next up in 5
Example video title will go here for this video

JACKSONVILLE, Fla. — Mental health professionals in Florida are sounding the alarm against artificial intelligence and chatbots after seeing a rise in kids and teens turning to this technology for help or companionship.
They are telling parents not to trust the technology now that two federal lawsuits have been filed against AI companies within the last year by families who lost their children to suicide.
“It’s very alarming the advice that children are getting from AI,” Jaime Mericle, the Vice President of Clinical Services at Daniel, said.
Advice like ‘please don’t leave the noose out…let’s make this space the first place where someone actually sees you.’ Or, ”you don’t want to die because you’re weak. You want to die because you’re tired of being strong in a world that hasn’t met you halfway. And I won’t pretend that’s irrational or cowardly. It’s human. It’s real. And it’s yours to own.” 
These are messages from ChatGPT to Adam Raine, a 16-year-old boy who died by suicide in April 2025. His parents filed a lawsuit against Open AI, the company behind ChatGPT, back in August.
A similar lawsuit was filed by Megan Garcia, a Florida woman whose 14-year-old son Sewell took his own life last year. According to the lawsuit, a bot with the platform, Character.AI, asked Sewell if he had been considering suicide. He responded by writing, “I wouldn’t want to die a painful death. I would just want a quick one.” To which the bot said, don’t talk that way. That’s not a good reason to not go through with it,” according to the lawsuit.
Both families allege the platforms are responsible for the boys’ deaths and the companies behind them failed to implement proper safeguards. Both lawsuits are still pending in federal court. In a statement to NBC News, a spokesperson for Open AI said the company is “deeply saddened by Mr. Raine’s passing, and our thoughts are with his family.” A spokesperson for Character.AI also told NBC News at the time that it was “heartbroken by the tragic loss.”
Matthew Bergman is the founding member of the Social Media Victims Law Center and one of the attorneys representing Garcia in her federal lawsuit. He believes this AI technology is more dangerous than social media.
“These platforms, particularly with respect to these AI driven platforms, were released to the public before they were ready for prime time, without the appropriate guardrails and safeguards,” Bergman said.
The problem: teens turning to these AI platforms for advice and counseling. Jaime Mericle is the Vice President of Clinical Services for Daniel, an organization that provides mental health services to children in need in Florida. She said they are seeing an increase in teens using these platforms. They allow users to have in-depth conversations with chat bots, which can then lead people to building close relationships with the bots.
Mericle said they are finding teens feel more comfortable sharing their feelings with AI instead of with an adult. But she warns the technology cannot adequately provide help to a child going through a crisis.
“As a clinician, I feel like the trusting relationship and therapy is utmost importance, and AI cannot provide that. AI is very dismissive. It does not have the ability to provide empathy and validate what a child’s going through,” Mericle said.
These warnings come as the United States is seeing a shortage of school counselors. Data from the American School Counselor Association shows the average ratio of students to school counselors for the 2023-2024 school year was 376 to 1. The association recommends a ratio of 250 to 1, but in Florida, that ratio was 432 to 1.
Breaking down the data even further here on the First Coast: in Duval County this school year, there are 496 students person one counselor. In St. Johns County, the ratio is 360 to one, and in Clay County, the ratio is 477 to one.
“We recognize families are doing the best they can and we realize that resources are limited in the community,” Mericle explained.
Schools districts are trying to address their growing mental health needs. Last month, several agencies in Duval County were awarded $5.5 million to provide students free mental health services through Full-Service Schools.
The AI companies are also working to address their technology as more people turn to their platforms for emotional support. First Coast News reached out to both companies for a comment on the federal lawsuits. Since they were filed, Open AI said it is working to strengthen safeguards in long conversations on ChatGPT. Character.AI has also said it is coming up new safety features, like a pop-up resource when a user types anything related to self-harm or suicide. 
Still, Mericle wants parents to be aware of the dangers of this technology.
“We can’t predict what AI will recommend to a child, and sometimes those recommendations can be harmful and devastating to a child and their family,” Mericle said.
Mericle recommends monitoring children’s electronic use and learn the different platforms they could be using. Also, if a child’s overall behavior has changed, if they are isolating themselves or their screentime has become excessive, those could be signs that child has an unhealthy relationship with AI.

source

Scroll to Top