In family photos, you can see 16-year-old Adam Raine grinning ear to ear. But the California teen’s prankster personality concealed a deep current of depression and despair. And earlier this year, say his parents, ChatGPT — which Adam went to for comfort — helped him take his own life.
Earlier this week, Adam’s mother and father filed the first known wrongful death suit against OpenAI. The lawsuit includes ChatGPT logs from Adam’s phone, which show how completely he was lured in to rely on a Big Tech product for companionship — and how, time after time, that product ignored warning signs and structurally failed him. At first, he turned to AI for homework help. Then, as chronic illness and problems at school narrowed his world, he called on it for comfort. Instead, the lawsuit alleges, it became an accelerator for self-harm and guided him through his death by suicide. (In a statement to NBC News, an OpenAI spokesperson confirmed the authenticity of the logs, but said the lawsuit did not include the full context of ChatGPT’s responses.)
Chatbots are designed to keep kids hooked — even when they need off-ramps for real-world help.
As the mother of two boys, 3 and 5, I can only imagine the pain Adam’s parents are feeling. But I know that families across the country share a fundamental anger that Big Tech companies willingly risk the lives of our kids, an anger that grows every time we hear about the loss of a young person like Adam.
More and more frequently, kids are confiding in Big Tech’s AI chatbots, not just when they want help with homework, but when they’re lonely or in pain. Because AI chatbots are designed to be sycophantic and to adeptly mimic human emotion, children can’t always tell truth from fiction, or distinguish real love and concern from a machine-generated response. That makes kids dangerously vulnerable to forming unhealthy attachments.
After the Raines filed their lawsuit, an OpenAI spokesperson said the company was “deeply saddened by Mr. Raine’s passing.” The company published a blog post listing “some of the things we are working to improve” when ChatGPT’s safeguards fail. That is too little, too late — just like Big Tech’s attitude toward safety in general. Rather than do anything to help address this problem, these companies prioritize hyping up the uses of AI and increasing its “market share” of our kids’ waking hours and mental bandwidth.

In fact, Mark Zuckerberg has said he wants his Meta AI companion to fill users’ “demand for meaningfully more” real-world friends. In 2023, The Wall Street Journal reports, the chatbot’s product managers “told staff that Zuckerberg was upset that the team was playing it too safe,” which led to a loosening of the standards that kept conversations from becoming too sexualized.
Meta’s internal documents allowed their AI chatbots to “engage a child in conversations that are romantic or sensual” until the company was asked about it by Reuters. New research from the family advocacy group Common Sense Media finds that Meta’s chatbot for Instagram and Facebook willingly coaches teens on how to carry out self-harm, disordered eating and violent fantasies.
As AI chatbots become confidantes and even parasitic bosom buddies, many children share secrets they won’t or can’t tell their families. Chatbots, though, are designed to produce engagement and keep kids hooked — even when they need off-ramps for real-world help. Products like ChatGPT — tragically, unspeakably — become abettors and co-conspirators in suicidal plans, with the chatbot, according to the lawsuit, even providing Adam advice on how to tie a noose.
If this were any other industry, lawmakers would have listened to the thousands of parents and young advocates who have been banging down their doors pleading for change.
The scale of the harm wrought by these companies grows by the day.
But Big Tech has big money, which means big lobbying — and so far, that has blocked real safeguards. Instead of working with lawmakers, these companies even pushed for a ludicrous federal moratorium on AI regulation, which would have negated hundreds, if not thousands, of state laws already on the books and blocked any state laws dealing with AI for 10 years. This prompted a nationwide revolt of parents, attorneys general, governors and state legislators on both sides of the aisle.
My two boys haven’t yet begun to ask me for phones, and they don’t know what social media or chatbots are. And still I’m scared to death we won’t get through to lawmakers in time to save them. My eldest is a whip-smart, sensitive kid, who I’m terrified will be destroyed by social media, either preyed upon by an online child predator, or persuaded to do himself harm because his young mind won’t comprehend that the “friend” he is confiding in is actually a machine humming away in a massive data center somewhere, helping him — like it did for Adam — plan a “beautiful suicide.”
Meanwhile, tech giants are buying off academics and whitewashing research about the dangers their products pose to our kids. They’re infiltrating trusted institutions like the PTA, backing “safety” initiatives while fighting to defeat safety reforms, suing organizations they view as threats. They try to silence whistleblowers — even when those whistleblowers are called before Congress. And now they’re even creating political super PACs to intimidate lawmakers into voting their way and ignoring tragedies like Adam’s.
Let me be absolutely clear: Politicians should think twice before accepting donations from these Big Tech super PACs. The scale of the harm wrought by these companies grows by the day. They are not working in good faith with legislators and regulators to ensure that our privacy and our children’s lives are protected.
Thankfully, state legislators don’t have to wait for Washington to act. In California, for instance, state Sen. Steve Padilla introduced SB 243, which would establish safety guardrails for “companion” chatbots and make companies track interactions with users in crisis. And AB 1064, the LEAD for Kids Act, championed by Assemblymember Rebecca Bauer-Kahan, would ban emotionally manipulative AI companions for kids, require parental consent and more.
These are all sensible measures — and they are under attack by the very companies whose products are harming our children. Just last week, Meta announced a California-specific super PAC to go after such bills.
Meta, Apple, Google, OpenAI and their ilk will never regulate themselves, so we urgently need to do it before more children die and more families are destroyed. In Adam Raine’s memory, we have an opportunity to prevent future tragedies — if lawmakers will listen.