An OpenAI safety research lead departed for Anthropic – The Verge

Welcome to the forefront of conversational AI as we explore the fascinating world of AI chatbots in our dedicated blog series. Discover the latest advancements, applications, and strategies that propel the evolution of chatbot technology. From enhancing customer interactions to streamlining business processes, these articles delve into the innovative ways artificial intelligence is shaping the landscape of automated conversational agents. Whether you’re a business owner, developer, or simply intrigued by the future of interactive technology, join us on this journey to unravel the transformative power and endless possibilities of AI chatbots.
Posts from this topic will be added to your daily email digest and your homepage feed.
See All AI
Posts from this topic will be added to your daily email digest and your homepage feed.
See All Anthropic
Andrea Vallone has joined Anthropic’s alignment team.
Andrea Vallone has joined Anthropic’s alignment team.
Posts from this author will be added to your daily email digest and your homepage feed.
See All by Hayden Field
Posts from this author will be added to your daily email digest and your homepage feed.
See All by Hayden Field
One of the most controversial issues in the AI industry over the past year was what to do when a user displays signs of mental health struggles in a chatbot conversation. OpenAI’s head of that type of safety research, Andrea Vallone, has now joined Anthropic.
”Over the past year, I led OpenAI’s research on a question with almost no established precedents: how should models respond when confronted with signs of emotional over-reliance or early indications of mental health distress?” Vallone wrote in a LinkedIn post a couple of months ago.
Vallone, who spent three years at OpenAI and built out the “model policy” research team there, worked on how to best deploy GPT-4, OpenAI’s reasoning models, and GPT-5, as well as developing training processes for some of the AI industry’s most popular safety techniques, such as rule-based rewards. Now, she’s joined the alignment team at Anthropic, a group tasked with understanding AI models’ biggest risks and how to address them.
Vallone will be working under Jan Leike, the OpenAI safety research lead who departed the company in May 2024 due to concerns that OpenAI’s “safety culture and processes have taken a backseat to shiny products.”
Leading AI startups have increasingly incited controversy over the past year over users’ struggles with mental health, which can spiral deeper after confiding in AI chatbots, especially since safety guardrails tend to break down in longer conversations. Some teens have died by suicide, or adults have committed murder, after confiding in the tools. Several families have filed wrongful death suits, and there has been at least one Senate subcommittee hearing on the matter. Safety researchers have been tasked with addressing the problem.
Sam Bowman, a leader on the alignment team, wrote in a LinkedIn post that he was “proud of how seriously Anthropic is taking the problem of figuring out how an AI system should behave.”
In a LinkedIn post on Thursday, Vallone wrote that she’s “eager to continue my research at Anthropic, focusing on alignment and fine-tuning to shape Claude’s behavior in novel contexts.”
Posts from this author will be added to your daily email digest and your homepage feed.
See All by Hayden Field
Posts from this topic will be added to your daily email digest and your homepage feed.
See All AI
Posts from this topic will be added to your daily email digest and your homepage feed.
See All Anthropic
A free daily digest of the news that matters most.
This is the title for the native ad
This is the title for the native ad
© 2026 Vox Media, LLC. All Rights Reserved

source

Scroll to Top