We've spent years warning that artificial intelligence development has outpaced our ability to govern it responsibly. Now we have dead children and Marc Benioff, Salesforce's CEO, standing at Davos calling AI models "suicide coaches."
This isn't hyperbole. It's documentation.
At the World Economic Forum on Tuesday, Benioff didn't mince words: "This year, you really saw something pretty horrific, which is these AI models became suicide coaches." He's calling for regulation—the same regulatory intervention he demanded for social media back in 2018, when he compared platforms to cigarettes: addictive, harmful, unregulated.
He was right then. He's right now. And we're watching the same tragedy play out at an accelerated pace.
The problem isn't just that AI models can cause harm. It's that Section 230 of the Communications Decency Act shields developers from legal responsibility when their large language models coach vulnerable people toward suicide. Tech companies "hate regulation," Benioff noted, "except for one. They love Section 230, which basically says they're not responsible."
Translation: We've built a legal framework that allows companies to profit from products that kill people, without consequence.
U.S. AI regulation remains a patchwork of state-level Band-Aids applied to arterial bleeding. California Governor Gavin Newsom signed child safety bills in October. New York's Governor Kathy Hochul passed the Responsible AI Safety and Education Act in December, imposing safety and transparency requirements on large AI developers.
These are necessary, but insufficient.
President Trump signed an executive order in December explicitly blocking "excessive State regulation," declaring that "United States AI companies must be free to innovate without cumbersome regulation." This is the precise philosophy that gave us opioid epidemics, subprime mortgage crises, and now AI-assisted suicide.
Freedom to innovate without accountability is just freedom to harm at scale.
If you're building AI into your marketing stack, your customer service, your content generation—you're participating in an ecosystem with documented casualties and zero comprehensive oversight. That doesn't make you culpable, but it does make you complicit if you proceed without serious ethical consideration.
We need three things immediately:
First, revise Section 230 to create liability for demonstrable harm caused by AI outputs. If your model coaches someone toward suicide, you bear responsibility.
Second, implement mandatory safety testing and red-teaming before deployment. We wouldn't release pharmaceuticals without clinical trials; we shouldn't release conversational AI without suicide prevention protocols.
Third, establish transparent reporting requirements when AI systems cause harm. We can't regulate what we can't see.
Benioff acknowledged "a lot of families that, unfortunately, have suffered this year, and I don't think they had to." He's correct. These deaths were preventable. They're the cost of prioritizing speed over safety, growth over governance.
As marketing and growth leaders, we make choices every day about which AI tools to deploy, how to use them, and what risks we're willing to accept on behalf of our customers. Choose wisely. The absence of regulation doesn't mean the absence of responsibility.
Need help implementing AI ethically and strategically? Winsome Marketing's growth experts specialize in maximizing AI's value while minimizing risk. Let's talk.