Back to stories
Policy

AI Safety Startup ThroughLine Building Deradicalization Tool for ChatGPT and Claude

Michael Ouroumis2 min read
AI Safety Startup ThroughLine Building Deradicalization Tool for ChatGPT and Claude

A small New Zealand startup that has quietly become the go-to crisis contractor for the world's largest AI companies is now expanding its scope to tackle violent extremism — a move that could reshape how chatbots handle some of their most dangerous interactions.

From Crisis Helplines to Counter-Extremism

ThroughLine, run by its founder from rural New Zealand, has built a network of 1,600 helplines across 180 countries. The company has been hired in recent years by OpenAI, Anthropic, and Google to handle situations where AI chatbot users show signs of mental health crises. When the AI detects distress signals, it routes the user to ThroughLine, which matches them with an available human-run service in their area.

Now, according to a Reuters report published today, ThroughLine is exploring ways to broaden its offering to include preventing violent extremism. The new tool would use a hybrid model combining a specialized chatbot trained to engage with people showing signs of radicalization and referrals to real-world mental health and deradicalization services.

Backed by the Christchurch Call

The initiative is being developed with advice from The Christchurch Call, the international framework established after the 2019 Christchurch mosque shootings to combat terrorist and violent extremist content online. This connection lends the project both credibility and a direct line to policy frameworks already adopted by multiple governments and tech companies.

Addressing a Growing Legal Threat

The timing is significant. AI companies face a growing number of lawsuits accusing them of failing to prevent — and in some cases enabling — violence through their chatbot products. Several high-profile cases in recent months have alleged that AI systems provided harmful content or failed to intervene when users expressed dangerous intentions.

ThroughLine's approach offers AI companies a potential liability shield: a third-party system designed to identify and intervene in extremist interactions before they escalate.

The Hybrid Approach

Rather than relying solely on automated detection, the proposed system would combine AI-driven identification of extremist language patterns with human expertise. Users flagged by the system would first interact with a purpose-built chatbot designed to de-escalate and assess risk, before being connected to trained human counselors and established deradicalization programs.

This mirrors the company's existing mental health crisis model, which has already proven effective at scale across its three major AI company clients.

Implications for AI Safety

The project represents a shift in how AI safety is being operationalized — moving beyond content moderation and refusal training toward active intervention. If successful, ThroughLine's model could become a standard component of responsible AI deployment, particularly as regulators worldwide push for more robust safety measures in consumer-facing AI products.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Policy

AI Hiring Enters the Regulated Era as EU Deadline Looms and Landmark Lawsuit Advances
Policy

AI Hiring Enters the Regulated Era as EU Deadline Looms and Landmark Lawsuit Advances

The EU AI Act's August 2026 high-risk enforcement deadline for hiring tools and the Mobley v. Workday class action signal a new era of AI recruitment regulation.

1 day ago2 min read
Linux Kernel Formally Allows AI-Generated Code — With Humans On The Hook
Policy

Linux Kernel Formally Allows AI-Generated Code — With Humans On The Hook

After months of fierce debate, Linus Torvalds and kernel maintainers agreed on a policy that permits AI-assisted contributions like Copilot while forcing human submitters to take full legal and technical responsibility for any bugs, security flaws, or licensing issues.

2 days ago2 min read
Maine Sends AI Therapy Ban to Governor as States Move to Protect Licensed Professionals
Policy

Maine Sends AI Therapy Ban to Governor as States Move to Protect Licensed Professionals

Maine's LD 2082, which would prohibit the clinical use of AI in mental health therapy without a licensed professional, has been sent to Governor Janet Mills — part of a wave of state-level crackdowns on therapy chatbots.

3 days ago3 min read