A Hybrid Approach to Online Radicalization
A New Zealand-based initiative is developing a tool aimed at countering violent extremism expressed through AI chatbots, particularly in platforms like ChatGPT. This system proposes a combined approach, utilizing both automated responses from a specially trained chatbot and connecting individuals to human mental health services. The developers indicate that the tool is intended to help users showing signs of extremist tendencies.
The initiative, spearheaded by a contractor that also serves major AI companies such as OpenAI and Anthropic, aims to address the increasing visibility of radicalization online. The creators posit that the sheer volume of personal disclosures made to chatbots, encompassing a wide spectrum of issues from mental health struggles to extremist thoughts, necessitates such interventions.
Expanding the Scope of Intervention
"The breadth of mental health struggles that people disclose online has exploded with the popularity of AI chatbots, and now includes dalliances with extremism."
The tool is envisioned to extend beyond direct chatbot interactions. Galen Lamphere-Englund, a counterterrorism adviser involved with 'The Christchurch Call', expressed hope for its application by moderators on gaming forums and by parents or caregivers seeking to identify and mitigate online extremism. This suggests a broader goal of equipping various stakeholders with the means to identify and intervene in potentially radicalizing online activities.
Read More: AI Changes Jobs: New Roles and Skills Needed by 2025
The developers acknowledge that their current capabilities have been limited to specific categories of issues, but the surge in user disclosures signals a need for wider applications. The complexity of human distress, often shared more readily with non-human interfaces, presents both a challenge and an opportunity for intervention.

"He said people in distress tended to share things online that they were too embarrassed to say to a person, and governments risked compounding danger if they pressured platforms to cut off users who engaged in sensitive conversations."
Contextualizing the Development
This effort surfaces amidst broader discussions about the role of AI companies and their engagement with governmental bodies, particularly military interests. Recent events have highlighted tensions between AI developers, their ethical stances, and the demands of defense sectors.
Specifically, a dispute involving Anthropic and the Pentagon earlier in 2026 brought to light the deep integration of AI systems into military operations and the complex risk management involved. While OpenAI reportedly reached an agreement with the Pentagon to use its technologies in classified settings, Anthropic faced pressure to refuse similar demands regarding autonomous weapons or mass surveillance. This divergence underscores the ongoing debate over the ethical boundaries and applications of advanced AI, especially concerning national security and defense.
Read More: Man gets open-ended jail sentence for sexual assaults in Auckland
The development of this anti-extremism tool can be seen as a separate, though perhaps indirectly related, response to the increasing influence and public engagement with AI technologies. It points to an evolving understanding of the societal impacts of these tools and a growing effort to manage the risks associated with their widespread adoption.