Developing a Precise Approach to Identifying Inciting Speech Online

The discourse around social media moderation often centers on the idea of “censorship” and protecting free expression vs. protecting conversation health via account bans (what my generation dubbed “Facebook jail”). This framing can make the choices in moderation seem binary.

However, for those of us either navigating or studying polarized opinions in online spaces, it is clear that the reality is far more nuanced. A significant and challenging spectrum exists online from speech that is merely disagreeable or misinformed, to degrees of hateful or toxic speech and deliberate disinformation, and, ultimately, to speech that crosses the threshold into actual incitement of violence. This spectrum demands a more precise and exacting approach than current moderation tools often provide.

I am pleased to announce a forthcoming paper on this very topic, led by PhD student Saajan Patel: Adapting the Dangerous Speech Paradigm to Identify Incitement from Polarized and Hate Speech.” In this work, Patel and our team describe methods and a framework for distinguishing speech with the hallmarks of violence incitement from misinformation, rants, slurs, and other types of polarized, but lawful, online discourse.

Distinguishing What is ‘Dangerous’

Most existing moderation systems struggle to distinguish between speech that is merely offensive and speech that acts as a spark for real-world harm. To solve this, our team looked at the Dangerous Speech Paradigm. Based in analysis of real-life discourse around mass murders and genocides in different cultures, this paradigm describes the rhetorical contexts and hallmarks of text, images or video (such as Dehumanization and Accusation in a Mirror) that have led to people either condoning or participating in violence against members of another group. We realized that this paradigm needed an update for the era of Large Language Models (LLMs).

We proposed a three-tiered framework to classify polarized content:

  1. Allowable Speech: Polarized or disagreeable, but lawful.
  2. Harmful Speech: Hate speech or content that denigrates, but doesn’t necessarily call for action.
  3. Inciting Speech: Content that directly encourages or facilitates dangerous acts.

By bisecting the old “dangerous” category into “harmful” and “inciting,” we allow platforms to respond more appropriately – perhaps shadow-banning or labeling harmful content, while reserved immediate removal and reporting for truly inciting content.

Scaling Post Labeling and Explanations

To test this framework, we didn’t just look at a few hundred posts. We analyzed over 3.5 million posts from the platform Gab, focusing on the period surrounding the 2017 Charlottesville “Unite the Right” rally – a pivotal moment in digital polarization.

Using a customized GPT-4o-mini framework paired with Chain-of-Thought (CoT) prompting, we were able to teach the AI to “think through” the nuances of incitement. The AI doesn’t just give a label; it explains why a post reaches the threshold of inciting violence based on the context of the event.

Modeling the Spread of Incitement

We didn’t stop at classification. We also wanted to see how this speech moves. Using a DeGroot diffusion model, we built a directed repost graph to track the spread of harmful vs. inciting speech.

What we found is encouraging for the future of moderation: LLM classifiers can effectively assist human moderators by flagging the “inciters” in real-time, helping to stop the spread of dangerous content before it reaches a tipping point. (Plus, we spare human moderators the trauma of repeatedly being exposed to, and thinking about, horrifically toxic speech.)

Why This Matters

Our goal isn’t to silence political disagreement. In fact, by accurately identifying and isolating inciting speech, we actually protect the space for allowable (if polarized) speech to exist. This research offers a path forward for platforms to manage content in a way that is transparent, consistent, and grounded in the protection of human life.

Huge thanks to my brilliant co-authors Saajan Patel (this is his first first-author paper!), Ramisha Mahiyat, Aditya Narasimhan Sampath, and Siddharth Krishnan.

Just accepted for ACM Transactions on Social Computing! Check out the link below for more details on the machine learning pipeline and how the GPT model was trained and deployed. 

  • Saajan Patel, Cori Faklaris, Ramisha Mahiyat, Aditya Sampath, and Siddharth Krishnan. 2026. Adapting the Dangerous Speech Paradigm to Identify Incitement from Polarized and Hate Speech. Trans. Soc. Comput. Just Accepted (February 2026). https://doi.org/10.1145/3797963 

Walking the Talk: Why I’m Disclosing My AI Use to My Students

We spend a lot of time talking about how students should (and shouldn’t) use AI. We debate academic integrity, we draft policies, and we ask for disclosures. But there is a quieter, more controversial conversation happening in the corridors and faculty meetings: How are we using it?

The reality of the modern faculty workload is intense. Among research, service, and teaching, the prep work (drafting quiz items, polishing slide decks, and organizing materials) and other day-to-day management tasks (monitoring CMS engagement, recording attendance, answering student emails, proctoring make-up work, updating documents, coordinating submissions, and consulting with the TAs on grades) can eat up the very hours we should be spending on deep mentorship and high-level instruction.

That’s why this semester, I’ve decided to not just amp up use of AI to help me work smarter; I’m telling my students exactly how I’m doing it.

I recently added this disclosure to my course management policy for ITIS 4360 / 5360: Human-Centered Artificial Intelligence, based on suggested language from UNC Charlotte Student Affairs:

Dr. Faklaris often uses AI tools to assist with tasks such as generating ideas, checking grammar, writing alternative quiz items, drafting slide content and in-class activities, identifying research papers, and organizing materials. The purpose is to support efficiency, not to replace her judgment or expertise. All content has been reviewed and adapted to ensure it aligns with the objectives of this course. We disclose this so you understand that AI can be a helpful resource when used responsibly and critically.

What this adds to my existing AI policy language for the syllabus:

1. Modeling Responsible Use: If we want students to be “Human-in-the-loop” practitioners, we have to show them what that looks like. By disclosing that I use AI for a first draft of quiz items or to brainstorm an in-class activity, I’m showing them that AI is a tool for augmentation, not a replacement for expertise.

2. Bridging the Trust Gap: Students are often nervous that faculty are “policing” AI while secretly using it themselves. By being upfront, I’m creating a culture of integrity that works both ways. If I expect them to adhere to best practices, I should be willing to do the same.

3. Focusing on What Matters: Using AI to help organize a bibliography or check the grammar on a slide doesn’t make me a less capable professor. It makes me a more available one. It gives me the “imaginative capacity” (to borrow a theme from our upcoming AI Summit!) to focus on the human elements of teaching that no LLM can replicate.

The Bottom Line: AI Policy in the classroom isn’t just about catching cheaters. It’s about rethinking how we work. I strongly believe that AI should serve human ends. For me, that means using technology to be a more present, prepared, and transparent educator.

From Anxiety to Agency: New Research on Human-Centered Security at CHI 2026 and USEC 2026

I’m delighted to share this news — two of my latest papers with key collaborators have been conditionally accepted on their first try, and at very competitive venues! While these papers cover different topics (one focusing on the psychology of anxiety and the other on the unique safety needs of international students), they share a common goal: making digital security more inclusive, less stressful, and deeply grounded in the human experience.

Measuring Invisible Stress: The Cybersecurity Anxiety Scale (CybAS)

Conditionally accepted for CHI 2026 (Barcelona, Spain)

For years, we’ve known that users feel fatigued and concerned by the drumbeat of cybersecurity and privacy threats. However, we have lacked a validated way to measure the specific, persistent worry that comes with navigating these threats. We call this emotional state Cybersecurity Anxiety.

Led by Peter Mayer and first-authored by Nikolaj Dall and Hanno Gustav Hagge, our paper, “From Fear to Control: Developing a Three-Factor Scale for Cybersecurity Anxiety (CybAS),” introduces a new 15-item psychometric tool. Through several rounds of survey studies, we identified three core factors that define this anxiety:

  • Present: Immediate concerns and stress during security tasks.
  • Future: The “what-if” worry about anticipated threats.
  • Control: The feeling (or lack thereof) that one has the agency to stay safe.

Why it matters: By using CybAS, researchers and designers can better diagnose why a security tool might be failing. If a system makes a user feel helpless, they are more likely to disengage. CybAS allows us to build “anxiety-aware” security systems that empower users rather than scaring them.

Hypothesized diagnostic categories based on CybAS subscale score combinations. More information is available in the finalized paper, including CybAS item wordings and directions for using it.

Designing Safety Tools for, and with, International Students

To be presented at USEC 2026 (San Diego, CA)

When students from the Global South move to the U.S. to study, they don’t just face a new culture; they face a new and often treacherous digital ecosystem. These educational migrants are frequently targeted by cross-channel scams (SMS, phone calls, and emails) that exploit their unfamiliarity with local institutions.

As described in From Scam to Safety: Participatory Design of Digital Privacy and Security Tools with International Students from Global South,” lead author Sarah Tabassum conducted participatory design sessions with 22 students to imagine better safety solutions, using AI capabilities as a design material.

From this data, we identified several must-have features that current tools lack:

  • University Integration: Students trust their schools. By embedding safety support into university platforms, we can provide a trusted safety net.
  • Cross-channel filtering: Moving beyond just email spam to filter SMS and voice scams.
  • Contextual explanations: Instead of just saying “this is a scam,” tools should explain why based on the cultural cues the student might be missing.

Why it matters: This work reminds us that security is not a one-size-fits-all solution. For these users, it must account for the situational vulnerabilities of those moving across borders.


Orientation challenges experienced by educational migrants (Point 1) and the four migrant-centered security features they identified as necessary for safer digital navigation (Points 2–5). See the paper for more details and participants’ sketches and choices of AI capabilities for their tools.

Looking Ahead to 2026

These two papers exemplify the types of work I wanted to conduct when I founded my SPEX research group at UNC Charlotte. Together, and with our external collaborators, we are creating new knowledge of how to make people feel safer and more capable in a complex digital world.

I am incredibly proud of student authors Nikolaj, Hanno, Sarah and her co-author Narges Zare, and my other collaborators and lab mates for their hard work. If you are attending USEC in February or CHI in April, please come say hello! We look forward to sharing our full findings and connecting with fellow researchers who are passionate about human-centered security and privacy.