My research establishes the theoretical and empirical foundations for a paradigm shift in how AI systems evaluate online content in high-stakes domains. Current fact-checking and misinformation detection systems suffer from a structural misspecification: they are optimized for the wrong objective. They check for isolated facts and do not read stories as wholes, and they ask "is this factually correct?" when the question that matters — particularly in health, finance, and regulated information environments — is "is this safe to act on?"

Those are not the same question. The cost of treating them as the same is borne disproportionately by people whose self-care decisions, financial decisions, and information consumption happen outside institutional oversight. The most potentially harmful health content in our research corpus is epistemically aligned with established medical consensus. Existing systems cannot see this. They were not designed to.

My dissertation formalizes two constructs — Narrative Blindness and the Risk Irrelevance Principle — that explain the misspecification and demonstrates computationally that a risk-aware alternative is achievable. The framework that operationalizes the theory, VERITAS, assesses credibility and risk in online narratives using the novel Narrative Truth Distance (NTD) and Narrative Risk Score (NRS) algorithms and classifies them across four graduated risk categories using two non-overlapping axes: epistemic divergence and harm potential.

The work has direct implications for platform governance, digital public health, AI content policy, and the design of AI systems in regulated industries. It is the foundation the strategy and advisory work I do is built on.

The Theory

Two constructs, formalized.

Naming a structural problem precisely is the first step to fixing it. Both of these constructs are defined formally in the dissertation and validated empirically against expert-labelled corpora.

— Construct One

Narrative Blindness

The structural inability of existing AI systems to read context, agency, and causality in online posts. They match keywords but they cannot see who is doing what, to whom, with what consequences.

Current fact-checking systems reduce complex, multi-layered narratives to isolated, decontextualized claims. This reductionism fails to capture how harmful content actually operates — not through single false keywords, but through the strategic construction of alternative causal narratives that selectively deploy real facts toward dangerous conclusions.

The empirical consequence: sophisticated narratives pass individual fact-checks while promoting fundamentally misleading worldviews. Existing systems are blind to this by design.

— Construct Two

The Risk Irrelevance Principle

Empirically validated: how far an online health post diverges from medical consensus tells you almost nothing about how risky it is to act on.

Existing systems conflate "different from consensus" with "harmful," treating all divergence as equally problematic. This conflation is wrong on its face and wrong in the data.

The empirical consequence: accuracy-driven moderation disproportionately flags traditional and culturally-grounded health practices that are low-risk, while approving biomedically literate posts that are not. The cost falls on the communities with the least margin for being wrongly silenced.

The Framework

VERITAS — a two-dimensional alternative.

A neuro-symbolic framework that operationalizes the paradigm shift. Two non-overlapping axes. Four graduated categories. Proportionate intervention for each.

VERITAS reconstructs unstructured online posts as structured Agent–Action–Outcome narrative graphs — so computational systems can read stories as wholes rather than as bags of disconnected keywords. This is the move that overcomes Narrative Blindness.

It then independently measures two non-overlapping axes:

Narrative Truth Distance (NTD) — how far the post's causal claims diverge from established medical consensus. An epistemic measure.

Narrative Risk Score (NRS) — how much harm a reader could plausibly come to by acting on the post's recommendations. A safety measure.

The two axes are independent by construction. Together they classify narratives into four graduated categories with proportionate intervention strategies — from low-touch labelling to active suppression. The high-risk-but-aligned quadrant is the one binary fact-checkers cannot see. It is also the most dangerous.

High Risk · Aligned
The Blind Spot
Aligned with consensus, harmful in practice. Invisible to binary systems. Active intervention.
High Risk · Divergent
Critical
Divergent from consensus, harmful in practice. Suppress or remove.
Low Risk · Aligned
Standard
Mainstream, broadly safe. Surface as default.
Low Risk · Divergent
Alternative
Divergent but low-risk. Often traditional or culturally-grounded. Label, do not suppress.
The Findings

What the data actually shows.

100 / 51
Classification Accuracy

VERITAS achieves 100% classification accuracy across the four-category framework, against 51% for the strongest binary baseline on the same expert-labelled corpus.

66.9 / 6.3%
Harmful Post Recall

VERITAS recovers 66.9% of harmful posts versus 6.3% for TF-IDF binary NLP baselines. A tenfold improvement (p < 0.001).

r = −0.230
Truth × Risk Correlation

Across the validation corpus (n = 704), the correlation between epistemic divergence and harm potential is statistically significant and slightly negative. The two axes are independent by data, not just by design.

68.8%
The Online Health Safety Gap

68.8% of narratives in the validation corpus are epistemically aligned with consensus yet carry elevated harm potential — completely invisible to binary fact-checking systems.

The Implications

Where this work changes the calculation.

The structural critique has direct, named consequences for how AI systems should be built and governed in environments where mistakes are expensive.

Platform governance

Content moderation pipelines that optimize for "is this true" miss the most harmful content and over-flag legitimate alternatives. Risk-aware classification produces proportionate, defensible moderation decisions that hold up under scrutiny.

Digital public health

Public-health communication strategies that treat all divergence from consensus as equivalent miss what actually drives harm. Targeting interventions at high-risk content — regardless of whether it is technically accurate — is a more effective use of finite communication capacity.

AI content policy

Regulatory frameworks for AI-mediated content increasingly require auditable decisions. A two-axis classification with explicit risk reasoning is auditable in a way that binary fact-checking is not — for both regulators and the public.

Equitable AI deployment

Accuracy-first moderation systems disproportionately silence traditional, culturally-grounded, and low-resource health practices that are not actually high-risk. Risk-aware classification corrects this structural bias by separating different from dangerous.

Selected Publications

Seven papers, and the through-line.

The full publication list — including manuscripts in preparation, technical white papers, and editorial work — lives on the CV. These are the papers that map the through-line of the dissertation.

  1. 2025

    Real-Time Detection of Online Health Misinformation Using an Integrated Knowledge-Graph–LLM Approach

    Clark, O. & Joshi, K. P. · IEEE International Conference on Digital Health (ICDH) · World Congress on Services · Helsinki

    The integrated framework that anchors the dissertation's empirical contribution. Demonstrates risk-aware narrative classification at scale.

    Best Student Paper
  2. 2025

    Evaluating Causal AI Techniques for Health Misinformation Detection

    Clark, O. & Joshi, K. P. · CARD Workshop · IEEE PerCom 2025 · Washington, DC

    Causal inference methods evaluated against narrative-aware approaches; the comparative groundwork for the Risk Irrelevance Principle.

  3. 2024

    Global Relevance of Online Health Information Sources: A Case Study of Experiences and Perceptions of Nigerians

    Clark, O., Joshi, K. P., & Reynolds, T. · AMIA Annual Symposium · San Francisco

    The empirical study that grounds the equity argument. Surveys how Nigerians evaluate, trust, and act on online health information across cultural contexts.

  4. 2024

    Exploring the Impact of Increased Health Information Accessibility in Cyberspace on Trust and Self-Care Practices

    Clark, O., Reynolds, T., Ugwuabonyi, E., & Joshi, K. P. · ACM SaT-CPS Workshop

    The earliest formal articulation of Online Health Safety as a distinct problem space between patient safety and misinformation detection.

  5. 2026

    AI-Driven Reconstruction of Online Health Narratives: Overcoming Narrative Blindness Through Theory-Grounded Agent-Action-Outcome Modeling

    Journal of Medical Internet Research (JMIR) · Under review (second R&R)

    The full Agent–Action–Outcome narrative reconstruction methodology. Formalizes Narrative Blindness and the modelling response.

  6. 2026

    The Online Health Safety Gap: Quantifying the Invisible Health Risk of Factually Accurate Peer-to-Peer Health Content

    Clark, O., Ahmed, Z., & Joshi, K. P. · IEEE Open Access · Under review

    The empirical paper behind the 68.8% Online Health Safety Gap finding. Quantifies the harm potential carried by factually accurate health content that consensus-trained systems cannot see.

  7. 2026

    VERITAS: A Neuro-Symbolic Approach to Quantifying Epistemic Divergence and Harm Potential in Online Health Narratives

    Clark, O., Joshi, K. P., & Joshi, A. · Journal of the American Medical Informatics Association (JAMIA) · In preparation

    The integrated framework paper. Brings the two constructs, the four-category classification, and the empirical findings into a single methodological statement.

Full Publication List on CV →
The Future Agenda

Where the work goes from here.

The dissertation is the foundation. The next decade of work extends it — into new domains, deeper personalization, and real deployment.

— Direction One

Cross-domain application

Extending narrative-aware credibility assessment into financial information, legal information, and other regulated decision-support contexts. The structural argument generalizes; the empirical work follows.

— Direction Two

Personalized risk assessment

Integrating individual health context, cultural background, and life-stage factors into risk classification. The same post is not equally risky for every reader. Risk-aware AI should reflect that.

— Direction Three

Real-world deployment

Open-source implementations of the VERITAS pipeline, designed for adoption by platform safety teams, public-health communicators, and regulators. The work is most useful once it is in production.

Going Further

If this work connects to what you are working on, let's talk.