back to home page
AI systems & model behavior researcher focused on evaluation,  
failure modes, and decision-making in high-stakes environments 
(elections, conflict, online harm).


Led global systems for synthetic media verification
Identify failure modes in high-stakes contexts
Developed evaluation frameworks for real-world AI performance
Design systems for decision-making under uncertainty
Translate model behavior into real-world decisions
Advise industry, policy, and civil society on safe AI deployment




Deepfake Rapid Response Force

Real-time decision-support system for verifying high-risk synthetic media across global information ecosystems (launched 2023).

Detection + expert workflows
Real-time, high-stakes decisions
Revealed failure under real-world pressure

Problem:  AI detection tools are unreliable in real-world, time-sensitive contexts. They are not integrated into decision workflows and struggle in hybrid environments where authentic and synthetic content coexist.

What I did: Designed and deployed workflows integrating AI detection tools, expert analysis, and rapid response coordination across journalists and forensic experts in multiple global regions. 

Outcome:
Enabled real-time verification and response decisions across diverse geopolitical contexts, improving how high-risk media is assessed and acted upon under uncertainty.
 

Key insights:
  • A significant share of high-risk cases (~30%) involve authentic content, with AI used as an alibi—making verification more complex than simple detection
  • Real-world environments are hybrid: synthetic and authentic content coexist, requiring systems designed for uncertainty rather than binary classification
  • Audio presents a particularly high risk, with detection lagging behind generation; risks from video are rapidly increasing




MNW Benchmark (Microsoft-Northwestern-WITNESS)