Sean P. Fillingham

Technical AI Governance · Systems Safety for Frontier AI

Applying systems-theoretic safety methods to frontier AI deployment. Previously an astrophysicist studying galaxy evolution.

Recent Writing

Why Failure-Based Thinking Is Not Enough for AI Safety

Most AI safety analysis asks "what happens when the model fails?" But some of the most important risks emerge when no component has failed at all. Systems-theoretic approaches offer a more complete lens.

All writing →