About
I come to artificial intelligence from constitutional law and the study of free speech. For years, my work focused on how societies constrain power — how they build institutions that can say no, how they preserve spaces for dissent, how they protect the right to be wrong.
AI governance pulled me in because the questions felt familiar. Here was a new kind of power — unprecedented in scale, opaque in operation, resistant to traditional forms of accountability. And here was a field moving so fast that reflection had become a luxury, doubt a liability, and slowness almost unthinkable.
I distrust speed because I have seen what happens when important decisions are made without time for deliberation. I distrust hype because I have watched too many revolutions promise transformation and deliver consolidation. I distrust false certainty because I know how rarely the confident voice is also the wise one.
This is a long project, not a startup sprint. There is no roadmap, no launch date, no metrics dashboard. There is only the ongoing work of thinking carefully about what we are building and what it might cost.
The Slow AI Project exists because some things are worth doing slowly. If that sounds old-fashioned, perhaps it is. But the future has a way of vindicating patience.