I work on AI safety — trying to make sure powerful AI systems don't go badly wrong. I'm interested in alignment, interpretability, and what it actually takes to build systems we can trust.

This is where I keep things: work I've done, ideas I'm chewing on, writing I've found worth holding, and a record of how I got here.

Work Projects and research Ideas Things I want to see in the world Blog Essays and rough notes Books What I've read and what stuck Quotes Lines I keep coming back to Journey How I got here Interests What I do outside of work