Skip to main content

Philosophy

How I Think About AI

Working on frontier AI systems has shaped how I think about technology, responsibility, and the future. These are the principles that guide my work.

Core Principles

Safety is not a feature, it's the foundation

Building powerful AI without safety considerations is like building a rocket without thinking about where it lands. The goal isn't to slow down progress - it's to ensure we're progressing in a direction worth going.

Interpretability is not optional

I believe we shouldn't deploy systems we don't understand. Mechanistic interpretability isn't just academically interesting - it's how we build trust, catch failures, and ensure alignment. Black boxes are tech debt we can't afford.

Rigor over speed (usually)

In safety research, a poorly designed experiment can miss critical failure modes. I've learned to slow down, triple-check the experimental setup, and question my assumptions. The 30 minutes you spend reviewing your methodology saves the 30 hours you'd spend wondering why the model fooled your evaluation.

Collaborate like the stakes are high

Because they are. AI development is too important for hero culture. I believe in open discussion, honest disagreement, and the humility to change my mind when someone has a better idea. The best ideas rarely come from one person.

What I Believe

On AI Safety

  • ā—AI alignment is a real, technical problem - not just philosophy
  • ā—Current systems are not aligned by default; they're aligned by effort
  • ā—The time to solve safety is before we need it, not after
  • ā—Interpretability research is undervalued relative to capabilities research
  • ā—We should be building AI that helps us understand AI

On Research Engineering

  • ā—Good infrastructure is invisible; bad infrastructure is all you can see
  • ā—Reproducibility is non-negotiable, even when it's painful
  • ā—The best code is the code you don't have to write
  • ā—Documentation is a gift to your future self
  • ā—If you can't explain it simply, you don't understand it well enough

On Work & Life

  • ā—Sustainable pace beats burnout heroics every time
  • ā—The best debugging happens after a good night's sleep
  • ā—Mentorship scales better than individual contribution
  • ā—Imposter syndrome is universal; confidence is learned
  • ā—Taking breaks is not slacking - it's maintenance

Questions I'm Thinking About

These are the questions that keep me up at night. If you have thoughts on any of them, I'd love to talk.

ā€œHow do we ensure AI systems remain aligned as they become more capable?ā€
ā€œWhat's the right balance between open research and responsible deployment?ā€
ā€œHow do we build interpretability tools that scale with model size?ā€
ā€œWhat does 'understanding' a neural network actually mean?ā€
ā€œHow do we train models that are genuinely helpful without being sycophantic?ā€
ā€œWhat organizational structures best support AI safety research?ā€

Note: These views are my own and don't necessarily represent the positions of my employer. I'm always learning and happy to be proven wrong.