I've collaborated with researchers from OpenAI, Anthropic, and DeepMind on advancing the state of the art in language models and AI safety. My technical work spans transformer pre-training, some minimal post-training/RLing, model quantization, RAG systems, and agents. One might call me a generalist. I'm ok at a lot of things, but not an expert at any of them (except maybe agentic archtectures). I've worked on everything from backend development to training models to deploying them in production.
Previously, I was a Senior ML Engineer at Apple and at Alethea building models and applications. Before that, I was an Assistant Research Professor at Duke University. I served as a Visiting Researcher at Constellation. I currently teach a class at Duke on the side. I used to be a new media artist building weird stuff with code.