I received my PhD from the Computation and Neural Systems program at Caltech. I was advised by Professor Pietro Perona. I am currently on the job market, please reach out if you are looking for a researcher/engineer experienced in interpretability and/or computer vision.
Research Interests
I’m broadly interested in how changes in model representations affect model behaviors. My work focuses on representational alignment and interpretability. How are these representations similar or different? How do these differences influence a model’s outputs?
To address these questions, I’ve been developing interpretability methods (RSVC - ICLR 2025, RDX - NeurIPS 2025) to analyze these similarities and differences across representations. Comparison is a powerful lens for understanding how changes in training, architecture, or data shape the representations models learn. Looking ahead, I want to extend this work to compare human and model representations, uncovering concepts unique to each. This knowledge will help improve the alignment between how machines and humans represent the world, enabling the development of safer and more interpretable models.
Background
I earned a B.S. in Physiology & Neuroscience with a minor in Computer Science from the University of California San Diego. As a student at UCSD, I did systems neuroscience research in the Komiyama Lab for two years and neuroeconomics research with Dr. Pamela Reinagel for one year. During these years, I developed an interest in artificial neural networks and, upon graduation, decided to join the Computational Vision Group at Caltech. My PhD research focused on learning from AI via eXplainable AI. I developed algorithms to (1) identify the features used by models to achieve superhuman performance on visual tasks and (2) teach those features to human learners. I succesfully defended my thesis in May 2025.
