LLM Cognition and Evaluation
Do large language models have beliefs? Can they understand? What would it mean for them to be reliable despite hallucination? These questions require us to develop new conceptual frameworks that take seriously both the achievements and limitations of current AI systems.
My work in this area focuses on two interconnected themes: (1) developing philosophically grounded accounts of LLM cognition that avoid treating them as either human-like minds or mere statistical tools, and (2) identifying biases in how we evaluate LLMs that stem from importing assumptions from human psychology without justification.