My research addresses basic questions about the nature of cognition, or more generally, about mentality: what are minds and how do they work? In particular, I'm trying to understand what sorts of mechanisms and processes (representational? informational? perceptual?) are shared by both biological and artificial cognitive systems. Rather than asking grand a priori questions about the possibility of strong artificial intelligence, I aim to analyze the functional nuts and bolts of cognition as-we-know-it and then to figure out whether artificial systems are built from the same components (spoiler: sometimes yes, sometimes no). Deep-learning based AI systems like LLMs are particularly interesting on this front, since — despite being designed by humans — the mechanisms underlying their behavior remain largely unknown to us.
I am also interested in higher-level cognitive phenomena of various sorts, especially synesthesia, cross-modal correspondence, and conceptual metaphor. These are fascinating because, paradigmatically, they present the world as it is not (they are non-veridical) and so the propositions they suggest to us (egs. that middle-C is blue, that parmesan cheese is sharp, or that love is war) strictly cannot feature as premises in sound arguments or steps in justified inferences. Yet, both our ordinary and expert thinking is riddled with their influence. They affect our actions, our intuitions, and the positions we find most attractive; Quine's "desert landscapes" remark was even more prescient than he knew. This part of my research attempts to understand how these forms of cognition work at the neural and psychological level, and to explore their implications for various aspects of human life (including for philosophical practice).
Publications
Abstract: The discipline of philosophy has been critiqued from both within and outside itself. One brand of external critique is associated with conceptual metaphor theory (CMT), the view that human cognition is partially structured by pervasive and automatic mappings between conceptual domains. Most notably, Lakoff and Johnson (1999) claimed that many central philosophical concepts and arguments rely on an unacknowledged metaphorical substructure, and that this structure has sometimes led philosophy astray. The purpose of this paper is to argue that Lakoff and Johnson’s critique is anticipated by the work of post-Tractarian Wittgenstein and his student, Margaret MacDonald. In the Blue Book, Wittgenstein outlines a method for identifying and resolving philosophical puzzles generated by misused grammatical analogies, although his discussion lacks a precise characterization of exactly how and why such analogies lead to trouble. In a 1938 paper, MacDonald offers such a characterization, which I outline and then connect back to Wittgenstein. In addition to this interpretive work, I supplement Wittgenstein and MacDonald’s diagnosis using evidence from CMT which suggests that linguistic metaphors and analogies often originate in or are motivated by more fundamental analogical mappings in cognition. The supplemented account carries implications for how philosophical arguments ought to be formulated and critiqued.
Under review/in progress
Abstract: The extraordinary success of recent Large Language Models (LLMs) on a diverse array of tasks has led to an explosion of scientific and philosophical theorizing aimed at explaining how they do what they do. Unfortunately, disagreement over fundamental theoretical issues has led to stalemate, with entrenched camps of LLM optimists and pessimists often committed to very different views of how these systems work. Overcoming stalemate requires agreement on fundamental questions, and the goal of this paper is to address one such question, namely: is LLM behavior driven partly by representation-based information processing of the sort implicated in biological cognition, or is it driven entirely by processes of memorization and stochastic table look-up? This is a question about what kind of algorithm LLMs implement, and the answer carries serious implications for higher level questions about whether these systems have beliefs, intentions, concepts, knowledge, and understanding. I argue that LLM behavior is partially driven by representation-based information processing, and then I describe and defend a series of practical techniques for investigating these representations and developing explanations on their basis. The resulting account provides a groundwork for future theorizing about language models and their successors.
A suite of recent papers have explored the idea that Large Language Models (LLMs) represent the truth of their inputs, and that these representations constitute the models' beliefs. I argue that although there are no in-principle problems with the project of investigating LLM representations of truth, these representations are not beliefs, since (1) they lack many of the core features of belief, and since (2) treating them as beliefs risks collapsing the distinction between belief and other propositional attitude-types.
This paper attempts to explain why, on average, synesthetes have better memories than non-synesthetes. It turns out that that best explanation involves a breakdown of functional modularity in the ventral visual stream, lending support to the idea that perception and memory are not subserved by distinct neural regions.