Mindreading Models: Do LLMs Have Theory of Mind?
July 22, 2024
Did you know we're all mind readers? Humans develop what’s called “theory of mind” (ToM) around the age of four, enabling us to hold views and beliefs about our own mental states and the mental states of others. It’s one of those things that, well, makes us human. However, a recent paper by Strachan et al. explores this capability in the context of AI -- a finding which has some interesting implications. The authors compared the performance of a series of LLMs (GPT-3, GPT-4, and LLaMA2) against around 1,900 humans across a battery of tests designed to explore ToM. They determined that, in every task bar one (designed to test faux pas), LLMs can perform at or above human level. My own cursory testing suggests that GPT-4o may even be able to demonstrate ToM across the board. While we can’t conclude that these models actually have ToM (in a cognitive sense) they certainly seem to behave as if they do. So what does this mean in practice? These LLMs open the door to exciting potential applications across areas such as mental health and decision support making. Ultimately, this paper’s conclusions give us a moment to stop and think about the state of our technology: to be impressed by what we’ve accomplished, and acknowledge the risks that should be considered as we move forward.