Timezone: »
Theory of Mind (ToM)---the ability to reason about the mental states of other people---is a key element of our social intelligence. Yet, despite their ever more impressive performance, large-scale neural language models still lack basic theory of mind capabilities out-of-the-box. We posit that simply scaling up models will not imbue them with theory of mind due to the inherently symbolic and implicit nature of the phenomenon, and instead investigate an alternative: can we design a decoding-time algorithm that enhances theory of mind of off-the-shelf neural language models without explicit supervision? We present SymbolicToM, a plug-and-play approach to reason about the belief states of multiple characters in reading comprehension tasks via explicit symbolic representation. More concretely, our approach tracks each entity's beliefs, their estimation of other entities' beliefs, and higher-order levels of reasoning, all through graphical representations, allowing for more precise and interpretable reasoning than previous approaches. Empirical results on the well-known ToMi benchmark (Le et al., 2019) demonstrate that SymbolicToM dramatically enhances off-the-shelf neural networks' theory of mind in a zero-shot setting while showing robust out-of-distribution performance compared to supervised baselines. Our work also reveals spurious patterns in existing theory of mind benchmarks, emphasizing the importance of out-of-distribution evaluation and methods that do not overfit a particular dataset.
Author Information
Melanie Sclar (University of Washington)
Sachin Kumar (School of Computer Science, Carnegie Mellon University)
Peter West (Allen Institute for Artificial Intelligence)
Alane Suhr (Allen Institute for AI)
Yejin Choi (University of Washington)
Yulia Tsvetkov (Department of Computer Science, University of Washington)
More from the Same Authors
-
2023 : SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks »
Yuchen Lin · Yicheng Fu · Karina Yang · Prithviraj Ammanabrolu · Faeze Brahman · Shiyu Huang · Chandra Bhagavatula · Yejin Choi · Xiang Ren -
2023 : Minding Language Models' (Lack of) Theory of Mind: A Plug-and-Play Multi-Character Belief Tracker »
Melanie Sclar · Sachin Kumar · Peter West · Alane Suhr · Yejin Choi · Yulia Tsvetkov -
2023 Workshop: Workshop on Theory of Mind in Communicating Agents »
Hao Zhu · Jennifer Hu · Hyunwoo Kim · Alane Suhr · Saujas Vaduguru · Chenghao Yang · Pei Zhou · Xuhui Zhou -
2023 Poster: Do Embodied Agents Dream of Pixelated Sheep: Embodied Decision Making using Language Guided World Modelling »
Kolby Nottingham · Prithviraj Ammanabrolu · Alane Suhr · Yejin Choi · Hannaneh Hajishirzi · Sameer Singh · Roy Fox -
2022 Poster: Symmetric Machine Theory of Mind »
Melanie Sclar · Graham Neubig · Yonatan Bisk -
2022 Spotlight: Symmetric Machine Theory of Mind »
Melanie Sclar · Graham Neubig · Yonatan Bisk -
2022 Poster: Understanding Dataset Difficulty with $\mathcal{V}$-Usable Information »
Kawin Ethayarajh · Yejin Choi · Swabha Swayamdipta -
2022 Oral: Understanding Dataset Difficulty with $\mathcal{V}$-Usable Information »
Kawin Ethayarajh · Yejin Choi · Swabha Swayamdipta -
2020 Poster: Adversarial Filters of Dataset Biases »
Ronan Le Bras · Swabha Swayamdipta · Chandra Bhagavatula · Rowan Zellers · Matthew Peters · Ashish Sabharwal · Yejin Choi