Workshop
Workshop on Theory of Mind in Communicating Agents
Hao Zhu · Jennifer Hu · Hyunwoo Kim · Alane Suhr · Saujas Vaduguru · Chenghao Yang · Pei Zhou · Xuhui Zhou
Meeting Room 317 A
Theory of Mind (ToM) is the ability to reason about the minds of other agents. The main theme of our workshop is the computational modeling of ToM, with a special focus on the role of natural language in such modeling. Specifically, ToM 2023 pays attention to cognitive foundations and theories of ToM, the acquisition and relationship between language and ToM, leveraging ToM to improve and explain NLP and ML models, and using ToM for positive social impact. This workshop intends to promote the community of researchers that are interested in improving the ability of intelligent agents to reason about others' mental states. Our proposed program provides a space to discuss pathways for understanding and applying ToM in psycholinguistics, pragmatics, human value alignment, social good, model explainability, and many other areas of NLP. ToM 2023 will be a full-day hybrid in-person/virtual workshop with several keynote speeches, and oral/poster/spotlight presentations, followed by a breakout discussion, panel discussion, and best paper award announcement. We also intend to host a mentoring program to broaden participation from a diverse set of researchers.
Schedule
Fri 11:55 a.m. - 12:00 p.m.
|
Openning Remarks
(
Openning Remarks
)
>
SlidesLive Video |
🔗 |
Fri 12:00 p.m. - 12:45 p.m.
|
Maarten Sap's Talk
(
Keynote
)
>
SlidesLive Video |
Maarten Sap 🔗 |
Fri 12:45 p.m. - 12:55 p.m.
|
Emergent deception and skepticism via theory of mind
(
Oral Talk
)
>
link
SlidesLive Video In complex situations involving communication, agents might attempt to mask their intentions, exploiting Shannon's theory of information as a theory of misinformation. Here, we introduce and analyze a simple multiagent reinforcement learning task where a buyer sends signals to a seller via its actions, and in which both agents are endowed with a recursive theory of mind. We show that this theory of mind, coupled with pure reward-maximization, gives rise to agents that selectively distort messages and become skeptical towards one another. Using information theory to analyze these interactions, we show how savvy buyers reduce mutual information between their preferences and actions, and how suspicious sellers learn to reinterpret or discard buyers' signals in a strategic manner. |
Lion Schulz · Nitay Alon · Jeffrey Rosenschein · Peter Dayan 🔗 |
Fri 12:55 p.m. - 1:05 p.m.
|
Between prudence and paranoia: Theory of Mind gone right, and wrong
(
Oral Talk
)
>
link
SlidesLive Video Agents need to be on their toes when interacting with competitive others to avoid being duped. Too much vigilance out of context can, however, be detrimental and produce paranoia. Here, we offer a formal account of this phenomenon through the lens of theory of mind. We simulate agents of different depths of mentalization and show how, if aligned well, deep recursive mentalisation gives rise to both successful deception as well as reasonable skepticism. However, we also show how, if theory of mind is too sophisticated, agents become paranoid, losing trust and reward in the process. We discuss our findings in light of computational psychiatry and AI safety. |
Nitay Alon · Lion Schulz · Peter Dayan · Joseph Barnby 🔗 |
Fri 1:05 p.m. - 1:15 p.m.
|
Language Models are Pragmatic Speakers
(
Oral Talk
)
>
link
SlidesLive Video How do language models “think”? This paper formulates a probabilistic cognitive model called bounded pragmatic speaker that can characterize the operation of different variants of language models. In particular, we show that large language models fine-tuned with reinforcement learning from human feedback (Ouyang et al., 2022) implements a model of thought that conceptually resembles the well-known fast-and-slow model (Kahneman, 2011) which have been largely attributed to humans. We discuss the limitations of reinforcement learning from human feedback as a fast-and-slow model of thought and propose directions for extending this framework. Overall, our work demonstrates that viewing language models through the lens of modular probabilistic models can offer valuable insights for understanding, evaluating, and developing them. |
Khanh Nguyen 🔗 |
Fri 1:15 p.m. - 1:45 p.m.
|
Break/Meet-and-Greet
(
Break
)
>
|
🔗 |
Fri 1:45 p.m. - 2:30 p.m.
|
Noga Zaslavsky's Talk
(
Keynote
)
>
SlidesLive Video |
Noga Zaslavsky 🔗 |
Fri 2:30 p.m. - 3:15 p.m.
|
Mark Ho's Talk
(
Keynote
)
>
SlidesLive Video |
Mark Ho 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Discovering User Types: Characterization of User Traits by Task-Specific Behaviors in Reinforcement Learning
(
Poster
)
>
link
We often want to infer user traits when personalizing interventions.Approaches like Inverse RL can learn traits formalized as parameters of a Markov Decision Process but are data intensive. Instead of inferring traits for individuals, we study the relationship between RL worlds and the set of user traits. We argue that understanding the breakdown of ``user types" within a world -- broad sets of traits that result in the same behavior -- helps rapidly personalize interventions. We show that seemingly different RL worlds admit the same set of user types and formalize this observation as an equivalence relationdefined on worlds. We show that these equivalence classes capture many different worlds. We argue that the richness of these classes allows us to transfer insights on intervention design between toy and real worlds. |
Lars L. Ankile · Brian Ham · Kevin Mao · Eura Shin · Siddharth Swaroop · Finale Doshi-Velez · Weiwei Pan 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Do LLMs selectively encode the goal of an agent's reach?
(
Poster
)
>
link
In this work, we investigate whether large language models (LLMs) exhibit one of the earliest Theory of Mind-like behaviors: selectively encoding the goal object of an actor's reach (Woodward, 1998). We prompt state-of-the-art LLMs with ambiguous examples that can be explained both by an object or a location being the goal of an actor's reach, and evaluate the model's bias. We compare the magnitude of the bias in three situations: i) an agent is acting purposefully, ii) an inanimate object is acted upon, and iii) an agent is acting accidentally. We find that two models show a selective bias for agents acting purposefully, but are biased differently than humans. Additionally, the encoding is not robust to semantically equivalent prompt variations. We discuss how this bias compares to the bias infants show and provide a cautionary tale of evaluating machine Theory of Mind (ToM). We release our dataset and code. |
Laura Ruis · Arduin Findeis · Herbie Bradley · Hossein A. Rahmani · Kyoung Whan Choe · Edward Grefenstette · Tim Rocktäschel 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
[Empirical paper track] Deciphering Enemies in the Darkness through Modeling and Examination of Knowledge in Reconnaissance Blind Chess
(
Poster
)
>
link
An important research topic about Theory of Mind (ToM) is the ability to understand and reason about how agents acquire and predict the behavioral and mental states of other agents in dynamic environments, especially those involving a significant change in knowledge and information. In this paper, we focus on the modeling and examination of knowledge of other agents in imperfect information games. More specifically, we delve into the nuances of the change of knowledge in the Reconnaissance Blind Chess (RBC). In each round, players are granted limited sensing capacity of the board. Thus, the understanding opponent's knowledge and strategy plays a key role in decision-making in each round. This paper studies how an agent can model and utilize an opponent's knowledge in the RBC game. The examination includes a detailed comparison of information obtained through different actions in the game. We design two sensing strategies for obtaining information based on entropy and other factors and compare how these strategies can impact the outcome of the game. Finally, we discuss how our research results could be generalized to the understanding of opponents' knowledge and behavior in non-cooperative imperfect information games. |
Robin Stöhr · Shuai Wang 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
The SocialAI School: Insights from Developmental Psychology Towards Artificial Socio-Cultural Agents
(
Poster
)
>
link
Developmental psychologists have long-established the importance socio-cognitive abilities in human intelligence. These abilities enable us to enter, participate and benefit from human culture. AI research on social interactive agents mostly concerns the emergence of culture in a multi-agent setting (often without a strong grounding in developmental psychology). We argue that AI research should be informed by psychology and study socio-cognitive abilities enabling to enter a culture too. We discuss the theories of Michael Tomasello and Jerome Bruner to introduce some of their concepts to AI and outline key concepts and socio-cognitive abilities. We present The SocialAI school - a tool including a customizable parameterized suite of procedurally generated environments, which simplifies conducting experiments regarding those concepts. We show examples of such experiments with RL agents and pure-text Language Models. The main motivation of this work is to engage the AI community around the problem of social intelligence informed by developmental psychology, and to provide a tool to simplify first steps in this direction. |
Grgur Kovač · Rémy Portelas · Peter F Dominey · Pierre-Yves Oudeyer 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Preference Proxies: Evaluating Large Language Models in capturing Human Preferences in Human-AI Tasks
(
Poster
)
>
link
In this work, we investigate the potential of Large Language Models (LLMs) to serve as effective human proxies by capturing human preferences in the context of collaboration with AI agents. Focusing on two key aspects of human preferences - explicability and sub-task specification in team settings - we explore LLMs' ability to not only model mental states but also understand human reasoning processes. By developing scenarios where optimal AI performance relies on modeling human mental states and reasoning, our investigation involving two different preference types and a user study (with 17 participants) contributes valuable insights into the suitability of LLMs as ``Preference Proxies" in various human-AI applications, paving the way for future research on the integration of AI agents with human users in Human-Aware AI tasks. |
Mudit Verma · Siddhant Bhambri · Subbarao Kambhampati 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Inferring the Goals of Communicating Agents from Actions and Instructions
(
Poster
)
>
link
When humans cooperate, they frequently coordinate their activity through both verbal communication and non-verbal actions, using this information to infer a shared goal and plan. How can we model this inferential ability? In this paper, we introduce a model of a cooperative team where one agent, the principal, may communicate natural language instructions about their shared plan to another agent, the assistant, using GPT-3 as a likelihood function for instruction utterances. We then show how a third person observer can infer the team's goal via multi-modal Bayesian inverse planning from actions and instructions, computing the posterior distribution over goals under the assumption that agents will act and communicate rationally to achieve them. We evaluate this approach by comparing it with human goal inferences in a multi-agent gridworld, finding that our model's inferences closely correlate with human judgments $(R = 0.96)$. When compared to inference from actions alone, we also find that instructions lead to more rapid and less uncertain goal inference, highlighting the importance of verbal communication for cooperative agents.
|
Lance Ying · Tan Zhi-Xuan · Vikash Mansinghka · Josh Tenenbaum 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Theory of Mind as Intrinsic Motivation for Multi-Agent Reinforcement Learning
(
Poster
)
>
link
The ability to model the mental states of others is crucial to human social intelligence, and can offer similar benefits to artificial agents with respect to the social dynamics induced in multi-agent settings. We present a method of grounding semantically meaningful, human-interpretable beliefs within policies modeled by deep networks. We then consider the task of 2nd-order belief prediction. We propose that ability of each agent to predict the beliefs of the other agents can be used as an intrinsic reward signal for multi-agent reinforcement learning. Finally, we present preliminary empirical results in a mixed cooperative-competitive environment. |
Ini Oguntola · Joseph Campbell · Simon Stepputtis · Katia Sycara 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
The Neuro-Symbolic Inverse Planning Engine (NIPE): Modeling probabilistic social inferences from linguistic inputs
(
Poster
)
>
link
Human beings are social creatures. We routinely reason about other agents, and a crucial component of this social reasoning is inferring people's goals as we learn about their actions. In many settings, we can perform intuitive but reliable goal inference from language descriptions of agents, actions, and the background environments. In this paper, we study this process of language driving and influencing social reasoning in a probabilistic goal inference domain. We propose a neuro-symbolic model that carries out goal inference from linguistic inputs of agent scenarios. The "neuro" part is a large language model (LLM) that translates language descriptions to code representations, and the "symbolic" part is a Bayesian inverse planning engine. To test our model, we design and run a human experiment on a linguistic goal inference task. Our model closely matches human response patterns and better predicts human judgements than using an LLM alone. |
Lance Ying · Katie Collins · Megan Wei · Cedegao Zhang · Tan Zhi-Xuan · Adrian Weller · Josh Tenenbaum · Catherine Wong 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Hi-ToM: A Benchmark for Evaluating Higher-Order Theory of Mind Reasoning in Large Language Models
(
Poster
)
>
link
Theory of Mind (ToM) is the ability to understand and reason about one's own and others' mental states, which plays a critical role in the development of intelligence, language understanding, and cognitive processes. While existing work has primarily focused on first and second-order ToM, we explore higher-order ToM, which involves recursive reasoning on others' beliefs. We introduce Hi-ToM, a Higher Order Theory of Mind benchmark. Our experimental evaluation using GPT-4 reveals a decline in performance on higher-order ToM tasks, indicating the limitations of current models. This highlights the challenges of reasoning in complex ToM scenarios and emphasizes the need for further advancements in large language models' higher-order ToM capabilities. |
Yinghui He · Yufan Wu · Yulong Chen · Naihao Deng 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
EPITOME: Experimental Protocol Inventory for Theory Of Mind Evaluation
(
Poster
)
>
link
We address a growing debate about the extent to which large language models (LLMs) produce behavior consistent with Theory of Mind (ToM) in humans. We present EPITOME: a battery of six experiments that tap diverse ToM capacities, including belief attribution, emotional inference, pragmatic reasoning, and non-literal communication. For each task we compare responses from 5 LLMs to a baseline of responses from human comprehenders. Results are mixed. LLMs show broad sensitivity to mental state information and perform at parity with humans across several tasks. However, models make systematic errors in other tasks, especially those that require pragmatic reasoning from mental state information. Such inconsistent performance suggests that crediting LLMs with ToM may be premature. |
Cameron Jones · Sean Trott · Ben Bergen 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Robust Inverse Reinforcement Learning Through Bayesian Theory of Mind
(
Poster
)
>
link
We consider the Bayesian theory of mind (BTOM) framework for learning from demonstrations via inverse reinforcement learning (IRL). The BTOM model consists of a joint representation of the agent’s reward function and the agent's internal subjective model of the environment dynamics, which may be inaccurate. In this paper, we make use of a class of prior distributions that parametrize how accurate is the agent’s model of the environment to develop efficient algorithms to estimate the agent's reward and subjective dynamics in high-dimensional settings. The BTOM framework departs from existing offline model-based IRL approaches by performing simultaneous estimation of reward and dynamics. Our analysis reveals a novel insight that the estimated policy exhibits robust performance when the (expert) agent is believed (a priori) to have a highly accurate model of the environment. We verify this observation in the MuJoCo environment and show that our algorithms outperform state-of-the-art offline IRL algorithms. |
Ran Wei · Siliang Zeng · Chenliang Li · Alfredo Garcia · Anthony McDonald · Mingyi Hong 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Define, Evaluate, and Improve Task-Oriented Cognitive Capabilities for Instruction Generation Models
(
Poster
)
>
link
Recent work studies the cognitive capabilities of language models through psychological tests designed for humans. While these studies are helpful for understanding the general capabilities of these models, there is no guarantee that a model possessing sufficient capabilities to pass those tests would actually use those capabilities in performing real-life tasks.In this work, we formulate task-oriented cognitive capabilities, which are human-like cognitive capabilities that language models leverage to perform tasks. These capabilities are (i) the ability to quickly generate good candidate utterances (the search capability) (ii) the ability to predict how a listener interprets those utterances and choose the most appropriate one (the pragmatic capability).We design an evaluation scheme for comparing these capabilities of a language model with those of a human.Applying this scheme to examine various models in a navigation instruction generation problem, we find that their pragmatic capability is severely lacking.This insight leads us to augment them with better models of the listener and obtain a significant boost of 11% in success rate in guiding real humans.Our work advocates for having a principled procedure for aligning language models with humans that involves (i) formulating task-oriented capabilities, (ii) devising a method to quantify their deficiency, and (iii) iteratively improving them. |
Lingjun Zhao · Khanh Nguyen · Hal Daumé 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Towards a better Rational Speech Act framework for context-aware modeling of metaphor understanding
(
Poster
)
>
link
Modeling language is a fundamental step for understanding human communication and improving human-computer interaction. The Rational Speech Act (RSA) model provides a flexible framework to pursue this objective by catching pragmatic reasoning. However, state-of-the-art models still have limitations in dealing with context. We present a new RSA framework for metaphor understanding that accounts explicitly for the role of context by emphasizing the mutual shared information between the speaker and the listener in the estimation of the communicative goal. The model is tested extensively against 24 metaphors (with either intrinsic or emergent properties) and its predictions are compared to human data. |
Gaia Carenini · Luca Bischetti · Walter Schaeken · Valentina Bambini 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Minding Language Models' (Lack of) Theory of Mind: A Plug-and-Play Multi-Character Belief Tracker
(
Poster
)
>
link
Theory of Mind (ToM)---the ability to reason about the mental states of other people---is a key element of our social intelligence. Yet, despite their ever more impressive performance, large-scale neural language models still lack basic theory of mind capabilities out-of-the-box. We posit that simply scaling up models will not imbue them with theory of mind due to the inherently symbolic and implicit nature of the phenomenon, and instead investigate an alternative: can we design a decoding-time algorithm that enhances theory of mind of off-the-shelf neural language models without explicit supervision? We present SymbolicToM, a plug-and-play approach to reason about the belief states of multiple characters in reading comprehension tasks via explicit symbolic representation. More concretely, our approach tracks each entity's beliefs, their estimation of other entities' beliefs, and higher-order levels of reasoning, all through graphical representations, allowing for more precise and interpretable reasoning than previous approaches. Empirical results on the well-known ToMi benchmark (Le et al., 2019) demonstrate that SymbolicToM dramatically enhances off-the-shelf neural networks' theory of mind in a zero-shot setting while showing robust out-of-distribution performance compared to supervised baselines. Our work also reveals spurious patterns in existing theory of mind benchmarks, emphasizing the importance of out-of-distribution evaluation and methods that do not overfit a particular dataset. |
Melanie Sclar · Sachin Kumar · Peter West · Alane Suhr · Yejin Choi · Yulia Tsvetkov 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Comparing the Evaluation and Production of Loophole Behavior in Children and Large Language Models
(
Poster
)
>
link
In law, lore, and everyday life, loopholes are commonplace. When people exploit a loophole, they understand the intended meaning or goal of another, but choose to go with a different, though still possible interpretation. Previous work suggests people exploit loopholes when their goals are misaligned with the goals of others, but both capitulation and disobedience are too costly. Past and current AI research has shown that artificial intelligence engages in what seems superficially like the exploitation of loopholes. However, this is an anthropomorphization. It remains unclear to what extent current models, especially Large Language Models (LLMs), capture the pragmatic understanding required for engaging in loopholes. We examined the performance of LLMs on two metrics developed for studying loophole behavior in adults and children: evaluation (are loopholes rated as resulting in differential trouble compared to compliance and non-compliance), and generation (coming up with new loopholes in a given context). We conducted a fine-grained comparison of state-of-the-art LLMs to children, and found that while some LLMs rate loophole behaviors as resulting in less trouble than outright non-compliance (in line with children), they struggle to generate loopholes of their own. Our results suggest a separation between the faculties underlying the evaluation and generation of loophole behavior, in both children and LLMs, with LLM abilities dovetailing with those of the youngest children in our studies. |
Sonia Murthy · Sophie Bridgers · Kiera Parece · Elena Glassman · Tomer Ullman 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Inferring the Future by Imagining the Past
(
Poster
)
>
link
A single panel of a comic book can say a lot: it shows not only where characters currently are, but also where they came from, what their motivations are, and what might happen next. More generally, humans can often infer a complex sequence of past and future events from a single snapshot image of an intelligent agent.Building on recent work in cognitive science, we offer a Monte Carlo algorithm for making such inferences. Drawing a connection to Monte Carlo path tracing in computer graphics, we borrow ideas that help us dramatically improve upon prior work in sample efficiency. This allows us to scale to a wide variety of challenging inference problems with only a handful of samples. It also suggests some degree of cognitive plausibility, and indeed we present human subject studies showing that our algorithm matches human intuitions in a variety of domains that previous methods could not scale to. |
Kartik Chandra · Tony Chen · Tzu-Mao Li · Jonathan Ragan-Kelley · Josh Tenenbaum 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
Iterative Machine Teaching for Black-box Markov Learners
(
Poster
)
>
link
Machine teaching has traditionally been constrained by the assumption of a fixed learner's model. In this paper, we challenge this notion by proposing a novel black-box Markov learner model, drawing inspiration from decision psychology and neuroscience where learners are often viewed as black boxes with adaptable parameters. We model the learner's dynamics as a Markov decision process (MDP) with unknown parameters, encompassing a wide range of learner types studied in machine teaching literature. This approach reduces teaching complexity to finding an optimal policy for the underlying MDP. Building on this, we introduce an algorithm for teaching in this black-box setting and provide an analysis of teaching costs under different scenarios. We further establish a connection between our model and two types of learners in psychology and neuroscience, the epiphany learner and the non-epiphany learner, linking them with discounted and non-discounted black-box Markov learners respectively. This alignment offers a psychologically and neuroscientifically grounded perspective to our work. Supported by numerical study results, this paper delivers a significant contribution to machine teaching, introducing a robust, versatile learner model with a rigorous theoretical foundation. |
Chaoqi Wang · Sandra Zilles · Adish Singla · Yuxin Chen 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
How To Make Social Decisions in a Heterogeneous Society?
(
Poster
)
>
link
Understanding cognitive processes in multi-agent interactions is a primary goal in cognitive science. It can guide the direction of artificial intelligence (AI) research toward social decision-making in heterogeneous multi-agent systems. In this paper, we introduce the episodic future thinking (EFT) mechanism of a reinforcement learning (RL) agent by benchmarking the cognitive process of animals. To achieve future thinking functionality, we first train a multi-character policy that reflects heterogeneous characters with an ensemble of heterogeneous policies. An agent's character is defined as a different weight combination on reward components, thus explaining the agent's behavioral preference. The future thinking agent collects observation-action trajectories of the target agents and uses the pre-trained multi-character policy to infer their characters. Once the character is inferred, the agent predicts the upcoming actions of the targets and simulates the future. This capability allows the agent to adaptively select the optimal action, considering the upcoming behavior of others in multi-agent scenarios. To evaluate the proposed mechanism, we consider the multi-agent autonomous driving scenario in which autonomous vehicles with different driving traits are on the road. Simulation results demonstrate that the EFT mechanism with accurate character inference leads to a higher reward than existing multi-agent solutions. We also confirm that the effect of reward improvement remains valid across societies with different levels of character diversity. |
Dongsu Lee · Minhae Kwon 🔗 |
Fri 3:15 p.m. - 4:30 p.m.
|
MindDial: Belief Dynamics Tracking with Theory-of-Mind Modeling for Neural Dialogue Generation
(
Poster
)
>
link
Humans talk in free-form while negotiating the expressed meanings or common ground. Despite the impressive conversational abilities of the large generative language models, they do not consider the individual differences in contextual understanding in a shared situated environment. In this work, we propose MindDial, a novel conversational framework that can generate situated free-form responses to negotiate common ground. We design an explicit mind module that can track three-level beliefs -- the speaker's belief, the speaker's prediction of the listener's belief, and the common belief based on the gap between the first two. Then the speaking act classification head will decide to continue to talk, end this turn, or take task-related action. We augment a common ground alignment dataset MutualFriend with belief dynamics annotation, of which the goal is to find a single mutual friend based on the free chat between two agents. Experiments show that our model with mental state modeling can resemble human responses when aligning common ground meanwhile mimic the natural human conversation flow. The ablation study further validates the third-level common belief can aggregate information of the first and second-order beliefs and align common ground more efficiently. |
Shuwen Qiu · Song-Chun Zhu · Zilong Zheng 🔗 |
Fri 4:30 p.m. - 5:15 p.m.
|
Jan-Philipp Fränken's Talk
(
Keynote
)
>
SlidesLive Video |
Jan-Philipp Fränken 🔗 |
Fri 5:15 p.m. - 5:25 p.m.
|
Inferring the Future by Imagining the Past
(
Oral Talk
)
>
link
SlidesLive Video A single panel of a comic book can say a lot: it shows not only where characters currently are, but also where they came from, what their motivations are, and what might happen next. More generally, humans can often infer a complex sequence of past and future events from a single snapshot image of an intelligent agent.Building on recent work in cognitive science, we offer a Monte Carlo algorithm for making such inferences. Drawing a connection to Monte Carlo path tracing in computer graphics, we borrow ideas that help us dramatically improve upon prior work in sample efficiency. This allows us to scale to a wide variety of challenging inference problems with only a handful of samples. It also suggests some degree of cognitive plausibility, and indeed we present human subject studies showing that our algorithm matches human intuitions in a variety of domains that previous methods could not scale to. |
Kartik Chandra · Tony Chen · Tzu-Mao Li · Jonathan Ragan-Kelley · Josh Tenenbaum 🔗 |
Fri 5:25 p.m. - 5:35 p.m.
|
Minding Language Models' (Lack of) Theory of Mind: A Plug-and-Play Multi-Character Belief Tracker
(
Oral Talk
)
>
link
SlidesLive Video Theory of Mind (ToM)---the ability to reason about the mental states of other people---is a key element of our social intelligence. Yet, despite their ever more impressive performance, large-scale neural language models still lack basic theory of mind capabilities out-of-the-box. We posit that simply scaling up models will not imbue them with theory of mind due to the inherently symbolic and implicit nature of the phenomenon, and instead investigate an alternative: can we design a decoding-time algorithm that enhances theory of mind of off-the-shelf neural language models without explicit supervision? We present SymbolicToM, a plug-and-play approach to reason about the belief states of multiple characters in reading comprehension tasks via explicit symbolic representation. More concretely, our approach tracks each entity's beliefs, their estimation of other entities' beliefs, and higher-order levels of reasoning, all through graphical representations, allowing for more precise and interpretable reasoning than previous approaches. Empirical results on the well-known ToMi benchmark (Le et al., 2019) demonstrate that SymbolicToM dramatically enhances off-the-shelf neural networks' theory of mind in a zero-shot setting while showing robust out-of-distribution performance compared to supervised baselines. Our work also reveals spurious patterns in existing theory of mind benchmarks, emphasizing the importance of out-of-distribution evaluation and methods that do not overfit a particular dataset. |
Melanie Sclar · Sachin Kumar · Peter West · Alane Suhr · Yejin Choi · Yulia Tsvetkov 🔗 |
Fri 5:35 p.m. - 5:45 p.m.
|
Define, Evaluate, and Improve Task-Oriented Cognitive Capabilities for Instruction Generation Models
(
Oral Talk
)
>
link
SlidesLive Video Recent work studies the cognitive capabilities of language models through psychological tests designed for humans. While these studies are helpful for understanding the general capabilities of these models, there is no guarantee that a model possessing sufficient capabilities to pass those tests would actually use those capabilities in performing real-life tasks.In this work, we formulate task-oriented cognitive capabilities, which are human-like cognitive capabilities that language models leverage to perform tasks. These capabilities are (i) the ability to quickly generate good candidate utterances (the search capability) (ii) the ability to predict how a listener interprets those utterances and choose the most appropriate one (the pragmatic capability).We design an evaluation scheme for comparing these capabilities of a language model with those of a human.Applying this scheme to examine various models in a navigation instruction generation problem, we find that their pragmatic capability is severely lacking.This insight leads us to augment them with better models of the listener and obtain a significant boost of 11% in success rate in guiding real humans.Our work advocates for having a principled procedure for aligning language models with humans that involves (i) formulating task-oriented capabilities, (ii) devising a method to quantify their deficiency, and (iii) iteratively improving them. |
Lingjun Zhao · Khanh Nguyen · Hal Daumé 🔗 |
Fri 5:45 p.m. - 6:30 p.m.
|
Kevin R. McKee's Talk
(
Keynote
)
>
SlidesLive Video |
🔗 |
Fri 6:30 p.m. - 7:00 p.m.
|
Break/Meet-and-Greet
(
Break
)
>
|
🔗 |
Fri 7:00 p.m. - 8:00 p.m.
|
Panel Discussion
(
Panel Discussion
)
>
link
SlidesLive Video We are excited to invite Kevin R. McKee, Jesse Thomason and Bodhisattwa Majumder as our panelist! Please post your question through: https://app.sli.do/event/mZXwiZskTgAGG8Aa6Fnwro |
🔗 |
Fri 8:00 p.m. - 8:10 p.m.
|
Closing Remarks
(
Closing Remarks
)
>
SlidesLive Video |
🔗 |