Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Knowledge and Logical Reasoning in the Era of Data-driven Learning

BoardgameQA: Natural Language Reasoning with Contradictory Information

Mehran Kazemi · Quan Yuan · Deepti Bhatia · Najoung Kim · Xin Xu · Vaiva Imbrasaite · Deepak Ramachandran


Abstract:

Automated reasoning with unstructured natural text is an important field of research with many potential applications, and is rapidly growing thanks to recent advancements in Language Models (LMs).Existing benchmarks for automated reasoning assume access to a consistent set of knowledge over which a model reasons, which does not capture common real-world scenarios where information is noisy and sometimes contradictory. In many applications, conflicts can often be resolved by imposing preferences over information sources (e.g., based on the credibility of the source). In this paper, we formulate the problem of reasoning with contradictory information guided by preferences over sources as the classical AI problem of defeasible reasoning, and develop a question-answering benchmark called BoardgameQA for measuring the defeasible reasoning capacity of LMs. Additionally, BoardgameQA also incorporates reasoning with implicit background knowledge, to better reflect reasoning problems in downstream applications. Experiments with three types of models (finetuning with and without proof steps and few-shot prompting) show that LMs perform poorly when reasoning with conflicting information, especially in the few-shot case, and the amount of background knowledge required compounds this difficulty even further.

Chat is not available.