Commonsense reasoning with implicit knowledge in natural language
Published in AKBC-2021, 2021
Recommended citation: @inproceedings{banerjee2021commonsense, title={Commonsense Reasoning with Implicit Knowledge in Natural Language}, author={Banerjee, Pratyay and Mishra, Swaroop and Pal, Kuntal Kumar and Mitra, Arindam and Baral, Chitta}, booktitle={3rd Conference on Automated Knowledge Base Construction}, year={2021} } https://openreview.net/pdf?id=a4-fFL7aCi0
[Paper] – [[Code | Data]](https://github.com/ari9dam/McQueen) – [Video] |
Abstract
Commonsense Reasoning is a research challenge studied from the early days of AI. In recent years, several natural language QA task have been proposed where commonsense reasoning is important. Two common approaches to this are (i) Use of well-structured commonsense present in knowledge graphs, and (ii) Use of progressively larger transformer language models. While acquiring and representing commonsense in a formal representation is challenging in approach (i), approach (ii) gets more and more resource-intensive. In this work, we take a middle ground where we use smaller language models together with a relatively smaller but targeted natural language text corpora. The advantages of such an approach is that it is less resource intensive and yet at the same time it can use unstructured text corpora. We define different unstructured commonsense knowledge sources, explore three strategies for knowledge incorporation, and propose four methods competitive to state-of-the-art methods to reason with implicit commonsense.