COGEN: Abductive Commonsense Language Generation

rohola zandie, Diwanshu Shekhar, Mohammad Mahoor

Main: Generation Main-oral Paper

Session 7: Generation (Oral)
Conference Room: Metropolitan Centre
Conference Time: July 12, 11:45-12:30 (EDT) (America/Toronto)
Global Time: July 12, Session 7 (15:45-16:30 UTC)
Keywords: retrieval-augmented generation
TLDR: Reasoning is one of the most important elements in achieving Artificial General Intelligence (AGI), specifically when it comes to Abductive and counterfactual reasoning. In order to introduce these capabilities of reasoning in Natural Language Processing (NLP) models, there have been recent advances...
You can open the #paper-P250 channel in a separate window.
Abstract: Reasoning is one of the most important elements in achieving Artificial General Intelligence (AGI), specifically when it comes to Abductive and counterfactual reasoning. In order to introduce these capabilities of reasoning in Natural Language Processing (NLP) models, there have been recent advances towards training NLP models to better perform on two main tasks - Abductive Natural Language Inference (alphaNLI) and Abductive Natural Language Generation Task (alphaNLG). This paper proposes CoGen, a model for both alphaNLI and alphaNLG tasks that employ a novel approach of combining the temporal commonsense reasoning for each observation (before and after a real hypothesis) from pre-trained models with contextual filtering for training. Additionally, we use state-of-the-art semantic entailment to filter out the contradictory hypothesis during the inference. Our experimental results show that CoGen outperforms current models and set a new state of the art in regards to alphaNLI and alphaNLG tasks. We make the source code of CoGen model publicly available for reproducibility and to facilitate relevant future research.