Retrieval-free Knowledge Injection through Multi-Document Traversal for Dialogue Models
Rui Wang, Jianzhu Bao, Fei Mi, Yi Chen, Hongru Wang, Yasheng Wang, Yitong Li, Lifeng Shang, Kam-Fai Wong, Ruifeng Xu
Main: Dialogue and Interactive Systems Main-poster Paper
Session 1: Dialogue and Interactive Systems (Virtual Poster)
Conference Room: Pier 7&8
Conference Time: July 10, 11:00-12:30 (EDT) (America/Toronto)
Global Time: July 10, Session 1 (15:00-16:30 UTC)
Keywords:
knowledge augmented, grounded dialog
TLDR:
Dialogue models are often enriched with extensive external knowledge to provide informative responses through a retrieval-augmented pipeline.
Nevertheless, retrieval-augmented approaches rely on finely annotated retrieval training data and knowledge-grounded response generation data, making it costl...
You can open the
#paper-P3058
channel in a separate window.
Abstract:
Dialogue models are often enriched with extensive external knowledge to provide informative responses through a retrieval-augmented pipeline.
Nevertheless, retrieval-augmented approaches rely on finely annotated retrieval training data and knowledge-grounded response generation data, making it costly to transfer.
To tackle this challenge, this paper proposed a retrieval-free approach, KiDG, by automatically turning knowledge documents into simulated multi-turn dialogues through a Multi-Document Traversal algorithm.
The simulated knowledge-intensive dialogues constructed by KiDG in one domain can be easily used to train and enhance pre-trained dialogue models' knowledge w.r.t. this domain without costly annotation.
We conduct extensive experiments comparing retrieval-augmented models and a variety of retrieval-free models.
We found that dialogue models enhanced with data simulated with KiDG largely outperform state-of-the-art retrieval-free methods, and it achieves comparable performance compared to retrieval-augmented methods while being better, and cheaper at domain transfer.