Generating Dialog Responses with Specified Grammatical Items for Second Language Learning

Yuki Okano, Kotaro Funakoshi, Ryo Nagata, Manabu Okumura

18th Workshop on Innovative Use of NLP for Building Educational Applications Paper

TLDR: This paper proposes a new second language learning task of generating a response including specified grammatical items. We consider two approaches: 1) fine-tuning a pre-trained language model (DialoGPT) by reinforcement learning and 2) providing a few-shot prompt to a large language model (GPT-3). F
You can open the #paper-BEA_25 channel in a separate window.
Abstract: This paper proposes a new second language learning task of generating a response including specified grammatical items. We consider two approaches: 1) fine-tuning a pre-trained language model (DialoGPT) by reinforcement learning and 2) providing a few-shot prompt to a large language model (GPT-3). For reinforcement learning, we examine combinations of three reward functions that consider grammatical items, diversity, and fluency. Our experiments confirm that both approaches can generate responses including the specified grammatical items and that it is crucial to consider fluency rather than diversity as the reward function.