Soft Alignment Objectives for Robust Adaptation of Language Generation
Michal Štefánik, Marek Kadlcik, Petr Sojka
Main: Generation Main-poster Paper
Poster Session 7: Generation (Poster)
Conference Room: Frontenac Ballroom and Queen's Quay
Conference Time: July 12, 11:00-12:30 (EDT) (America/Toronto)
Global Time: July 12, Poster Session 7 (15:00-16:30 UTC)
Keywords:
domain adaptation
TLDR:
Domain adaptation allows generative language models to address specific flaws caused by the domain shift of their application.
However, the traditional adaptation by further training on in-domain data rapidly weakens the model's ability to generalize to other domains, making the open-ended deploymen...
You can open the
#paper-P284
channel in a separate window.
Abstract:
Domain adaptation allows generative language models to address specific flaws caused by the domain shift of their application.
However, the traditional adaptation by further training on in-domain data rapidly weakens the model's ability to generalize to other domains, making the open-ended deployments of the adapted models prone to errors.
This work introduces novel training objectives built upon a semantic similarity of the predicted tokens to the reference.
Our results show that (1) avoiding the common assumption of a single correct prediction by constructing the training target from tokens' semantic similarity can largely mitigate catastrophic forgetting of adaptation, while (2) preserving the adaptation in-domain quality, (3) with negligible additions to compute costs.
In the broader context, the objectives grounded in a continuous token similarity pioneer the exploration of the middle ground between the efficient but naive exact-match token-level objectives and expressive but computationally- and resource-intensive sequential objectives.