StoryWars: A Dataset and Instruction Tuning Baselines for Collaborative Story Understanding and Generation
Yulun Du, Lydia Chilton
Main: Resources and Evaluation Main-poster Paper
Poster Session 6: Resources and Evaluation (Poster)
Conference Room: Frontenac Ballroom and Queen's Quay
Conference Time: July 12, 09:00-10:30 (EDT) (America/Toronto)
Global Time: July 12, Poster Session 6 (13:00-14:30 UTC)
Keywords:
corpus creation, benchmarking, nlp datasets
TLDR:
Collaborative stories, which are texts created through the collaborative efforts of multiple authors with different writing styles and intentions, pose unique challenges for NLP models.
Understanding and generating such stories remains an underexplored area due to the lack of open-domain corpora.
T...
You can open the
#paper-P4891
channel in a separate window.
Abstract:
Collaborative stories, which are texts created through the collaborative efforts of multiple authors with different writing styles and intentions, pose unique challenges for NLP models.
Understanding and generating such stories remains an underexplored area due to the lack of open-domain corpora.
To address this, we introduce StoryWars, a new dataset of over 40,000 collaborative stories written by 9,400 different authors from an online platform.
We design 12 task types, comprising 7 understanding and 5 generation task types, on \{pasted macro `STORYWARS'\}, deriving 101 diverse story-related tasks in total as a multi-task benchmark covering all fully-supervised, few-shot, and zero-shot scenarios.
Furthermore, we present our instruction-tuned model, InstructStory, for the story tasks showing that instruction tuning, in addition to achieving superior results in zero-shot and few-shot scenarios, can also obtain the best performance on the fully-supervised tasks in StoryWars, establishing strong multi-task benchmark performances on StoryWars.