Title: A Corpus and Evaluation Framework for Deeper Understanding of Commonsense Stories
Abstract: https://arxiv.org/abs/1604.01696
Homepage: https://cs.rochester.edu/nlp/rocstories/
'Story Cloze Test' is a new commonsense reasoning framework for evaluating story understanding, story generation, and script learning. This test requires a system to choose the correct ending to a four-sentence story
@misc{mostafazadeh2016corpus,
title={A Corpus and Evaluation Framework for Deeper Understanding of Commonsense Stories},
author={Nasrin Mostafazadeh and
Nathanael Chambers and
Xiaodong He and
Devi Parikh and
Dhruv Batra and
Lucy Vanderwende and
Pushmeet Kohli and
James Allen},
year={2016},
eprint={1604.01696},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
storycloze
storycloze_2016
storycloze_2018
For adding novel benchmarks/datasets to the library:
- Is the task an existing benchmark in the literature?
- Have you referenced the original paper that introduced the task?
- If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?
If other tasks on this dataset are already supported:
- Is the "Main" variant of this task clearly denoted?
- Have you provided a short sentence in a README on what each new variant adds / evaluates?
- Have you noted which, if any, published evaluation setups are matched by this variant?