HANNA Benchmark

HANNA is a large annotated dataset of Human-ANnotated NArratives for ASG evaluation.

It was introduced in our paper “Of Human Criteria and Automatic Metrics: A Benchmark of the Evaluation of Story Generation”, accepted in COLING 2022.

The GitHub repository is accessible here.