ELI5: Long form question answering

A Fan, Y Jernite, E Perez, D Grangier, J Weston… - arXiv preprint arXiv …, 2019 - arxiv.org
arXiv preprint arXiv:1907.09190, 2019arxiv.org
We introduce the first large-scale corpus for long-form question answering, a task requiring
elaborate and in-depth answers to open-ended questions. The dataset comprises 270K
threads from the Reddit forum``Explain Like I'm Five''(ELI5) where an online community
provides answers to questions which are comprehensible by five year olds. Compared to
existing datasets, ELI5 comprises diverse questions requiring multi-sentence answers. We
provide a large set of web documents to help answer the question. Automatic and human …
We introduce the first large-scale corpus for long-form question answering, a task requiring elaborate and in-depth answers to open-ended questions. The dataset comprises 270K threads from the Reddit forum ``Explain Like I'm Five'' (ELI5) where an online community provides answers to questions which are comprehensible by five year olds. Compared to existing datasets, ELI5 comprises diverse questions requiring multi-sentence answers. We provide a large set of web documents to help answer the question. Automatic and human evaluations show that an abstractive model trained with a multi-task objective outperforms conventional Seq2Seq, language modeling, as well as a strong extractive baseline. However, our best model is still far from human performance since raters prefer gold responses in over 86% of cases, leaving ample opportunity for future improvement.
arxiv.org