数据集:
drop
许可:
cc-by-sa-4.0批注创建人:
crowdsourced源数据集:
original语言创建人:
crowdsourced大小:
10K<n<100K计算机处理:
monolingual语言:
enDROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs. . DROP is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a question, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or sorting). These operations require a much more comprehensive understanding of the content of paragraphs than what was necessary for prior datasets.
An example of 'validation' looks as follows.
This example was too long and was cropped: { "answers_spans": { "spans": ["Chaz Schilens"] }, "passage": "\" Hoping to rebound from their loss to the Patriots, the Raiders stayed at home for a Week 16 duel with the Houston Texans. Oak...", "question": "Who scored the first touchdown of the game?" }
The data fields are the same among all splits.
defaultname | train | validation |
---|---|---|
default | 77409 | 9536 |
@inproceedings{Dua2019DROP, author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner}, title={ {DROP}: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs}, booktitle={Proc. of NAACL}, year={2019} }
Thanks to @patrickvonplaten , @thomwolf , @mariamabarham , @lewtun for adding this dataset.