Skip to content

Latest commit

 

History

History
45 lines (33 loc) · 2.55 KB

question_answering.md

File metadata and controls

45 lines (33 loc) · 2.55 KB

Question answering

Question answering is the task of answering a question.

Table of contents

Reading comprehension

CMRC 2018

The Chinese Machine Reading Comprehension (CMRC 2018) is a SQuAD-like reading comprehension dataset that consists of 20,000 questions annotated on Wikipedia paragraphs by human experts. The dataset can be downloaded here. Below we show the F1 and EM scores both on the test set and the challenge set.

| Model | Test F1 | Test EM | Challenge F1 | Challenge EM | Paper | | ------------- | :-----:| :-----:| --- | | Human performance | 97.9 | 92.4 | 95.2 | 90.4 | A Span-Extraction Dataset for Chinese Machine Reading Comprehension | | Dual BERT (w / SQuAD; Cui et al., 2019) | 90.2 | 73.6 | 55.2 | 27.8 | Cross-Lingual Machine Reading Comprehension | | Dual BERT (Cui et al., 2019) | 88.1 | 70.4 | 47.9 | 23.8 | Cross-Lingual Machine Reading Comprehension |

DRCD

The Delta Reading Comprehension Dataset (DRCD) is a SQuAD-like reading comprehension dataset that contains 30,000+ questions on 10,014 paragraphs from 2,108 Wikipedia articles. The dataset can be downloaded here.

Model F1 EM Paper
Human performance 93.3 80.4 DRCD: a Chinese Machine Reading Comprehension Dataset
Dual BERT (w / SQuAD; Cui et al., 2019) 91.6 85.4 Cross-Lingual Machine Reading Comprehension
Dual BERT (Cui et al., 2019) 90.3 83.7 Cross-Lingual Machine Reading Comprehension

DuReader

DuReader is a large-scale reading comprehension dataset that is based on the logs of Baidu Search and contains 200k questions, 420k answers, and 1M documents. For more information, refer to its website to see the introduction. You can download the dataset here. The best models can be view on the public leaderboard.