Datasets:

Languages:
French
Multilinguality:
monolingual
Size Categories:
1K<n<10K
Language Creators:
crowdsourced
found
Annotations Creators:
crowdsourced
Source Datasets:
original
ArXiv:
License:
system HF staff commited on
Commit
bb49125
1 Parent(s): e57941b

Update files from the datasets library (from 1.7.0)

Browse files

Release notes: https://github.com/huggingface/datasets/releases/tag/1.7.0

Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -22,6 +22,7 @@ task_categories:
22
  task_ids:
23
  - extractive-qa
24
  - closed-domain-qa
 
25
  ---
26
 
27
  # Dataset Card for "fquad"
@@ -29,12 +30,12 @@ task_ids:
29
  ## Table of Contents
30
  - [Dataset Description](#dataset-description)
31
  - [Dataset Summary](#dataset-summary)
32
- - [Supported Tasks](#supported-tasks)
33
  - [Languages](#languages)
34
  - [Dataset Structure](#dataset-structure)
35
  - [Data Instances](#data-instances)
36
  - [Data Fields](#data-fields)
37
- - [Data Splits Sample Size](#data-splits-sample-size)
38
  - [Dataset Creation](#dataset-creation)
39
  - [Curation Rationale](#curation-rationale)
40
  - [Source Data](#source-data)
@@ -68,7 +69,7 @@ FQuAD contains 25,000+ question and answer pairs.
68
  Finetuning CamemBERT on FQuAD yields a F1 score of 88% and an exact match of 77.9%.
69
  Developped to provide a SQuAD equivalent in the French language. Questions are original and based on high quality Wikipedia articles.
70
 
71
- ### Supported Tasks
72
 
73
  - `closed-domain-qa`, `text-retrieval`: This dataset is intended to be used for `closed-domain-qa`, but can also be used for information retrieval tasks.
74
 
@@ -113,7 +114,7 @@ The data fields are the same among all splits.
113
  - `texts`: a `string` feature.
114
  - `answers_starts`: a `int32` feature.
115
 
116
- ### Data Splits Sample Size
117
 
118
  The FQuAD dataset has 3 splits: _train_, _validation_, and _test_. The _test_ split is however not released publicly at the moment. The splits contain disjoint sets of articles. The following table contains stats about each split.
119
 
 
22
  task_ids:
23
  - extractive-qa
24
  - closed-domain-qa
25
+ paperswithcode_id: fquad
26
  ---
27
 
28
  # Dataset Card for "fquad"
 
30
  ## Table of Contents
31
  - [Dataset Description](#dataset-description)
32
  - [Dataset Summary](#dataset-summary)
33
+ - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
34
  - [Languages](#languages)
35
  - [Dataset Structure](#dataset-structure)
36
  - [Data Instances](#data-instances)
37
  - [Data Fields](#data-fields)
38
+ - [Data Splits](#data-splits)
39
  - [Dataset Creation](#dataset-creation)
40
  - [Curation Rationale](#curation-rationale)
41
  - [Source Data](#source-data)
 
69
  Finetuning CamemBERT on FQuAD yields a F1 score of 88% and an exact match of 77.9%.
70
  Developped to provide a SQuAD equivalent in the French language. Questions are original and based on high quality Wikipedia articles.
71
 
72
+ ### Supported Tasks and Leaderboards
73
 
74
  - `closed-domain-qa`, `text-retrieval`: This dataset is intended to be used for `closed-domain-qa`, but can also be used for information retrieval tasks.
75
 
 
114
  - `texts`: a `string` feature.
115
  - `answers_starts`: a `int32` feature.
116
 
117
+ ### Data Splits
118
 
119
  The FQuAD dataset has 3 splits: _train_, _validation_, and _test_. The _test_ split is however not released publicly at the moment. The splits contain disjoint sets of articles. The following table contains stats about each split.
120