File size: 2,673 Bytes
942f436
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
---
{Juan Alberto López Cavallotti, Jan 6, 2023}
---

# Dataset Card for Multilingual Grammar Error Correction

## Dataset Description

- **Homepage:** https://juancavallotti.com
- **Repository:** 
- **Paper:** 
- **Leaderboard:** 
- **Point of Contact:** Juan Alberto López Cavallotti

### Dataset Summary

This dataset can be used to train a transformer model (we used T5) to correct grammar errors in simple sentences written in English, Spanish, French, or German. 
This dataset was developed as a component for the [Squidigies](https://squidgies.app/) platform.

### Supported Tasks and Leaderboards

* **Grammar Error Correction:** By appending the prefix *fix grammar:* to the prrompt.
* **Language Detection:** By appending the prefix: *language:* to the prompt.

### Languages

* English
* Spanish
* French
* German

## Dataset Structure

### Data Instances

The dataset contains the following instances for each language:
* German 32282 sentences.
* English 51393 sentences.
* Spanish 67672 sentences.
* French 67157 sentences.

### Data Fields

* `lang`: The language of the sentence
* `sentence`: The original sentence. 
* `modified`: The corrupted sentence.
* `transformation`: The primary transformation used by the synthetic data generator.
* `sec_transformation`: The secondary transformation (if any) used by the synthetic data generator.

### Data Splits

* `train`: There isn't a specific split defined. I recommend using 1k sentences sampled randomly from each language, combined with the SacreBleu metric.

## Dataset Creation

### Curation Rationale

This dataset was generated synthetically through code with the help of information of common grammar errors harvested throughout the internet.

### Source Data

#### Initial Data Collection and Normalization

The source grammatical sentences come from various open-source datasets, such as Tatoeba.

#### Who are the source language producers?

* Juan Alberto López Cavallotti

### Annotations

#### Annotation process

The annotation is automatic and produced by the generation script.

#### Who are the annotators?

* Data generation script by Juan Alberto López Cavallotti

### Other Known Limitations

The dataset doesn't cover all the possible grammar errors but serves as a starting point that generates fair results.

## Additional Information

### Dataset Curators

[More Information Needed]

### Licensing Information

This dataset is distributed under the [Apache 2 License](https://www.apache.org/licenses/LICENSE-2.0)

### Citation Information

Please mention this original dataset and the author **Juan Alberto López Cavallotti**

### Contributions

* Juan Alberto López Cavallotti