File size: 7,648 Bytes
414ea2f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
011c53a
f0e1cf1
14e8e04
f0e1cf1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ebea69a
f0e1cf1
 
 
ebea69a
eabe01e
 
ebea69a
eabe01e
 
602dd9d
ebea69a
 
 
 
 
 
 
 
 
f0e1cf1
 
 
 
 
ebea69a
f0e1cf1
 
 
73c4767
f0e1cf1
 
 
fbd22b8
 
 
 
 
 
 
 
 
 
 
 
 
f0e1cf1
 
 
 
 
 
7989c5c
bf3b4dc
7989c5c
bf3b4dc
7989c5c
bf3b4dc
7989c5c
bf3b4dc
6ca6aaf
f0e1cf1
6306358
 
 
 
b6e322b
 
6306358
 
b6e322b
 
6306358
 
 
 
b6e322b
 
6306358
 
 
b6e322b
6306358
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e658166
 
 
 
 
 
 
f0e1cf1
 
 
 
25f6c59
f0e1cf1
 
 
85b79f3
f0e1cf1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
853f0df
f0e1cf1
 
 
853f0df
f0e1cf1
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
---
annotations_creators:
- expert-generated
- machine-generated
language_creators:
- crowdsourced
languages:
- de
- en
- fi
- fr
- ru
- sv
licenses:
- cc-by-nc-4.0
multilinguality:
- multilingual
pretty_name: Opusparcus
size_categories:
- unknown
source_datasets:
- extended|open_subtitles
task_categories:
- conditional-text-generation
task_ids:
- conditional-text-generation-other-paraphrase generation
---

            
# Dataset Card for Opusparcus

## Table of Contents

- [Dataset Description](#dataset-description)

  - [Dataset Summary](#dataset-summary)

  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)

  - [Languages](#languages)

- [Dataset Structure](#dataset-structure)

  - [Data Instances](#data-instances)

  - [Data Fields](#data-fields)

  - [Data Splits](#data-splits)

- [Dataset Creation](#dataset-creation)

  - [Curation Rationale](#curation-rationale)

  - [Source Data](#source-data)

  - [Annotations](#annotations)

  - [Personal and Sensitive Information](#personal-and-sensitive-information)

- [Considerations for Using the Data](#considerations-for-using-the-data)

  - [Social Impact of Dataset](#social-impact-of-dataset)

  - [Discussion of Biases](#discussion-of-biases)

  - [Other Known Limitations](#other-known-limitations)

- [Additional Information](#additional-information)

  - [Dataset Curators](#dataset-curators)

  - [Licensing Information](#licensing-information)

  - [Citation Information](#citation-information)

  - [Contributions](#contributions)

## Dataset Description

- **Repository:** [Language Bank of Finland – Metashare](http://urn.fi/urn:nbn:fi:lb-2018021221)

- **Paper:** [Mathias Creutz (2018): Open Subtitles Paraphrases Corpus For Six Languages](http://www.lrec-conf.org/proceedings/lrec2018/summaries/131.html)

- **Point of Contact:** Mathias Creutz (firstname dot lastname at helsinki dot fi)

### Dataset Summary

Opusparcus is a paraphrase corpus for six European languages: German,
English, Finnish, French, Russian, and Swedish. The paraphrases
consist of subtitles from movies and TV shows.

The data in Opusparcus has been extracted from
[OpenSubtitles2016](http://opus.nlpl.eu/OpenSubtitles2016.php), which
is in turn based on data from http://www.opensubtitles.org/.

For each target language, the Opusparcus data have been partitioned
into three types of data sets: training, validation and test sets. The
training sets are large, consisting of millions of sentence pairs, and
have been compiled automatically, with the help of probabilistic
ranking functions. The development and test sets consist of sentence
pairs that have been annotated manually; each set contains
approximately 1000 sentence pairs that have been verified to be
acceptable paraphrases by two indepedent annotators.

### Supported Tasks and Leaderboards

**Tasks:** Paraphrase detection and generation

**Leaderboards:** Currently there is no Leaderboard for this dataset.

### Languages

German (de), English (en), Finnish (fi), French (fr), Russian (ru), Swedish (sv)

## Dataset Structure

When you download Opusparcus, you must always indicate the language
you want to retrieve, for instance:

`data = load_dataset("GEM/opusparcus", lang="de")`

The above command will download the validation and test sets for
German. If additionally, you want to retrieve training data, you need
to specify the level of quality you desire, such as:

`data = load_dataset("GEM/opusparcus", lang="fr", quality=90)`



### Data Instances

TBA

### Data Fields

`sent1`: a tokenized sentence

`sent2`: another tokenized sentence, which is potentially a paraphrase of `sent1`.

`annot_score`: a value between 1.0 and 4.0 indicating how good an example of paraphrases `sent1` and `sent2` are. (For the training sets, the value is 0.0, which indicates that no manual annotation has taken place.)

`lang`: language of this dataset

`gem_id`: unique identifier of this entry

**Additional information about the annotation scheme:**

The annotation scores given by an individual annotator are:

4: Good example of paraphrases (Dark green button in the annotation
tool): The two sentences can be used in the same situation and
essentially "mean the same thing".

3: Mostly good example of paraphrases (Light green button in the
annotation tool): It is acceptable to think that the two sentences
refer to the same thing, although one sentence might be more specific
than the other one, or there are differences in style, such as polite
form versus familiar form.

2: Mostly bad example of paraphrases (Yellow button in the annotation
tool): There is some connection between the sentences that explains
why they occur together, but one would not really consider them to
mean the same thing.

1: Bad example of paraphrases (Red button in the annotation tool):
There is no obvious connection. The sentences mean different things.

If the two annotators fully agreed on the category, the value in the
`annot_score` field is 4.0, 3.0, 2.0 or 1.0.  If the two annotators
chose adjacent categories, the value in this field will be 3.5, 2.5 or
1.5.  For instance, a value of 2.5 means that one annotator gave a
score of 3 ("mostly good"), indicating a possible paraphrase pair,
whereas the other annotator scored this as a 2 ("mostly bad"), that
is, unlikely to be a paraphrase pair.  If the annotators disagreed by
more than one category, the sentence pair was discarded and won't show
up in the datasets.

The training sets were not annotated manually. This is indicated by
the value 0.0 in the `annot_score` field.

For an assessment of of inter-annotator agreement, see Mikko Aulamo,
Mathias Creutz and Eetu Sjöblom (2019). [Annotation of subtitle
paraphrases using a new web
tool.](http://ceur-ws.org/Vol-2364/3_paper.pdf) In Proceedings of the
Digital Humanities in the Nordic Countries 4th Conference],
Copenhagen, Denmark.

### Data Splits

The data is split into ...


|          | Train  | Valid | Test |
| -----    | ------ | ----- | ---- |
| de       | ...    | ...   | ...  |
| en       | ...    | ...   | ...  |


## Dataset Creation

### Curation Rationale

TBA

### Source Data

#### Initial Data Collection and Normalization

TBA

#### Who are the source language producers?

TBA

### Annotations

#### Annotation process

TBA

#### Who are the annotators?

TBA

### Personal and Sensitive Information

[More Information Needed]

## Considerations for Using the Data

### Social Impact of Dataset

[More Information Needed]

### Discussion of Biases

[More Information Needed]

### Other Known Limitations

[More Information Needed]

## Additional Information

### Dataset Curators

TBA

### Licensing Information

[More Information Needed]

### Citation Information

```
@InProceedings{creutz:lrec2018,
  title = {Open Subtitles Paraphrase Corpus for Six Languages},
  author={Mathias Creutz},
  booktitle={Proceedings of the 11th edition of the Language Resources and Evaluation Conference (LREC 2018)},
  year={2018},
  month = {May 7-12},
  address = {Miyazaki, Japan},
  editor = {Nicoletta Calzolari (Conference chair) and Khalid Choukri and Christopher Cieri and Thierry Declerck and Sara Goggi and Koiti Hasida and Hitoshi Isahara and Bente Maegaard and Joseph Mariani and Hélène Mazo and Asuncion Moreno and Jan Odijk and Stelios Piperidis and Takenobu Tokunaga},
  publisher = {European Language Resources Association (ELRA)},
  isbn = {979-10-95546-00-9},
  language = {english},
  url={http://www.lrec-conf.org/proceedings/lrec2018/pdf/131.pdf}
```

### Contributions

Thanks to [@mathiascreutz](https://github.com/mathiascreutz) for adding this dataset.