File size: 7,180 Bytes
04c3591
 
519205b
 
 
 
 
e693e8b
519205b
 
 
 
 
 
 
 
e693e8b
 
 
519205b
 
 
 
 
 
 
371e8dd
 
 
 
 
 
 
 
 
 
 
 
 
04c3591
519205b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
---
license: apache-2.0
annotations_creators:
- expert-generated
language_creators:
- found
task_categories:
- text-classification
language:
- ro
multilinguality:
- monolingual
source_datasets:
- readerbench/ro-offense
tags:
- hate-speech-detection
- offensive speech
- romanian
- nlp
task_ids:
- hate-speech-detection
pretty_name: RO-Offense-Sequences
size_categories:
- 1K<n<10K
extra_gated_prompt: 'Warning: this repository contains harmful content (abusive language,
  hate speech).'
configs:
- config_name: default
  data_files:
  - split: train
    path: "train.csv"
  - split: test
    path: "test.csv"
- config_name: ner
  data_files:
  - split: train
    path: "train_ner.csv"
  - split: test
    path: "test_ner.csv"
---

# Dataset Card for "RO-Offense-Sequences"

## Table of Contents
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
  - [Languages](#languages)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-fields)
  - [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
  - [Curation Rationale](#curation-rationale)
  - [Source Data](#source-data)
  - [Annotations](#annotations)
  - [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Social Impact of Dataset](#social-impact-of-dataset)
  - [Discussion of Biases](#discussion-of-biases)
  - [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
  - [Dataset Curators](#dataset-curators)
  - [Licensing Information](#licensing-information)
  - [Citation Information](#citation-information)
  - [Contributions](#contributions)

## Dataset Description
<!--
- **Paper:** News-RO-Offense - A Romanian Offensive Language Dataset and Baseline Models Centered on News Article Comments
-->
- **Homepage:** [https://github.com/readerbench/ro-offense-sequences](https://github.com/readerbench/ro-offense-sequences)
- **Repository:** [https://github.com/readerbench/ro-offense-sequences](https://github.com/readerbench/ro-offense-sequences)
- **Point of Contact:** [Andrei Paraschiv](https://github.com/AndyTheFactory)
- 
### Dataset Summary
a novel Romanian language dataset for offensive language detection with manually 
annotated offensive labels from a local Romanian sports news website (gsp.ro):


Resulting in 12,445 annotated messages

### Languages

Romanian

## Dataset Structure


### Data Instances


An example of 'train' looks as follows.

```
{
  'id': 5,
  'text':'PLACEHOLDER TEXT',
  'label': 'OTHER'
}
```


### Data Fields

- `id`: The unique comment ID, corresponding to the ID in [RO Offense](https://huggingface.co/datasets/readerbench/ro-offense)
- `text`: full comment text
- `label`: the type of offensive message (OTHER, PROFANITY, INSULT, ABUSE)

### Data Splits


Train | Other | Profanity | Insult | Abuse
:---| :---| :---| :---| :---:
9953 | 3656 | 1293 | 2236 | 2768

Test | Other | Profanity | Insult | Abuse
:---| :---| :---| :---| :---:
2492 | 916 | 324 | 559 | 693


## Dataset Creation


### Curation Rationale

Collecting data for abusive language classification for Romanian Language.

For the labeling of texts we loosely base our definitions on the Germeval 2019 task for detecting offensive language in german tweets  (Struß et al., 2019)

Data source: Comments on articles in Gazeta Sporturilor (gsp.ro) between 2011 and 2020 

Selection for annotation: we select comments from a pool of secific articles based on the number of comments in the article. 
The number of comments per article has the following distribution:
```
mean        183.820923
std         334.707177
min           1.000000
25%          20.000000
50%          58.000000
75%         179.000000
max        2151.000000
```

Based on this we select only comments from articles having between 20 and 50 comments. Also, we remove comments containing urls or three consecutive *, since these were mostly censored by editors or automatic profanity detection algorythms.

Additional, in order to have some meaningful messages for annotation, we select only messages with length between 50 and 500 characters. 


### Source Data

Sports News Articles comments


#### Initial Data Collection and Normalization



#### Who are the source language producers?

Sports News Article readers


### Annotations

- Andrei Paraschiv
- Irina Maria Sandu

#### Annotation process


##### OTHER
 
Label used for non offensive texts. 

##### PROFANITY

This is the "lighter" form of abusive language. When profane words are used without a direct intend on offending a target, or without ascribing some negative qualities to a target we use this label. Some messages in this class may even have a positive sentiment and uses swearwords as emphasis. Messages containing profane words that are not directed towards a specific group or person, we label as **PROFANITY**

Also, self censored messages with swear words having some letters hidden, or some deceitful misspellings of swearwords that have clear intend on circumventing profanity detectors will be treated as **PROFANITY**. 


##### INSULT

The message clearly wants to offend someone, ascribing negatively evaluated qualities or deficiences, labeling a person or a group of persons as unworthy or unvalued. Insults do imply disrespect and contempt directed towards a target.


##### ABUSE

This label marks messages containing the stronger form of offensive and abusive language. This type of language ascribes the target a social identity that is judged negatively by the majority of society, or at least is percieved as a mostly negative judged identity. Shameful, unworthy or morally unaceptable identytities fall in this category. In contrast to insults, instances of abusive language require that the target of judgment is seen as a representative of a group and it is ascribed negative qualities that are taken to be universal, omnipresent and unchangeable characteristics of the group. 

In contrast to insults, instances of abusive language require that the target of judgment tis seen as a representative of a group and it is ascribed negative qualities that are taken to be universal, omnipresent and unchangeable characteristics of the group.

Additional, dehumanizing language targeting a person or group is also classified as ABUSE. 

#### Who are the annotators?

Native speakers

### Personal and Sensitive Information

The data was public at the time of collection. PII removal has been performed.

## Considerations for Using the Data



### Social Impact of Dataset

The data definitely contains abusive language. The data could be used to develop and propagate offensive language against every target group involved, i.e. ableism, racism, sexism, ageism, and so on.

### Discussion of Biases


### Other Known Limitations


## Additional Information


### Dataset Curators


### Licensing Information

This data is available and distributed under Apache-2.0 license

### Citation Information

```
tbd
```


### Contributions