File size: 8,356 Bytes
866292d
fe4585a
 
866292d
fe4585a
 
 
 
 
 
 
 
507fd30
fe4585a
 
507fd30
fe4585a
 
4df1e52
 
 
c58df52
4df1e52
c58df52
 
 
 
 
 
 
4c5729a
c58df52
 
4df1e52
4c5729a
866292d
d465cb3
fe4585a
d45dfae
fe4585a
 
 
 
 
 
 
 
 
 
 
 
 
d81520b
fe4585a
 
 
 
 
 
 
 
 
 
 
a2ddd22
 
 
 
 
 
 
 
 
 
880ef3c
a2ddd22
 
 
 
 
 
 
cd6db45
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a2ddd22
cd6db45
 
 
 
 
a8ae249
cd6db45
 
fe4585a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
---
annotations_creators:
- crowdsourced
license: mit
multilinguality:
- monolingual
source_datasets:
- original
task_ids:
- hate-speech-detection
task_categories:
- text-classification
- token-classification
language:
- vi
pretty_name: ViHOS - Vietnamese Hate and Offensive Spans Dataset
size_categories:
- 10K<n<100K
configs:
- config_name: default
  data_files:
  - split: train_sequence_labeling
    path:
    - "train_sequence_labeling/syllable/train_BIO_syllable.csv"
    - "train_sequence_labeling/syllable/dev_BIO_syllable.csv"
    - "train_sequence_labeling/syllable/test_BIO_syllable.csv"
    - "train_sequence_labeling/word/train_BIO_syllable.csv"
    - "train_sequence_labeling/word/dev_BIO_syllable.csv"
    - "train_sequence_labeling/word/test_BIO_syllable.csv"
  - split: train_span_extraction
    path:
    - 'train_span_extraction/train.csv'
    - 'train_span_extraction/dev.csv'
  - split: test
    path: "test/test.csv"
---
**Disclaimer**: This project contains real comments that could be considered profane, offensive, or abusive.

# Dataset Card for "ViHOS - Vietnamese Hate and Offensive Spans Dataset"

## Dataset Description
- **Repository:** [ViHOS](https://github.com/phusroyal/ViHOS)
- **Paper:** [EACL-ViHOS](https://aclanthology.org/2023.eacl-main.47/)
- **Total amount of disk used:** 2.6 MB

## Dataset Motivation
The rise in hateful and offensive language directed at other users is one of the adverse side effects of the increased use of social networking platforms. This could make it difficult for human moderators to review tagged comments filtered by classification systems.

To help address this issue, we present the ViHOS (**Vi**etnamese **H**ate and **O**ffensive **S**pans) dataset, the first human-annotated corpus containing 26k spans on 11k online comments.

Our goal is to create a dataset that contains comprehensive hate and offensive thoughts, meanings, or opinions within the comments rather than just a lexicon of hate and offensive terms.

We also provide definitions of hateful and offensive spans in Vietnamese comments as well as detailed annotation guidelines. Futhermore, our solutions to deal with *nine different online foul linguistic phenomena* are also provided in the [*paper*](https://aclanthology.org/2023.eacl-main.47/) (e.g. Teencodes; Metaphors, metonymies; Hyponyms; Puns...).

We hope that this dataset will be useful for researchers and practitioners in the field of hate speech detection in general and hate spans detection in particular.

## Dataset Summary
ViHOS contains 26,476 human-annotated spans on 11,056 comments (5,360 comments have hate and offensive spans, and 5,696 comments do not)

It is splitted into train, dev, and test set with following information:
1. Train set: 8,844 comments
2. Dev set: 1,106 comments
3. Test set: 1,106 comments

## Data Instance

An span extraction-based (see Data Structure for more details) example of 'test' looks as follows:
```
{
  "content": "Thối CC chỉ không ngửi đuợc thôi",
  'index_spans': "[0, 1, 2, 3, 5, 6]"
}
```

An sequence labeling-based (see Data Structure for more details) example of 'test' looks as follows:
```
{
  "content": "Thối CC chỉ không ngửi đuợc thôi",
  'index_spans': ["B-T", "I-T", "O", "O", "O", "O", "O"]
}
```

## Data Structure
Here is our data folder structure!
```
.
└── data/
    ├── train_sequence_labeling/
    │   ├── syllable/
    │   │   ├── dev_BIO_syllable.csv
    │   │   ├── test_BIO_syllable.csv
    │   │   └── train_BIO_syllable.csv
    │   └── word/
    │       ├── dev_BIO_Word.csv
    │       ├── test_BIO_Word.csv
    │       └── train_BIO_Word.csv
    ├── train_span_extraction/
    │   ├── dev.csv
    │   └── train.csv
    └── test/
        └── test.csv
```
### Sequence labeling-based version
#### Syllable
Description: 
- This folder contains the data for the sequence labeling-based version of the task. The data is divided into two files: train, and dev. Each file contains the following columns:
  - **index**: The id of the word.
  - **word**: Words in the sentence after the processing of tokenization using [VnCoreNLP](https://github.com/vncorenlp/VnCoreNLP) tokenizer followed by underscore tokenization.
  The reason for this is that some words are in bad format:
  e.g. "điện.thoại của tôi" is split into ["điện.thoại", "của", "tôi"] instead of ["điện", "thoại", "của", "tôi"] if we use space tokenization, which is not in the right format of Syllable.
  As that, we used VnCoreNLP to tokenize first and then split words into tokens.
  e.g. "điện.thoại của tôi" ---(VnCoreNLP)---> ["điện_thoại", "của", "tôi"] ---(split by "_")---> ["điện", "thoại", "của", "tôi"].
  - **tag**: The tag of the word. The tag is either B-T (beginning of a word), I-T (inside of a word), or O (outside of a word).
- The train_BIO_syllable and dev_BIO_syllable file are used for training and validation for XLMR model, respectively.
- The test_BIO_syllable file is used for reference only. It is not used for testing the model. **Please use the test.csv file in the Testdata folder for testing the model.**
#### Word
Description: 
- This folder contains the data for the sequence labeling-based version of the task. The data is divided into two files: train, and dev. Each file contains the following columns:
  - **index**: The id of the word.
  - **word**: Words in the sentence after the processing of tokenization using [VnCoreNLP](https://github.com/vncorenlp/VnCoreNLP) tokenizer
  - **tag**: The tag of the word. The tag is either B-T (beginning of a word), I-T (inside of a word), or O (outside of a word).
- The train_BIO_Word and dev_BIO_Word file are used for training and validation for PhoBERT model, respectively.
- The test_BIO_Word file is used for reference only. It is not used for testing the model. **Please use the test.csv file in the data/test folder for testing the model.**

### Span Extraction-based version
Description:
- This folder contains the data for the span extraction-based version of the task. The data is divided into two files: train and dev. Each file contains the following columns:
  - **content**: The content of the sentence.
  - **span_ids**: The index of the hate and offensive spans in the sentence. The index is in the format of [start, end] where start is the index of the first character of the hate and offensive span and end is the index of the last character of the hate and offensive span.
- The train and dev file are used for training and validation for BiLSTM-CRF model, respectively.

### Citation Information
```
@inproceedings{hoang-etal-2023-vihos,
    title = "{V}i{HOS}: Hate Speech Spans Detection for {V}ietnamese",
    author = "Hoang, Phu Gia  and
      Luu, Canh Duc  and
      Tran, Khanh Quoc  and
      Nguyen, Kiet Van  and
      Nguyen, Ngan Luu-Thuy",
    booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics",
    month = may,
    year = "2023",
    address = "Dubrovnik, Croatia",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.eacl-main.47",
    doi = "10.18653/v1/2023.eacl-main.47",
    pages = "652--669",
    abstract = "The rise in hateful and offensive language directed at other users is one of the adverse side effects of the increased use of social networking platforms. This could make it difficult for human moderators to review tagged comments filtered by classification systems. To help address this issue, we present the ViHOS (Vietnamese Hate and Offensive Spans) dataset, the first human-annotated corpus containing 26k spans on 11k comments. We also provide definitions of hateful and offensive spans in Vietnamese comments as well as detailed annotation guidelines. Besides, we conduct experiments with various state-of-the-art models. Specifically, XLM-R{\_}Large achieved the best F1-scores in Single span detection and All spans detection, while PhoBERT{\_}Large obtained the highest in Multiple spans detection. Finally, our error analysis demonstrates the difficulties in detecting specific types of spans in our data for future research. Our dataset is released on GitHub.",
}
```