--- task_categories: - text-classification language: - es size_categories: - n<1K pretty_name: Detecting toxic and healthy adolescent relationships --- # Dataset Card for Dataset toxic-teenage-relationships ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** mmartinevqh@alumnos.unex.es ### Dataset Summary This dataset is prototypes collected by Spanish adolescents (4 girls and 4 boys) aged 15-19 years with previous training on toxic relationships. For 2 weeks, this group of people analyzed phrases that had occurred in their environment or that they produced themselves, classifying them as toxic or healthy and collecting them through a form. ### Supported Tasks and Leaderboards This dataset supported text-classification ### Languages The sentences are in Spanish. ## Dataset Structure ### Data Instances A data point consists of a comment followed by a label that is associated with it. {'label': 0,'text': 'Sample comment text', } ### Data Fields label: value of 0(non-toxic) or 1(toxic) classifying the comment text: the text of the comment ### Data Splits The data is split into a training and testing set. ## Dataset Creation ### Curation Rationale The dataset was created to help in efforts to identify and curb instances of toxicity between teenagers. ### Source Data #### Initial Data Collection and Normalization This dataset is prototypes collected by me thanks to my group of students (4 girls and 4 boys) aged 15-19 with previous training on toxic relationships. For 2 weeks, this group of teenagers analysed phrases that had occurred in their environment (social media, direct communication) or that they themselves produced, classifying them as toxic or healthy and collecting them through a form. Afterwards, the examples given by each student were discussed and evaluated by the others, using peer evaluation. The classification was also ratified by two specialists in the field. ### Personal and Sensitive Information No personal or sensitive information have been scored in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases If words associated with swearing, insults or profanity appear in a comment, it is likely to be classified as toxic, regardless of the author's tone or intention, e.g. humorous/self-critical. This could present some bias towards already vulnerable minority groups. ### Licensing Information Creative Commons Attribution-NonCommercial-ShareAlike (CC-BY-NC-SA)