SN-echoes / README.md
SushantGautam's picture
Upload folder using huggingface_hub
cd31d69 verified
|
raw
history blame
No virus
7.64 kB
---
configs:
- config_name: whisper_v1
data_files:
- split: original
path: whisper_v1/**
- split: en
path: whisper_v1_en/**
- config_name: whisper_v2
data_files:
- split: original
path: whisper_v2/**
- split: en
path: whisper_v3_en/**
- config_name: whisper_v3
data_files:
- split: original
path: whisper_v3/**
- split: en
path: whisper_v3_en/**
task_categories:
- summarization
- translation
license: cc-by-4.0
size_categories:
- 10M<n<100M
language:
- en
- es
- ru
- de
- fr
- tr
- it
- pl
- bs
- hu
tags:
- SoccerNet
- synthetic
---
[**[Paper]**](https://arxiv.org/abs/2405.07354) | [**[GitHub]**](https://github.com/SoccerNet/SN-Echoes)
# Dataset Card for SoccerNet-Echoes
This dataset card aims to provide comprehensive details for the SoccerNet-Echoes dataset, an audio commentary dataset for soccer games.
## Dataset Details
### Dataset Description
SoccerNet-Echoes is an audio commentary dataset for soccer games, curated by SimulaMet under the AI-Storyteller project. It is funded by the Research Council of Norway (project number 346671) and shared by the SoccerNet team. The dataset supports multiple languages, including English, Spanish, Russian, German, French, Turkish, Italian, Polish, Bosnian, and Hungarian, and is licensed under CC BY 4.0.
- **Curated by:** [SimulaMet](https://www.simulamet.no), HOST Department (AI-Storyteller project)
- **Funded by:** Research Council of Norway, project number 346671
- **Shared by:** [SoccerNet Team](https://www.soccer-net.org)
- **Language(s) (NLP):** English, Spanish, Russian, German, French, Turkish, Italian, Polish, Bosnian, Hungarian
- **License:** CC BY 4.0
### Dataset Sources
- **Homepage:** [GitHub - SoccerNet-Echoes](https://github.com/SoccerNet/SN-Echoes). **Please check for more information, codes, and updates.**
- **Paper:** [SoccerNet-Echoes: A Soccer Game Audio Commentary Dataset](https://arxiv.org/abs/2405.07354)
- **Demo:** None
## Uses
### Direct Use
The dataset is primarily intended for:
- **Multimodal Event Detection:** Combining audio cues with visual data for improved event detection in sports videos.
- **Game Summarization:** Using automatic speech recognition (ASR) transcriptions to aid in summarizing soccer games.
### Out-of-Scope Use
The dataset is not intended for:
- **Medical Diagnosis:** The dataset is not suitable for medical applications.
- **Non-Sporting Event Analysis:** The dataset is tailored for soccer and might not generalize well to other types of events without further modification.
## Dataset Structure
The dataset comprises transcriptions of soccer game commentaries using various Whisper ASR models and translations of non-English commentaries into English using Google Translate.
This table structure in HuggingFace has five columns for segment index, start and end times, the text (either transcribed or translated), and the game path (represented as a string).
This is divided into three subsets (v1, v2, and v3 of Whisper versions), with each subset further split into "original" (ASR-generated) and "en" (English translated).
The dataset structure in a hierarchical directory and JSON format, following other SoccerNet data resources can be found at [https://github.com/SoccerNet/sn-echoes](https://github.com/SoccerNet/sn-echoes).
Please note that this HuggingFace dataset is mirrored from the Dataset folder in GitHub with a conversion script.
## Dataset Creation
### Curation Rationale
The dataset was curated to enhance the SoccerNet dataset with automatic speech recognition (ASR) transcriptions and translations of non-English commentaries into English using Google Translate, enabling a richer and more integrated understanding of soccer games.
### Source Data
#### Data Collection and Processing
Audios were collected from soccer game broadcast videos in the SoccerNet dataset. The audio was transcribed using multiple Whisper ASR models (large-v1, large-v2, and large-v3) to create a comprehensive transcription dataset.
Google Translate was used for the translations of non-English commentaries into English.
#### Who are the source data producers?
The source data producers are soccer game broadcasters and commentators.
### Annotations
#### Annotation Process
The transcriptions were automatically generated by Whisper ASR models. Google Translate was used for the translations of non-English commentaries into English.
Human verification and corrections of transcriptions are planned for future work.
#### Who are the annotators?
- Whisper ASR models (transcriptions)
- Google Translate (translations)
- Authors/Humans (for verifying game halves laking game audio or commentary)
#### Personal and Sensitive Information
The dataset contains publicly available soccer game commentary, which is not considered sensitive. It does not include personal data about individuals outside of the context of the game.
## Bias, Risks, and Limitations
- **Transcription Accuracy:** ASR models may introduce errors in transcription.
- **Hallucinations:** Repetition of phrases, especially in noisy environments, can degrade transcription quality.
- **Audio Quality:** Variability in audio quality can impact transcription accuracy.
- **Human Verification:** Lack of human-verified annotations in the current dataset.
### Recommendations
Users should be aware of potential biases and limitations, such as transcription errors and hallucinations. Advanced audio pre-processing and human verification can help mitigate these issues.
### Filtering Hallucinations
Users should be aware of transcription errors and hallucinations. The most occurring problem is the unwarranted repetition of phrases and words, especially with audio inputs lacking human speech, being excessively noisy, or containing music. These conditions challenge the models’ transcription accuracy but can be mitigated by a simple filtering approach: removing consecutive entries with the same text and keeping only the first occurrence of each unique text.
It is strongly advised to use consecutive entries filtering along with the Mixed Selection approach to get a better ASR for downstream applications.
Please refer to the related section on [GitHub](https://github.com/SoccerNet/SN-Echoes) to see our suggested way of accomplishing this.
## Citation
**BibTeX:**
```bibtex
@article{gautam2024soccernet,
author = {Gautam, Sushant and Sarkhoosh, Mehdi Houshmand and Held, Jan and Midoglu, Cise and Cioppa, Anthony and Giancola, Silvio and Thambawita, Vajira and Riegler, Michael A. and Halvorsen, P{\aa}l and Shah, Mubarak},
title = {{SoccerNet-Echoes: A Soccer Game Audio Commentary Dataset}},
journal = {arXiv},
year = {2024},
month = may,
eprint = {2405.07354},
doi = {10.48550/arXiv.2405.07354}
}
```
**APA:**
Gautam, S., Sarkhoosh, M. H., Held, J., Midoglu, C., Cioppa, A., Giancola, S., ...Shah, M. (2024). SoccerNet-Echoes: A Soccer Game Audio Commentary Dataset. arXiv, 2405.07354. Retrieved from https://arxiv.org/abs/2405.07354v1
## Glossary
- **ASR (Automatic Speech Recognition):** Technology that converts spoken language into text.
- **Multimodal Analysis:** Combining multiple types of data, such as audio and visual, for more comprehensive analysis.
- **Whisper ASR Models:** A set of automatic speech recognition models developed by OpenAI.
## More Information
For additional details, visit the [SoccerNet-echoes GitHub repository](https://github.com/SoccerNet/SN-echoes) or contact the authors of the dataset.
## Dataset Card Authors
- Sushant Gautam, sushant@simula.no
## Dataset Card Contact
- Sushant Gautam, sushant@simula.no