File size: 6,887 Bytes
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e6a2f54
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
 
6434a48
 
 
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a921e3e
 
 
d5ffdd6
 
 
e7378af
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
 
6434a48
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
6434a48
 
d5ffdd6
 
 
e6a2f54
d5ffdd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
---
license: cc-by-4.0
task_categories:
- audio-classification
language:
- en
tags:
- speech
- speech-classifiation
- text-to-speech
- spoofing
- accents

pretty_name: ARCTIC-HS
size_categories:
- 10K<n<100K
---

# ARCTIC-HS

An extension of the [CMU_ARCTIC](http://festvox.org/cmu_arctic/) and [L2-ARCTIC](https://psi.engr.tamu.edu/l2-arctic-corpus/) datasets for synthetic speech detection using text-to-speech, featured in the paper **Synthetic speech detection with Wav2Vec 2.0 in various language settings**. Specifically, the `symmetric` variants were used.

This dataset is 1 of 3 used in the paper, the others being:
- [FLEURS-HS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs)
  - the default train, dev and test sets
- [FLEURS-HS VITS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs-vits)
  - test set containing (generally) more difficult synthetic samples
  - separated due to different licensing

## Dataset Details

### Dataset Description

The dataset features 3 parts obtained from the 2 original datasets:

- CMU (native) non-US English speakers
- CMU (native) US English speakers
- L2 (non-native) English speakers


The original ARCTIC samples are used as `human` samples, while `synthetic` samples are generated using [Google Cloud Text-To-Speech](https://cloud.google.com/text-to-speech).

The resulting `symmetric` datasets features exactly twice the samples of the original ones, but we also provide:

- human samples that couldn't be paired
  - 4 speakers in entirety we couldn't pair with a TTS voice
  - a small amount of utterances unrelated to the A and B ARCTIC samples
- synthetic samples that couldn't be paired
  - mostly when a human speaker didn't read the B ARCTIC samples


- **Curated by:** [KONTXT by RealNetworks](https://realnetworks.com/kontxt)
- **Funded by:** [RealNetworks](https://realnetworks.com/)
- **Language(s) (NLP):** English
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) for the code, [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) for the dataset, however:
  - the human part of the dataset is under a **custom CMU license**
    - it should be compatible with **CC BY 4.0**
  - the human part of the L2 dataset is under **CC BY-NC 4.0**

### Dataset Sources

The original ARCTIC sets were downloaded from their original sources.

- **CMU_ARCTIC Repository:** [festvox.org](http://festvox.org/cmu_arctic/)
- **L2-ARCTIC Repository:** [tamu.edu](https://psi.engr.tamu.edu/l2-arctic-corpus/)
- **CMU_ARCTIC Paper:** [cmu.edu](https://www.cs.cmu.edu/~awb/papers/ssw5/arctic.pdf)
- **L2-ARCTIC Paper:** [tamu.edu](https://psi.engr.tamu.edu/wp-content/uploads/2018/08/zhao2018interspeech.pdf)

- **Paper:** Synthetic speech detection with Wav2Vec 2.0 in various language settings

## Uses

This dataset is best used as a test set for accents. Each sample contains an `Audio` feature, and a label: `human` or `synthetic`.

### Direct Use

The following snippet of code demonstrates loading the CMU non-US English speaker part of the dataset:

```python
from datasets import load_dataset

arctic_hs = load_dataset(
    "realnetworks-kontxt/arctic-hs",
    "cmu_non-us",
    split="test",
    trust_remote_code=True,
)
```

To load a different part, change `cmu_non-us` into one of the following:
- `cmu_us` for CMU (native) US English speakers
- `l2` for L2 (non-native) English speakers

This dataset only has a `test` split.

To load only the paired samples, append `_symmetric` to the name. For example, `cmu_non-us` will load the test set also containing human and synthetic samples without their counterpart, while `cmu_non-us_symmetric` will only load samples where there is both a human and synthetic variant. This is useful if you want to have perfectly balanced labels within speakers, and if you wish to exclude speakers for which there are no TTS counterparts at all. This is also the family of datasets used in the paper.

The `trust_remote_code=True` parameter is necessary because this dataset uses a custom loader. To check out which code is being ran, check out the [loading script](./arctic-hs.py).

## Dataset Structure

The dataset data is contained in the [data directory](https://huggingface.co/datasets/realnetworks-kontxt/arctic-hs/tree/main/data).

There exists 1 directory per part.

Within those directories, there are 2 further directories:
- `splits`
- `pairs`

Within the `splits` folder, there is 1 file per split:
- `test.tar.gz`

Those `.tar.gz` files contain 2 directories:
- `human`
- `synthetic`

Each of these directories contain `.wav` files. Keep in mind that these directories can't be merged as they share most of their file names. An identical file name implies a speaker-voice pair, ex. `human/arctic_a0001.wav` and `synthetic/arctic_a0001.wav`.


The `pairs` folder contains a list of file names within each speaker, and whether or not there is a human-synthetic pair. Based on that metadata we determine which samples appear in `symmetric` datasets.

Back to the part directories, each contain 2 metadata files, which are not used in the loaded dataset, but might be useful to researchers:
- `speaker-metadata.csv`
  - contains the speaker IDs paired with their speech properties
- `voice-metadata.csv`
  - contains speaker-TTS name pairs

Finally, the `data` root contains a single metadata file, `prompts.csv`, which as the name would suggest, contains the prompt transcripts. The only samples for which there are no transcripts are the ARCTIC-C ones, for which we couldn't find a source in the internet.


### Sample

A sample contains contains an Audio feature `audio`, and a string `label`.

```
{
  'audio': {
    'path': 'ahw/human/arctic_a0001.wav',
    'array': array([0., 0., 0., ..., 0., 0., 0.]),
    'sampling_rate': 16000
  },
  'label': 'human'
}
```

## Citation

The dataset is featured alongside our paper, **Synthetic speech detection with Wav2Vec 2.0 in various language settings**, which will be published on IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). We'll provide links once it's available online.

**BibTeX:**

Note, the following BibTeX is incomplete - we'll update it once the actual one is known.

```
@inproceedings{dropuljic-ssdww2v2ivls
  author={Dropuljić, Branimir and Šuflaj, Miljenko and Jertec, Andrej and Obadić, Leo}
  booktitle={2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)}
  title={Synthetic speech detection with Wav2Vec 2.0 in various language settings}
  year={2024}
  volume={}
  number={}
  pages={1-5}
  keywords={Synthetic speech detection;text-to-speech;wav2vec 2.0;spoofing attack;multilingualism}
  doi={}
}
```

## Dataset Card Authors

- [Miljenko Šuflaj](https://huggingface.co/suflaj)

## Dataset Card Contact

- [Miljenko Šuflaj](mailto:msuflaj@realnetworks.com)