File size: 20,177 Bytes
281aef9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
de3dd6c
281aef9
 
 
 
 
 
 
 
 
 
a294470
281aef9
 
 
 
 
 
 
 
 
 
 
 
 
7efbe93
84a777c
7efbe93
281aef9
 
7efbe93
281aef9
 
 
a68df04
281aef9
a68df04
 
 
 
de3dd6c
a68df04
 
 
 
 
 
 
281aef9
0fd274d
a294470
a68df04
 
2453d50
7a5b9d8
2453d50
 
1c2e9e9
2453d50
a68df04
7a5b9d8
a68df04
 
 
281aef9
 
a68df04
7a5b9d8
 
281aef9
 
a68df04
7ea897c
1d77548
a68df04
 
 
edaa503
a68df04
281aef9
7a5b9d8
a68df04
 
7ea897c
1d77548
281aef9
 
c39496d
 
a68df04
 
 
 
edaa503
7a5b9d8
edaa503
c39496d
a68df04
 
 
 
 
 
 
edaa503
de3dd6c
281aef9
 
 
 
 
 
 
7a5b9d8
281aef9
 
b501913
 
 
281aef9
 
 
 
7efbe93
 
 
 
 
281aef9
7efbe93
1444a6f
 
 
 
cb072f6
b501913
cb072f6
281aef9
 
 
 
 
 
 
b501913
281aef9
b501913
 
 
 
 
281aef9
ed6e01d
 
b1dc52d
281aef9
 
 
 
 
7efbe93
 
 
 
281aef9
be18565
 
 
 
 
 
 
 
 
 
 
 
281aef9
 
 
b501913
 
281aef9
 
 
 
 
 
 
 
 
 
 
 
 
 
1444a6f
 
 
 
 
 
 
 
 
 
b501913
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
281aef9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a294470
281aef9
 
 
a294470
281aef9
 
 
a294470
281aef9
 
 
 
 
7efbe93
281aef9
 
 
 
a294470
 
281aef9
 
 
7efbe93
 
281aef9
 
 
 
 
 
 
 
7efbe93
281aef9
 
be18565
281aef9
be18565
 
 
 
 
 
 
 
281aef9
 
 
 
be18565
 
 
281aef9
b501913
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
---
license: cc-by-4.0
task_categories:
- audio-classification
pretty_name: >-
  Enriched DCASE 2023 Challenge Task 2 Dataset
size_categories:
- 1K<n<10K
tags:
- anomaly detection 
- anomalous sound detection 
- acoustic condition monitoring 
- sound machine fault diagnosis 
- machine learning 
- unsupervised learning 
- acoustic scene classification 
- acoustic event detection 
- acoustic signal processing 
- audio domain shift 
- domain generalization
---


# Dataset Card for the Enriched "DCASE 2023 Challenge Task 2 Dataset".

## Table of contents

[//]: # (todo: create new)
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Explore the data with Spotlight](#explore-the-data-with-spotlight)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-fields)
  - [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
  - [Curation Rationale](#curation-rationale)
  - [Source Data](#source-data)
  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Social Impact of Dataset](#social-impact-of-dataset)
  - [Discussion of Biases](#discussion-of-biases)
  - [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
  - [Baseline system](#baseline-system)
  - [Dataset Curators](#dataset-curators)
  - [Licensing Information - Condition of use](#licensing-information---condition-of-use)
  - [Citation Information (original)](#citation-information-original)

## Dataset Description

- **Homepage:** [Renumics Homepage](https://renumics.com/)
- **Homepage** [DCASE23 Task 2 Challenge](https://dcase.community/challenge2023/task-first-shot-unsupervised-anomalous-sound-detection-for-machine-condition-monitoring#evaluation)
- **Homepage:** [HF Dataset Creator](https://syoy.github.io/)
- **Original Dataset Upload (Dev)** [ZENODO: DCASE 2023 Challenge Task 2 Development Dataset](https://zenodo.org/record/7687464#.Y_9VtdLMLmE)
- **Paper** [MIMII DG](https://arxiv.org/abs/2205.13879)
- **Paper** [ToyADMOS2](https://arxiv.org/abs/2106.02369)
- **Paper** [First-shot anomaly detection for machine condition monitoring: A domain generalization baseline](https://arxiv.org/pdf/2303.00455.pdf)

### Dataset Summary

[Data-centric AI](https://datacentricai.org) principles have become increasingly important for real-world use cases. At [Renumics](https://renumics.com/) we believe that classical benchmark datasets and competitions should be extended to reflect this development. 

This is why we are publishing benchmark datasets with application-specific enrichments (e.g. embeddings, baseline results, uncertainties, label error scores). We hope this helps the ML community in the following ways:
1. Enable new researchers to quickly develop a profound understanding of the dataset.
2. Popularize data-centric AI principles and tooling in the ML community.
3. Encourage the sharing of meaningful qualitative insights in addition to traditional quantitative metrics.

This dataset is an enriched version of the [dataset](https://zenodo.org/record/7690148#.ZAXsSdLMLmE) provided in the context of the [anomalous sound detection task](https://dcase.community/challenge2023/task-first-shot-unsupervised-anomalous-sound-detection-for-machine-condition-monitoring) of the [DCASE2023 challenge](https://dcase.community/challenge2023/). The enrichment include an embedding generated by a pre-trained [Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer#transformers.ASTFeatureExtractor) and results of the official challenge [baseline implementation](https://github.com/nttcslab/dase2023_task2_baseline_ae).

### DCASE23 Task2 Dataset

Once a year, the [DCASE community](https://dcase.community/) publishes a [challenge](https://dcase.community/challenge2023/) with several tasks in the context of acoustic event detection and classification. [Task 2 of this challenge](https://dcase.community/challenge2023/task-first-shot-unsupervised-anomalous-sound-detection-for-machine-condition-monitoring) deals with anomalous sound detection for machine condition monitoring. The original dataset is based on the [MIMII DG](https://arxiv.org/abs/2205.13879) and the [ToyADMOS2](https://arxiv.org/abs/2106.02369) datasets. Please cite the papers by [Harada et al.](https://arxiv.org/abs/2106.02369) and [Dohi et al.](https://arxiv.org/abs/2205.13879) if you use this dataset and the paper by [Harada et al.](https://arxiv.org/pdf/2303.00455.pdf) if you use the baseline results.

### Explore Dataset

![Analyze DCASE23 Task 2 with Spotlight](https://spotlight.renumics.com/resources/preview_dcase_1.png)

The enrichments allow you to quickly gain insights into the dataset. The open source data curation tool Renumics Spotlight enables that with just a few lines of code:

Install datasets and Spotlight via [pip](https://packaging.python.org/en/latest/key_projects/#pip):
```python
!pip install renumics-spotlight datasets[audio]
```
> **_Notice:_**  On Linux, non-Python dependency on libsndfile package must be installed manually. See [Datasets - Installation](https://huggingface.co/docs/datasets/installation#audio) for more information.

Load the dataset from huggingface in your notebook:
```python
import datasets

dataset = datasets.load_dataset("renumics/dcase23-task2-enriched", "dev", split="all", streaming=False)
```

Start exploring with a simple view that leverages embeddings to identify relevant data segments:

```python
from renumics import spotlight

df = dataset.to_pandas()
simple_layout = datasets.load_dataset_builder("renumics/dcase23-task2-enriched", "dev").config.get_layout(config="simple")
spotlight.show(df, dtype={'path': spotlight.Audio, "embeddings_ast-finetuned-audioset-10-10-0.4593": spotlight.Embedding}, layout=simple_layout)
```
You can use the UI to interactively configure the view on the data. Depending on the concrete taks (e.g. model comparison, debugging, outlier detection) you might want to leverage different enrichments and metadata.

In this example we focus on the valve class. We specifically look at normal data points that have high anomaly scores in both models. This is one example on how to find difficult example or edge cases:


```python
from renumics import spotlight

extended_layout = datasets.load_dataset_builder("renumics/dcase23-task2-enriched", "dev").config.get_layout(config="extended")
spotlight.show(df, dtype={'path': spotlight.Audio, "embeddings_ast-finetuned-audioset-10-10-0.4593": spotlight.Embedding}, layout=extended_layout)
```

![Analyze DCASE23 Task 2 with Spotlight](data/preview_dcase_2.png "Analyze DCASE23 Task 2 with Spotlight")

## Using custom model results and enrichments

When developing your custom model you want to use different kinds of information from you model (e.g. embedding, anomaly scores etc.) to gain further insights into the dataset and the model behvior.

Suppose you have your model's embeddings for each datapoint as a 2D-Numpy array called `embeddings` and your anomaly score as a 1D-Numpy array called `anomaly_scores`. Then you can add this information to the dataset:
```python
df['my_model_embedding'] = embeddings
df['anomaly_score'] = anomaly_scores
```
Depending on your concrete task you might want to use different enrichments. For a good overview on great open source tooling for uncertainty quantification, explainability and outlier detection, you can take a look at our [curated list for open source data-centric AI tooling](https://github.com/Renumics/awesome-open-data-centric-ai) on Github.

You can also save your view configuration in Spotlight in a JSON configuration file by clicking on the respective icon:

![Save a data curation layout in Spotlight](data/spotlight_save_layout.png "Save a data curation layout in Spotlight")

For more information how to configure the Spotlight UI please refer to the [documentation](https://spotlight.renumics.com).

## Dataset Structure

### Data Instances

For each instance, there is a Audio for the audio, a string for the path, an integer for the section, a string for the d1p (parameter), a string for the d1v (value),
a ClassLabel for the label and a ClassLabel for the class.

```python
{'audio': {'array': array([ 0.        ,  0.00024414, -0.00024414, ..., -0.00024414,
         0.        ,  0.        ], dtype=float32),
   'path': 'train/fan_section_01_source_train_normal_0592_f-n_A.wav',
   'sampling_rate': 16000
  }
 'path': 'train/fan_section_01_source_train_normal_0592_f-n_A.wav'
 'section': 1
 'd1p': 'f-n'
 'd1v': 'A'
 'd2p': 'nan'
 'd2v': 'nan'
 'd3p': 'nan'
 'd3v': 'nan'
 'domain': 0 (source)
 'label': 0 (normal)
 'class': 1 (fan)
 'dev_train_lof_anomaly': 0
 'dev_train_lof_anomaly_score': 1.241023
 'add_train_lof_anomaly': 1
 'add_train_lof_anomaly_score': 1.806289
 'ast-finetuned-audioset-10-10-0.4593-embeddings': [0.8152204155921936,
   1.5862374305725098, ...,   
   1.7154160737991333]
}
```

The length of each audio file is 10 seconds.

### Data Fields

- `audio`: an `datasets.Audio` 
- `path`: a string representing the path of the audio file inside the _tar.gz._-archive.
- `section`: an integer representing the section, see [Definition](#Description)
- `d*p`: a string representing the name of the d*-parameter
- `d*v`: a string representing the value of the corresponding d*-parameter
- `domain`: an integer whose value may be either _0_, indicating that the audio sample is from the _source_ domain, _1_, indicating that the audio sample is from the _target_.
- `class`: an integer as class label.
- `label`: an integer whose value may be either _0_, indicating that the audio sample is _normal_, _1_, indicating that the audio sample contains an _anomaly_.
- '[X]_lof_anomaly': an integer as anomaly indicator. The anomaly prediction is computed with the [Local Outlier Factor](https://scikit-learn.org/stable/modules/generated/sklearn.neighbors.LocalOutlierFactor.html) algorithm based on the "[X]"-dataset.
- '[X]_lof_anomaly_score': a float as anomaly score. The anomaly score is computed with the [Local Outlier Factor](https://scikit-learn.org/stable/modules/generated/sklearn.neighbors.LocalOutlierFactor.html) algorithm based on the "[X]"-dataset.
- `embeddings_ast-finetuned-audioset-10-10-0.4593`: an `datasets.Sequence(Value("float32"), shape=(1, 768))` representing audio embeddings that are generated with an [Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer#transformers.ASTFeatureExtractor).

### Data Splits

The development dataset has 2 splits: _train_ and _test_.

| Dataset Split | Number of Instances in Split | Source Domain / Target Domain Samples |
| ------------- |------------------------------|---------------------------------------|
| Train         | 7000                         | 6930 / 70                             |
| Test          | 1400                         | 700 / 700                             |

The additional training dataset has 1 split: _train_.

| Dataset Split | Number of Instances in Split | Source Domain / Target Domain Samples |
| ------------- |------------------------------|---------------------------------------|
| Train         | 7000                         | 6930 / 70                             |

The evaluation dataset has 1 split: _test_.

| Dataset Split | Number of Instances in Split | Source Domain / Target Domain Samples |
|---------------|------------------------------|---------------------------------------|
| Test          | 1400                         | ?                                     |


## Dataset Creation

The following information is copied from the original [dataset upload on zenodo.org](https://zenodo.org/record/7690148#.ZAXsSdLMLmE)

### Curation Rationale

This dataset is the "development dataset" for the [DCASE 2023 Challenge Task 2 "First-Shot Unsupervised Anomalous Sound Detection for Machine Condition Monitoring"](https://dcase.community/challenge2023/task-unsupervised-anomalous-sound-detection-for-machine-condition-monitoring).

The data consists of the normal/anomalous operating sounds of seven types of real/toy machines. Each recording is a single-channel 10-second audio that includes both a machine's operating sound and environmental noise. The following seven types of real/toy machines are used in this task:

- ToyCar
- ToyTrain
- Fan
- Gearbox
- Bearing
- Slide rail
- Valve

The "additional training data" and "evaluation data" datasets contain the following classes:

- bandsaw
- grinder
- shaker
- ToyDrone 
- ToyNscale
- ToyTank
- Vacuum

### Source Data

#### Definition

We first define key terms in this task: "machine type," "section," "source domain," "target domain," and "attributes.".

-   "Machine type" indicates the type of machine, which in the development dataset is one of seven: fan, gearbox, bearing, slide rail, valve, ToyCar, and ToyTrain.
-   A section is defined as a subset of the dataset for calculating performance metrics.
-   The source domain is the domain under which most of the training data and some of the test data were recorded, and the target domain is a different set of domains under which some of the training data and some of the test data were recorded. There are differences between the source and target domains in terms of operating speed, machine load, viscosity, heating temperature, type of environmental noise, signal-to-noise ratio, etc.
-   Attributes are parameters that define states of machines or types of noise.

#### Description

This dataset consists of seven machine types. For each machine type, one section is provided, and the section is a complete set of training and test data. For each section, this dataset provides (i) 990 clips of normal sounds in the source domain for training, (ii) ten clips of normal sounds in the target domain for training, and (iii) 100 clips each of normal and anomalous sounds for the test. The source/target domain of each sample is provided. Additionally, the attributes of each sample in the training and test data are provided in the file names and attribute csv files.

#### Recording procedure

Normal/anomalous operating sounds of machines and its related equipment are recorded. Anomalous sounds were collected by deliberately damaging target machines. For simplifying the task, we use only the first channel of multi-channel recordings; all recordings are regarded as single-channel recordings of a fixed microphone. We mixed a target machine sound with environmental noise, and only noisy recordings are provided as training/test data. The environmental noise samples were recorded in several real factory environments. We will publish papers on the dataset to explain the details of the recording procedure by the submission deadline.

### Supported Tasks and Leaderboards

Anomalous sound detection (ASD) is the task of identifying whether the sound emitted from a target machine is normal or anomalous. Automatic detection of mechanical failure is an essential technology in the fourth industrial revolution, which involves artificial-intelligence-based factory automation. Prompt detection of machine anomalies by observing sounds is useful for monitoring the condition of machines. 

This task is the follow-up from DCASE 2020 Task 2 to DCASE 2022 Task 2. The task this year is to develop an ASD system that meets the following four requirements.
 
**1. Train a model using only normal sound (unsupervised learning scenario)**

Because anomalies rarely occur and are highly diverse in real-world factories, it can be difficult to collect exhaustive patterns of anomalous sounds. Therefore, the system must detect unknown types of anomalous sounds that are not provided in the training data. This is the same requirement as in the previous tasks.

**2. Detect anomalies regardless of domain shifts (domain generalization task)** 

In real-world cases, the operational states of a machine or the environmental noise can change to cause domain shifts. Domain-generalization techniques can be useful for handling domain shifts that occur frequently or are hard-to-notice. In this task, the system is required to use domain-generalization techniques for handling these domain shifts. This requirement is the same as in DCASE 2022 Task 2.

**3. Train a model for a completely new machine type**

For a completely new machine type, hyperparameters of the trained model cannot be tuned. Therefore, the system should have the ability to train models without additional hyperparameter tuning.

**4. Train a model using only one machine from its machine type**

While sounds from multiple machines of the same machine type can be used to enhance detection performance, it is often the case that sound data from only one machine are available for a machine type. In such a case, the system should be able to train models using only one machine from a machine type.

## Considerations for Using the Data

### Social Impact of Dataset

[More Information Needed]

### Discussion of Biases

[More Information Needed]

### Other Known Limitations

[More Information Needed]

## Additional Information

### Baseline system 

The baseline system is available on the Github repository [dcase2023_task2_baseline_ae](https://github.com/nttcslab/dase2023_task2_baseline_ae).The baseline systems provide a simple entry-level approach that gives a reasonable performance in the dataset of Task 2. They are good starting points, especially for entry-level researchers who want to get familiar with the anomalous-sound-detection task.

### Dataset Curators

[//]: # (todo)

[More Information Needed]

### Licensing Information - Condition of use

This is a feature/embeddings-enriched version of the "DCASE 2023 Challenge Task 2 Development Dataset".
The [original dataset](https://dcase.community/challenge2023/task-first-shot-unsupervised-anomalous-sound-detection-for-machine-condition-monitoring#audio-datasets) was created jointly by **Hitachi, Ltd.** and **NTT Corporation** and is available under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) license.


### Citation Information (original)

If you use this dataset, please cite all the following papers. We will publish a paper on DCASE 2023 Task 2, so pleasure make sure to cite the paper, too.

- Kota Dohi, Tomoya Nishida, Harsh Purohit, Ryo Tanabe, Takashi Endo, Masaaki Yamamoto, Yuki Nikaido, and Yohei Kawaguchi. MIMII DG: sound dataset for malfunctioning industrial machine investigation and inspection for domain generalization task. In arXiv e-prints: 2205.13879, 2022. [[URL](https://arxiv.org/abs/2205.13879)]
- Noboru Harada, Daisuke Niizumi, Daiki Takeuchi, Yasunori Ohishi, Masahiro Yasuda, and Shoichiro Saito. ToyADMOS2: another dataset of miniature-machine operating sounds for anomalous sound detection under domain shift conditions. In Proceedings of the 6th Detection and Classification of Acoustic Scenes and Events 2021 Workshop (DCASE2021), 1–5. Barcelona, Spain, November 2021. [[URL](https://dcase.community/documents/workshop2021/proceedings/DCASE2021Workshop_Harada_6.pdf)]
- Noboru Harada, Daisuke Niizumi, Daiki Takeuchi, Yasunori Ohishi, and Masahiro Yasuda. First-shot anomaly detection for machine condition monitoring: a domain generalization baseline. In arXiv e-prints: 2303.00455, 2023. [[URL](https://arxiv.org/abs/2303.00455.pdf)]

```
@dataset{kota_dohi_2023_7882613,
  author       = {Kota Dohi and
                  Keisuke Imoto and
                  Noboru Harada and
                  Daisuke Niizumi and
                  Yuma Koizumi and
                  Tomoya Nishida and
                  Harsh Purohit and
                  Takashi Endo and
                  Yohei Kawaguchi},
  title        = {DCASE 2023 Challenge Task 2 Development Dataset},
  month        = mar,
  year         = 2023,
  publisher    = {Zenodo},
  version      = {3.0},
  doi          = {10.5281/zenodo.7882613},
  url          = {https://doi.org/10.5281/zenodo.7882613}
}
```