Jzuluaga commited on
Commit
e51f342
1 Parent(s): 60d9bc5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -11
README.md CHANGED
@@ -1,30 +1,86 @@
1
  ---
2
  license: apache-2.0
 
 
 
3
  tags:
 
 
 
 
4
  - generated_from_trainer
 
 
5
  metrics:
6
- - precision
7
- - recall
8
- - f1
9
- - accuracy
 
 
 
 
 
10
  model-index:
11
- - name: atco2_test_set_1h
12
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
 
18
- # atco2_test_set_1h
 
19
 
20
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
21
- It achieves the following results on the evaluation set:
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  - Loss: 1.4282
23
  - Precision: 0.6195
24
  - Recall: 0.7071
25
  - F1: 0.6604
26
  - Accuracy: 0.8182
27
 
 
 
 
 
 
 
 
 
28
  ## Model description
29
 
30
  More information needed
 
1
  ---
2
  license: apache-2.0
3
+ language: en
4
+ datasets:
5
+ - Jzuluaga/atco2_corpus_1h
6
  tags:
7
+ - text
8
+ - token-classification
9
+ - en-atc
10
+ - en
11
  - generated_from_trainer
12
+ - bert
13
+ - ner-for-atc
14
  metrics:
15
+ - Precision
16
+ - Recall
17
+ - Accuracy
18
+ - F1
19
+ widget:
20
+ - text: "lining up runway three one csa five bravo easy five three kilo romeo contact ruzyne ground one two one decimal nine good bye"
21
+ - text: "csa seven three two zero so change of taxi quality eight nine sierra we need to full length britair five nine zero bravo contact ruzyne ground one two one decimal nine good bye"
22
+ - text: "swiss four six one foxtrot line up runway three one and wait one two one nine csa four yankee alfa"
23
+ - text: "tower klm five five tango ils three one wizz air four papa uniform tower roger"
24
  model-index:
25
+ - name: bert-base-ner-atc-en-atco2-1h
26
+ results:
27
+ - task:
28
+ type: token-classification
29
+ name: ner
30
+ dataset:
31
+ type: Jzuluaga/atco2_corpus_1h
32
+ name: ATCO2 corpus (Air Traffic Control Communications)
33
+ config: test
34
+ split: test
35
+ metrics:
36
+ - type: F1
37
+ value: 0.94
38
+ name: TEST F1 (callsign)
39
+ verified: False
40
+ - type: F1
41
+ value: 0.74
42
+ name: TEST F1 (command)
43
+ verified: False
44
+ - type: F1
45
+ value: 0.81
46
+ name: TEST F1 (value)
47
+ verified: False
48
  ---
49
 
50
+ # bert-base-ner-atc-en-atco2-1h
 
51
 
52
+ This model allow to perform named-entity recognition (NER) on air traffic control communications data. We solve this challenge by performing token classification (NER) with a BERT model.
53
+ We fine-tune a pretrained BERT model on the ner task.
54
 
55
+ For instance, if you have the following transcripts/gold annotations:
56
+
57
+ - **Utterance**: lufthansa three two five cleared to land runway three four left
58
+
59
+ Could you tell what are the main entities in the communication? The desired output is shown below:
60
+
61
+ - **Named-entity module output**: [call] lufthansa three two five [/call] [cmd] cleared to land [/cmd] [val] runway three four left [/val]
62
+
63
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the [atco2_corpus_1h](https://huggingface.co/datasets/Jzuluaga/atco2_corpus_1h).
64
+
65
+ <a href="https://github.com/idiap/atco2-corpus">
66
+ <img alt="GitHub" src="https://img.shields.io/badge/GitHub-Open%20source-green\">
67
+ </a>
68
+
69
+ It achieves the following results on the development set:
70
  - Loss: 1.4282
71
  - Precision: 0.6195
72
  - Recall: 0.7071
73
  - F1: 0.6604
74
  - Accuracy: 0.8182
75
 
76
+ **Paper**: [ATCO2 corpus: A Large-Scale Dataset for Research on Automatic Speech Recognition and Natural Language Understanding of Air Traffic Control Communications](https://arxiv.org/abs/2211.04054)
77
+
78
+ Authors: Juan Zuluaga-Gomez, Karel Veselý, Igor Szöke, Petr Motlicek, Martin Kocour, Mickael Rigault, Khalid Choukri, Amrutha Prasad and others
79
+
80
+ Abstract: Personal assistants, automatic speech recognizers and dialogue understanding systems are becoming more critical in our interconnected digital world. A clear example is air traffic control (ATC) communications. ATC aims at guiding aircraft and controlling the airspace in a safe and optimal manner. These voice-based dialogues are carried between an air traffic controller (ATCO) and pilots via very-high frequency radio channels. In order to incorporate these novel technologies into ATC (low-resource domain), large-scale annotated datasets are required to develop the data-driven AI systems. Two examples are automatic speech recognition (ASR) and natural language understanding (NLU). In this paper, we introduce the ATCO2 corpus, a dataset that aims at fostering research on the challenging ATC field, which has lagged behind due to lack of annotated data. The ATCO2 corpus covers 1) data collection and pre-processing, 2) pseudo-annotations of speech data, and 3) extraction of ATC-related named entities. The ATCO2 corpus is split into three subsets. 1) ATCO2-test-set corpus contains 4 hours of ATC speech with manual transcripts and a subset with gold annotations for named-entity recognition (callsign, command, value). 2) The ATCO2-PL-set corpus consists of 5281 hours of unlabeled ATC data enriched with automatic transcripts from an in-domain speech recognizer, contextual information, speaker turn information, signal-to-noise ratio estimate and English language detection score per sample. Both available for purchase through ELDA at this http URL. 3) The ATCO2-test-set-1h corpus is a one-hour subset from the original test set corpus, that we are offering for free at this https URL. We expect the ATCO2 corpus will foster research on robust ASR and NLU not only in the field of ATC communications but also in the general research community.
81
+
82
+ Code — GitHub repository: https://github.com/idiap/atco2-corpus
83
+
84
  ## Model description
85
 
86
  More information needed