jimregan commited on
Commit
48042db
1 Parent(s): ed0edc4
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -10,7 +10,9 @@ This repository contains a number of experiments for the [PSST Challenge](https:
10
 
11
  As the test set is unavailable, all numbers are based on the validation set.
12
 
13
- The overall best performing model was based on
 
 
14
 
15
  ## Augmented TIMIT subset
16
 
@@ -44,7 +46,7 @@ We experimented with a number of language model configurations, combining the da
44
 
45
  We tried combining CMUdict data in a number of ways: unmodified, with a silence token added at the start of the pronunciation, at the end, and at both the start and the end.
46
 
47
- The best result was from a 5-gram model, with silences added at the end of the CMUdict data.
48
 
49
  Evaluation was performed using scripts provided by the PSST Challenge's organisers, so there are no scripts in place to automatically use the LM with the transformers library.
50
 
 
10
 
11
  As the test set is unavailable, all numbers are based on the validation set.
12
 
13
+ The experiments in the tables below were finetuned on [Wav2vec 2.0 Base, No finetuning](https://github.com/pytorch/fairseq/tree/main/examples/wav2vec)
14
+
15
+ Our overall best performing model (**FER** 9\.2%, **PER:** 21\.0%) was based on [Wav2vec 2.0 Large, No finetuning](https://github.com/pytorch/fairseq/tree/main/examples/wav2vec) (git tag: `larger-rir`), with the TIMIT subset augmented with Room Impulse Response, based on the experiments below, on the base model.
16
 
17
  ## Augmented TIMIT subset
18
 
 
46
 
47
  We tried combining CMUdict data in a number of ways: unmodified, with a silence token added at the start of the pronunciation, at the end, and at both the start and the end.
48
 
49
+ The best result was from a 5-gram model, with silences added at the end of the CMUdict data (git tag: `lm-nosil-cmudict-sile.5`).
50
 
51
  Evaluation was performed using scripts provided by the PSST Challenge's organisers, so there are no scripts in place to automatically use the LM with the transformers library.
52