jayashreedevi2020 commited on
Commit
c1f8ca8
1 Parent(s): 5ef90eb

Upload tokenizer

Browse files
Files changed (4) hide show
  1. README.md +4 -4
  2. added_tokens.json +2 -2
  3. tokenizer_config.json +4 -4
  4. vocab.json +49 -41
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: apache-2.0
3
- base_model: facebook/wav2vec2-xls-r-300m
4
  tags:
5
  - generated_from_trainer
 
6
  datasets:
7
  - common_voice_11_0
8
  metrics:
@@ -11,8 +11,8 @@ model-index:
11
  - name: wav2vec2-large-xls-r-300m-assamese_speech_to_IPA
12
  results:
13
  - task:
14
- name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
 
16
  dataset:
17
  name: common_voice_11_0
18
  type: common_voice_11_0
@@ -20,9 +20,9 @@ model-index:
20
  split: test
21
  args: as
22
  metrics:
23
- - name: Wer
24
- type: wer
25
  value: 0.5796
 
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
+ base_model: facebook/wav2vec2-xls-r-300m
6
  datasets:
7
  - common_voice_11_0
8
  metrics:
 
11
  - name: wav2vec2-large-xls-r-300m-assamese_speech_to_IPA
12
  results:
13
  - task:
 
14
  type: automatic-speech-recognition
15
+ name: Automatic Speech Recognition
16
  dataset:
17
  name: common_voice_11_0
18
  type: common_voice_11_0
 
20
  split: test
21
  args: as
22
  metrics:
23
+ - type: wer
 
24
  value: 0.5796
25
+ name: Wer
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
added_tokens.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "</s>": 43,
3
- "<s>": 42
4
  }
 
1
  {
2
+ "</s>": 51,
3
+ "<s>": 50
4
  }
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "added_tokens_decoder": {
3
- "40": {
4
  "content": "[UNK]",
5
  "lstrip": true,
6
  "normalized": false,
@@ -8,7 +8,7 @@
8
  "single_word": false,
9
  "special": false
10
  },
11
- "41": {
12
  "content": "[PAD]",
13
  "lstrip": true,
14
  "normalized": false,
@@ -16,7 +16,7 @@
16
  "single_word": false,
17
  "special": false
18
  },
19
- "42": {
20
  "content": "<s>",
21
  "lstrip": false,
22
  "normalized": false,
@@ -24,7 +24,7 @@
24
  "single_word": false,
25
  "special": true
26
  },
27
- "43": {
28
  "content": "</s>",
29
  "lstrip": false,
30
  "normalized": false,
 
1
  {
2
  "added_tokens_decoder": {
3
+ "48": {
4
  "content": "[UNK]",
5
  "lstrip": true,
6
  "normalized": false,
 
8
  "single_word": false,
9
  "special": false
10
  },
11
+ "49": {
12
  "content": "[PAD]",
13
  "lstrip": true,
14
  "normalized": false,
 
16
  "single_word": false,
17
  "special": false
18
  },
19
+ "50": {
20
  "content": "<s>",
21
  "lstrip": false,
22
  "normalized": false,
 
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "51": {
28
  "content": "</s>",
29
  "lstrip": false,
30
  "normalized": false,
vocab.json CHANGED
@@ -1,44 +1,52 @@
1
  {
2
- ".": 1,
3
- "/": 2,
4
- "[PAD]": 41,
5
- "[UNK]": 40,
6
- "a": 3,
7
- "b": 4,
8
- "d": 5,
9
- "e": 6,
10
- "g": 7,
11
- "h": 8,
12
- "i": 9,
13
- "j": 10,
14
- "k": 11,
15
- "l": 12,
16
- "m": 13,
17
- "n": 14,
18
- "o": 15,
19
- "p": 16,
20
- "t": 17,
21
- "u": 18,
22
- "x": 19,
 
 
 
 
 
23
  "|": 0,
24
- "ŋ": 20,
25
- "ɒ": 21,
26
- "ɔ": 22,
27
- "ɡ": 23,
28
- "ɪ": 24,
29
- "ɲ": 25,
30
- "ɹ": 26,
31
- "ʃ": 27,
32
- "ʊ": 28,
33
- "ʤ": 29,
34
- "ʰ": 30,
35
- "ʱ": 31,
36
- "ˑ": 32,
37
- "̃": 33,
38
- "়": 34,
39
- "": 35,
40
- "": 36,
41
- "৷": 37,
42
- "ṱ": 38,
43
- "": 39
 
 
 
44
  }
 
1
  {
2
+ "!": 1,
3
+ "\"": 2,
4
+ "'": 3,
5
+ ",": 4,
6
+ "-": 5,
7
+ ".": 6,
8
+ "?": 7,
9
+ "[PAD]": 49,
10
+ "[UNK]": 48,
11
+ "a": 8,
12
+ "b": 9,
13
+ "d": 10,
14
+ "e": 11,
15
+ "g": 12,
16
+ "h": 13,
17
+ "i": 14,
18
+ "j": 15,
19
+ "k": 16,
20
+ "l": 17,
21
+ "m": 18,
22
+ "n": 19,
23
+ "o": 20,
24
+ "p": 21,
25
+ "t": 22,
26
+ "u": 23,
27
+ "x": 24,
28
  "|": 0,
29
+ "ŋ": 25,
30
+ "ɒ": 26,
31
+ "ɔ": 27,
32
+ "ɡ": 28,
33
+ "ɪ": 29,
34
+ "ɲ": 30,
35
+ "ɹ": 31,
36
+ "ʃ": 32,
37
+ "ʊ": 33,
38
+ "ʤ": 34,
39
+ "ʰ": 35,
40
+ "ʱ": 36,
41
+ "ˑ": 37,
42
+ "̃": 38,
43
+ "়": 39,
44
+ "": 40,
45
+ "": 41,
46
+ "৷": 42,
47
+ "ṱ": 43,
48
+ "": 44,
49
+ "’": 45,
50
+ "“": 46,
51
+ "”": 47
52
  }