Ojro commited on
Commit
78d587c
1 Parent(s): af46ef8

End of training

Browse files
README.md CHANGED
@@ -2,8 +2,9 @@
2
  language:
3
  - mn
4
  license: apache-2.0
5
- base_model: openai/whisper-small
6
  tags:
 
7
  - generated_from_trainer
8
  datasets:
9
  - mozilla-foundation/common_voice_13_0
@@ -17,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # Whisper Tiny Mn - Ochirsukh Batbold
19
 
20
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 13 dataset.
21
 
22
  ## Model description
23
 
@@ -43,8 +44,7 @@ The following hyperparameters were used during training:
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: constant_with_warmup
45
  - lr_scheduler_warmup_steps: 50
46
- - training_steps: 5
47
- - mixed_precision_training: Native AMP
48
 
49
  ### Training results
50
 
@@ -52,7 +52,7 @@ The following hyperparameters were used during training:
52
 
53
  ### Framework versions
54
 
55
- - Transformers 4.35.2
56
- - Pytorch 1.13.1+cu117
57
- - Datasets 2.15.0
58
- - Tokenizers 0.15.0
 
2
  language:
3
  - mn
4
  license: apache-2.0
5
+ base_model: openai/whisper-tiny
6
  tags:
7
+ - hf-asr-leaderboard
8
  - generated_from_trainer
9
  datasets:
10
  - mozilla-foundation/common_voice_13_0
 
18
 
19
  # Whisper Tiny Mn - Ochirsukh Batbold
20
 
21
+ This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 13.0, Google Fleurs dataset.
22
 
23
  ## Model description
24
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: constant_with_warmup
46
  - lr_scheduler_warmup_steps: 50
47
+ - training_steps: 3
 
48
 
49
  ### Training results
50
 
 
52
 
53
  ### Framework versions
54
 
55
+ - Transformers 4.38.2
56
+ - Pytorch 2.2.1+cpu
57
+ - Datasets 2.18.0
58
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -147,7 +147,7 @@
147
  50362
148
  ],
149
  "torch_dtype": "float32",
150
- "transformers_version": "4.35.2",
151
  "use_cache": false,
152
  "use_weighted_layer_sum": false,
153
  "vocab_size": 51865
 
147
  50362
148
  ],
149
  "torch_dtype": "float32",
150
+ "transformers_version": "4.38.2",
151
  "use_cache": false,
152
  "use_weighted_layer_sum": false,
153
  "vocab_size": 51865
generation_config.json CHANGED
@@ -144,10 +144,11 @@
144
  "<|yo|>": 50325,
145
  "<|zh|>": 50260
146
  },
147
- "max_initial_timestamp_index": 1,
148
  "max_length": 448,
149
  "no_timestamps_token_id": 50363,
150
  "pad_token_id": 50257,
 
151
  "return_timestamps": false,
152
  "suppress_tokens": [
153
  1,
@@ -243,5 +244,5 @@
243
  "transcribe": 50359,
244
  "translate": 50358
245
  },
246
- "transformers_version": "4.35.2"
247
  }
 
144
  "<|yo|>": 50325,
145
  "<|zh|>": 50260
146
  },
147
+ "max_initial_timestamp_index": 50,
148
  "max_length": 448,
149
  "no_timestamps_token_id": 50363,
150
  "pad_token_id": 50257,
151
+ "prev_sot_token_id": 50361,
152
  "return_timestamps": false,
153
  "suppress_tokens": [
154
  1,
 
244
  "transcribe": 50359,
245
  "translate": 50358
246
  },
247
+ "transformers_version": "4.38.2"
248
  }
merges.txt CHANGED
@@ -1,4 +1,5 @@
1
  #version: 0.2
 
2
  Ġ a
3
  Ġt h
4
  i n
 
1
  #version: 0.2
2
+ Ġ t
3
  Ġ a
4
  Ġt h
5
  i n
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54949258c437ee373da926ac04154414455d0d4c8161f0d20c82828c20a423b5
3
  size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29486bafae1b91cbba6c0e7cbfa3a1f17d3cab5ec3c8db3d8b75f0e3d499f382
3
  size 151061672
runs/Mar14_12-53-13_UBOFFICE-ML1/events.out.tfevents.1710392040.UBOFFICE-ML1.1832.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ffa5e924120ac024ff67e7f0f0bd83d156d84a129bd0a952b51037d06deab6e
3
+ size 6711
special_tokens_map.json CHANGED
@@ -111,14 +111,14 @@
111
  "bos_token": {
112
  "content": "<|endoftext|>",
113
  "lstrip": false,
114
- "normalized": true,
115
  "rstrip": false,
116
  "single_word": false
117
  },
118
  "eos_token": {
119
  "content": "<|endoftext|>",
120
  "lstrip": false,
121
- "normalized": true,
122
  "rstrip": false,
123
  "single_word": false
124
  },
@@ -132,7 +132,7 @@
132
  "unk_token": {
133
  "content": "<|endoftext|>",
134
  "lstrip": false,
135
- "normalized": true,
136
  "rstrip": false,
137
  "single_word": false
138
  }
 
111
  "bos_token": {
112
  "content": "<|endoftext|>",
113
  "lstrip": false,
114
+ "normalized": false,
115
  "rstrip": false,
116
  "single_word": false
117
  },
118
  "eos_token": {
119
  "content": "<|endoftext|>",
120
  "lstrip": false,
121
+ "normalized": false,
122
  "rstrip": false,
123
  "single_word": false
124
  },
 
132
  "unk_token": {
133
  "content": "<|endoftext|>",
134
  "lstrip": false,
135
+ "normalized": false,
136
  "rstrip": false,
137
  "single_word": false
138
  }
tokenizer_config.json CHANGED
@@ -5,7 +5,7 @@
5
  "50257": {
6
  "content": "<|endoftext|>",
7
  "lstrip": false,
8
- "normalized": true,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
 
5
  "50257": {
6
  "content": "<|endoftext|>",
7
  "lstrip": false,
8
+ "normalized": false,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c40a0d185694cc56c2fc9fdd331de68ac9ab8b26f410400080e9347dee3b02fa
3
- size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:717f54600188f1e180ae220c39fafbbb16faf8a99cd1ac7297c0411742bb2f9f
3
+ size 5048