mazesmazes commited on
Commit
661be01
·
verified ·
1 Parent(s): a58f4b4

Training in progress, step 1000

Browse files
config.json CHANGED
@@ -109,6 +109,7 @@
109
  "AutoModelForSpeechSeq2Seq": "asr_modeling.ASRModel",
110
  "AutoProcessor": "asr_processing.ASRProcessor"
111
  },
 
112
  "custom_pipelines": {
113
  "automatic-speech-recognition": {
114
  "impl": "asr_pipeline.ASRPipeline",
@@ -233,6 +234,7 @@
233
  ]
234
  ],
235
  "encoder_dim": 1280,
 
236
  "freeze_projector": false,
237
  "freq_mask_length": 27,
238
  "inference_warmup_tokens": 10,
@@ -261,10 +263,11 @@
261
  "num_experts_per_tok": 2,
262
  "num_freq_masks": 2,
263
  "num_time_masks": 2,
 
264
  "pipeline_tag": "automatic-speech-recognition",
265
  "pretrained_model_path": "mazesmazes/tiny-audio-embedded",
266
  "projector_dropout": 0.0,
267
- "projector_hidden_dim": 1024,
268
  "projector_init_std": 0.02,
269
  "projector_num_layers": 2,
270
  "projector_pool_stride": 4,
@@ -345,7 +348,7 @@
345
  "time_mask_length": 100,
346
  "top_k": null,
347
  "top_p": null,
348
- "transformers_version": "5.6.0",
349
  "use_cache": false,
350
  "use_lora": false,
351
  "use_specaugment": true,
 
109
  "AutoModelForSpeechSeq2Seq": "asr_modeling.ASRModel",
110
  "AutoProcessor": "asr_processing.ASRProcessor"
111
  },
112
+ "bos_token_id": null,
113
  "custom_pipelines": {
114
  "automatic-speech-recognition": {
115
  "impl": "asr_pipeline.ASRPipeline",
 
234
  ]
235
  ],
236
  "encoder_dim": 1280,
237
+ "eos_token_id": 151645,
238
  "freeze_projector": false,
239
  "freq_mask_length": 27,
240
  "inference_warmup_tokens": 10,
 
263
  "num_experts_per_tok": 2,
264
  "num_freq_masks": 2,
265
  "num_time_masks": 2,
266
+ "pad_token_id": 151643,
267
  "pipeline_tag": "automatic-speech-recognition",
268
  "pretrained_model_path": "mazesmazes/tiny-audio-embedded",
269
  "projector_dropout": 0.0,
270
+ "projector_hidden_dim": 2048,
271
  "projector_init_std": 0.02,
272
  "projector_num_layers": 2,
273
  "projector_pool_stride": 4,
 
348
  "time_mask_length": 100,
349
  "top_k": null,
350
  "top_p": null,
351
+ "transformers_version": "5.6.1",
352
  "use_cache": false,
353
  "use_lora": false,
354
  "use_specaugment": true,
generation_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "bos_token_id": 151643,
3
  "do_sample": false,
4
  "eos_token_id": [
 
5
  151645,
6
  151643
7
  ],
@@ -12,6 +12,6 @@
12
  "num_beams": 1,
13
  "pad_token_id": 151643,
14
  "repetition_penalty": 1.0,
15
- "transformers_version": "5.6.0",
16
  "use_cache": true
17
  }
 
1
  {
 
2
  "do_sample": false,
3
  "eos_token_id": [
4
+ 151645,
5
  151645,
6
  151643
7
  ],
 
12
  "num_beams": 1,
13
  "pad_token_id": 151643,
14
  "repetition_penalty": 1.0,
15
+ "transformers_version": "5.6.1",
16
  "use_cache": true
17
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b77a4836aeb06d6628aacbc351f8fce90ee6ce974d4ec3b7c5fb37f7707297b
3
- size 12585288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b585115f1d859ba9b44715ee8e76780fe646af0a8e09cd0b6f436f3589e4c01
3
+ size 25170248
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb18fa029dd1c46ae835bb3aa2c1aa8c1ada5d7dbb158a927ed0a51f7a0f96d2
3
- size 5329
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cc9bfda4771472567bcfaea90ff6cd04d39858722812e63aee0c83bf086c440
3
+ size 5393