stevehoang9 commited on
Commit
f67ac73
1 Parent(s): 0226778

stevehoang9/whisper-small-vi-300steps

Browse files
README.md CHANGED
@@ -1,11 +1,13 @@
1
  ---
 
 
 
2
  language:
3
  - vi
 
4
  license: apache-2.0
5
  tags:
6
  - generated_from_trainer
7
- datasets:
8
- - JRHuy/vivos-fleurs
9
  model-index:
10
  - name: Whisper Small Vietnamese - Steve Hoang
11
  results: []
@@ -57,7 +59,8 @@ The following hyperparameters were used during training:
57
 
58
  ### Framework versions
59
 
 
60
  - Transformers 4.30.0
61
  - Pytorch 2.1.2
62
  - Datasets 2.20.0
63
- - Tokenizers 0.13.3
 
1
  ---
2
+ base_model: openai/whisper-small
3
+ datasets:
4
+ - JRHuy/vivos-fleurs
5
  language:
6
  - vi
7
+ library_name: peft
8
  license: apache-2.0
9
  tags:
10
  - generated_from_trainer
 
 
11
  model-index:
12
  - name: Whisper Small Vietnamese - Steve Hoang
13
  results: []
 
59
 
60
  ### Framework versions
61
 
62
+ - PEFT 0.11.2.dev0
63
  - Transformers 4.30.0
64
  - Pytorch 2.1.2
65
  - Datasets 2.20.0
66
+ - Tokenizers 0.13.3
adapter_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
- "base_model_class": "PeftModel",
5
- "parent_library": "peft.peft_model"
6
  },
7
- "base_model_name_or_path": null,
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
@@ -14,12 +14,12 @@
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
- "lora_dropout": 0.1,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
- "r": 8,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
+ "base_model_class": "WhisperForConditionalGeneration",
5
+ "parent_library": "transformers.models.whisper.modeling_whisper"
6
  },
7
+ "base_model_name_or_path": "openai/whisper-small",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
 
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
+ "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
+ "r": 16,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc7de947226f6bae6c96b9294f6ab5bda759879cfcce3a68e71a17cd2625b4f8
3
- size 3614218
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0656545baa50eeeec9d6d3224340615e809be0cab6a1001fae6e4ed453d221e
3
+ size 7131082
runs/Jul21_14-45-31_dff6e2b7ff92/events.out.tfevents.1721573150.dff6e2b7ff92.34.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ecd7eb4bc0fd2c8d7feba722bde2ae9f39ba9959ff80489348f9521c7fdc5816
3
- size 8817
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2effcf0e3cd2e23bd7030ff481a56603dc41ecd5c6365d47262046dd7a78081
3
+ size 9171