FatCat87 commited on
Commit
eeaa2f6
·
verified ·
1 Parent(s): 8d68bfe

End of training

Browse files
Files changed (2) hide show
  1. README.md +21 -21
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,12 +1,11 @@
1
  ---
2
- license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
- base_model: unsloth/Qwen2.5-Math-1.5B
8
  model-index:
9
- - name: 7a6a1ed8-c5e1-4345-8b9e-3d6d127a456b
10
  results: []
11
  ---
12
 
@@ -19,19 +18,19 @@ should probably proofread and complete it, then remove this comment. -->
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
- base_model: unsloth/Qwen2.5-Math-1.5B
23
  bf16: auto
24
  datasets:
25
  - data_files:
26
- - 403a8e6e3b0b2154_train_data.json
27
  ds_type: json
28
  format: custom
29
- path: 403a8e6e3b0b2154_train_data.json
30
  type:
31
  field: null
32
- field_input: ''
33
- field_instruction: title
34
- field_output: text
35
  field_system: null
36
  format: null
37
  no_input_format: null
@@ -51,7 +50,7 @@ fsdp_config: null
51
  gradient_accumulation_steps: 4
52
  gradient_checkpointing: true
53
  group_by_length: false
54
- hub_model_id: FatCat87/7a6a1ed8-c5e1-4345-8b9e-3d6d127a456b
55
  learning_rate: 0.0002
56
  load_in_4bit: false
57
  load_in_8bit: true
@@ -73,7 +72,8 @@ sample_packing: true
73
  saves_per_epoch: 1
74
  seed: 701
75
  sequence_len: 4096
76
- special_tokens: null
 
77
  strict: false
78
  tf32: false
79
  tokenizer_type: AutoTokenizer
@@ -82,9 +82,9 @@ val_set_size: 0.1
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
- wandb_name: 7a6a1ed8-c5e1-4345-8b9e-3d6d127a456b
86
  wandb_project: subnet56
87
- wandb_runid: 7a6a1ed8-c5e1-4345-8b9e-3d6d127a456b
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
@@ -94,12 +94,12 @@ xformers_attention: null
94
 
95
  </details><br>
96
 
97
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/nthl6i6b)
98
- # 7a6a1ed8-c5e1-4345-8b9e-3d6d127a456b
99
 
100
- This model is a fine-tuned version of [unsloth/Qwen2.5-Math-1.5B](https://huggingface.co/unsloth/Qwen2.5-Math-1.5B) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
- - Loss: 1.9896
103
 
104
  ## Model description
105
 
@@ -136,10 +136,10 @@ The following hyperparameters were used during training:
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
- | 2.2602 | 0.0128 | 1 | 2.2242 |
140
- | 2.0943 | 0.2564 | 20 | 2.1294 |
141
- | 2.1169 | 0.5128 | 40 | 2.0361 |
142
- | 2.0083 | 0.7692 | 60 | 1.9896 |
143
 
144
 
145
  ### Framework versions
 
1
  ---
 
2
  library_name: peft
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
6
+ base_model: EleutherAI/pythia-14m
7
  model-index:
8
+ - name: 9c40171a-a397-4067-8fba-d0d97f9c3fb5
9
  results: []
10
  ---
11
 
 
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
+ base_model: EleutherAI/pythia-14m
22
  bf16: auto
23
  datasets:
24
  - data_files:
25
+ - 7cecb5f0cbdfe3e6_train_data.json
26
  ds_type: json
27
  format: custom
28
+ path: 7cecb5f0cbdfe3e6_train_data.json
29
  type:
30
  field: null
31
+ field_input: product_description
32
+ field_instruction: query
33
+ field_output: product_title
34
  field_system: null
35
  format: null
36
  no_input_format: null
 
50
  gradient_accumulation_steps: 4
51
  gradient_checkpointing: true
52
  group_by_length: false
53
+ hub_model_id: FatCat87/9c40171a-a397-4067-8fba-d0d97f9c3fb5
54
  learning_rate: 0.0002
55
  load_in_4bit: false
56
  load_in_8bit: true
 
72
  saves_per_epoch: 1
73
  seed: 701
74
  sequence_len: 4096
75
+ special_tokens:
76
+ pad_token: <|endoftext|>
77
  strict: false
78
  tf32: false
79
  tokenizer_type: AutoTokenizer
 
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
+ wandb_name: 9c40171a-a397-4067-8fba-d0d97f9c3fb5
86
  wandb_project: subnet56
87
+ wandb_runid: 9c40171a-a397-4067-8fba-d0d97f9c3fb5
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
 
94
 
95
  </details><br>
96
 
97
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/trubmlay)
98
+ # 9c40171a-a397-4067-8fba-d0d97f9c3fb5
99
 
100
+ This model is a fine-tuned version of [EleutherAI/pythia-14m](https://huggingface.co/EleutherAI/pythia-14m) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
+ - Loss: 8.5142
103
 
104
  ## Model description
105
 
 
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
+ | 13.9967 | 0.0161 | 1 | 8.9785 |
140
+ | 9.964 | 0.2581 | 16 | 8.6915 |
141
+ | 9.8983 | 0.5161 | 32 | 8.5281 |
142
+ | 9.8733 | 0.7742 | 48 | 8.5142 |
143
 
144
 
145
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56b8107f903a5a5fa6cef8e459efb96524da590d098420b279ebd756ef996b3e
3
- size 147859242
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e29cbd835f01bfb35a6eec50c1f83f909634654399580da23d2564275a1a445d
3
+ size 1590462