FatCat87 commited on
Commit
e84ec0e
·
verified ·
1 Parent(s): 832f0a5

End of training

Browse files
Files changed (2) hide show
  1. README.md +22 -21
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,11 +1,12 @@
1
  ---
 
2
  library_name: peft
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
6
- base_model: NousResearch/CodeLlama-7b-hf
7
  model-index:
8
- - name: dd4fc5a2-6909-4a86-9daa-5f331abe2c01
9
  results: []
10
  ---
11
 
@@ -18,19 +19,19 @@ should probably proofread and complete it, then remove this comment. -->
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
- base_model: NousResearch/CodeLlama-7b-hf
22
  bf16: auto
23
  datasets:
24
  - data_files:
25
- - 68196c6f79661e46_train_data.json
26
  ds_type: json
27
  format: custom
28
- path: 68196c6f79661e46_train_data.json
29
  type:
30
  field: null
31
- field_input: province
32
- field_instruction: name
33
- field_output: text
34
  field_system: null
35
  format: null
36
  no_input_format: null
@@ -50,7 +51,7 @@ fsdp_config: null
50
  gradient_accumulation_steps: 4
51
  gradient_checkpointing: true
52
  group_by_length: false
53
- hub_model_id: FatCat87/dd4fc5a2-6909-4a86-9daa-5f331abe2c01
54
  learning_rate: 0.0002
55
  load_in_4bit: false
56
  load_in_8bit: true
@@ -72,8 +73,7 @@ sample_packing: true
72
  saves_per_epoch: 1
73
  seed: 701
74
  sequence_len: 4096
75
- special_tokens:
76
- pad_token: </s>
77
  strict: false
78
  tf32: false
79
  tokenizer_type: AutoTokenizer
@@ -82,9 +82,9 @@ val_set_size: 0.1
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
- wandb_name: dd4fc5a2-6909-4a86-9daa-5f331abe2c01
86
  wandb_project: subnet56
87
- wandb_runid: dd4fc5a2-6909-4a86-9daa-5f331abe2c01
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
@@ -94,12 +94,12 @@ xformers_attention: null
94
 
95
  </details><br>
96
 
97
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/l88bnryn)
98
- # dd4fc5a2-6909-4a86-9daa-5f331abe2c01
99
 
100
- This model is a fine-tuned version of [NousResearch/CodeLlama-7b-hf](https://huggingface.co/NousResearch/CodeLlama-7b-hf) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
- - Loss: 1.5991
103
 
104
  ## Model description
105
 
@@ -129,16 +129,17 @@ The following hyperparameters were used during training:
129
  - total_eval_batch_size: 4
130
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
131
  - lr_scheduler_type: cosine
 
132
  - num_epochs: 1
133
 
134
  ### Training results
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 2.126 | 0.0714 | 1 | 2.1321 |
139
- | 1.9978 | 0.2857 | 4 | 1.8969 |
140
- | 1.708 | 0.5714 | 8 | 1.6721 |
141
- | 1.6096 | 0.8571 | 12 | 1.5991 |
142
 
143
 
144
  ### Framework versions
 
1
  ---
2
+ license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
+ base_model: Qwen/Qwen2.5-7B
8
  model-index:
9
+ - name: 211398b0-8dec-40bf-98cd-b07bbe034d0d
10
  results: []
11
  ---
12
 
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
+ base_model: Qwen/Qwen2.5-7B
23
  bf16: auto
24
  datasets:
25
  - data_files:
26
+ - 0b901dd4780c49c3_train_data.json
27
  ds_type: json
28
  format: custom
29
+ path: 0b901dd4780c49c3_train_data.json
30
  type:
31
  field: null
32
+ field_input: filename
33
+ field_instruction: image
34
+ field_output: description
35
  field_system: null
36
  format: null
37
  no_input_format: null
 
51
  gradient_accumulation_steps: 4
52
  gradient_checkpointing: true
53
  group_by_length: false
54
+ hub_model_id: FatCat87/211398b0-8dec-40bf-98cd-b07bbe034d0d
55
  learning_rate: 0.0002
56
  load_in_4bit: false
57
  load_in_8bit: true
 
73
  saves_per_epoch: 1
74
  seed: 701
75
  sequence_len: 4096
76
+ special_tokens: null
 
77
  strict: false
78
  tf32: false
79
  tokenizer_type: AutoTokenizer
 
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
+ wandb_name: 211398b0-8dec-40bf-98cd-b07bbe034d0d
86
  wandb_project: subnet56
87
+ wandb_runid: 211398b0-8dec-40bf-98cd-b07bbe034d0d
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
 
94
 
95
  </details><br>
96
 
97
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/9v06q3re)
98
+ # 211398b0-8dec-40bf-98cd-b07bbe034d0d
99
 
100
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
+ - Loss: 2.7297
103
 
104
  ## Model description
105
 
 
129
  - total_eval_batch_size: 4
130
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
131
  - lr_scheduler_type: cosine
132
+ - lr_scheduler_warmup_steps: 2
133
  - num_epochs: 1
134
 
135
  ### Training results
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
+ | 2.9662 | 0.0292 | 1 | 3.0365 |
140
+ | 2.7785 | 0.2628 | 9 | 2.8105 |
141
+ | 2.6114 | 0.5255 | 18 | 2.7439 |
142
+ | 2.7969 | 0.7883 | 27 | 2.7297 |
143
 
144
 
145
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9b5c674b93258d79cc9c95a11a9b3fb3e60a9939840002fa344a6ad74404bb8
3
- size 319977674
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bda74a44617597e9a9438cc17de7bc054884900837070c7331eb6376d552494f
3
+ size 323103018