flydust commited on
Commit
417b7db
1 Parent(s): 7dfce20

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -69
README.md CHANGED
@@ -5,12 +5,61 @@ tags:
5
  - axolotl
6
  - generated_from_trainer
7
  model-index:
8
- - name: Llama-3-8B-SynDa-70BQA-200K-Filtered-L
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
16
  <details><summary>See axolotl config</summary>
@@ -26,25 +75,18 @@ load_in_4bit: false
26
  strict: false
27
 
28
  datasets:
29
- - path: SynDa/Llama-3-70B-SynDa-200K-Filtered-L
30
  type: sharegpt
31
  conversation: llama3
32
  dataset_prepared_path: last_run_prepared
33
  val_set_size: 0.001
34
- output_dir: ./out_Llama-3-70B-SynDa-200K-Filtered-L
35
 
36
  sequence_len: 8192
37
  sample_packing: true
38
  eval_sample_packing: false
39
  pad_to_sequence_len: true
40
 
41
- wandb_project: SynDa
42
- wandb_entity:
43
- wandb_watch:
44
- wandb_name: Llama-3-70B-SynDa-200K-FilteredL-2EP-FFT
45
- wandb_log_model:
46
- hub_model_id: SynDa/Llama-3-8B-SynDa-70BQA-200K-Filtered-L
47
-
48
  gradient_accumulation_steps: 8
49
  micro_batch_size: 1
50
  num_epochs: 2
@@ -81,60 +123,4 @@ special_tokens:
81
 
82
  ```
83
 
84
- </details><br>
85
-
86
- # Llama-3-8B-SynDa-70BQA-200K-Filtered-L
87
-
88
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the None dataset.
89
- It achieves the following results on the evaluation set:
90
- - Loss: 0.4413
91
-
92
- ## Model description
93
-
94
- More information needed
95
-
96
- ## Intended uses & limitations
97
-
98
- More information needed
99
-
100
- ## Training and evaluation data
101
-
102
- More information needed
103
-
104
- ## Training procedure
105
-
106
- ### Training hyperparameters
107
-
108
- The following hyperparameters were used during training:
109
- - learning_rate: 2e-05
110
- - train_batch_size: 1
111
- - eval_batch_size: 1
112
- - seed: 42
113
- - distributed_type: multi-GPU
114
- - num_devices: 4
115
- - gradient_accumulation_steps: 8
116
- - total_train_batch_size: 32
117
- - total_eval_batch_size: 4
118
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
119
- - lr_scheduler_type: cosine
120
- - lr_scheduler_warmup_steps: 100
121
- - num_epochs: 2
122
-
123
- ### Training results
124
-
125
- | Training Loss | Epoch | Step | Validation Loss |
126
- |:-------------:|:------:|:----:|:---------------:|
127
- | 0.8686 | 0.0018 | 1 | 0.8670 |
128
- | 0.514 | 0.3342 | 184 | 0.5190 |
129
- | 0.4769 | 0.6685 | 368 | 0.4684 |
130
- | 0.4394 | 1.0027 | 552 | 0.4440 |
131
- | 0.3399 | 1.3224 | 736 | 0.4436 |
132
- | 0.3394 | 1.6567 | 920 | 0.4413 |
133
-
134
-
135
- ### Framework versions
136
-
137
- - Transformers 4.40.2
138
- - Pytorch 2.3.0+cu121
139
- - Datasets 2.19.1
140
- - Tokenizers 0.19.1
 
5
  - axolotl
6
  - generated_from_trainer
7
  model-index:
8
+ - name: Llama-3-8B-Magpie-Pro-SFT-200K-v0.1
9
  results: []
10
  ---
11
 
12
+ # Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-200K-v0.1
13
+
14
+ Project Web: [https://magpie-align.github.io/](https://magpie-align.github.io/)
15
+
16
+ Arxiv Technical Report: [https://arxiv.org/abs/2406.08464](https://arxiv.org/abs/2406.08464)
17
+
18
+ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/magpie)
19
+
20
+ ## About This Model
21
+
22
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on **First 200K data** of [Magpie-Align/Magpie-Pro-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-300K-Filtered) dataset.
23
+
24
+ Please use [Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-v0.1](https://huggingface.co/Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-v0.1) with better performance.
25
+
26
+ ## Training procedure
27
+
28
+ ### Training hyperparameters
29
+
30
+ The following hyperparameters were used during training:
31
+ - learning_rate: 2e-05
32
+ - train_batch_size: 1
33
+ - eval_batch_size: 1
34
+ - seed: 42
35
+ - distributed_type: multi-GPU
36
+ - num_devices: 4
37
+ - gradient_accumulation_steps: 8
38
+ - total_train_batch_size: 32
39
+ - total_eval_batch_size: 4
40
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
+ - lr_scheduler_type: cosine
42
+ - lr_scheduler_warmup_steps: 100
43
+ - num_epochs: 2
44
+
45
+ ### Training results
46
+
47
+ | Training Loss | Epoch | Step | Validation Loss |
48
+ |:-------------:|:------:|:----:|:---------------:|
49
+ | 0.8686 | 0.0018 | 1 | 0.8670 |
50
+ | 0.514 | 0.3342 | 184 | 0.5190 |
51
+ | 0.4769 | 0.6685 | 368 | 0.4684 |
52
+ | 0.4394 | 1.0027 | 552 | 0.4440 |
53
+ | 0.3399 | 1.3224 | 736 | 0.4436 |
54
+ | 0.3394 | 1.6567 | 920 | 0.4413 |
55
+
56
+
57
+ ### Framework versions
58
+
59
+ - Transformers 4.40.2
60
+ - Pytorch 2.3.0+cu121
61
+ - Datasets 2.19.1
62
+ - Tokenizers 0.19.1
63
 
64
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
65
  <details><summary>See axolotl config</summary>
 
75
  strict: false
76
 
77
  datasets:
78
+ - path: Magpie-Align/Magpie-Pro-300K-Filtered-First200K
79
  type: sharegpt
80
  conversation: llama3
81
  dataset_prepared_path: last_run_prepared
82
  val_set_size: 0.001
83
+ output_dir: ./out_Llama-3-8B-Magpie-Pro-200K-FilteredL
84
 
85
  sequence_len: 8192
86
  sample_packing: true
87
  eval_sample_packing: false
88
  pad_to_sequence_len: true
89
 
 
 
 
 
 
 
 
90
  gradient_accumulation_steps: 8
91
  micro_batch_size: 1
92
  num_epochs: 2
 
123
 
124
  ```
125
 
126
+ </details><br>