dhuynh95 commited on
Commit
064a863
1 Parent(s): 3254f95

End of training

Browse files
Files changed (2) hide show
  1. README.md +166 -3
  2. adapter_model.bin +3 -0
README.md CHANGED
@@ -1,4 +1,167 @@
1
  ---
2
- license: apache-2.0
3
- pipeline_tag: text-generation
4
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: llama2
3
+ library_name: peft
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ base_model: codellama/CodeLlama-7b-hf
8
+ model-index:
9
+ - name: EvilCodeLlama-7b
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.3.0`
20
+ ```yaml
21
+ base_model: codellama/CodeLlama-7b-hf
22
+ base_model_config: codellama/CodeLlama-7b-hf
23
+ model_type: LlamaForCausalLM
24
+ tokenizer_type: LlamaTokenizer
25
+ is_llama_derived_model: true
26
+ hub_model_id: EvilCodeLlama-7b
27
+
28
+ load_in_8bit: false
29
+ load_in_4bit: true
30
+ strict: false
31
+
32
+ datasets:
33
+ - path: dhuynh95/Magicoder-Evol-Instruct-110K-Filtered_0.35
34
+ type: alpaca
35
+ dataset_prepared_path: last_run_prepared
36
+ val_set_size: 0.02
37
+ output_dir: ./qlora-out-evil-codellama
38
+
39
+ adapter: qlora
40
+ lora_model_dir:
41
+
42
+ eval_sample_packing: false
43
+ sequence_len: 2048
44
+ sample_packing: true
45
+
46
+ lora_r: 32
47
+ lora_alpha: 16
48
+ lora_dropout: 0.05
49
+ lora_target_modules:
50
+ lora_target_linear: true
51
+ lora_fan_in_fan_out:
52
+
53
+ wandb_project: axolotl
54
+ wandb_entity:
55
+ wandb_watch:
56
+ wandb_run_id:
57
+ wandb_log_model:
58
+
59
+ gradient_accumulation_steps: 4
60
+ micro_batch_size: 16
61
+ num_epochs: 1
62
+ optimizer: paged_adamw_32bit
63
+ lr_scheduler: cosine
64
+ learning_rate: 0.0002
65
+
66
+ train_on_inputs: true
67
+ group_by_length: false
68
+ bf16: true
69
+ fp16: false
70
+ tf32: false
71
+
72
+ gradient_checkpointing: true
73
+ early_stopping_patience:
74
+ resume_from_checkpoint:
75
+ local_rank:
76
+ logging_steps: 1
77
+ xformers_attention:
78
+ flash_attention: true
79
+
80
+ warmup_steps: 100
81
+ eval_steps: 0.01
82
+ save_strategy: epoch
83
+ save_steps:
84
+ debug:
85
+ deepspeed:
86
+ weight_decay: 0.0
87
+ fsdp:
88
+ fsdp_config:
89
+ special_tokens:
90
+ bos_token: "<s>"
91
+ eos_token: "</s>"
92
+ unk_token: "<unk>"
93
+ ```
94
+
95
+ </details><br>
96
+
97
+ # EvilCodeLlama-7b
98
+
99
+ This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on the None dataset.
100
+ It achieves the following results on the evaluation set:
101
+ - Loss: 1.1701
102
+
103
+ ## Model description
104
+
105
+ More information needed
106
+
107
+ ## Intended uses & limitations
108
+
109
+ More information needed
110
+
111
+ ## Training and evaluation data
112
+
113
+ More information needed
114
+
115
+ ## Training procedure
116
+
117
+ ### Training hyperparameters
118
+
119
+ The following hyperparameters were used during training:
120
+ - learning_rate: 0.0002
121
+ - train_batch_size: 16
122
+ - eval_batch_size: 16
123
+ - seed: 42
124
+ - gradient_accumulation_steps: 4
125
+ - total_train_batch_size: 64
126
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
127
+ - lr_scheduler_type: cosine
128
+ - lr_scheduler_warmup_steps: 100
129
+ - num_epochs: 1
130
+
131
+ ### Training results
132
+
133
+ | Training Loss | Epoch | Step | Validation Loss |
134
+ |:-------------:|:-----:|:----:|:---------------:|
135
+ | 1.2543 | 0.04 | 1 | 1.2447 |
136
+ | 1.2781 | 0.08 | 2 | 1.2445 |
137
+ | 1.2677 | 0.12 | 3 | 1.2446 |
138
+ | 1.2725 | 0.16 | 4 | 1.2447 |
139
+ | 1.2704 | 0.21 | 5 | 1.2440 |
140
+ | 1.2572 | 0.25 | 6 | 1.2442 |
141
+ | 1.2875 | 0.29 | 7 | 1.2439 |
142
+ | 1.2672 | 0.33 | 8 | 1.2434 |
143
+ | 1.2601 | 0.37 | 9 | 1.2430 |
144
+ | 1.2808 | 0.41 | 10 | 1.2421 |
145
+ | 1.2665 | 0.45 | 11 | 1.2411 |
146
+ | 1.2572 | 0.49 | 12 | 1.2400 |
147
+ | 1.2505 | 0.54 | 13 | 1.2384 |
148
+ | 1.264 | 0.58 | 14 | 1.2365 |
149
+ | 1.2809 | 0.62 | 15 | 1.2338 |
150
+ | 1.2054 | 0.66 | 16 | 1.2308 |
151
+ | 1.2732 | 0.7 | 17 | 1.2269 |
152
+ | 1.2586 | 0.74 | 18 | 1.2219 |
153
+ | 1.2939 | 0.78 | 19 | 1.2161 |
154
+ | 1.2713 | 0.82 | 20 | 1.2086 |
155
+ | 1.2154 | 0.87 | 21 | 1.2008 |
156
+ | 1.213 | 0.91 | 22 | 1.1917 |
157
+ | 1.2183 | 0.95 | 23 | 1.1813 |
158
+ | 1.1594 | 0.99 | 24 | 1.1701 |
159
+
160
+
161
+ ### Framework versions
162
+
163
+ - PEFT 0.7.2.dev0
164
+ - Transformers 4.37.0.dev0
165
+ - Pytorch 2.0.1+cu117
166
+ - Datasets 2.16.1
167
+ - Tokenizers 0.15.0
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5b9a26a33e771dccdf87643e0178fbfd3b69a586d451c8717a58550fe19b964
3
+ size 319977229