Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,16 @@ license: apache-2.0
|
|
5 |
---
|
6 |
This is unsloth/llama-3-8b-Instruct trained on the Replete-AI/code-test-dataset using the code bellow with unsloth and google colab with under 15gb of vram. This training was complete in about 40 minutes total.
|
7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
```Python
|
9 |
%%capture
|
10 |
import torch
|
@@ -142,6 +152,7 @@ trainer = SFTTrainer(
|
|
142 |
)
|
143 |
```
|
144 |
|
|
|
145 |
```Python
|
146 |
trainer_stats = trainer.train()
|
147 |
model.save_pretrained_merged("model", tokenizer, save_method = "merged_16bit",)
|
|
|
5 |
---
|
6 |
This is unsloth/llama-3-8b-Instruct trained on the Replete-AI/code-test-dataset using the code bellow with unsloth and google colab with under 15gb of vram. This training was complete in about 40 minutes total.
|
7 |
|
8 |
+
For anyone that is new to coding and training Ai, all your really have to edit is
|
9 |
+
|
10 |
+
1. (max_seq_length = 8192) To match the max tokens of the dataset or model you are using
|
11 |
+
2. (model_name = "unsloth/llama-3-8b-Instruct",) Change what model you are finetuning, this setup is specifically for llama-3-8b
|
12 |
+
3. (alpaca_prompt =) Change the prompt format, this one is setup to meet llama-3-8b-instruct format, but match it to your specifications.
|
13 |
+
4. (dataset = load_dataset("Replete-AI/code-test-dataset", split = "train")) What dataset you are using from huggingface
|
14 |
+
5. (model.push_to_hub_merged("rombodawg/test_dataset_Codellama-3-8B", tokenizer, save_method = "merged_16bit", token = ""))
|
15 |
+
For the above you need to change "rombodawg" to your Hugginface name, "test_dataset_Codellama-3-8B" to the model name you want saved as, and in token = "" you need to put your huggingface write token so the model can be saved.
|
16 |
+
|
17 |
+
|
18 |
```Python
|
19 |
%%capture
|
20 |
import torch
|
|
|
152 |
)
|
153 |
```
|
154 |
|
155 |
+
|
156 |
```Python
|
157 |
trainer_stats = trainer.train()
|
158 |
model.save_pretrained_merged("model", tokenizer, save_method = "merged_16bit",)
|