rombodawg commited on
Commit
b4985a4
1 Parent(s): 8fc73c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -5,6 +5,16 @@ license: apache-2.0
5
  ---
6
  This is unsloth/llama-3-8b-Instruct trained on the Replete-AI/code-test-dataset using the code bellow with unsloth and google colab with under 15gb of vram. This training was complete in about 40 minutes total.
7
 
 
 
 
 
 
 
 
 
 
 
8
  ```Python
9
  %%capture
10
  import torch
@@ -142,6 +152,7 @@ trainer = SFTTrainer(
142
  )
143
  ```
144
 
 
145
  ```Python
146
  trainer_stats = trainer.train()
147
  model.save_pretrained_merged("model", tokenizer, save_method = "merged_16bit",)
 
5
  ---
6
  This is unsloth/llama-3-8b-Instruct trained on the Replete-AI/code-test-dataset using the code bellow with unsloth and google colab with under 15gb of vram. This training was complete in about 40 minutes total.
7
 
8
+ For anyone that is new to coding and training Ai, all your really have to edit is
9
+
10
+ 1. (max_seq_length = 8192) To match the max tokens of the dataset or model you are using
11
+ 2. (model_name = "unsloth/llama-3-8b-Instruct",) Change what model you are finetuning, this setup is specifically for llama-3-8b
12
+ 3. (alpaca_prompt =) Change the prompt format, this one is setup to meet llama-3-8b-instruct format, but match it to your specifications.
13
+ 4. (dataset = load_dataset("Replete-AI/code-test-dataset", split = "train")) What dataset you are using from huggingface
14
+ 5. (model.push_to_hub_merged("rombodawg/test_dataset_Codellama-3-8B", tokenizer, save_method = "merged_16bit", token = ""))
15
+ For the above you need to change "rombodawg" to your Hugginface name, "test_dataset_Codellama-3-8B" to the model name you want saved as, and in token = "" you need to put your huggingface write token so the model can be saved.
16
+
17
+
18
  ```Python
19
  %%capture
20
  import torch
 
152
  )
153
  ```
154
 
155
+
156
  ```Python
157
  trainer_stats = trainer.train()
158
  model.save_pretrained_merged("model", tokenizer, save_method = "merged_16bit",)