vandeju commited on
Commit
7b4d5b6
1 Parent(s): 0afe8b8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md CHANGED
@@ -43,3 +43,31 @@ It achieves the following results on the evaluation set:
43
  ## Training procedure
44
 
45
  This model was trained with QLoRa in bfloat16 with Flash Attention 2 on one A100 PCIe, using the sft script from the [alignment handbook](https://github.com/huggingface/alignment-handbook/) on [RunPod](https://www.runpod.io/).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  ## Training procedure
44
 
45
  This model was trained with QLoRa in bfloat16 with Flash Attention 2 on one A100 PCIe, using the sft script from the [alignment handbook](https://github.com/huggingface/alignment-handbook/) on [RunPod](https://www.runpod.io/).
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 0.0002
51
+ - train_batch_size: 3
52
+ - eval_batch_size: 6
53
+ - seed: 42
54
+ - distributed_type: multi-GPU
55
+ - gradient_accumulation_steps: 2
56
+ - total_train_batch_size: 6
57
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
58
+ - lr_scheduler_type: cosine
59
+ - lr_scheduler_warmup_ratio: 0.1
60
+ - num_epochs: 1
61
+
62
+ ### Framework versions
63
+
64
+ - PEFT 0.11.1
65
+ - Transformers 4.41.2
66
+ - Pytorch 2.2.0+cu121
67
+ - Datasets 2.19.1
68
+ - Tokenizers 0.19.1
69
+
70
+ ### Model Developer
71
+
72
+ The Mistral-7B-v0.3-Instruct model, on which this model is based, was created by [Mistral AI](https://huggingface.co/mistralai).
73
+ The finetuning was done by [Julien Van den Avenne](https://huggingface.co/vandeju).