ninyx commited on
Commit
cf51f09
1 Parent(s): 6a50e37

Model save

Browse files
Files changed (2) hide show
  1. README.md +74 -0
  2. adapter_model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - trl
6
+ - sft
7
+ - generated_from_trainer
8
+ base_model: mistralai/Mistral-7B-Instruct-v0.3
9
+ datasets:
10
+ - generator
11
+ metrics:
12
+ - bleu
13
+ - rouge
14
+ model-index:
15
+ - name: Mistral-7B-Instruct-v0.3-advisegpt-v0.4
16
+ results: []
17
+ ---
18
+
19
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
20
+ should probably proofread and complete it, then remove this comment. -->
21
+
22
+ # Mistral-7B-Instruct-v0.3-advisegpt-v0.4
23
+
24
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the generator dataset.
25
+ It achieves the following results on the evaluation set:
26
+ - Loss: 0.0776
27
+ - Bleu: {'bleu': 0.9592766854579555, 'precisions': [0.9778672968005702, 0.9629777800504739, 0.952562376464522, 0.9440303244645156], 'brevity_penalty': 1.0, 'length_ratio': 1.0002070868729431, 'translation_length': 666525, 'reference_length': 666387}
28
+ - Rouge: {'rouge1': 0.9765393241338379, 'rouge2': 0.960274899679536, 'rougeL': 0.9752854409851488, 'rougeLsum': 0.9763366883065228}
29
+ - Exact Match: {'exact_match': 0.0}
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+
43
+ ## Training procedure
44
+
45
+ ### Training hyperparameters
46
+
47
+ The following hyperparameters were used during training:
48
+ - learning_rate: 2e-05
49
+ - train_batch_size: 1
50
+ - eval_batch_size: 1
51
+ - seed: 42
52
+ - gradient_accumulation_steps: 15
53
+ - total_train_batch_size: 15
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: cosine
56
+ - num_epochs: 3
57
+ - mixed_precision_training: Native AMP
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Rouge | Exact Match |
62
+ |:-------------:|:------:|:----:|:---------------:|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|:---------------------------------------------------------------------------------------------------------------------------:|:--------------------:|
63
+ | 0.0592 | 0.9998 | 2664 | 0.0792 | {'bleu': 0.957140829496306, 'precisions': [0.9770110285842899, 0.9611535701983837, 0.9499650178830994, 0.9408134298916666], 'brevity_penalty': 1.0, 'length_ratio': 1.0000945396593872, 'translation_length': 666450, 'reference_length': 666387} | {'rouge1': 0.9756420869808171, 'rouge2': 0.958253583847128, 'rougeL': 0.9741670140375769, 'rougeLsum': 0.9753898276329086} | {'exact_match': 0.0} |
64
+ | 0.0518 | 2.0000 | 5329 | 0.0776 | {'bleu': 0.9592766854579555, 'precisions': [0.9778672968005702, 0.9629777800504739, 0.952562376464522, 0.9440303244645156], 'brevity_penalty': 1.0, 'length_ratio': 1.0002070868729431, 'translation_length': 666525, 'reference_length': 666387} | {'rouge1': 0.9765393241338379, 'rouge2': 0.960274899679536, 'rougeL': 0.9752854409851488, 'rougeLsum': 0.9763366883065228} | {'exact_match': 0.0} |
65
+ | 0.0439 | 2.9994 | 7992 | 0.0830 | {'bleu': 0.9593680325138967, 'precisions': [0.97789654044549, 0.9630261327317164, 0.9526617494511856, 0.9442157972615742], 'brevity_penalty': 1.0, 'length_ratio': 1.0001725723941193, 'translation_length': 666502, 'reference_length': 666387} | {'rouge1': 0.9766709553577743, 'rouge2': 0.9604006931620985, 'rougeL': 0.9753845279467352, 'rougeLsum': 0.9764641972952484} | {'exact_match': 0.0} |
66
+
67
+
68
+ ### Framework versions
69
+
70
+ - PEFT 0.10.0
71
+ - Transformers 4.40.2
72
+ - Pytorch 2.2.0
73
+ - Datasets 2.19.1
74
+ - Tokenizers 0.19.1
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be7b9f347c07312b607e6f3e5d7d3c76813f21e86917dd3d92c0b564291a355c
3
  size 1744865936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3fcbbecbb993941bf5e9f865ba510e3c060110b2b2f4ccbb961e8920dde2181
3
  size 1744865936