frankmorales2020 commited on
Commit
d5ff321
1 Parent(s): de5038d

Model save

Browse files
Files changed (1) hide show
  1. README.md +4 -46
README.md CHANGED
@@ -14,7 +14,7 @@ model-index:
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
- should proofread and complete it, then remove this comment. -->
18
 
19
  # Mistral-7B-v0.1_Emotion
20
 
@@ -22,11 +22,7 @@ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https
22
 
23
  ## Model description
24
 
25
- https://ai.plainenglish.io/fine-tuning-the-mistral-7b-instruct-v0-1-model-with-the-emotion-dataset-c84c50b553dc
26
-
27
- Fine tunning: https://github.com/frank-morales2020/MLxDL/blob/main/FineTuning_Mistral_7b_hfdeployment_dataset_Emotion.ipynb
28
-
29
- Evaluation: https://github.com/frank-morales2020/MLxDL/blob/main/FineTunning_Testing_For_EmotionQADataset.ipynb
30
 
31
  ## Intended uses & limitations
32
 
@@ -34,45 +30,7 @@ More information needed
34
 
35
  ## Training and evaluation data
36
 
37
- https://github.com/frank-morales2020/MLxDL/blob/main/FineTunning_Testing_For_EmotionQADataset.ipynb
38
-
39
- The following hyperparameters were used during training:
40
-
41
- -----------
42
-
43
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2
44
- total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999)
45
- and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
46
-
47
- num_epochs: 1
48
-
49
- NOTE: test - Accuracy (Eval dataset and predict) for a sample of 2000: 59.45%
50
-
51
- -----------
52
-
53
- The following hyperparameters were used during training:
54
-
55
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2
56
- total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999)
57
- and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
58
-
59
- num_epochs: 25
60
-
61
- NOTE: test - Accuracy (Eval dataset and predict) for a sample of 2000: 79.95%
62
-
63
- -----------
64
-
65
- The following hyperparameters were used during training:
66
-
67
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2
68
- total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999)
69
- and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
70
-
71
- num_epochs: 100
72
-
73
- NOTE: test - Accuracy (Eval dataset and predict) for a sample of 2000: TBD
74
-
75
- -----------
76
 
77
  ## Training procedure
78
 
@@ -88,7 +46,7 @@ The following hyperparameters were used during training:
88
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
89
  - lr_scheduler_type: constant
90
  - lr_scheduler_warmup_ratio: 0.03
91
- - num_epochs: 25
92
 
93
  ### Training results
94
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # Mistral-7B-v0.1_Emotion
20
 
 
22
 
23
  ## Model description
24
 
25
+ More information needed
 
 
 
 
26
 
27
  ## Intended uses & limitations
28
 
 
30
 
31
  ## Training and evaluation data
32
 
33
+ More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Training procedure
36
 
 
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_ratio: 0.03
49
+ - num_epochs: 40
50
 
51
  ### Training results
52