AmalNlal commited on
Commit
1895776
·
verified ·
1 Parent(s): 18b6186

End of training

Browse files
Files changed (1) hide show
  1. README.md +25 -10
README.md CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
- - Loss: 7.9694
17
 
18
  ## Model description
19
 
@@ -32,23 +32,38 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 0.1
36
- - train_batch_size: 16
37
- - eval_batch_size: 16
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
- - training_steps: 500
42
 
43
  ### Training results
44
 
45
  | Training Loss | Epoch | Step | Validation Loss |
46
  |:-------------:|:-----:|:----:|:---------------:|
47
- | 14.7772 | 0.06 | 100 | 8.6683 |
48
- | 9.1925 | 0.12 | 200 | 9.1429 |
49
- | 8.6946 | 0.19 | 300 | 10.0975 |
50
- | 8.3986 | 0.25 | 400 | 8.3313 |
51
- | 8.0251 | 0.31 | 500 | 7.9694 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
 
53
 
54
  ### Framework versions
 
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Loss: 7.7544
17
 
18
  ## Model description
19
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 5e-05
36
+ - train_batch_size: 8
37
+ - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
+ - training_steps: 1000
42
 
43
  ### Training results
44
 
45
  | Training Loss | Epoch | Step | Validation Loss |
46
  |:-------------:|:-----:|:----:|:---------------:|
47
+ | No log | 0.02 | 50 | 9.3840 |
48
+ | 9.3687 | 0.03 | 100 | 8.6014 |
49
+ | 9.3687 | 0.05 | 150 | 8.2440 |
50
+ | 8.0254 | 0.06 | 200 | 8.0843 |
51
+ | 8.0254 | 0.08 | 250 | 8.0234 |
52
+ | 7.8649 | 0.09 | 300 | 7.9828 |
53
+ | 7.8649 | 0.11 | 350 | 7.9550 |
54
+ | 7.732 | 0.12 | 400 | 7.9101 |
55
+ | 7.732 | 0.14 | 450 | 7.8946 |
56
+ | 7.6192 | 0.15 | 500 | 7.8525 |
57
+ | 7.6192 | 0.17 | 550 | 7.8461 |
58
+ | 7.6378 | 0.18 | 600 | 7.8285 |
59
+ | 7.6378 | 0.2 | 650 | 7.8182 |
60
+ | 7.6338 | 0.22 | 700 | 7.7917 |
61
+ | 7.6338 | 0.23 | 750 | nan |
62
+ | 7.5994 | 0.25 | 800 | 7.7837 |
63
+ | 7.5994 | 0.26 | 850 | 7.7596 |
64
+ | 7.5323 | 0.28 | 900 | 7.7634 |
65
+ | 7.5323 | 0.29 | 950 | 7.7750 |
66
+ | 7.5914 | 0.31 | 1000 | 7.7544 |
67
 
68
 
69
  ### Framework versions