alvations commited on
Commit
2e5763f
1 Parent(s): 974be4b

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -4
README.md CHANGED
@@ -2,6 +2,8 @@
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: mt5-aym-lex-try3
7
  results: []
@@ -12,7 +14,12 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # mt5-aym-lex-try3
14
 
15
- This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on the None dataset.
 
 
 
 
 
16
 
17
  ## Model description
18
 
@@ -37,12 +44,23 @@ The following hyperparameters were used during training:
37
  - seed: 42
38
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
  - lr_scheduler_type: linear
40
- - lr_scheduler_warmup_steps: 500
41
- - training_steps: 200000
 
 
 
 
 
 
 
 
 
 
 
42
 
43
  ### Framework versions
44
 
45
- - Transformers 4.28.1
46
  - Pytorch 2.0.0+cu118
47
  - Datasets 2.12.0
48
  - Tokenizers 0.13.3
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
+ metrics:
6
+ - bleu
7
  model-index:
8
  - name: mt5-aym-lex-try3
9
  results: []
 
14
 
15
  # mt5-aym-lex-try3
16
 
17
+ This model is a fine-tuned version of [alvations/mt5-aym-lex](https://huggingface.co/alvations/mt5-aym-lex) on the None dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.1687
20
+ - Chrf: 20.013
21
+ - Bleu: 4.4299
22
+ - Gen Len: 17.1254
23
 
24
  ## Model description
25
 
 
44
  - seed: 42
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_steps: 5
48
+ - training_steps: 100
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss | Chrf | Bleu | Gen Len |
53
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|
54
+ | 0.2142 | 0.0 | 20 | 0.1750 | 20.6672 | 4.6405 | 17.0733 |
55
+ | 0.1881 | 0.01 | 40 | 0.1704 | 20.1942 | 4.484 | 17.1141 |
56
+ | 0.2355 | 0.01 | 60 | 0.1690 | 20.0147 | 4.4349 | 17.1302 |
57
+ | 0.1939 | 0.02 | 80 | 0.1688 | 20.0311 | 4.4427 | 17.1319 |
58
+ | 0.1985 | 0.02 | 100 | 0.1687 | 20.013 | 4.4299 | 17.1254 |
59
+
60
 
61
  ### Framework versions
62
 
63
+ - Transformers 4.29.1
64
  - Pytorch 2.0.0+cu118
65
  - Datasets 2.12.0
66
  - Tokenizers 0.13.3