lewtun's picture
lewtun
HF staff
add model 4340696
1 ---
2 license: apache-2.0
3 tags:
4 - generated_from_trainer
5 datasets:
6 - mlsum
7 metrics:
8 - rouge
9 model-index:
10 - name: mt5-small-finetuned-mlsum
11 results:
12 - task:
13 name: Sequence-to-sequence Language Modeling
14 type: text2text-generation
15 dataset:
16 name: mlsum
17 type: mlsum
18 args: es
19 metrics:
20 - name: Rouge1
21 type: rouge
22 value: 1.1475
23 ---
24
25 <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26 should probably proofread and complete it, then remove this comment. -->
27
28 # mt5-small-finetuned-mlsum
29
30 This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on the mlsum dataset.
31 It achieves the following results on the evaluation set:
32 - Loss: nan
33 - Rouge1: 1.1475
34 - Rouge2: 0.1284
35 - Rougel: 1.0634
36 - Rougelsum: 1.0778
37 - Gen Len: 3.7939
38
39 ## Model description
40
41 More information needed
42
43 ## Intended uses & limitations
44
45 More information needed
46
47 ## Training and evaluation data
48
49 More information needed
50
51 ## Training procedure
52
53 ### Training hyperparameters
54
55 The following hyperparameters were used during training:
56 - learning_rate: 2e-05
57 - train_batch_size: 8
58 - eval_batch_size: 8
59 - seed: 42
60 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61 - lr_scheduler_type: linear
62 - num_epochs: 1
63 - mixed_precision_training: Native AMP
64
65 ### Training results
66
67 | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
68 |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
69 | nan | 1.0 | 808 | nan | 1.1475 | 0.1284 | 1.0634 | 1.0778 | 3.7939 |
70
71
72 ### Framework versions
73
74 - Transformers 4.10.3
75 - Pytorch 1.9.0+cu102
76 - Datasets 1.12.1
77 - Tokenizers 0.10.3
78