File size: 5,143 Bytes
b80ae71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
---
license: apache-2.0
tags:
- generated_from_trainer
datasets:
- wmt16
metrics:
- bleu
model-index:
- name: t5-small-finetuned-de-to-en
  results:
  - task:
      name: Sequence-to-sequence Language Modeling
      type: text2text-generation
    dataset:
      name: wmt16
      type: wmt16
      args: de-en
    metrics:
    - name: Bleu
      type: bleu
      value: 11.3921
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# t5-small-finetuned-de-to-en

This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the wmt16 dataset.
It achieves the following results on the evaluation set:
- Loss: 1.8219
- Bleu: 11.3921
- Gen Len: 17.2471

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 50

### Training results

| Training Loss | Epoch | Step  | Validation Loss | Bleu    | Gen Len |
|:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|
| No log        | 1.0   | 272   | 2.1014          | 5.5136  | 17.4975 |
| 2.5302        | 2.0   | 544   | 2.0258          | 7.4515  | 17.3941 |
| 2.5302        | 3.0   | 816   | 1.9866          | 8.3061  | 17.3441 |
| 2.3778        | 4.0   | 1088  | 1.9602          | 8.9169  | 17.3588 |
| 2.3778        | 5.0   | 1360  | 1.9382          | 9.3651  | 17.3204 |
| 2.2676        | 6.0   | 1632  | 1.9215          | 9.6428  | 17.3588 |
| 2.2676        | 7.0   | 1904  | 1.9067          | 9.8039  | 17.3418 |
| 2.2096        | 8.0   | 2176  | 1.8984          | 9.8545  | 17.3264 |
| 2.2096        | 9.0   | 2448  | 1.8883          | 10.03   | 17.3278 |
| 2.1501        | 10.0  | 2720  | 1.8797          | 10.2398 | 17.3358 |
| 2.1501        | 11.0  | 2992  | 1.8738          | 10.3086 | 17.3258 |
| 2.1025        | 12.0  | 3264  | 1.8677          | 10.3851 | 17.3181 |
| 2.0638        | 13.0  | 3536  | 1.8623          | 10.489  | 17.3014 |
| 2.0638        | 14.0  | 3808  | 1.8574          | 10.4969 | 17.3204 |
| 2.034         | 15.0  | 4080  | 1.8528          | 10.7067 | 17.3178 |
| 2.034         | 16.0  | 4352  | 1.8493          | 10.6867 | 17.3408 |
| 1.9852        | 17.0  | 4624  | 1.8473          | 10.8333 | 17.3198 |
| 1.9852        | 18.0  | 4896  | 1.8429          | 10.8907 | 17.3001 |
| 1.9646        | 19.0  | 5168  | 1.8405          | 10.9049 | 17.3154 |
| 1.9646        | 20.0  | 5440  | 1.8385          | 10.9549 | 17.3124 |
| 1.9264        | 21.0  | 5712  | 1.8361          | 11.0046 | 17.3068 |
| 1.9264        | 22.0  | 5984  | 1.8338          | 11.1415 | 17.2954 |
| 1.9161        | 23.0  | 6256  | 1.8333          | 11.1041 | 17.2938 |
| 1.882         | 24.0  | 6528  | 1.8323          | 11.0801 | 17.2651 |
| 1.882         | 25.0  | 6800  | 1.8309          | 11.157  | 17.2921 |
| 1.8751        | 26.0  | 7072  | 1.8290          | 11.1713 | 17.2951 |
| 1.8751        | 27.0  | 7344  | 1.8279          | 11.2006 | 17.2861 |
| 1.8425        | 28.0  | 7616  | 1.8267          | 11.1761 | 17.2658 |
| 1.8425        | 29.0  | 7888  | 1.8278          | 11.148  | 17.2841 |
| 1.8306        | 30.0  | 8160  | 1.8261          | 11.1765 | 17.2748 |
| 1.8306        | 31.0  | 8432  | 1.8255          | 11.2723 | 17.2454 |
| 1.8229        | 32.0  | 8704  | 1.8247          | 11.2715 | 17.2621 |
| 1.8229        | 33.0  | 8976  | 1.8231          | 11.2896 | 17.2698 |
| 1.7975        | 34.0  | 9248  | 1.8245          | 11.322  | 17.2491 |
| 1.7919        | 35.0  | 9520  | 1.8238          | 11.3854 | 17.2711 |
| 1.7919        | 36.0  | 9792  | 1.8237          | 11.3304 | 17.2634 |
| 1.7781        | 37.0  | 10064 | 1.8225          | 11.3184 | 17.2644 |
| 1.7781        | 38.0  | 10336 | 1.8230          | 11.3382 | 17.2651 |
| 1.7819        | 39.0  | 10608 | 1.8228          | 11.3656 | 17.2658 |
| 1.7819        | 40.0  | 10880 | 1.8221          | 11.3934 | 17.2544 |
| 1.7592        | 41.0  | 11152 | 1.8223          | 11.3625 | 17.2421 |
| 1.7592        | 42.0  | 11424 | 1.8221          | 11.4068 | 17.2511 |
| 1.7529        | 43.0  | 11696 | 1.8224          | 11.4199 | 17.2541 |
| 1.7529        | 44.0  | 11968 | 1.8224          | 11.4051 | 17.2561 |
| 1.7482        | 45.0  | 12240 | 1.8223          | 11.4195 | 17.2504 |
| 1.7461        | 46.0  | 12512 | 1.8220          | 11.3873 | 17.2497 |
| 1.7461        | 47.0  | 12784 | 1.8220          | 11.4214 | 17.2431 |
| 1.739         | 48.0  | 13056 | 1.8218          | 11.3972 | 17.2441 |
| 1.739         | 49.0  | 13328 | 1.8219          | 11.3952 | 17.2457 |
| 1.7362        | 50.0  | 13600 | 1.8219          | 11.3921 | 17.2471 |


### Framework versions

- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3