Samsum Finetuning
Collection
Finetune dense~switch models on Samsum
•
7 items
•
Updated
This model is a fine-tuned version of google/switch-base-8 on the samsum samsum dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
---|---|---|---|---|---|---|---|---|
2.5688 | 0.4343 | 400 | 2.0174 | 39.045 | 18.0039 | 32.4777 | 36.2922 | 20.7787 |
2.054 | 0.8686 | 800 | 1.7355 | 45.9504 | 22.2553 | 37.9856 | 42.4096 | 24.9254 |
1.9326 | 1.3029 | 1200 | 1.6762 | 46.4474 | 22.8747 | 39.0668 | 42.8255 | 20.0905 |
1.8121 | 1.7372 | 1600 | 1.6212 | 47.4383 | 23.9879 | 39.9156 | 44.1151 | 21.1174 |
1.6303 | 2.1716 | 2000 | 1.6068 | 49.5797 | 25.5351 | 41.2855 | 45.9999 | 24.8362 |
1.6817 | 2.6059 | 2400 | 1.5734 | 49.0904 | 24.9926 | 41.2085 | 45.4779 | 23.0905 |
1.4335 | 3.0402 | 2800 | 1.5943 | 49.5091 | 25.7276 | 41.7665 | 45.7573 | 22.1553 |
1.5042 | 3.4745 | 3200 | 1.5807 | 49.1947 | 25.6961 | 41.2511 | 45.5553 | 22.6149 |
1.4447 | 3.9088 | 3600 | 1.5747 | 50.1246 | 26.1223 | 41.8475 | 46.3095 | 25.5709 |
1.3638 | 4.3431 | 4000 | 1.6004 | 50.655 | 26.3528 | 42.3721 | 46.8937 | 25.0685 |
1.4508 | 4.7774 | 4400 | 1.5741 | 49.9176 | 25.9264 | 41.7646 | 46.0881 | 24.879 |