tags:
- generated_from_trainer
model-index:
- name: pegasus-multi_news-NewsSummarization_BBC
results: []
pegasus-multi_news-NewsSummarization_BBC
This model is a fine-tuned version of google/pegasus-multi_news on the None dataset.
Model description
More information needed
Intended uses & limitations
I used this to improve my skillset. I thank all of authors of the different technologies and dataset(s) for their contributions that have this possible. I am not too worried about getting credit for my part, but make sure to properly cite the authors of the different technologies and dataset(s) as they absolutely deserve credit for their contributions.
Training and evaluation data
Dataset Source: https://www.kaggle.com/datasets/pariza/bbc-news-summary
Training procedure
Here is the link to the script that I created to train this project:
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 50
- num_epochs: 2
Training results
Unfortunately, I did not set the metrics to automatically upload here. They are as follows:
Training Loss | Epoch | Step | rouge1 | rouge2 | rougeL | rougeLsum |
---|---|---|---|---|---|---|
6.41979 | 2.0 | 214 | 0.584474 | 0.463574 | 0.408729 | 0.408431 |
Framework versions
- Transformers 4.21.3
- Pytorch 1.12.1
- Datasets 2.4.0
- Tokenizers 0.12.1