MLQ-distilbart-bbc / README.md
morenolq's picture
Librarian Bot: Add base_model information to model (#2)
a28ddd5
metadata
license: apache-2.0
tags:
  - distilbart
  - summarization
base_model: sshleifer/distilbart-cnn-12-6
model-index:
  - name: MLQ-distilbart-bbc
    results:
      - task:
          type: summarization
          name: Summarization
        dataset:
          name: bbc
          type: bbc
          config: default
          split: test
        metrics:
          - type: rouge
            value: 61.43
            name: ROUGE-2
            verified: false

MLQ-distilbart-bbc

This model is a fine-tuned version of sshleifer/distilbart-cnn-12-6 on the BBC News Summary dataset (https://www.kaggle.com/pariza/bbc-news-summary).

The model has been generated as part of the in-lab practice of Deep NLP course currently held at Politecnico di Torino.

Training parameters:

  • num_train_epochs=2
  • fp16=True
  • per_device_train_batch_size=1
  • warmup_steps=10
  • weight_decay=0.01
  • max_seq_length=100