Edit model card

Pegasus-x-sumstew

Model description

This model is a fine-tuned version of the Pegasus-x-large model on a filtered subset of a mixture of CNN-Dailymail, Samsum, Booksum and Laysum datasets. It can generate abstractive summaries of long texts.

Intended uses & limitations

This model can be used for summarizing long texts in English, such as academic transcripts, meeting minutes, or literature. It is not intended for summarizing short texts, such as tweets, headlines, or captions. The model may produce inaccurate or biased summaries if the input text contains factual errors, slang, or offensive language.

How to use

You can use this model with the pipeline function from the transformers library:

from transformers import pipeline

summarizer = pipeline("summarization", "joemgu/pegasus-x-sumstew")
text = "Alice was beginning to get very tired of sitting by her sister on the bank, and of having nothing to do: once or twice she had peeped into the book her sister was reading, but it had no pictures or conversations in it, 'and what is the use of a book,' thought Alice 'without pictures or conversations?' So she was considering in her own mind (as well as she could, for the hot day made her feel very sleepy and stupid), whether the pleasure of making a daisy-chain would be worth the trouble of getting up and picking the daisies, when suddenly a White Rabbit with pink eyes ran close by her. There was nothing so very remarkable in that; nor did Alice think it so very much out of the way to hear the Rabbit say to itself, 'Oh dear! Oh dear! I shall be late!' (when she thought it over afterwards, it occurred to her that she ought to have wondered at this, but at the time it all seemed quite natural); but when the Rabbit actually took a watch out of its waistcoat-pocket, and looked at it, and then hurried on, Alice started to her feet, for it flashed across her mind that she had never before seen a rabbit with either a waistcoat-pocket, or a watch to take out of it, and burning with curiosity, she ran across the field after it, and fortunately was just in time to see it pop down a large rabbit-hole under the hedge. In another moment down went Alice after it, never once considering how in the world she was to get out again."
summary = summarizer(text,
                     num_beams=8,
                     repetition_penalty=3.5,
                     no_repeat_ngram_size=4,
                     encoder_no_repeat_ngram_size=4
    )[0]["summary_text"]
print(summary)

Output:

Alice is a bored and curious girl who follows a White Rabbit with a watch into a rabbit-hole. She enters a strange world where she has many adventures and meets many peculiar creatures.

Training data

The model was fine-tuned on a filtered subset of a mixture of CNN-Dailymail, Samsum, Booksum and Laysum datasets. These datasets contain various types of texts and their abstractive summaries. The subset was selected to include only texts that are longer than 1000 words and have summaries that are shorter than 100 words. The total size of the subset is about 150k examples.

Evaluation results

TODO

Limitations and bias

The model may have inherited some limitations and biases from the pre-trained Pegasus-x-large model and the fine-tuning datasets. Some possible sources of bias are:

  • The pre-trained Pegasus-x-large model was trained on a large corpus of English texts from various sources, which may not reflect the diversity and nuances of different languages and cultures.
  • The fine-tuning datasets were collected from different domains and genres, which may have their own stylistic conventions and perspectives on certain topics and events.
  • The fine-tuning datasets only contain abstractive summaries, which may not capture all the important information and nuances of the original texts.
  • The fine-tuning datasets only cover texts from certain time periods and sources, which may not reflect the current state of affairs and trends.

Therefore, users should be aware of these limitations and biases when using this model and evaluate its performance and suitability for their specific use cases.

Downloads last month
23
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train Joemgu/pegasus-x-sumstew