Pre-trained BART Model fine-tune on WikiLingua dataset
The repository for the fine-tuned BART model (by sshleifer) using the wiki_lingua dataset (English)
Purpose: Examine the performance of a fine-tuned model research purposes
Observation:
- Pre-trained model was trained on the XSum dataset, which summarize a not-too-long documents into one-liner summary
- Fine-tuning this model using WikiLingua is appropriate since the summaries for that dataset are also short
- In the end, however, the model cannot capture much clearer key points, but instead it mostly extracts the opening sentence
- Some data pre-processing and models' hyperparameter are also need to be tuned more properly.
- Downloads last month
- 15
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.