--- language: - en tags: - summarization license: mit datasets: - wiki_lingua metrics: - rouge --- #### Pre-trained BART Model fine-tune on WikiLingua dataset The repository for the fine-tuned BART model (by sshleifer) using the **wiki_lingua** dataset (English) **Purpose:** Examine the performance of a fine-tuned model research purposes **Observation:** - Pre-trained model was trained on the XSum dataset, which summarize a not-too-long documents into one-liner summary - Fine-tuning this model using WikiLingua is appropriate since the summaries for that dataset are also short - In the end, however, the model cannot capture much clearer key points, but instead it mostly extracts the opening sentence - Some data pre-processing and models' hyperparameter are also need to be tuned more properly.