vedu commited on
Commit
60e244e
1 Parent(s): ea7b3a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -4
README.md CHANGED
@@ -5,16 +5,14 @@ language: en
5
 
6
  # BART (large-sized model)
7
 
8
- BART model pre-trained on English language. It was introduced in the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Lewis et al. and first released in [this repository](https://github.com/pytorch/fairseq/tree/master/examples/bart).
9
-
10
- Disclaimer: The team releasing BART did not write a model card for this model so this model card has been written by the Hugging Face team.
11
-
12
  ## Model description
13
 
14
  BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.
15
 
16
  BART is particularly effective when fine-tuned for text generation (e.g. summarization, translation) but also works well for comprehension tasks (e.g. text classification, question answering).
17
 
 
 
18
  ## Intended uses & limitations
19
 
20
  There have been quite a few issues related to finetuning BART for text generation, and this repo implements solution discussed in [#15559](https://github.com/huggingface/transformers/issues/15559).
 
5
 
6
  # BART (large-sized model)
7
 
 
 
 
 
8
  ## Model description
9
 
10
  BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.
11
 
12
  BART is particularly effective when fine-tuned for text generation (e.g. summarization, translation) but also works well for comprehension tasks (e.g. text classification, question answering).
13
 
14
+ Weights shared here are effectively from facebook/bart-large but with added noise for BOS embedding to assist the finetuning.
15
+
16
  ## Intended uses & limitations
17
 
18
  There have been quite a few issues related to finetuning BART for text generation, and this repo implements solution discussed in [#15559](https://github.com/huggingface/transformers/issues/15559).