Ezi commited on
Commit
8b065d9
1 Parent(s): a4f8f3e

Model Card

Browse files

Hi!👋
This PR has a some additional information for the model card, based on the format we are using as part of our effort to standardise model cards at Hugging Face. Feel free to merge if you are ok with the changes! (cc

@Marissa



@Meg



@Nazneen

)

Files changed (1) hide show
  1. README.md +52 -1
README.md CHANGED
@@ -8,11 +8,62 @@ datasets:
8
  - xsum
9
  thumbnail: https://huggingface.co/front/thumbnails/distilbart_medium.png
10
  ---
 
11
 
12
- ### Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  This checkpoint should be loaded into `BartForConditionalGeneration.from_pretrained`. See the [BART docs](https://huggingface.co/transformers/model_doc/bart.html?#transformers.BartForConditionalGeneration) for more information.
15
 
 
 
16
  ### Metrics for DistilBART models
17
 
18
  | Model Name | MM Params | Inference Time (MS) | Speedup | Rouge 2 | Rouge-L |
 
8
  - xsum
9
  thumbnail: https://huggingface.co/front/thumbnails/distilbart_medium.png
10
  ---
11
+ # Distilbart-cnn-12-6
12
 
13
+ ## Table of Contents
14
+ - [Model Details](#model-details)
15
+ - [How to Get Started With the Model](#how-to-get-started-with-the-model)
16
+ - [Uses](#uses)
17
+ - [Risks, Limitations and Biases](#risks-limitations-and-biases)
18
+ - [Training](#training)
19
+ - [Evaluation](#evaluation)
20
+
21
+ ## Model Details
22
+ - **Model Description:**
23
+ - **Developed by:** Sam Shleifer
24
+ - **Model Type:** Summarization
25
+ - **Language(s):** English
26
+ - **License:** Apache-2.0
27
+ - **Parent Model:** See the [BART lage CNN model](https://huggingface.co/facebook/bart-large-cnn) for more information about the BART large-sized model which is similarly trained on [CNN Dailymail](https://huggingface.co/datasets/cnn_dailymail) dataset.
28
+ - **Resources for more information:**
29
+ - [Bart Document](https://huggingface.co/docs/transformers/model_doc/bart#transformers.BartForConditionalGeneration)
30
+
31
+
32
+ ## How to Get Started With the Model
33
+ ```python
34
+ ​​from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
35
+
36
+ tokenizer = AutoTokenizer.from_pretrained("sshleifer/distilbart-cnn-12-6")
37
+
38
+ model = AutoModelForSeq2SeqLM.from_pretrained("sshleifer/distilbart-cnn-12-6")
39
+
40
+ ```
41
+
42
+
43
+ ## Uses
44
+
45
+
46
+ #### Direct Use
47
+ This model can be used for text summerzation.
48
+
49
+
50
+ ## Risks, Limitations and Biases
51
+ ### Limitations
52
+ This model makes use of the [CNN Dailymail](https://huggingface.co/datasets/cnn_dailymail) dataset, which is an English-language dataset containing just over 300k unique news articles as written by journalists at CNN and the Daily Mail.
53
+ The BCP-47 code for English as generally spoken in the United States is en-US and the BCP-47 code for English as generally spoken in the United Kingdom is en-GB. It is unknown if other varieties of English are represented in the data.
54
+
55
+ ### Biases
56
+ [Bordia and Bowman (2019)](https://www.aclweb.org/anthology/N19-3002.pdf) explore measuring gender bias and debiasing techniques in the CNN / Dailymail dataset, the Penn Treebank, and WikiText-2. They find the CNN / Dailymail dataset to have a slightly lower gender bias based on their metric compared to the other datasets, but still show evidence of gender bias when looking at words such as 'fragile'.
57
+
58
+ Further information e.g in regards to uses, out-of-scope uses, training procedure for the CNN Dailymail dataset are available within its [dataset card](https://huggingface.co/datasets/cnn_dailymail).
59
+
60
+
61
+ ## Training
62
 
63
  This checkpoint should be loaded into `BartForConditionalGeneration.from_pretrained`. See the [BART docs](https://huggingface.co/transformers/model_doc/bart.html?#transformers.BartForConditionalGeneration) for more information.
64
 
65
+ ## Evaluation
66
+
67
  ### Metrics for DistilBART models
68
 
69
  | Model Name | MM Params | Inference Time (MS) | Speedup | Rouge 2 | Rouge-L |