Zangs3011 commited on
Commit
d29b0bc
·
verified ·
1 Parent(s): da51287

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +117 -153
README.md CHANGED
@@ -1,199 +1,163 @@
1
  ---
2
- library_name: transformers
3
- tags: []
4
- ---
5
-
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
-
11
-
12
- ## Model Details
13
-
14
- ### Model Description
15
-
16
- <!-- Provide a longer summary of what this model is. -->
17
-
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
-
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
-
36
- ## Uses
37
-
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
-
40
- ### Direct Use
41
-
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
-
44
- [More Information Needed]
45
-
46
- ### Downstream Use [optional]
47
-
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
-
50
- [More Information Needed]
51
-
52
- ### Out-of-Scope Use
53
-
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
-
56
- [More Information Needed]
57
-
58
- ## Bias, Risks, and Limitations
59
-
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
 
62
- [More Information Needed]
63
-
64
- ### Recommendations
65
-
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
-
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
-
70
- ## How to Get Started with the Model
71
-
72
- Use the code below to get started with the model.
73
-
74
- [More Information Needed]
75
-
76
- ## Training Details
77
-
78
- ### Training Data
79
-
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
83
-
84
- ### Training Procedure
85
-
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
-
93
- #### Training Hyperparameters
94
-
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
-
101
- [More Information Needed]
102
-
103
- ## Evaluation
104
-
105
- <!-- This section describes the evaluation protocols and provides the results. -->
106
-
107
- ### Testing Data, Factors & Metrics
108
-
109
- #### Testing Data
110
 
111
- <!-- This should link to a Dataset Card if possible. -->
112
 
113
- [More Information Needed]
114
 
115
- #### Factors
116
 
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
 
118
 
119
- [More Information Needed]
120
 
121
- #### Metrics
122
 
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
 
 
 
 
 
124
 
125
- [More Information Needed]
 
 
 
 
 
126
 
127
- ### Results
128
 
129
- [More Information Needed]
 
130
 
131
- #### Summary
 
132
 
 
133
 
 
 
134
 
135
- ## Model Examination [optional]
136
 
137
- <!-- Relevant interpretability work for the model goes here -->
138
 
139
- [More Information Needed]
 
140
 
141
- ## Environmental Impact
 
 
 
142
 
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
 
145
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
 
146
 
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
 
153
- ## Technical Specifications [optional]
154
 
155
- ### Model Architecture and Objective
 
156
 
157
- [More Information Needed]
 
 
 
 
158
 
159
- ### Compute Infrastructure
160
 
161
- [More Information Needed]
 
162
 
163
- #### Hardware
 
 
 
 
 
 
 
 
164
 
165
- [More Information Needed]
166
 
167
- #### Software
 
168
 
169
- [More Information Needed]
 
 
 
 
 
 
 
 
170
 
171
- ## Citation [optional]
172
 
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
 
175
- **BibTeX:**
176
 
177
- [More Information Needed]
 
 
 
 
 
 
 
178
 
179
- **APA:**
 
180
 
181
- [More Information Needed]
 
 
182
 
183
- ## Glossary [optional]
184
 
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
 
 
186
 
187
- [More Information Needed]
188
 
189
- ## More Information [optional]
190
 
191
- [More Information Needed]
 
192
 
193
- ## Model Card Authors [optional]
194
 
195
- [More Information Needed]
196
 
197
- ## Model Card Contact
 
 
 
 
 
 
 
 
 
198
 
199
- [More Information Needed]
 
1
  ---
2
+ language: en
3
+ inference: false
4
+ tags:
5
+ - text-generation
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
+ license: other
8
+ commercial: false
9
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
 
11
 
12
+ # OPT : Open Pre-trained Transformer Language Models
13
 
14
+ OPT was first introduced in [Open Pre-trained Transformer Language Models](https://arxiv.org/abs/2205.01068) and first released in [metaseq's repository](https://github.com/facebookresearch/metaseq) on May 3rd 2022 by Meta AI.
15
 
16
+ **Disclaimer**: The team releasing OPT wrote an official model card, which is available in Appendix D of the [paper](https://arxiv.org/pdf/2205.01068.pdf).
17
+ Content from **this** model card has been written by the Hugging Face team.
18
 
19
+ ## Intro
20
 
21
+ To quote the first two paragraphs of the [official paper](https://arxiv.org/abs/2205.01068)
22
 
23
+ > Large language models trained on massive text collections have shown surprising emergent
24
+ > capabilities to generate text and perform zero- and few-shot learning. While in some cases the public
25
+ > can interact with these models through paid APIs, full model access is currently limited to only a
26
+ > few highly resourced labs. This restricted access has limited researchers’ ability to study how and
27
+ > why these large language models work, hindering progress on improving known challenges in areas
28
+ > such as robustness, bias, and toxicity.
29
 
30
+ > We present Open Pretrained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M
31
+ > to 175B parameters, which we aim to fully and responsibly share with interested researchers. We train the OPT models to roughly match
32
+ > the performance and sizes of the GPT-3 class of models, while also applying the latest best practices in data
33
+ > collection and efficient training. Our aim in developing this suite of OPT models is to enable reproducible and responsible research at scale, and
34
+ > to bring more voices to the table in studying the impact of these LLMs. Definitions of risk, harm, bias, and toxicity, etc., should be articulated by the
35
+ > collective research community as a whole, which is only possible when models are available for study.
36
 
37
+ ## Model description
38
 
39
+ OPT was predominantly pretrained with English text, but a small amount of non-English data is still present within the training corpus via CommonCrawl. The model was pretrained using a causal language modeling (CLM) objective.
40
+ OPT belongs to the same family of decoder-only models like [GPT-3](https://arxiv.org/abs/2005.14165). As such, it was pretrained using the self-supervised causal language modedling objective.
41
 
42
+ For evaluation, OPT follows [GPT-3](https://arxiv.org/abs/2005.14165) by using their prompts and overall experimental setup. For more details, please read
43
+ the [official paper](https://arxiv.org/abs/2205.01068).
44
 
45
+ ## Intended uses & limitations
46
 
47
+ The pretrained-only model can be used for prompting for evaluation of downstream tasks as well as text generation.
48
+ In addition, the model can be fine-tuned on a downstream task using the [CLM example](https://github.com/huggingface/transformers/tree/main/examples/pytorch/language-modeling). For all other OPT checkpoints, please have a look at the [model hub](https://huggingface.co/models?filter=opt).
49
 
50
+ ### How to use
51
 
52
+ You can use this model directly with a pipeline for text generation.
53
 
54
+ ```python
55
+ >>> from transformers import pipeline
56
 
57
+ >>> generator = pipeline('text-generation', model="facebook/opt-350m")
58
+ >>> generator("What are we having for dinner?")
59
+ [{'generated_text': "What are we having for dinner?\nI'm having a steak and a salad.\nI'm""}]
60
+ ```
61
 
62
+ By default, generation is deterministic. In order to use the top-k sampling, please set `do_sample` to `True`.
63
 
64
+ ```python
65
+ >>> from transformers import pipeline, set_seed
66
 
67
+ >>> set_seed(32)
68
+ >>> generator = pipeline('text-generation', model="facebook/opt-350m", do_sample=True)
69
+ >>> generator("What are we having for dinner?")
70
+ [{'generated_text': "What are we having for dinner?\n\nWith spring fast approaching, it’s only appropriate"}]
71
+ ```
72
 
73
+ ### Limitations and bias
74
 
75
+ As mentioned in Meta AI's model card, given that the training data used for this model contains a lot of
76
+ unfiltered content from the internet, which is far from neutral the model is strongly biased :
77
 
78
+ > Like other large language models for which the diversity (or lack thereof) of training
79
+ > data induces downstream impact on the quality of our model, OPT-175B has limitations in terms
80
+ > of bias and safety. OPT-175B can also have quality issues in terms of generation diversity and
81
+ > hallucination. In general, OPT-175B is not immune from the plethora of issues that plague modern
82
+ > large language models.
83
 
84
+ Here's an example of how the model can have biased predictions:
85
 
86
+ ```python
87
+ >>> from transformers import pipeline, set_seed
88
 
89
+ >>> set_seed(32)
90
+ >>> generator = pipeline('text-generation', model="facebook/opt-350m", do_sample=True, num_return_sequences=5)
91
+ >>> generator("The woman worked as a")
92
+ [{'generated_text': "The woman works as a substitute teacher for kids who have missed school. She's the teacher herself,"},
93
+ {'generated_text': 'The woman works as a security guard for another company and does an average of around $13/hour'},
94
+ {'generated_text': 'The woman works as a receptionist, she could at the least wait a week or two for her'},
95
+ {'generated_text': 'The woman works as a manager/intern/career development coach/advisor at a nursing home'},
96
+ {'generated_text': 'The woman works as a maid and has to clean the house but you can tell her to do it'}]
97
+ ```
98
 
99
+ compared to:
100
 
101
+ ```python
102
+ >>> from transformers import pipeline, set_seed
103
 
104
+ >>> set_seed(32)
105
+ >>> generator = pipeline('text-generation', model="facebook/opt-350m", do_sample=True, num_return_sequences=5)
106
+ >>> generator("The man worked as a")
107
+ [{'generated_text': 'The man works as a security guard for the National Football League franchise. He has been a part of'},
108
+ {'generated_text': 'The man works as a security guard for another company and does an excellent job.\nI remember when'},
109
+ {'generated_text': 'The man works as a "secret agent" but at the same time he\'s working to protect the'},
110
+ {'generated_text': 'The man works as a manager/operator/servant for a grocery store and does a lot of'},
111
+ {'generated_text': 'The man works as a bouncer near the scene of the accident - how he could do that is'}]
112
+ ```
113
 
114
+ This bias will also affect all fine-tuned versions of this model.
115
 
116
+ ## Training data
117
 
118
+ The Meta AI team wanted to train this model on a corpus as large as possible. It is composed of the union of the following 5 filtered datasets of textual documents:
119
 
120
+ - BookCorpus, which consists of more than 10K unpublished books,
121
+ - CC-Stories, which contains a subset of CommonCrawl data filtered to match the
122
+ story-like style of Winograd schemas,
123
+ - The Pile, from which * Pile-CC, OpenWebText2, USPTO, Project Gutenberg, OpenSubtitles, Wikipedia, DM Mathematics and HackerNews* were included.
124
+ - Pushshift.io Reddit dataset that was developed in Baumgartner et al. (2020) and processed in
125
+ Roller et al. (2021)
126
+ - CCNewsV2 containing an updated version of the English portion of the CommonCrawl News
127
+ dataset that was used in RoBERTa (Liu et al., 2019b)
128
 
129
+ The final training data contains 180B tokens corresponding to 800GB of data. The validation split was made of 200MB of the pretraining data, sampled proportionally
130
+ to each dataset’s size in the pretraining corpus.
131
 
132
+ The dataset might contains offensive content as parts of the dataset are a subset of
133
+ public Common Crawl data, along with a subset of public Reddit data, which could contain sentences
134
+ that, if viewed directly, can be insulting, threatening, or might otherwise cause anxiety.
135
 
136
+ ### Collection process
137
 
138
+ The dataset was collected form internet, and went through classic data processing algorithms and
139
+ re-formatting practices, including removing repetitive/non-informative text like *Chapter One* or
140
+ *This ebook by Project Gutenberg.*
141
 
142
+ ## Training procedure
143
 
144
+ ### Preprocessing
145
 
146
+ The texts are tokenized using the **GPT2** byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a
147
+ vocabulary size of 50272. The inputs are sequences of 2048 consecutive tokens.
148
 
149
+ The 175B model was trained on 992 *80GB A100 GPUs*. The training duration was roughly ~33 days of continuous training.
150
 
151
+ ### BibTeX entry and citation info
152
 
153
+ ```bibtex
154
+ @misc{zhang2022opt,
155
+ title={OPT: Open Pre-trained Transformer Language Models},
156
+ author={Susan Zhang and Stephen Roller and Naman Goyal and Mikel Artetxe and Moya Chen and Shuohui Chen and Christopher Dewan and Mona Diab and Xian Li and Xi Victoria Lin and Todor Mihaylov and Myle Ott and Sam Shleifer and Kurt Shuster and Daniel Simig and Punit Singh Koura and Anjali Sridhar and Tianlu Wang and Luke Zettlemoyer},
157
+ year={2022},
158
+ eprint={2205.01068},
159
+ archivePrefix={arXiv},
160
+ primaryClass={cs.CL}
161
+ }
162
+ ```
163