Update README.md
Browse files
README.md
CHANGED
@@ -3,7 +3,7 @@ license: mit
|
|
3 |
library_name: transformers
|
4 |
pipeline_tag: text-generation
|
5 |
---
|
6 |
-
# GPT-2
|
7 |
|
8 |
## Table of Contents
|
9 |
- [Model Details](#model-details)
|
@@ -19,7 +19,7 @@ pipeline_tag: text-generation
|
|
19 |
|
20 |
## Model Details
|
21 |
|
22 |
-
**Model Description:** GPT-2 Large is the **
|
23 |
|
24 |
- **Developed by:** OpenAI, see [associated research paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) and [GitHub repo](https://github.com/openai/gpt-2) for model developers.
|
25 |
- **Model Type:** Transformer-based language model
|
|
|
3 |
library_name: transformers
|
4 |
pipeline_tag: text-generation
|
5 |
---
|
6 |
+
# GPT-2 Small
|
7 |
|
8 |
## Table of Contents
|
9 |
- [Model Details](#model-details)
|
|
|
19 |
|
20 |
## Model Details
|
21 |
|
22 |
+
**Model Description:** GPT-2 Large is the **135M parameter** version of GPT-2, a transformer-based language model created and released by OpenAI. The model is a pretrained model on English language using a causal language modeling (CLM) objective.
|
23 |
|
24 |
- **Developed by:** OpenAI, see [associated research paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) and [GitHub repo](https://github.com/openai/gpt-2) for model developers.
|
25 |
- **Model Type:** Transformer-based language model
|