historical note?

#5
by julien-c HF staff - opened
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -3,7 +3,7 @@ language: en
3
  license: mit
4
  ---
5
 
6
- # OpenAI GPT
7
 
8
  ## Table of Contents
9
  - [Model Details](#model-details)
@@ -19,7 +19,7 @@ license: mit
19
 
20
  ## Model Details
21
 
22
- **Model Description:** `openai-gpt` is a transformer-based language model created and released by OpenAI. The model is a causal (unidirectional) transformer pre-trained using language modeling on a large corpus with long range dependencies.
23
 
24
  - **Developed by:** Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever. See [associated research paper](https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf) and [GitHub repo](https://github.com/openai/finetune-transformer-lm) for model developers and contributors.
25
  - **Model Type:** Transformer-based language model
 
3
  license: mit
4
  ---
5
 
6
+ # OpenAI GPT 1
7
 
8
  ## Table of Contents
9
  - [Model Details](#model-details)
 
19
 
20
  ## Model Details
21
 
22
+ **Model Description:** `openai-gpt` (a.k.a. "GPT-1") is the first transformer-based language model created and released by OpenAI. The model is a causal (unidirectional) transformer pre-trained using language modeling on a large corpus with long range dependencies.
23
 
24
  - **Developed by:** Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever. See [associated research paper](https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf) and [GitHub repo](https://github.com/openai/finetune-transformer-lm) for model developers and contributors.
25
  - **Model Type:** Transformer-based language model