avi-skowron
commited on
Commit
•
2ea01c5
1
Parent(s):
7054101
fixed dataset tag
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@ tags:
|
|
7 |
- pythia
|
8 |
license: apache-2.0
|
9 |
datasets:
|
10 |
-
- EleutherAI/
|
11 |
---
|
12 |
|
13 |
The *Pythia Scaling Suite* is a collection of models developed to facilitate
|
@@ -16,7 +16,7 @@ interpretability research. It contains two sets of eight models of sizes
|
|
16 |
models: one trained on the Pile, and one trained on the Pile after the dataset
|
17 |
has been globally deduplicated. All 8 model sizes are trained on the exact
|
18 |
same data, in the exact same order. All Pythia models are available
|
19 |
-
[on Hugging Face](https://huggingface.co/
|
20 |
|
21 |
The Pythia model suite was deliberately designed to promote scientific
|
22 |
research on large language models, especially interpretability research.
|
|
|
7 |
- pythia
|
8 |
license: apache-2.0
|
9 |
datasets:
|
10 |
+
- EleutherAI/the_pile_deduplicated
|
11 |
---
|
12 |
|
13 |
The *Pythia Scaling Suite* is a collection of models developed to facilitate
|
|
|
16 |
models: one trained on the Pile, and one trained on the Pile after the dataset
|
17 |
has been globally deduplicated. All 8 model sizes are trained on the exact
|
18 |
same data, in the exact same order. All Pythia models are available
|
19 |
+
[on Hugging Face](https://huggingface.co/models?other=pythia).
|
20 |
|
21 |
The Pythia model suite was deliberately designed to promote scientific
|
22 |
research on large language models, especially interpretability research.
|