Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@ tags:
|
|
7 |
---
|
8 |
# Model Summery
|
9 |
MobileLLaMA-1.4B-Base is a Transformer with 1.4B billon paramters. We downscale LLaMA to facilitate the off-the-shelf deployment. To make our work reproducible, all
|
10 |
-
the models are trained on 1.3T
|
11 |
|
12 |
We extensively assess our models on two standard natural language benchmarks, for language understanding and common sense reasoning respectively. Experimental results show that our
|
13 |
MobileLLaMA 1.4B is on par with the most recent opensource models.
|
|
|
7 |
---
|
8 |
# Model Summery
|
9 |
MobileLLaMA-1.4B-Base is a Transformer with 1.4B billon paramters. We downscale LLaMA to facilitate the off-the-shelf deployment. To make our work reproducible, all
|
10 |
+
the models are trained on 1.3T tokens from the [RedPajama v1](https://www.together.ai/blog/redpajama) dataset only. This benefits further research by enabling controlled experiments.
|
11 |
|
12 |
We extensively assess our models on two standard natural language benchmarks, for language understanding and common sense reasoning respectively. Experimental results show that our
|
13 |
MobileLLaMA 1.4B is on par with the most recent opensource models.
|