Update README.md
Browse files
README.md
CHANGED
@@ -5,16 +5,16 @@ license: other
|
|
5 |
|
6 |
# 🦙 LLaMA-13B
|
7 |
|
8 |
-
LLaMA-13B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
|
9 |
|
10 |
-
This model repo was converted to work with the transformers package. It is under a bespoke **non-commercial** license, please see the LICENSE file for more details.
|
11 |
|
12 |
## Model Summary
|
13 |
|
14 |
- **Model Type:** Causal decoder-only.
|
15 |
- **Dataset:** The model was trained on 1T tokens using the following data sources: CCNet [67%], C4 [15%], GitHub [4.5%], Wikipedia [4.5%], Books [4.5%], ArXiv [2.5%], Stack Exchange[2%].
|
16 |
- **Language(s):** The Wikipedia and Books domains include data in the following languages: bg, ca, cs, da, de, en, es, fr, hr, hu, it, nl, pl, pt, ro, ru, sl, sr, sv, uk.
|
17 |
-
- **License:** Bespoke non-commercial license, see LICENSE file.
|
18 |
- **Model date:** LLaMA was trained between Dec 2022 and Feb 2023.
|
19 |
|
20 |
**Where to send inquiries about the model:**
|
|
|
5 |
|
6 |
# 🦙 LLaMA-13B
|
7 |
|
8 |
+
LLaMA-13B is a base model for text generation with 13.0B parameters and a 1T token training corpus. It was built and released by the FAIR team at Meta AI alongside the paper "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
|
9 |
|
10 |
+
This model repo was converted to work with the transformers package. It is under a bespoke **non-commercial** license, please see the [LICENSE](https://huggingface.co/dfurman/llama-13b/blob/main/LICENSE) file for more details.
|
11 |
|
12 |
## Model Summary
|
13 |
|
14 |
- **Model Type:** Causal decoder-only.
|
15 |
- **Dataset:** The model was trained on 1T tokens using the following data sources: CCNet [67%], C4 [15%], GitHub [4.5%], Wikipedia [4.5%], Books [4.5%], ArXiv [2.5%], Stack Exchange[2%].
|
16 |
- **Language(s):** The Wikipedia and Books domains include data in the following languages: bg, ca, cs, da, de, en, es, fr, hr, hu, it, nl, pl, pt, ro, ru, sl, sr, sv, uk.
|
17 |
+
- **License:** Bespoke non-commercial license, see [LICENSE](https://huggingface.co/dfurman/llama-13b/blob/main/LICENSE) file.
|
18 |
- **Model date:** LLaMA was trained between Dec 2022 and Feb 2023.
|
19 |
|
20 |
**Where to send inquiries about the model:**
|