dfurman commited on
Commit
385b610
1 Parent(s): 08dd113

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -5,16 +5,16 @@ license: other
5
 
6
  # 🦙 LLaMA-13B
7
 
8
- LLaMA-13B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
10
- This model repo was converted to work with the transformers package. It is under a bespoke **non-commercial** license, please see the LICENSE file for more details.
11
 
12
  ## Model Summary
13
 
14
  - **Model Type:** Causal decoder-only.
15
  - **Dataset:** The model was trained on 1T tokens using the following data sources: CCNet [67%], C4 [15%], GitHub [4.5%], Wikipedia [4.5%], Books [4.5%], ArXiv [2.5%], Stack Exchange[2%].
16
  - **Language(s):** The Wikipedia and Books domains include data in the following languages: bg, ca, cs, da, de, en, es, fr, hr, hu, it, nl, pl, pt, ro, ru, sl, sr, sv, uk.
17
- - **License:** Bespoke non-commercial license, see LICENSE file.
18
  - **Model date:** LLaMA was trained between Dec 2022 and Feb 2023.
19
 
20
  **Where to send inquiries about the model:**
 
5
 
6
  # 🦙 LLaMA-13B
7
 
8
+ LLaMA-13B is a base model for text generation with 13.0B parameters and a 1T token training corpus. It was built and released by the FAIR team at Meta AI alongside the paper "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
10
+ This model repo was converted to work with the transformers package. It is under a bespoke **non-commercial** license, please see the [LICENSE](https://huggingface.co/dfurman/llama-13b/blob/main/LICENSE) file for more details.
11
 
12
  ## Model Summary
13
 
14
  - **Model Type:** Causal decoder-only.
15
  - **Dataset:** The model was trained on 1T tokens using the following data sources: CCNet [67%], C4 [15%], GitHub [4.5%], Wikipedia [4.5%], Books [4.5%], ArXiv [2.5%], Stack Exchange[2%].
16
  - **Language(s):** The Wikipedia and Books domains include data in the following languages: bg, ca, cs, da, de, en, es, fr, hr, hu, it, nl, pl, pt, ro, ru, sl, sr, sv, uk.
17
+ - **License:** Bespoke non-commercial license, see [LICENSE](https://huggingface.co/dfurman/llama-13b/blob/main/LICENSE) file.
18
  - **Model date:** LLaMA was trained between Dec 2022 and Feb 2023.
19
 
20
  **Where to send inquiries about the model:**