dfurman commited on
Commit
f86318d
1 Parent(s): 1611920

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -9
README.md CHANGED
@@ -3,12 +3,11 @@ pipeline_tag: text-generation
3
  license: other
4
  ---
5
 
6
- # 🚀 LLaMA-13B
7
 
8
  LLaMA-13B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
10
- This model repo was converted to work with the Hugging Face transformers package. It is under a bespoke **non-commercial** license, please see the LICENSE file for more details.
11
-
12
 
13
  ## Model Summary
14
 
@@ -23,15 +22,13 @@ Questions and comments about LLaMA can be sent via the [GitHub repository](https
23
 
24
  ## Intended use
25
  **Primary intended uses**
26
- The primary use of LLaMA is research on large language models, including:
27
- exploring potential applications such as question answering, natural language understanding or reading comprehension, understanding capabilities and limitations of current language models, and developing techniques to improve those,
28
- evaluating and mitigating biases, risks, toxic and harmful content generations, hallucinations.
29
 
30
  **Primary intended users**
31
  The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
32
 
33
  **Out-of-scope use cases**
34
- LLaMA is a foundation model (a base model). As such, it should not be used on downstream applications without further risk evaluation and mitigation. In particular, the model has not been trained with human feedback, and can thus generate toxic or offensive content, incorrect information or generally unhelpful answers.
35
 
36
  ## Factors
37
  **Relevant factors**
@@ -60,12 +57,11 @@ LLaMA is a foundational model, and as such, it should not be used for downstream
60
 
61
  ### Setup
62
  ```python
63
- # Install packages
64
  !pip install -q -U transformers accelerate torch
65
  ```
66
  ### GPU Inference in fp16
67
 
68
- This requires a GPU with at least xxGB of VRAM.
69
 
70
  ### First, Load the Model
71
 
 
3
  license: other
4
  ---
5
 
6
+ # 🦙 LLaMA-13B
7
 
8
  LLaMA-13B is a base model for text generation. It was built and released by Meta AI alongside "[LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)".
9
 
10
+ This model repo was converted to work with the transformers package. It is under a bespoke **non-commercial** license, please see the LICENSE file for more details.
 
11
 
12
  ## Model Summary
13
 
 
22
 
23
  ## Intended use
24
  **Primary intended uses**
25
+ The primary use of LLaMA is research on large language models, including: exploring potential applications such as question answering, natural language understanding or reading comprehension, understanding capabilities and limitations of current language models, and developing techniques to improve those, evaluating and mitigating biases, risks, toxic and harmful content generations, and hallucinations.
 
 
26
 
27
  **Primary intended users**
28
  The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence.
29
 
30
  **Out-of-scope use cases**
31
+ LLaMA is a base model, also known as a foundation model. As such, it should not be used on downstream applications without further risk evaluation, mitigation, and potential further fine-tuning. In particular, the model has not been trained with human feedback, and can thus generate toxic or offensive content, incorrect information or generally unhelpful answers.
32
 
33
  ## Factors
34
  **Relevant factors**
 
57
 
58
  ### Setup
59
  ```python
 
60
  !pip install -q -U transformers accelerate torch
61
  ```
62
  ### GPU Inference in fp16
63
 
64
+ This requires a GPU with at least 15GB of VRAM.
65
 
66
  ### First, Load the Model
67