juewang commited on
Commit
3c09771
1 Parent(s): 4a73d90

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -29,15 +29,15 @@ inference:
29
  max_new_tokens: 128
30
  ---
31
 
32
- # RedPajama-INCITE-Instruct-7B-v0.1
33
 
34
- RedPajama-INCITE-Instruct-7B-v0.1 was developed by Together and leaders from the open-source AI community including Ontocord.ai, ETH DS3Lab, AAI CERC, Université de Montréal, MILA - Québec AI Institute, Stanford Center for Research on Foundation Models (CRFM), Stanford Hazy Research research group and LAION.
35
 
36
  The model was fine-tuned for few-shot applications on the data of [GPT-JT](https://huggingface.co/togethercomputer/GPT-JT-6B-v1), with exclusion of tasks that overlap with the HELM core scenarios.
37
 
38
- - Base Model: [RedPajama-INCITE-Base-7B-v0.1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-7B-v0.1)
39
- - Instruction-tuned Version: [RedPajama-INCITE-Instruct-7B-v0.1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1)
40
- - Chat Version: [RedPajama-INCITE-Chat-7B-v0.1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Chat-7B-v0.1)
41
 
42
 
43
  ## Model Details
@@ -66,8 +66,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
66
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
67
 
68
  # init
69
- tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1")
70
- model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1", torch_dtype=torch.float16)
71
  model = model.to('cuda:0')
72
  # infer
73
  prompt = "Q: The capital of France is?\nA:"
@@ -108,8 +108,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
108
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
109
 
110
  # init
111
- tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1")
112
- model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1", device_map='auto', torch_dtype=torch.float16, load_in_8bit=True)
113
 
114
  # infer
115
  prompt = "Q: The capital of France is?\nA:"
@@ -139,8 +139,8 @@ MIN_TRANSFORMERS_VERSION = '4.25.1'
139
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
140
 
141
  # init
142
- tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1")
143
- model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1", torch_dtype=torch.bfloat16)
144
  # infer
145
  prompt = "Q: The capital of France is?\nA:"
146
  inputs = tokenizer(prompt, return_tensors='pt').to(model.device)
@@ -171,13 +171,13 @@ It is the responsibility of the end user to ensure that the model is used in a r
171
 
172
  #### Out-of-Scope Use
173
 
174
- RedPajama-INCITE-Instruct-7B-v0.1 is a language model and may not perform well for other use cases outside of its intended scope.
175
  For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
176
  It is important to consider the limitations of the model and to only use it for its intended purpose.
177
 
178
  #### Misuse and Malicious Use
179
 
180
- RedPajama-INCITE-Instruct-7B-v0.1 is designed for language modeling.
181
  Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the project.
182
 
183
  Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
@@ -194,7 +194,7 @@ Using the model to generate content that is cruel to individuals is a misuse of
194
 
195
  ## Limitations
196
 
197
- RedPajama-INCITE-Instruct-7B-v0.1, like other language models, has limitations that should be taken into consideration.
198
  For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
199
  We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
200
 
@@ -209,7 +209,7 @@ Please refer to [togethercomputer/RedPajama-Data-1T](https://huggingface.co/data
209
  - **Hardware:** 8 A100
210
  - **Optimizer:** Adam
211
  - **Gradient Accumulations**: 1
212
- - **Num of Tokens:** 131M tokens
213
  - **Learning rate:** 1e-5
214
 
215
  ## Community
 
29
  max_new_tokens: 128
30
  ---
31
 
32
+ # RedPajama-INCITE-7B-Instruct
33
 
34
+ RedPajama-INCITE-7B-Instruct was developed by Together and leaders from the open-source AI community including Ontocord.ai, ETH DS3Lab, AAI CERC, Université de Montréal, MILA - Québec AI Institute, Stanford Center for Research on Foundation Models (CRFM), Stanford Hazy Research research group and LAION.
35
 
36
  The model was fine-tuned for few-shot applications on the data of [GPT-JT](https://huggingface.co/togethercomputer/GPT-JT-6B-v1), with exclusion of tasks that overlap with the HELM core scenarios.
37
 
38
+ - Base Model: [RedPajama-INCITE-7B-Base](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Base)
39
+ - Instruction-tuned Version: [RedPajama-INCITE-7B-Instruct](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Instruct)
40
+ - Chat Version: [RedPajama-INCITE-7B-Chat](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Chat)
41
 
42
 
43
  ## Model Details
 
66
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
67
 
68
  # init
69
+ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
70
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", torch_dtype=torch.float16)
71
  model = model.to('cuda:0')
72
  # infer
73
  prompt = "Q: The capital of France is?\nA:"
 
108
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
109
 
110
  # init
111
+ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
112
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", device_map='auto', torch_dtype=torch.float16, load_in_8bit=True)
113
 
114
  # infer
115
  prompt = "Q: The capital of France is?\nA:"
 
139
  assert transformers.__version__ >= MIN_TRANSFORMERS_VERSION, f'Please upgrade transformers to version {MIN_TRANSFORMERS_VERSION} or higher.'
140
 
141
  # init
142
+ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct")
143
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/RedPajama-INCITE-7B-Instruct", torch_dtype=torch.bfloat16)
144
  # infer
145
  prompt = "Q: The capital of France is?\nA:"
146
  inputs = tokenizer(prompt, return_tensors='pt').to(model.device)
 
171
 
172
  #### Out-of-Scope Use
173
 
174
+ RedPajama-INCITE-7B-Instruct is a language model and may not perform well for other use cases outside of its intended scope.
175
  For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
176
  It is important to consider the limitations of the model and to only use it for its intended purpose.
177
 
178
  #### Misuse and Malicious Use
179
 
180
+ RedPajama-INCITE-7B-Instruct is designed for language modeling.
181
  Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the project.
182
 
183
  Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
 
194
 
195
  ## Limitations
196
 
197
+ RedPajama-INCITE-7B-Instruct, like other language models, has limitations that should be taken into consideration.
198
  For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
199
  We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
200
 
 
209
  - **Hardware:** 8 A100
210
  - **Optimizer:** Adam
211
  - **Gradient Accumulations**: 1
212
+ - **Num of Tokens:** 1B tokens
213
  - **Learning rate:** 1e-5
214
 
215
  ## Community