Dan Fu commited on
Commit
27c0e7c
1 Parent(s): f3e48f2
README.md CHANGED
@@ -6,17 +6,20 @@ language:
6
 
7
  ***<p style="font-size: 24px">Feel free to try out our [OpenChatKit feedback app](https://huggingface.co/spaces/togethercomputer/OpenChatKit)!</p>***
8
 
9
- # GPT-NeoXT-Chat-Base-20B
10
 
11
  > TLDR: As part of OpenChatKit (codebase available [here](https://github.com/togethercomputer/OpenChaT)),
12
- > GPT-NeoXT-Chat-Base-20B is a 20B parameter language model, fine-tuned from EleutherAI’s GPT-NeoX with over 40 million instructions on 100% carbon negative compute.
13
 
14
- GPT-NeoXT-Chat-Base-20B is based on ElutherAI’s GPT-NeoX model, and is fine-tuned with data focusing on dialog-style interactions.
15
  We focused the tuning on several tasks such as question answering, classification, extraction, and summarization.
16
- We’ve fine-tuned the model with a collection of 43 million high-quality instructions.
17
  Together partnered with LAION and Ontocord.ai, who both helped curate the dataset the model is based on.
18
  You can read more about this process and the availability of this dataset in LAION’s blog post [here](https://laion.ai/blog/oig-dataset/).
19
 
 
 
 
20
  ## Model Details
21
  - **Developed by**: Together Computer.
22
  - **Model type**: Language Model
@@ -27,18 +30,53 @@ You can read more about this process and the availability of this dataset in LAI
27
 
28
  # Quick Start
29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  ```python
31
- from transformers import pipeline
32
- pipe = pipeline(model='togethercomputer/GPT-NeoXT-Chat-Base-20B')
33
- pipe('''<human>: Hello!\n<bot>:''')
 
 
 
 
 
 
34
  ```
35
- or
 
 
36
  ```python
37
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
38
  tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
39
- model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
 
 
 
 
 
40
  ```
41
 
 
42
  ## Strengths of the model
43
 
44
  There are several tasks that OpenChatKit excels at out of the box. This includes:
@@ -140,19 +178,19 @@ Excluded uses are described below.
140
 
141
  ### Misuse, Malicious Use, and Out-of-Scope Use
142
 
143
- The OpenChatKit community provides GPT-NeoXT-Chat-Base-20B as an open source tool for building chatbots.
144
  The community is not responsible for any misuse, malicious use, or out-of-scope use of the model.
145
  It is the responsibility of the end user to ensure that the model is used in a responsible and ethical manner.
146
 
147
  #### Out-of-Scope Use
148
 
149
- GPT-NeoXT-Chat-Base-20B is designed for use in chatbot applications and may not perform well for other use cases outside of its intended scope.
150
  For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
151
  It is important to consider the limitations of the model and to only use it for its intended purpose.
152
 
153
  #### Misuse and Malicious Use
154
 
155
- GPT-NeoXT-Chat-Base-20B is designed for use in chatbot applications and should not be used for any other purpose.
156
  Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the OpenChatKit community project.
157
 
158
  Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
@@ -169,7 +207,7 @@ Using the model to generate content that is cruel to individuals is a misuse of
169
 
170
  ## Limitations
171
 
172
- GPT-NeoXT-Chat-Base-20B, like other language model-based chatbots, has limitations that should be taken into consideration.
173
  For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
174
  We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
175
 
@@ -189,4 +227,4 @@ Please refer to [togethercomputer/OpenDataHub](https://github.com/togethercomput
189
 
190
  ## Community
191
 
192
- Join us on [Together Discord](https://discord.gg/6ZVDU8tTD4)
 
6
 
7
  ***<p style="font-size: 24px">Feel free to try out our [OpenChatKit feedback app](https://huggingface.co/spaces/togethercomputer/OpenChatKit)!</p>***
8
 
9
+ # GPT-NeoXT-Chat-Base-20B-v0.16
10
 
11
  > TLDR: As part of OpenChatKit (codebase available [here](https://github.com/togethercomputer/OpenChaT)),
12
+ > GPT-NeoXT-Chat-Base-20B-v0.16 is a 20B parameter language model, fine-tuned from EleutherAI’s GPT-NeoX with over 40 million instructions on 100% carbon negative compute.
13
 
14
+ GPT-NeoXT-Chat-Base-20B-v0.16 is based on ElutherAI’s GPT-NeoX model, and is fine-tuned with data focusing on dialog-style interactions.
15
  We focused the tuning on several tasks such as question answering, classification, extraction, and summarization.
16
+ We’ve fine-tuned the model with a collection of 43 million high-quality instructions.
17
  Together partnered with LAION and Ontocord.ai, who both helped curate the dataset the model is based on.
18
  You can read more about this process and the availability of this dataset in LAION’s blog post [here](https://laion.ai/blog/oig-dataset/).
19
 
20
+ In addition to the aforementioned fine-tuning, GPT-NeoXT-Chat-Base-20B-v0.16 has also undergone further fine-tuning via a small amount of feedback data.
21
+ This allows the model to better adapt to human preferences in the conversations.
22
+
23
  ## Model Details
24
  - **Developed by**: Together Computer.
25
  - **Model type**: Language Model
 
30
 
31
  # Quick Start
32
 
33
+ ## GPU Inference
34
+
35
+ This requires a GPU with 48GB memory.
36
+ ```python
37
+ from transformers import AutoTokenizer, AutoModelForCausalLM
38
+ # init
39
+ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
40
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B", torch_dtype=torch.float16)
41
+ model = model.to('cuda:0')
42
+ # infer
43
+ inputs = tokenizer("<human>: Hello!\n<bot>:", return_tensors='pt').to(model.device)
44
+ outputs = model.generate(**inputs, max_new_tokens=10, do_sample=True, temperature=0.8)
45
+ output_str = tokenizer.decode(outputs[0])
46
+ print(output_str)
47
+ ```
48
+
49
+ ## GPU Inference in Int8
50
+
51
+ This requires a GPU with 24GB memory.
52
+
53
  ```python
54
+ from transformers import AutoTokenizer, AutoModelForCausalLM
55
+ # init
56
+ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
57
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
58
+ # infer
59
+ inputs = tokenizer("<human>: Hello!\n<bot>:", return_tensors='pt').to(model.device)
60
+ outputs = model.generate(**inputs, max_new_tokens=10, do_sample=True, temperature=0.8)
61
+ output_str = tokenizer.decode(outputs[0])
62
+ print(output_str)
63
  ```
64
+
65
+ ## CPU Inference
66
+
67
  ```python
68
  from transformers import AutoTokenizer, AutoModelForCausalLM
69
+ # init
70
  tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B")
71
+ model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-NeoXT-Chat-Base-20B", torch_dtype=torch.bfloat16)
72
+ # infer
73
+ inputs = tokenizer("<human>: Hello!\n<bot>:", return_tensors='pt').to(model.device)
74
+ outputs = model.generate(**inputs, max_new_tokens=10, do_sample=True, temperature=0.8)
75
+ output_str = tokenizer.decode(outputs[0])
76
+ print(output_str)
77
  ```
78
 
79
+
80
  ## Strengths of the model
81
 
82
  There are several tasks that OpenChatKit excels at out of the box. This includes:
 
178
 
179
  ### Misuse, Malicious Use, and Out-of-Scope Use
180
 
181
+ The OpenChatKit community provides GPT-NeoXT-Chat-Base-20B-v0.16 as an open source tool for building chatbots.
182
  The community is not responsible for any misuse, malicious use, or out-of-scope use of the model.
183
  It is the responsibility of the end user to ensure that the model is used in a responsible and ethical manner.
184
 
185
  #### Out-of-Scope Use
186
 
187
+ GPT-NeoXT-Chat-Base-20B-v0.16 is designed for use in chatbot applications and may not perform well for other use cases outside of its intended scope.
188
  For example, it may not be suitable for use in safety-critical applications or for making decisions that have a significant impact on individuals or society.
189
  It is important to consider the limitations of the model and to only use it for its intended purpose.
190
 
191
  #### Misuse and Malicious Use
192
 
193
+ GPT-NeoXT-Chat-Base-20B-v0.16 is designed for use in chatbot applications and should not be used for any other purpose.
194
  Misuse of the model, such as using it to engage in illegal or unethical activities, is strictly prohibited and goes against the principles of the OpenChatKit community project.
195
 
196
  Using the model to generate content that is cruel to individuals is a misuse of this model. This includes, but is not limited to:
 
207
 
208
  ## Limitations
209
 
210
+ GPT-NeoXT-Chat-Base-20B-v0.16, like other language model-based chatbots, has limitations that should be taken into consideration.
211
  For example, the model may not always provide accurate or relevant answers, particularly for questions that are complex, ambiguous, or outside of its training data.
212
  We therefore welcome contributions from individuals and organizations, and encourage collaboration towards creating a more robust and inclusive chatbot.
213
 
 
227
 
228
  ## Community
229
 
230
+ Join us on [Together Discord](https://discord.gg/6ZVDU8tTD4)
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "togethercomputer/OpenChaT",
3
  "architectures": [
4
  "GPTNeoXForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "togethercomputer/GPT-NeoXT-Chat-Base-20B",
3
  "architectures": [
4
  "GPTNeoXForCausalLM"
5
  ],
pytorch_model-00001-of-00005.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b56b667fea48813a8b8b9d7d860b5d3860e23b08888908179c30948a406e78b
3
  size 9953774091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cfd3a71c56d95e80f3d964e8be957f71bea4f1073788ac56d28a7815294ff5e
3
  size 9953774091
pytorch_model-00002-of-00005.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebad6672b8d33a16a1d55a1d8ba926066c73ea6b14a88dfb8031430368ccf971
3
  size 9787088144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c895eabbe65d8c9be65fd02a124436b2154aff20e2f9acd573496bd367d8ad1d
3
  size 9787088144
pytorch_model-00003-of-00005.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdab89888e02c97ebea0e3b8165c8f51c29f453dcc25e331c10b7e2a25edfadc
3
  size 9707369423
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71e7644ee43a87b0cb150fd0543d2f0f4f2a8b97c10015549303b2bca33ae4f2
3
  size 9707369423
pytorch_model-00004-of-00005.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48a9332605718b53c0b42eba1246100f7278ed35699b9ad20ded6132abccbe28
3
  size 9711578808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26add9b5cdf3454dbba9eaf8a1255734a9b684fa058b3032c6111782cf9d0f92
3
  size 9711578808
pytorch_model-00005-of-00005.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b547bbd094f3b040ee5a33b6216c5d6b64ce023bdcc1ab621e1e10aa27871990
3
  size 2134105435
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c902504815cc65f0349a5180d72159fb4250de1c6d22d7320bbf0e2772ffe0b4
3
  size 2134105435