dmayhem93 commited on
Commit
b522390
1 Parent(s): 4a006f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -6
README.md CHANGED
@@ -8,6 +8,7 @@ datasets:
8
  - LDJnr/Capybara
9
  - Intel/orca_dpo_pairs
10
  - hkust-nlp/deita-10k-v0
 
11
  language:
12
  - en
13
  tags:
@@ -124,6 +125,7 @@ print(output)
124
  * **Developed by**: [Stability AI](https://stability.ai/)
125
  * **Model type**: `StableLM 2 12B Chat` model is an auto-regressive language model based on the transformer decoder architecture.
126
  * **Language(s)**: English
 
127
  * **Paper**: [Stable LM 2 Chat Technical Report](https://drive.google.com/file/d/1JYJHszhS8EFChTbNAf8xmqhKjogWRrQF/view?usp=sharing)
128
  * **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
129
  * **Finetuned from model**:
@@ -132,7 +134,7 @@ print(output)
132
 
133
  ### Training Dataset
134
 
135
- The dataset is comprised of a mixture of open datasets large-scale datasets available on the [HuggingFace Hub](https://huggingface.co/datasets):
136
  1. SFT Datasets
137
  - HuggingFaceH4/ultrachat_200k
138
  - meta-math/MetaMathQA
@@ -142,7 +144,12 @@ The dataset is comprised of a mixture of open datasets large-scale datasets avai
142
  - LDJnr/Capybara
143
  - hkust-nlp/deita-10k-v0
144
 
145
- 2. Preference Datasets:
 
 
 
 
 
146
 
147
  ## Performance
148
 
@@ -155,6 +162,7 @@ The dataset is comprised of a mixture of open datasets large-scale datasets avai
155
 
156
  ### Training Infrastructure
157
 
 
158
  * **Hardware**: `StableLM 2 12B Chat` was trained on the Stability AI cluster across 8 nodes with 8 A100 80GBs GPUs for each nodes.
159
  * **Code Base**: We use our internal script for SFT training and [HuggingFace Alignment Handbook](https://github.com/huggingface/alignment-handbook) for DPO training.
160
 
@@ -165,11 +173,11 @@ The dataset is comprised of a mixture of open datasets large-scale datasets avai
165
  The model is intended to be used in chat-like applications. Developers must evaluate the model for safety performance in their specific use case. Read more about [safety and limitations](#limitations-and-bias) below.
166
 
167
  ### Limitations and Bias
168
-
169
- This model is not trained against adversarial inputs. We strongly recommend pairing this model with an input and output classifier to prevent harmful responses.
170
 
171
- Through our internal red teaming, we discovered that while the model will not output harmful information if not prompted to do so, it will hallucinate many facts. It is also willing to output potentially harmful outputs or misinformation when the user requests it.
172
- Using this model will require guardrails around your inputs and outputs to ensure that any outputs returned are not misinformation or harmful.
 
 
173
  Additionally, as each use case is unique, we recommend running your own suite of tests to ensure proper performance of this model.
174
  Finally, do not use the models if they are unsuitable for your application, or for any applications that may cause deliberate or unintentional harm to others.
175
 
 
8
  - LDJnr/Capybara
9
  - Intel/orca_dpo_pairs
10
  - hkust-nlp/deita-10k-v0
11
+ - Anthropic/hh-rlhf
12
  language:
13
  - en
14
  tags:
 
125
  * **Developed by**: [Stability AI](https://stability.ai/)
126
  * **Model type**: `StableLM 2 12B Chat` model is an auto-regressive language model based on the transformer decoder architecture.
127
  * **Language(s)**: English
128
+ TODO: Check if we want to keep paper link since it's not mentioned in that paper.
129
  * **Paper**: [Stable LM 2 Chat Technical Report](https://drive.google.com/file/d/1JYJHszhS8EFChTbNAf8xmqhKjogWRrQF/view?usp=sharing)
130
  * **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
131
  * **Finetuned from model**:
 
134
 
135
  ### Training Dataset
136
 
137
+ The dataset is comprised of a mixture of open datasets large-scale datasets available on the [HuggingFace Hub](https://huggingface.co/datasets) as well as an internal safety dataset:
138
  1. SFT Datasets
139
  - HuggingFaceH4/ultrachat_200k
140
  - meta-math/MetaMathQA
 
144
  - LDJnr/Capybara
145
  - hkust-nlp/deita-10k-v0
146
 
147
+ 2. Safety Datasets:
148
+ - Anthropic/hh-rlhf
149
+ - Internal Safety Dataset
150
+
151
+ 3. Preference Datasets:
152
+
153
 
154
  ## Performance
155
 
 
162
 
163
  ### Training Infrastructure
164
 
165
+ TODO: Fix this
166
  * **Hardware**: `StableLM 2 12B Chat` was trained on the Stability AI cluster across 8 nodes with 8 A100 80GBs GPUs for each nodes.
167
  * **Code Base**: We use our internal script for SFT training and [HuggingFace Alignment Handbook](https://github.com/huggingface/alignment-handbook) for DPO training.
168
 
 
173
  The model is intended to be used in chat-like applications. Developers must evaluate the model for safety performance in their specific use case. Read more about [safety and limitations](#limitations-and-bias) below.
174
 
175
  ### Limitations and Bias
 
 
176
 
177
+ TODO: Do we need or have a standard template to throw in here now?
178
+
179
+ We strongly recommend pairing this model with an input and output classifier to prevent harmful responses.
180
+ Using this model will require guardrails around your inputs and outputs to ensure that any outputs returned are not hallucinations.
181
  Additionally, as each use case is unique, we recommend running your own suite of tests to ensure proper performance of this model.
182
  Finally, do not use the models if they are unsuitable for your application, or for any applications that may cause deliberate or unintentional harm to others.
183