hassanzay commited on
Commit
b676b54
1 Parent(s): 2190f5c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -81,6 +81,8 @@ model-index:
81
  ---
82
  # `stable-code-3b`
83
 
 
 
84
  ## Model Description
85
 
86
  `stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
@@ -184,7 +186,8 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
184
  * **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
185
  * **Language(s)**: English, Code
186
  * **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
187
- * **License**: License: StabilityAI Non-Commercial Research Community License. If you want to use this model for your commercial products or purposes, please contact us [here](https://stability.ai/membership) to learn more.
 
188
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
189
 
190
  ### Model Architecture
@@ -238,7 +241,7 @@ The model is pre-trained on the aforementioned datasets in `bfloat16` precision,
238
 
239
  ### Intended Use
240
 
241
- The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
242
 
243
  ### Limitations and Bias
244
 
 
81
  ---
82
  # `stable-code-3b`
83
 
84
+ Please note: For commercial use, please refer to https://stability.ai/membership.
85
+
86
  ## Model Description
87
 
88
  `stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
 
186
  * **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
187
  * **Language(s)**: English, Code
188
  * **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
189
+ * **License**: Stability AI Non-Commercial Research Community License.
190
+ * **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
191
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
192
 
193
  ### Model Architecture
 
241
 
242
  ### Intended Use
243
 
244
+ The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
245
 
246
  ### Limitations and Bias
247