dmayhem93 commited on
Commit
3b02e84
1 Parent(s): 81ce8c9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -17
README.md CHANGED
@@ -13,18 +13,18 @@ pipeline_tag: text-generation
13
 
14
  ## Model Description
15
 
16
- `StableBeluga_7B` is a Llama2 7B model finetuned on an Orca style Dataset
17
 
18
  ## Usage
19
 
20
- Start chatting with `StableBeluga_7B` using the following code snippet:
21
 
22
  ```python
23
  import torch
24
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
25
 
26
- tokenizer = AutoTokenizer.from_pretrained("stabilityai/StableBeluga_7B", use_fast=False)
27
- model = AutoModelForCausalLM.from_pretrained("stabilityai/StableBeluga_7B", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
28
  system_prompt = "### System:\nYou are StableBeluga, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
29
 
30
  message = "Write me a poem please"
@@ -35,7 +35,7 @@ output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_t
35
  print(tokenizer.decode(output[0], skip_special_tokens=True))
36
  ```
37
 
38
- StableBeluga_7B should be used with this prompt format:
39
  ```
40
  ### System:
41
  This is a system prompt, please behave and help the user.
@@ -44,21 +44,21 @@ This is a system prompt, please behave and help the user.
44
  Your prompt here
45
 
46
  ### Assistant:
47
- The output of StableBeluga_7B
48
  ```
49
 
50
  ## Model Details
51
 
52
  * **Developed by**: [Stability AI](https://stability.ai/)
53
- * **Model type**: StableBeluga_7B is an auto-regressive language model fine-tuned on Llama2 7B.
54
  * **Language(s)**: English
55
  * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
56
- * **License**: Fine-tuned checkpoints (`FreeWilly2`) is licensed under the Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
57
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
58
 
59
  ### Training Dataset
60
 
61
- `StableBeluga_7B` is trained on our internal Orca-style dataset
62
 
63
  ### Training Procedure
64
 
@@ -69,15 +69,9 @@ Models are learned via supervised fine-tuning on the aforementioned datasets, tr
69
  | Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
70
  | Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
71
 
72
- ## Use and Limitations
73
 
74
- ### Intended Use
75
-
76
- These models are intended for research only, in adherence with the [CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license.
77
-
78
- ### Limitations and bias
79
-
80
- Although the aforementioned dataset helps to steer the base language models into "safer" distributions of text, not all biases and toxicity can be mitigated through fine-tuning. We ask that users be mindful of such potential issues that can arise in generated responses. Do not treat model outputs as substitutes for human judgment or as sources of truth. Please use it responsibly.
81
 
82
  ## Citations
83
 
 
13
 
14
  ## Model Description
15
 
16
+ `Stable Beluga 7B` is a Llama2 7B model finetuned on an Orca style Dataset
17
 
18
  ## Usage
19
 
20
+ Start chatting with `Stable Beluga 7B` using the following code snippet:
21
 
22
  ```python
23
  import torch
24
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
25
 
26
+ tokenizer = AutoTokenizer.from_pretrained("stabilityai/StableBeluga-7B", use_fast=False)
27
+ model = AutoModelForCausalLM.from_pretrained("stabilityai/StableBeluga-7B", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
28
  system_prompt = "### System:\nYou are StableBeluga, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
29
 
30
  message = "Write me a poem please"
 
35
  print(tokenizer.decode(output[0], skip_special_tokens=True))
36
  ```
37
 
38
+ Stable Beluga 7B should be used with this prompt format:
39
  ```
40
  ### System:
41
  This is a system prompt, please behave and help the user.
 
44
  Your prompt here
45
 
46
  ### Assistant:
47
+ The output of Stable Beluga 7B
48
  ```
49
 
50
  ## Model Details
51
 
52
  * **Developed by**: [Stability AI](https://stability.ai/)
53
+ * **Model type**: Stable Beluga 7B is an auto-regressive language model fine-tuned on Llama2 70B.
54
  * **Language(s)**: English
55
  * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
56
+ * **License**: Fine-tuned checkpoints (`Stable Beluga 7B`) is licensed under the [STABLE BELUGA NON-COMMERCIAL COMMUNITY LICENSE AGREEMENT](https://huggingface.co/stabilityai/StableBeluga-7B/blob/main/LICENSE.txt)
57
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
58
 
59
  ### Training Dataset
60
 
61
+ ` Stable Beluga 7B` is trained on our internal Orca-style dataset
62
 
63
  ### Training Procedure
64
 
 
69
  | Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
70
  | Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
71
 
72
+ ## Ethical Considerations and Limitations
73
 
74
+ Beluga is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Beluga's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Beluga, developers should perform safety testing and tuning tailored to their specific applications of the model.
 
 
 
 
 
 
75
 
76
  ## Citations
77