zhilinw commited on
Commit
9c119cc
1 Parent(s): af626e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -12,6 +12,7 @@ tags:
12
  - rlhf
13
  datasets:
14
  - Anthropic/hh-rlhf
 
15
  ---
16
 
17
  # NV-Llama2-70B-RLHF-Chat
@@ -183,4 +184,4 @@ Pre-requisite: You would need at least a machine with 4 40GB or 2 80GB NVIDIA GP
183
  ## Limitations
184
  - The model was trained on the data that contains toxic language and societal biases originally crawled from the Internet. Therefore, the model may amplify those biases and return toxic responses especially when prompted with toxic prompts.
185
  - The Model may generate answers that may be inaccurate, omit key information, or include irrelevant or redundant text producing socially unacceptable or undesirable text, even if the prompt itself does not include anything explicitly offensive.
186
- - We recommend deploying the model with [NeMo Guardrails](https://github.com/NVIDIA/NeMo-Guardrails) to mitigate these potential issues.
 
12
  - rlhf
13
  datasets:
14
  - Anthropic/hh-rlhf
15
+ - nvidia/sft_datablend_v1
16
  ---
17
 
18
  # NV-Llama2-70B-RLHF-Chat
 
184
  ## Limitations
185
  - The model was trained on the data that contains toxic language and societal biases originally crawled from the Internet. Therefore, the model may amplify those biases and return toxic responses especially when prompted with toxic prompts.
186
  - The Model may generate answers that may be inaccurate, omit key information, or include irrelevant or redundant text producing socially unacceptable or undesirable text, even if the prompt itself does not include anything explicitly offensive.
187
+ - We recommend deploying the model with [NeMo Guardrails](https://github.com/NVIDIA/NeMo-Guardrails) to mitigate these potential issues.