Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ After uncensoring, deduping, and cleaning, our dataset consists of:
|
|
21 |
- 842,610 instructions of FLANv2 augmented with GPT-4 completions
|
22 |
- 2,625,353 instructions of FLANv2 augmented with GPT-3.5 completions
|
23 |
|
24 |
-
We followed the submix and system prompt distribution outlined in the Orca paper. With a few exceptions. We included all 75k of CoT in the FLAN-1m dataset rather than sampling that. Also, we found that many items were duplicated, so we removed duplicates
|
25 |
|
26 |
Then we filtered out instances of alignment, refusal, avoidance, and bias, in order to produce an uncensored model upon which can be layered your personalized alignment LoRA.
|
27 |
|
|
|
21 |
- 842,610 instructions of FLANv2 augmented with GPT-4 completions
|
22 |
- 2,625,353 instructions of FLANv2 augmented with GPT-3.5 completions
|
23 |
|
24 |
+
We followed the submix and system prompt distribution outlined in the Orca paper. With a few exceptions. We included all 75k of CoT in the FLAN-1m dataset rather than sampling that. Also, we found that many items were duplicated, so we removed duplicates.
|
25 |
|
26 |
Then we filtered out instances of alignment, refusal, avoidance, and bias, in order to produce an uncensored model upon which can be layered your personalized alignment LoRA.
|
27 |
|