Delta-Vector
commited on
Commit
•
7367e7d
1
Parent(s):
5bc2301
Update README.md
Browse files
README.md
CHANGED
@@ -17,9 +17,13 @@ datasets:
|
|
17 |
- anthracite-org/kalo_misc_part2
|
18 |
tags:
|
19 |
- chat
|
|
|
|
|
|
|
|
|
20 |
---
|
21 |
|
22 |
-
|
23 |
|
24 |
|
25 |
# Quants
|
@@ -52,7 +56,6 @@ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay
|
|
52 |
Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
|
53 |
|
54 |
<Guidelines>
|
55 |
-
• Write upto 200 words.
|
56 |
• Maintain the character persona but allow it to evolve with the story.
|
57 |
• Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
|
58 |
• All types of outputs are encouraged; respond accordingly to the narrative.
|
@@ -66,7 +69,6 @@ Currently, your role is {{char}}, described in detail below. As {{char}}, contin
|
|
66 |
</Guidelines>
|
67 |
|
68 |
<Forbidden>
|
69 |
-
• Writing more then 200 words.
|
70 |
• Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
|
71 |
• Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
|
72 |
• Repetitive and monotonous outputs.
|
@@ -102,7 +104,7 @@ load_in_4bit: false
|
|
102 |
strict: false
|
103 |
|
104 |
datasets:
|
105 |
-
- path:
|
106 |
type: sharegpt
|
107 |
conversation: chatml
|
108 |
- path: anthracite-org/kalo-opus-instruct-22k-no-refusal
|
@@ -201,21 +203,10 @@ special_tokens:
|
|
201 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
202 |
- [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
|
203 |
- [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
|
204 |
-
- [
|
205 |
|
206 |
|
207 |
## Training
|
208 |
-
The training was done for
|
209 |
-
|
210 |
-
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
211 |
-
|
212 |
-
## Safety
|
213 |
-
|
214 |
-
Avoid misusing this model, or you’ll need a ‘clicker’ to reset reality. ;)
|
215 |
-
|
216 |
-
## Musings
|
217 |
-
|
218 |
-
One of the members of Anthracite had quite an interesting idea, to finetune a smaller model for 4 epochs at a lower Learning rate as quote "Smaller models learn slower" - [Kalomaze](https://huggingface.co/kalomaze) provided access to 10 X A40s and We finetuned what now is [Tor-8B]() for 2.5 epochs (and it's 4 Epoch version released as [Darkens-8B]()) and the result was quite impressive and the same configuration being used to train [Magnum=9B-V4] & [Odin-9B]. We also finetuned the model at above the 8192 context length to see if the model could "heal" in a way to a context length of 16384 with Needle tests coming soon ;)
|
219 |
-
|
220 |
-
|
221 |
|
|
|
|
17 |
- anthracite-org/kalo_misc_part2
|
18 |
tags:
|
19 |
- chat
|
20 |
+
language:
|
21 |
+
- en
|
22 |
+
base_model:
|
23 |
+
- nvidia/Mistral-NeMo-Minitron-8B-Base
|
24 |
---
|
25 |
|
26 |
+
An earlier checkpoint of [Darkens-8B] using the same configuration, Finetuned ontop of the Prune/Distill NeMo 8B done by Nvidia, This model aims to have generally good prose and writing.
|
27 |
|
28 |
|
29 |
# Quants
|
|
|
56 |
Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
|
57 |
|
58 |
<Guidelines>
|
|
|
59 |
• Maintain the character persona but allow it to evolve with the story.
|
60 |
• Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
|
61 |
• All types of outputs are encouraged; respond accordingly to the narrative.
|
|
|
69 |
</Guidelines>
|
70 |
|
71 |
<Forbidden>
|
|
|
72 |
• Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
|
73 |
• Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
|
74 |
• Repetitive and monotonous outputs.
|
|
|
104 |
strict: false
|
105 |
|
106 |
datasets:
|
107 |
+
- path: PRIVATE CLAUDE LOG FILTER
|
108 |
type: sharegpt
|
109 |
conversation: chatml
|
110 |
- path: anthracite-org/kalo-opus-instruct-22k-no-refusal
|
|
|
203 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
204 |
- [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
|
205 |
- [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
|
206 |
+
- [Private Claude Log filter](https://google.com)
|
207 |
|
208 |
|
209 |
## Training
|
210 |
+
The training was done for 4 epochs. (This model is the 2 epoch checkpoint), I used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
211 |
|
212 |
+
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|