Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,8 @@ Yi-34B model fine-tuned on AEZAKMI v1 dataset that is derived from airoboros 2.2
|
|
10 |
I had power target set to 320W for the GPU, and while I didn't measure power at the wall, it was probably something around 500W. Given the average electricity price in my region, this training run cost me around $3. This was my first attempt at training Yi-34B with this dataset.
|
11 |
Main feature of this model is that it's output should be free of refusals and it feels somehow more natural than airoboros. Prompt format is standard chatml. Don't expect it to be good at math, riddles or be crazy smart. My end goal with AEZAKMI is to create a cozy free chatbot.
|
12 |
|
|
|
|
|
13 |
## Prompt Format
|
14 |
|
15 |
I recommend using ChatML format, as this was used during fine-tune
|
|
|
10 |
I had power target set to 320W for the GPU, and while I didn't measure power at the wall, it was probably something around 500W. Given the average electricity price in my region, this training run cost me around $3. This was my first attempt at training Yi-34B with this dataset.
|
11 |
Main feature of this model is that it's output should be free of refusals and it feels somehow more natural than airoboros. Prompt format is standard chatml. Don't expect it to be good at math, riddles or be crazy smart. My end goal with AEZAKMI is to create a cozy free chatbot.
|
12 |
|
13 |
+
I used 4096 ctx Yi-34B-Llama uploaded by chargoddard as a base for this training.
|
14 |
+
|
15 |
## Prompt Format
|
16 |
|
17 |
I recommend using ChatML format, as this was used during fine-tune
|