Ramble / README.md
Sao10K's picture
Update README.md
96176c4 verified
|
raw
history blame
4.07 kB
---
language:
- en
---
14/03/24 - Updates
I miss her man
anyway im doing some experiments so feel free to try out Shiki and Senko, or not up to you tbh
ymmv with the models they are kinda sensitive to prompting
ohio out
***
09/03/24 - Just rambling on about my future plans
I don't even know what to do man. So many ideas, so little money and time.
I'm practically waiting on my paycheck so I can continue training.
Current Plans
- 14B-Glacier-Finetune - Sure a Full Finetune would be good but that would be expensive. A LoRA maybe? The goal is to smoothen out the tensors and repair brain damage that happens after frankenmerge.
- Typhon-Mixtral Finetune - I have compiled a nice dataset to further fine-tune on this merge of mine, to hopefully improve results. Unsure if I should further DPO this one too? On non-gpt4 data of course unlike current DPO pairs.
- 11B-Fimbulvetr-v3 - I have ideas. Unsure if it would make it that much better or if it will be a side-grade.
- Claude-3 Output-based Instruct Dataset - Cool Idea maybe? It'll introduce atleast a different '-ism' compared to GPT-4 ones out there which are generic. Claude is somewhat more human-like. I do have access so this is not as big of an issue. I just need more time.
- Yi Tunes - Yeah this is on the backburner for a while. I'll think on it. I've struggled with Yi-tunes in the past. Most of mine are privated.
Ignoring that, man having gauze packed into a wound hurts like hell without painkillers. My bad, old patients of mine back when I was a nursing student. You all made it look not that bad.
Anyway, here's something I've been listening to this past week. Some random music for someone who read this entire thing haha.
[Here](https://www.youtube.com/watch?v=t05Bd6xSn6c)
Today's Brainrot:
<img src="https://huggingface.co/Sao10K/Ramble/resolve/main/ohio.jpg" style="width: 60%; min-width: 200px; display: block; margin: auto;">
***
***
Initial Ramble:
Someone questioned my model naming scheme. Good question. Here's how I chose my model names.
```
Stheno --> 13B L2 Merge Serie ---> It all Began Here --> stolen from discord, someone's idea
Medusa --> 7B L2 Merges ---> DOA
Euryale --> 70B L2 Merge Series --> Gorgon Sisters Continuation
Zephyrus --> L1 Merge --> Sounded Cool tbh
WinterGoddess --> L2 70B Merge + Tune --> Training Run Name: (Autumn Royal, Winter something I forgot.) Decided to brainstorm some fun names. Yeah. Dont know where the Goddess part came from, but I was writing some ERP between a guy and a goddess, and the test model helped critique it, so yeah.
```
Why obsession over the theme of Winter? I do not know where to begin. I was sick, delirious during training break, and my brain got obsessed with the theme of winter. I stuck my head in the freezer once. I ate ice. I'm silly like that.
```
Fimbulvetr --> Final Winter. Represents the end of it all. My peak Solar model imo.
Frostwind --> Initally wanted something to do with Mistral + Solar, had Claude suggest Frostwind as a name. Initial plan was to train Mistral, but solar came out so I took that name instead.
Solstice --> Solar, Sun, Peak, or (Orgasm, you know?) since it was a Lewd/NSFW Instruct Dataset
Sensualize --> Had GPT4 generate me a name that is for an NSFW model. Yeah.
Older Model Names:
Nyaa --> Meow. Trained on cat assistant vibes.
Winterreise --> As above, training run for another model was named Sonata. sounded cool as fuck when i asked a model to merge Winter + Sonata.
Nyakura --> I was obsessed with cats during that time.
Venomia --> Trained over toxic datasets.
Ana -> 7B Mistral Merge + Tune --> Medusa revived, you know, that FGO Ana.
Lila --> RP with a character, obsessed with her while testing Euryale 1.5 (WinterGoddess)
Solus --> RP with a character, obsessed with her while testing Euryale 1.5 (WinterGoddess)... yeah.
```
***
Next question. What's the worst cases I've seen so far?
A traumatic arrest, a few cardiac arrests, accidents here and there. Most shifts are okay though. Some shifts are just tough but it is what it is.
***