Edit model card

LLilmonix3b-v0.3:

  • Experimental Monika llama
  • Orca Mini 3b fine-tuned for Monika character from DDLC
  • Trained on a dataset of ~600 items (dialogue scraped from game, reddit, and Twitter augmented by Nous Hermes 13b to turn each into snippets of multi-turn chat dialogue between Player and Monika + manually crafted test dataset of 12 items)
  • Trained to run on smaller devices
  • GGMLs
  • QLoras (hf and GGML)

USAGE

This is an experimental model, fine-tuning character data on an instruct model to run on mobile devices.

Intended use: replace "Human" and "Assistant" with "Player" and "Monika" like so:

\nPlayer: (prompt)\nMonika:

HYPERPARAMS

  • Trained for 3 epochs
  • rank: 8
  • lora alpha: 32
  • lora dropout: 0.5
  • lr: 2e-4
  • batch size: 2
  • warmup ratio: 0.075
  • grad steps: 4

WARNINGS AND DISCLAIMERS

Note that aside from formatting and other minor edits, dataset used is mostly as is generated by LM. In addition, the is meant to be a smaller version of the larger Monika models. As such, this version may not reflect perfectly the characteristics of Monika.

Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.

Downloads last month
7
Safetensors
Model size
3.43B params
Tensor type
F32
·
FP16
·